/************************************************************************** * * Copyright 2007-2008 Tungsten Graphics, Inc., Cedar Park, Texas. * All Rights Reserved. * Copyright 2009-2010 VMware, Inc. All rights Reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sub license, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice (including the * next paragraph) shall be included in all copies or substantial portions * of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. * **************************************************************************/ /** * TGSI interpreter/executor. * * Flow control information: * * Since we operate on 'quads' (4 pixels or 4 vertices in parallel) * flow control statements (IF/ELSE/ENDIF, LOOP/ENDLOOP) require special * care since a condition may be true for some quad components but false * for other components. * * We basically execute all statements (even if they're in the part of * an IF/ELSE clause that's "not taken") and use a special mask to * control writing to destination registers. This is the ExecMask. * See store_dest(). * * The ExecMask is computed from three other masks (CondMask, LoopMask and * ContMask) which are controlled by the flow control instructions (namely: * (IF/ELSE/ENDIF, LOOP/ENDLOOP and CONT). * * * Authors: * Michal Krol * Brian Paul */ #include "pipe/p_compiler.h" #include "pipe/p_state.h" #include "pipe/p_shader_tokens.h" #include "tgsi/tgsi_dump.h" #include "tgsi/tgsi_parse.h" #include "tgsi/tgsi_util.h" #include "tgsi_exec.h" #include "util/u_memory.h" #include "util/u_math.h" #define FAST_MATH 1 static void micro_iabs(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->i[0] = src->i[0] >= 0 ? src->i[0] : -src->i[0]; dst->i[1] = src->i[1] >= 0 ? src->i[1] : -src->i[1]; dst->i[2] = src->i[2] >= 0 ? src->i[2] : -src->i[2]; dst->i[3] = src->i[3] >= 0 ? src->i[3] : -src->i[3]; } static void micro_ineg(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->i[0] = -src->i[0]; dst->i[1] = -src->i[1]; dst->i[2] = -src->i[2]; dst->i[3] = -src->i[3]; } static void micro_mov(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src->u[0]; dst->u[1] = src->u[1]; dst->u[2] = src->u[2]; dst->u[3] = src->u[3]; } #define TILE_TOP_LEFT 0 #define TILE_TOP_RIGHT 1 #define TILE_BOTTOM_LEFT 2 #define TILE_BOTTOM_RIGHT 3 #define CHAN_X 0 #define CHAN_Y 1 #define CHAN_Z 2 #define CHAN_W 3 enum tgsi_exec_datatype { TGSI_EXEC_DATA_FLOAT, TGSI_EXEC_DATA_INT, TGSI_EXEC_DATA_UINT }; /* * Shorthand locations of various utility registers (_I = Index, _C = Channel) */ #define TEMP_0_I TGSI_EXEC_TEMP_00000000_I #define TEMP_0_C TGSI_EXEC_TEMP_00000000_C #define TEMP_7F_I TGSI_EXEC_TEMP_7FFFFFFF_I #define TEMP_7F_C TGSI_EXEC_TEMP_7FFFFFFF_C #define TEMP_80_I TGSI_EXEC_TEMP_80000000_I #define TEMP_80_C TGSI_EXEC_TEMP_80000000_C #define TEMP_FF_I TGSI_EXEC_TEMP_FFFFFFFF_I #define TEMP_FF_C TGSI_EXEC_TEMP_FFFFFFFF_C #define TEMP_1_I TGSI_EXEC_TEMP_ONE_I #define TEMP_1_C TGSI_EXEC_TEMP_ONE_C #define TEMP_2_I TGSI_EXEC_TEMP_TWO_I #define TEMP_2_C TGSI_EXEC_TEMP_TWO_C #define TEMP_128_I TGSI_EXEC_TEMP_128_I #define TEMP_128_C TGSI_EXEC_TEMP_128_C #define TEMP_M128_I TGSI_EXEC_TEMP_MINUS_128_I #define TEMP_M128_C TGSI_EXEC_TEMP_MINUS_128_C #define TEMP_KILMASK_I TGSI_EXEC_TEMP_KILMASK_I #define TEMP_KILMASK_C TGSI_EXEC_TEMP_KILMASK_C #define TEMP_OUTPUT_I TGSI_EXEC_TEMP_OUTPUT_I #define TEMP_OUTPUT_C TGSI_EXEC_TEMP_OUTPUT_C #define TEMP_PRIMITIVE_I TGSI_EXEC_TEMP_PRIMITIVE_I #define TEMP_PRIMITIVE_C TGSI_EXEC_TEMP_PRIMITIVE_C #define TEMP_CC_I TGSI_EXEC_TEMP_CC_I #define TEMP_CC_C TGSI_EXEC_TEMP_CC_C #define TEMP_3_I TGSI_EXEC_TEMP_THREE_I #define TEMP_3_C TGSI_EXEC_TEMP_THREE_C #define TEMP_HALF_I TGSI_EXEC_TEMP_HALF_I #define TEMP_HALF_C TGSI_EXEC_TEMP_HALF_C #define TEMP_R0 TGSI_EXEC_TEMP_R0 #define TEMP_P0 TGSI_EXEC_TEMP_P0 #define IS_CHANNEL_ENABLED(INST, CHAN)\ ((INST).Dst[0].Register.WriteMask & (1 << (CHAN))) #define IS_CHANNEL_ENABLED2(INST, CHAN)\ ((INST).Dst[1].Register.WriteMask & (1 << (CHAN))) #define FOR_EACH_ENABLED_CHANNEL(INST, CHAN)\ for (CHAN = 0; CHAN < NUM_CHANNELS; CHAN++)\ if (IS_CHANNEL_ENABLED( INST, CHAN )) #define FOR_EACH_ENABLED_CHANNEL2(INST, CHAN)\ for (CHAN = 0; CHAN < NUM_CHANNELS; CHAN++)\ if (IS_CHANNEL_ENABLED2( INST, CHAN )) /** The execution mask depends on the conditional mask and the loop mask */ #define UPDATE_EXEC_MASK(MACH) \ MACH->ExecMask = MACH->CondMask & MACH->LoopMask & MACH->ContMask & MACH->Switch.mask & MACH->FuncMask static const union tgsi_exec_channel ZeroVec = { { 0.0, 0.0, 0.0, 0.0 } }; #define CHECK_INF_OR_NAN(chan) do {\ assert(!util_is_inf_or_nan((chan)->f[0]));\ assert(!util_is_inf_or_nan((chan)->f[1]));\ assert(!util_is_inf_or_nan((chan)->f[2]));\ assert(!util_is_inf_or_nan((chan)->f[3]));\ } while (0) #ifdef DEBUG static void print_chan(const char *msg, const union tgsi_exec_channel *chan) { debug_printf("%s = {%f, %f, %f, %f}\n", msg, chan->f[0], chan->f[1], chan->f[2], chan->f[3]); } #endif #ifdef DEBUG static void print_temp(const struct tgsi_exec_machine *mach, uint index) { const struct tgsi_exec_vector *tmp = &mach->Temps[index]; int i; debug_printf("Temp[%u] =\n", index); for (i = 0; i < 4; i++) { debug_printf(" %c: { %f, %f, %f, %f }\n", "XYZW"[i], tmp->xyzw[i].f[0], tmp->xyzw[i].f[1], tmp->xyzw[i].f[2], tmp->xyzw[i].f[3]); } } #endif /** * Check if there's a potential src/dst register data dependency when * using SOA execution. * Example: * MOV T, T.yxwz; * This would expand into: * MOV t0, t1; * MOV t1, t0; * MOV t2, t3; * MOV t3, t2; * The second instruction will have the wrong value for t0 if executed as-is. */ boolean tgsi_check_soa_dependencies(const struct tgsi_full_instruction *inst) { uint i, chan; uint writemask = inst->Dst[0].Register.WriteMask; if (writemask == TGSI_WRITEMASK_X || writemask == TGSI_WRITEMASK_Y || writemask == TGSI_WRITEMASK_Z || writemask == TGSI_WRITEMASK_W || writemask == TGSI_WRITEMASK_NONE) { /* no chance of data dependency */ return FALSE; } /* loop over src regs */ for (i = 0; i < inst->Instruction.NumSrcRegs; i++) { if ((inst->Src[i].Register.File == inst->Dst[0].Register.File) && (inst->Src[i].Register.Index == inst->Dst[0].Register.Index)) { /* loop over dest channels */ uint channelsWritten = 0x0; FOR_EACH_ENABLED_CHANNEL(*inst, chan) { /* check if we're reading a channel that's been written */ uint swizzle = tgsi_util_get_full_src_register_swizzle(&inst->Src[i], chan); if (channelsWritten & (1 << swizzle)) { return TRUE; } channelsWritten |= (1 << chan); } } } return FALSE; } /** * Initialize machine state by expanding tokens to full instructions, * allocating temporary storage, setting up constants, etc. * After this, we can call tgsi_exec_machine_run() many times. */ void tgsi_exec_machine_bind_shader( struct tgsi_exec_machine *mach, const struct tgsi_token *tokens, uint numSamplers, struct tgsi_sampler **samplers) { uint k; struct tgsi_parse_context parse; struct tgsi_exec_labels *labels = &mach->Labels; struct tgsi_full_instruction *instructions; struct tgsi_full_declaration *declarations; uint maxInstructions = 10, numInstructions = 0; uint maxDeclarations = 10, numDeclarations = 0; uint instno = 0; #if 0 tgsi_dump(tokens, 0); #endif util_init_math(); mach->Tokens = tokens; mach->Samplers = samplers; k = tgsi_parse_init (&parse, mach->Tokens); if (k != TGSI_PARSE_OK) { debug_printf( "Problem parsing!\n" ); return; } mach->Processor = parse.FullHeader.Processor.Processor; mach->ImmLimit = 0; labels->count = 0; declarations = (struct tgsi_full_declaration *) MALLOC( maxDeclarations * sizeof(struct tgsi_full_declaration) ); if (!declarations) { return; } instructions = (struct tgsi_full_instruction *) MALLOC( maxInstructions * sizeof(struct tgsi_full_instruction) ); if (!instructions) { FREE( declarations ); return; } while( !tgsi_parse_end_of_tokens( &parse ) ) { uint pointer = parse.Position; uint i; tgsi_parse_token( &parse ); switch( parse.FullToken.Token.Type ) { case TGSI_TOKEN_TYPE_DECLARATION: /* save expanded declaration */ if (numDeclarations == maxDeclarations) { declarations = REALLOC(declarations, maxDeclarations * sizeof(struct tgsi_full_declaration), (maxDeclarations + 10) * sizeof(struct tgsi_full_declaration)); maxDeclarations += 10; } if (parse.FullToken.FullDeclaration.Declaration.File == TGSI_FILE_OUTPUT) { unsigned reg; for (reg = parse.FullToken.FullDeclaration.Range.First; reg <= parse.FullToken.FullDeclaration.Range.Last; ++reg) { ++mach->NumOutputs; } } memcpy(declarations + numDeclarations, &parse.FullToken.FullDeclaration, sizeof(declarations[0])); numDeclarations++; break; case TGSI_TOKEN_TYPE_IMMEDIATE: { uint size = parse.FullToken.FullImmediate.Immediate.NrTokens - 1; assert( size <= 4 ); assert( mach->ImmLimit + 1 <= TGSI_EXEC_NUM_IMMEDIATES ); for( i = 0; i < size; i++ ) { mach->Imms[mach->ImmLimit][i] = parse.FullToken.FullImmediate.u[i].Float; } mach->ImmLimit += 1; } break; case TGSI_TOKEN_TYPE_INSTRUCTION: assert( labels->count < MAX_LABELS ); labels->labels[labels->count][0] = instno; labels->labels[labels->count][1] = pointer; labels->count++; /* save expanded instruction */ if (numInstructions == maxInstructions) { instructions = REALLOC(instructions, maxInstructions * sizeof(struct tgsi_full_instruction), (maxInstructions + 10) * sizeof(struct tgsi_full_instruction)); maxInstructions += 10; } memcpy(instructions + numInstructions, &parse.FullToken.FullInstruction, sizeof(instructions[0])); numInstructions++; break; case TGSI_TOKEN_TYPE_PROPERTY: break; default: assert( 0 ); } } tgsi_parse_free (&parse); if (mach->Declarations) { FREE( mach->Declarations ); } mach->Declarations = declarations; mach->NumDeclarations = numDeclarations; if (mach->Instructions) { FREE( mach->Instructions ); } mach->Instructions = instructions; mach->NumInstructions = numInstructions; } struct tgsi_exec_machine * tgsi_exec_machine_create( void ) { struct tgsi_exec_machine *mach; uint i; mach = align_malloc( sizeof *mach, 16 ); if (!mach) goto fail; memset(mach, 0, sizeof(*mach)); mach->Addrs = &mach->Temps[TGSI_EXEC_TEMP_ADDR]; mach->MaxGeometryShaderOutputs = TGSI_MAX_TOTAL_VERTICES; mach->Predicates = &mach->Temps[TGSI_EXEC_TEMP_P0]; /* Setup constants. */ for( i = 0; i < 4; i++ ) { mach->Temps[TEMP_0_I].xyzw[TEMP_0_C].u[i] = 0x00000000; mach->Temps[TEMP_7F_I].xyzw[TEMP_7F_C].u[i] = 0x7FFFFFFF; mach->Temps[TEMP_80_I].xyzw[TEMP_80_C].u[i] = 0x80000000; mach->Temps[TEMP_FF_I].xyzw[TEMP_FF_C].u[i] = 0xFFFFFFFF; mach->Temps[TEMP_1_I].xyzw[TEMP_1_C].f[i] = 1.0f; mach->Temps[TEMP_2_I].xyzw[TEMP_2_C].f[i] = 2.0f; mach->Temps[TEMP_128_I].xyzw[TEMP_128_C].f[i] = 128.0f; mach->Temps[TEMP_M128_I].xyzw[TEMP_M128_C].f[i] = -128.0f; mach->Temps[TEMP_3_I].xyzw[TEMP_3_C].f[i] = 3.0f; mach->Temps[TEMP_HALF_I].xyzw[TEMP_HALF_C].f[i] = 0.5f; } #ifdef DEBUG /* silence warnings */ (void) print_chan; (void) print_temp; #endif return mach; fail: align_free(mach); return NULL; } void tgsi_exec_machine_destroy(struct tgsi_exec_machine *mach) { if (mach) { FREE(mach->Instructions); FREE(mach->Declarations); } align_free(mach); } static void micro_abs( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { dst->f[0] = fabsf( src->f[0] ); dst->f[1] = fabsf( src->f[1] ); dst->f[2] = fabsf( src->f[2] ); dst->f[3] = fabsf( src->f[3] ); } static void micro_add( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1 ) { dst->f[0] = src0->f[0] + src1->f[0]; dst->f[1] = src0->f[1] + src1->f[1]; dst->f[2] = src0->f[2] + src1->f[2]; dst->f[3] = src0->f[3] + src1->f[3]; } static void micro_ceil( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { dst->f[0] = ceilf( src->f[0] ); dst->f[1] = ceilf( src->f[1] ); dst->f[2] = ceilf( src->f[2] ); dst->f[3] = ceilf( src->f[3] ); } static void micro_cos( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { dst->f[0] = cosf( src->f[0] ); dst->f[1] = cosf( src->f[1] ); dst->f[2] = cosf( src->f[2] ); dst->f[3] = cosf( src->f[3] ); } static void micro_ddx( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { dst->f[0] = dst->f[1] = dst->f[2] = dst->f[3] = src->f[TILE_BOTTOM_RIGHT] - src->f[TILE_BOTTOM_LEFT]; } static void micro_ddy( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { dst->f[0] = dst->f[1] = dst->f[2] = dst->f[3] = src->f[TILE_BOTTOM_LEFT] - src->f[TILE_TOP_LEFT]; } static void micro_div( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1 ) { if (src1->f[0] != 0) { dst->f[0] = src0->f[0] / src1->f[0]; } if (src1->f[1] != 0) { dst->f[1] = src0->f[1] / src1->f[1]; } if (src1->f[2] != 0) { dst->f[2] = src0->f[2] / src1->f[2]; } if (src1->f[3] != 0) { dst->f[3] = src0->f[3] / src1->f[3]; } } static void micro_eq( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1, const union tgsi_exec_channel *src2, const union tgsi_exec_channel *src3 ) { dst->f[0] = src0->f[0] == src1->f[0] ? src2->f[0] : src3->f[0]; dst->f[1] = src0->f[1] == src1->f[1] ? src2->f[1] : src3->f[1]; dst->f[2] = src0->f[2] == src1->f[2] ? src2->f[2] : src3->f[2]; dst->f[3] = src0->f[3] == src1->f[3] ? src2->f[3] : src3->f[3]; } static void micro_exp2( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { #if FAST_MATH dst->f[0] = util_fast_exp2( src->f[0] ); dst->f[1] = util_fast_exp2( src->f[1] ); dst->f[2] = util_fast_exp2( src->f[2] ); dst->f[3] = util_fast_exp2( src->f[3] ); #else #if DEBUG /* Inf is okay for this instruction, so clamp it to silence assertions. */ uint i; union tgsi_exec_channel clamped; for (i = 0; i < 4; i++) { if (src->f[i] > 127.99999f) { clamped.f[i] = 127.99999f; } else if (src->f[i] < -126.99999f) { clamped.f[i] = -126.99999f; } else { clamped.f[i] = src->f[i]; } } src = &clamped; #endif dst->f[0] = powf( 2.0f, src->f[0] ); dst->f[1] = powf( 2.0f, src->f[1] ); dst->f[2] = powf( 2.0f, src->f[2] ); dst->f[3] = powf( 2.0f, src->f[3] ); #endif } static void micro_float_clamp(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { uint i; for (i = 0; i < 4; i++) { if (src->f[i] > 0.0f) { if (src->f[i] > 1.884467e+019f) dst->f[i] = 1.884467e+019f; else if (src->f[i] < 5.42101e-020f) dst->f[i] = 5.42101e-020f; else dst->f[i] = src->f[i]; } else { if (src->f[i] < -1.884467e+019f) dst->f[i] = -1.884467e+019f; else if (src->f[i] > -5.42101e-020f) dst->f[i] = -5.42101e-020f; else dst->f[i] = src->f[i]; } } } static void micro_flr( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { dst->f[0] = floorf( src->f[0] ); dst->f[1] = floorf( src->f[1] ); dst->f[2] = floorf( src->f[2] ); dst->f[3] = floorf( src->f[3] ); } static void micro_frc( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { dst->f[0] = src->f[0] - floorf( src->f[0] ); dst->f[1] = src->f[1] - floorf( src->f[1] ); dst->f[2] = src->f[2] - floorf( src->f[2] ); dst->f[3] = src->f[3] - floorf( src->f[3] ); } static void micro_lg2( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { #if FAST_MATH dst->f[0] = util_fast_log2( src->f[0] ); dst->f[1] = util_fast_log2( src->f[1] ); dst->f[2] = util_fast_log2( src->f[2] ); dst->f[3] = util_fast_log2( src->f[3] ); #else dst->f[0] = logf( src->f[0] ) * 1.442695f; dst->f[1] = logf( src->f[1] ) * 1.442695f; dst->f[2] = logf( src->f[2] ) * 1.442695f; dst->f[3] = logf( src->f[3] ) * 1.442695f; #endif } static void micro_le( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1, const union tgsi_exec_channel *src2, const union tgsi_exec_channel *src3 ) { dst->f[0] = src0->f[0] <= src1->f[0] ? src2->f[0] : src3->f[0]; dst->f[1] = src0->f[1] <= src1->f[1] ? src2->f[1] : src3->f[1]; dst->f[2] = src0->f[2] <= src1->f[2] ? src2->f[2] : src3->f[2]; dst->f[3] = src0->f[3] <= src1->f[3] ? src2->f[3] : src3->f[3]; } static void micro_lt( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1, const union tgsi_exec_channel *src2, const union tgsi_exec_channel *src3 ) { dst->f[0] = src0->f[0] < src1->f[0] ? src2->f[0] : src3->f[0]; dst->f[1] = src0->f[1] < src1->f[1] ? src2->f[1] : src3->f[1]; dst->f[2] = src0->f[2] < src1->f[2] ? src2->f[2] : src3->f[2]; dst->f[3] = src0->f[3] < src1->f[3] ? src2->f[3] : src3->f[3]; } static void micro_max( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1 ) { dst->f[0] = src0->f[0] > src1->f[0] ? src0->f[0] : src1->f[0]; dst->f[1] = src0->f[1] > src1->f[1] ? src0->f[1] : src1->f[1]; dst->f[2] = src0->f[2] > src1->f[2] ? src0->f[2] : src1->f[2]; dst->f[3] = src0->f[3] > src1->f[3] ? src0->f[3] : src1->f[3]; } static void micro_min( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1 ) { dst->f[0] = src0->f[0] < src1->f[0] ? src0->f[0] : src1->f[0]; dst->f[1] = src0->f[1] < src1->f[1] ? src0->f[1] : src1->f[1]; dst->f[2] = src0->f[2] < src1->f[2] ? src0->f[2] : src1->f[2]; dst->f[3] = src0->f[3] < src1->f[3] ? src0->f[3] : src1->f[3]; } static void micro_mul( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1 ) { dst->f[0] = src0->f[0] * src1->f[0]; dst->f[1] = src0->f[1] * src1->f[1]; dst->f[2] = src0->f[2] * src1->f[2]; dst->f[3] = src0->f[3] * src1->f[3]; } #if 0 static void micro_imul64( union tgsi_exec_channel *dst0, union tgsi_exec_channel *dst1, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1 ) { dst1->i[0] = src0->i[0] * src1->i[0]; dst1->i[1] = src0->i[1] * src1->i[1]; dst1->i[2] = src0->i[2] * src1->i[2]; dst1->i[3] = src0->i[3] * src1->i[3]; dst0->i[0] = 0; dst0->i[1] = 0; dst0->i[2] = 0; dst0->i[3] = 0; } #endif #if 0 static void micro_umul64( union tgsi_exec_channel *dst0, union tgsi_exec_channel *dst1, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1 ) { dst1->u[0] = src0->u[0] * src1->u[0]; dst1->u[1] = src0->u[1] * src1->u[1]; dst1->u[2] = src0->u[2] * src1->u[2]; dst1->u[3] = src0->u[3] * src1->u[3]; dst0->u[0] = 0; dst0->u[1] = 0; dst0->u[2] = 0; dst0->u[3] = 0; } #endif #if 0 static void micro_movc( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1, const union tgsi_exec_channel *src2 ) { dst->u[0] = src0->u[0] ? src1->u[0] : src2->u[0]; dst->u[1] = src0->u[1] ? src1->u[1] : src2->u[1]; dst->u[2] = src0->u[2] ? src1->u[2] : src2->u[2]; dst->u[3] = src0->u[3] ? src1->u[3] : src2->u[3]; } #endif static void micro_neg( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { dst->f[0] = -src->f[0]; dst->f[1] = -src->f[1]; dst->f[2] = -src->f[2]; dst->f[3] = -src->f[3]; } static void micro_pow( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1 ) { #if FAST_MATH dst->f[0] = util_fast_pow( src0->f[0], src1->f[0] ); dst->f[1] = util_fast_pow( src0->f[1], src1->f[1] ); dst->f[2] = util_fast_pow( src0->f[2], src1->f[2] ); dst->f[3] = util_fast_pow( src0->f[3], src1->f[3] ); #else dst->f[0] = powf( src0->f[0], src1->f[0] ); dst->f[1] = powf( src0->f[1], src1->f[1] ); dst->f[2] = powf( src0->f[2], src1->f[2] ); dst->f[3] = powf( src0->f[3], src1->f[3] ); #endif } static void micro_rnd( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { dst->f[0] = floorf( src->f[0] + 0.5f ); dst->f[1] = floorf( src->f[1] + 0.5f ); dst->f[2] = floorf( src->f[2] + 0.5f ); dst->f[3] = floorf( src->f[3] + 0.5f ); } static void micro_sgn( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { dst->f[0] = src->f[0] < 0.0f ? -1.0f : src->f[0] > 0.0f ? 1.0f : 0.0f; dst->f[1] = src->f[1] < 0.0f ? -1.0f : src->f[1] > 0.0f ? 1.0f : 0.0f; dst->f[2] = src->f[2] < 0.0f ? -1.0f : src->f[2] > 0.0f ? 1.0f : 0.0f; dst->f[3] = src->f[3] < 0.0f ? -1.0f : src->f[3] > 0.0f ? 1.0f : 0.0f; } static void micro_trunc( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src0 ) { dst->f[0] = (float) (int) src0->f[0]; dst->f[1] = (float) (int) src0->f[1]; dst->f[2] = (float) (int) src0->f[2]; dst->f[3] = (float) (int) src0->f[3]; } static void micro_sin( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { dst->f[0] = sinf( src->f[0] ); dst->f[1] = sinf( src->f[1] ); dst->f[2] = sinf( src->f[2] ); dst->f[3] = sinf( src->f[3] ); } static void micro_sqrt( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src ) { dst->f[0] = sqrtf( src->f[0] ); dst->f[1] = sqrtf( src->f[1] ); dst->f[2] = sqrtf( src->f[2] ); dst->f[3] = sqrtf( src->f[3] ); } static void micro_sub( union tgsi_exec_channel *dst, const union tgsi_exec_channel *src0, const union tgsi_exec_channel *src1 ) { dst->f[0] = src0->f[0] - src1->f[0]; dst->f[1] = src0->f[1] - src1->f[1]; dst->f[2] = src0->f[2] - src1->f[2]; dst->f[3] = src0->f[3] - src1->f[3]; } static void fetch_src_file_channel( const struct tgsi_exec_machine *mach, const uint file, const uint swizzle, const union tgsi_exec_channel *index, union tgsi_exec_channel *chan ) { switch( swizzle ) { case TGSI_SWIZZLE_X: case TGSI_SWIZZLE_Y: case TGSI_SWIZZLE_Z: case TGSI_SWIZZLE_W: switch( file ) { case TGSI_FILE_CONSTANT: assert(mach->Consts); if (index->i[0] < 0) chan->f[0] = 0.0f; else chan->f[0] = mach->Consts[index->i[0]][swizzle]; if (index->i[1] < 0) chan->f[1] = 0.0f; else chan->f[1] = mach->Consts[index->i[1]][swizzle]; if (index->i[2] < 0) chan->f[2] = 0.0f; else chan->f[2] = mach->Consts[index->i[2]][swizzle]; if (index->i[3] < 0) chan->f[3] = 0.0f; else chan->f[3] = mach->Consts[index->i[3]][swizzle]; break; case TGSI_FILE_INPUT: case TGSI_FILE_SYSTEM_VALUE: chan->u[0] = mach->Inputs[index->i[0]].xyzw[swizzle].u[0]; chan->u[1] = mach->Inputs[index->i[1]].xyzw[swizzle].u[1]; chan->u[2] = mach->Inputs[index->i[2]].xyzw[swizzle].u[2]; chan->u[3] = mach->Inputs[index->i[3]].xyzw[swizzle].u[3]; break; case TGSI_FILE_TEMPORARY: assert(index->i[0] < TGSI_EXEC_NUM_TEMPS); chan->u[0] = mach->Temps[index->i[0]].xyzw[swizzle].u[0]; chan->u[1] = mach->Temps[index->i[1]].xyzw[swizzle].u[1]; chan->u[2] = mach->Temps[index->i[2]].xyzw[swizzle].u[2]; chan->u[3] = mach->Temps[index->i[3]].xyzw[swizzle].u[3]; break; case TGSI_FILE_IMMEDIATE: assert( index->i[0] < (int) mach->ImmLimit ); chan->f[0] = mach->Imms[index->i[0]][swizzle]; assert( index->i[1] < (int) mach->ImmLimit ); chan->f[1] = mach->Imms[index->i[1]][swizzle]; assert( index->i[2] < (int) mach->ImmLimit ); chan->f[2] = mach->Imms[index->i[2]][swizzle]; assert( index->i[3] < (int) mach->ImmLimit ); chan->f[3] = mach->Imms[index->i[3]][swizzle]; break; case TGSI_FILE_ADDRESS: chan->u[0] = mach->Addrs[index->i[0]].xyzw[swizzle].u[0]; chan->u[1] = mach->Addrs[index->i[1]].xyzw[swizzle].u[1]; chan->u[2] = mach->Addrs[index->i[2]].xyzw[swizzle].u[2]; chan->u[3] = mach->Addrs[index->i[3]].xyzw[swizzle].u[3]; break; case TGSI_FILE_PREDICATE: assert(index->i[0] < TGSI_EXEC_NUM_PREDS); assert(index->i[1] < TGSI_EXEC_NUM_PREDS); assert(index->i[2] < TGSI_EXEC_NUM_PREDS); assert(index->i[3] < TGSI_EXEC_NUM_PREDS); chan->u[0] = mach->Predicates[0].xyzw[swizzle].u[0]; chan->u[1] = mach->Predicates[0].xyzw[swizzle].u[1]; chan->u[2] = mach->Predicates[0].xyzw[swizzle].u[2]; chan->u[3] = mach->Predicates[0].xyzw[swizzle].u[3]; break; case TGSI_FILE_OUTPUT: /* vertex/fragment output vars can be read too */ chan->u[0] = mach->Outputs[index->i[0]].xyzw[swizzle].u[0]; chan->u[1] = mach->Outputs[index->i[1]].xyzw[swizzle].u[1]; chan->u[2] = mach->Outputs[index->i[2]].xyzw[swizzle].u[2]; chan->u[3] = mach->Outputs[index->i[3]].xyzw[swizzle].u[3]; break; default: assert( 0 ); } break; default: assert( 0 ); } } static void fetch_source(const struct tgsi_exec_machine *mach, union tgsi_exec_channel *chan, const struct tgsi_full_src_register *reg, const uint chan_index, enum tgsi_exec_datatype src_datatype) { union tgsi_exec_channel index; uint swizzle; /* We start with a direct index into a register file. * * file[1], * where: * file = Register.File * [1] = Register.Index */ index.i[0] = index.i[1] = index.i[2] = index.i[3] = reg->Register.Index; /* There is an extra source register that indirectly subscripts * a register file. The direct index now becomes an offset * that is being added to the indirect register. * * file[ind[2].x+1], * where: * ind = Indirect.File * [2] = Indirect.Index * .x = Indirect.SwizzleX */ if (reg->Register.Indirect) { union tgsi_exec_channel index2; union tgsi_exec_channel indir_index; const uint execmask = mach->ExecMask; uint i; /* which address register (always zero now) */ index2.i[0] = index2.i[1] = index2.i[2] = index2.i[3] = reg->Indirect.Index; /* get current value of address register[swizzle] */ swizzle = tgsi_util_get_src_register_swizzle( ®->Indirect, CHAN_X ); fetch_src_file_channel( mach, reg->Indirect.File, swizzle, &index2, &indir_index ); /* add value of address register to the offset */ index.i[0] += (int) indir_index.f[0]; index.i[1] += (int) indir_index.f[1]; index.i[2] += (int) indir_index.f[2]; index.i[3] += (int) indir_index.f[3]; /* for disabled execution channels, zero-out the index to * avoid using a potential garbage value. */ for (i = 0; i < QUAD_SIZE; i++) { if ((execmask & (1 << i)) == 0) index.i[i] = 0; } } /* There is an extra source register that is a second * subscript to a register file. Effectively it means that * the register file is actually a 2D array of registers. * * file[1][3] == file[1*sizeof(file[1])+3], * where: * [3] = Dimension.Index */ if (reg->Register.Dimension) { /* The size of the first-order array depends on the register file type. * We need to multiply the index to the first array to get an effective, * "flat" index that points to the beginning of the second-order array. */ switch (reg->Register.File) { case TGSI_FILE_INPUT: case TGSI_FILE_SYSTEM_VALUE: index.i[0] *= TGSI_EXEC_MAX_INPUT_ATTRIBS; index.i[1] *= TGSI_EXEC_MAX_INPUT_ATTRIBS; index.i[2] *= TGSI_EXEC_MAX_INPUT_ATTRIBS; index.i[3] *= TGSI_EXEC_MAX_INPUT_ATTRIBS; break; case TGSI_FILE_CONSTANT: index.i[0] *= TGSI_EXEC_MAX_CONST_BUFFER; index.i[1] *= TGSI_EXEC_MAX_CONST_BUFFER; index.i[2] *= TGSI_EXEC_MAX_CONST_BUFFER; index.i[3] *= TGSI_EXEC_MAX_CONST_BUFFER; break; default: assert( 0 ); } index.i[0] += reg->Dimension.Index; index.i[1] += reg->Dimension.Index; index.i[2] += reg->Dimension.Index; index.i[3] += reg->Dimension.Index; /* Again, the second subscript index can be addressed indirectly * identically to the first one. * Nothing stops us from indirectly addressing the indirect register, * but there is no need for that, so we won't exercise it. * * file[1][ind[4].y+3], * where: * ind = DimIndirect.File * [4] = DimIndirect.Index * .y = DimIndirect.SwizzleX */ if (reg->Dimension.Indirect) { union tgsi_exec_channel index2; union tgsi_exec_channel indir_index; const uint execmask = mach->ExecMask; uint i; index2.i[0] = index2.i[1] = index2.i[2] = index2.i[3] = reg->DimIndirect.Index; swizzle = tgsi_util_get_src_register_swizzle( ®->DimIndirect, CHAN_X ); fetch_src_file_channel( mach, reg->DimIndirect.File, swizzle, &index2, &indir_index ); index.i[0] += (int) indir_index.f[0]; index.i[1] += (int) indir_index.f[1]; index.i[2] += (int) indir_index.f[2]; index.i[3] += (int) indir_index.f[3]; /* for disabled execution channels, zero-out the index to * avoid using a potential garbage value. */ for (i = 0; i < QUAD_SIZE; i++) { if ((execmask & (1 << i)) == 0) index.i[i] = 0; } } /* If by any chance there was a need for a 3D array of register * files, we would have to check whether Dimension is followed * by a dimension register and continue the saga. */ } swizzle = tgsi_util_get_full_src_register_swizzle( reg, chan_index ); fetch_src_file_channel( mach, reg->Register.File, swizzle, &index, chan ); if (reg->Register.Absolute) { if (src_datatype == TGSI_EXEC_DATA_FLOAT) { micro_abs(chan, chan); } else { micro_iabs(chan, chan); } } if (reg->Register.Negate) { if (src_datatype == TGSI_EXEC_DATA_FLOAT) { micro_neg(chan, chan); } else { micro_ineg(chan, chan); } } } static void store_dest(struct tgsi_exec_machine *mach, const union tgsi_exec_channel *chan, const struct tgsi_full_dst_register *reg, const struct tgsi_full_instruction *inst, uint chan_index, enum tgsi_exec_datatype dst_datatype) { uint i; union tgsi_exec_channel null; union tgsi_exec_channel *dst; uint execmask = mach->ExecMask; int offset = 0; /* indirection offset */ int index; if (dst_datatype == TGSI_EXEC_DATA_FLOAT) { CHECK_INF_OR_NAN(chan); } /* There is an extra source register that indirectly subscripts * a register file. The direct index now becomes an offset * that is being added to the indirect register. * * file[ind[2].x+1], * where: * ind = Indirect.File * [2] = Indirect.Index * .x = Indirect.SwizzleX */ if (reg->Register.Indirect) { union tgsi_exec_channel index; union tgsi_exec_channel indir_index; uint swizzle; /* which address register (always zero for now) */ index.i[0] = index.i[1] = index.i[2] = index.i[3] = reg->Indirect.Index; /* get current value of address register[swizzle] */ swizzle = tgsi_util_get_src_register_swizzle( ®->Indirect, CHAN_X ); /* fetch values from the address/indirection register */ fetch_src_file_channel( mach, reg->Indirect.File, swizzle, &index, &indir_index ); /* save indirection offset */ offset = (int) indir_index.f[0]; } switch (reg->Register.File) { case TGSI_FILE_NULL: dst = &null; break; case TGSI_FILE_OUTPUT: index = mach->Temps[TEMP_OUTPUT_I].xyzw[TEMP_OUTPUT_C].u[0] + reg->Register.Index; dst = &mach->Outputs[offset + index].xyzw[chan_index]; #if 0 if (TGSI_PROCESSOR_GEOMETRY == mach->Processor) { fprintf(stderr, "STORING OUT[%d] mask(%d), = (", offset + index, execmask); for (i = 0; i < QUAD_SIZE; i++) if (execmask & (1 << i)) fprintf(stderr, "%f, ", chan->f[i]); fprintf(stderr, ")\n"); } #endif break; case TGSI_FILE_TEMPORARY: index = reg->Register.Index; assert( index < TGSI_EXEC_NUM_TEMPS ); dst = &mach->Temps[offset + index].xyzw[chan_index]; break; case TGSI_FILE_ADDRESS: index = reg->Register.Index; dst = &mach->Addrs[index].xyzw[chan_index]; break; case TGSI_FILE_LOOP: assert(reg->Register.Index == 0); assert(mach->LoopCounterStackTop > 0); assert(chan_index == CHAN_X); dst = &mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[chan_index]; break; case TGSI_FILE_PREDICATE: index = reg->Register.Index; assert(index < TGSI_EXEC_NUM_PREDS); dst = &mach->Predicates[index].xyzw[chan_index]; break; default: assert( 0 ); return; } if (inst->Instruction.Predicate) { uint swizzle; union tgsi_exec_channel *pred; switch (chan_index) { case CHAN_X: swizzle = inst->Predicate.SwizzleX; break; case CHAN_Y: swizzle = inst->Predicate.SwizzleY; break; case CHAN_Z: swizzle = inst->Predicate.SwizzleZ; break; case CHAN_W: swizzle = inst->Predicate.SwizzleW; break; default: assert(0); return; } assert(inst->Predicate.Index == 0); pred = &mach->Predicates[inst->Predicate.Index].xyzw[swizzle]; if (inst->Predicate.Negate) { for (i = 0; i < QUAD_SIZE; i++) { if (pred->u[i]) { execmask &= ~(1 << i); } } } else { for (i = 0; i < QUAD_SIZE; i++) { if (!pred->u[i]) { execmask &= ~(1 << i); } } } } switch (inst->Instruction.Saturate) { case TGSI_SAT_NONE: for (i = 0; i < QUAD_SIZE; i++) if (execmask & (1 << i)) dst->i[i] = chan->i[i]; break; case TGSI_SAT_ZERO_ONE: for (i = 0; i < QUAD_SIZE; i++) if (execmask & (1 << i)) { if (chan->f[i] < 0.0f) dst->f[i] = 0.0f; else if (chan->f[i] > 1.0f) dst->f[i] = 1.0f; else dst->i[i] = chan->i[i]; } break; case TGSI_SAT_MINUS_PLUS_ONE: for (i = 0; i < QUAD_SIZE; i++) if (execmask & (1 << i)) { if (chan->f[i] < -1.0f) dst->f[i] = -1.0f; else if (chan->f[i] > 1.0f) dst->f[i] = 1.0f; else dst->i[i] = chan->i[i]; } break; default: assert( 0 ); } } #define FETCH(VAL,INDEX,CHAN)\ fetch_source(mach, VAL, &inst->Src[INDEX], CHAN, TGSI_EXEC_DATA_FLOAT) #define STORE(VAL,INDEX,CHAN)\ store_dest(mach, VAL, &inst->Dst[INDEX], inst, CHAN, TGSI_EXEC_DATA_FLOAT) /** * Execute ARB-style KIL which is predicated by a src register. * Kill fragment if any of the four values is less than zero. */ static void exec_kil(struct tgsi_exec_machine *mach, const struct tgsi_full_instruction *inst) { uint uniquemask; uint chan_index; uint kilmask = 0; /* bit 0 = pixel 0, bit 1 = pixel 1, etc */ union tgsi_exec_channel r[1]; /* This mask stores component bits that were already tested. */ uniquemask = 0; for (chan_index = 0; chan_index < 4; chan_index++) { uint swizzle; uint i; /* unswizzle channel */ swizzle = tgsi_util_get_full_src_register_swizzle ( &inst->Src[0], chan_index); /* check if the component has not been already tested */ if (uniquemask & (1 << swizzle)) continue; uniquemask |= 1 << swizzle; FETCH(&r[0], 0, chan_index); for (i = 0; i < 4; i++) if (r[0].f[i] < 0.0f) kilmask |= 1 << i; } mach->Temps[TEMP_KILMASK_I].xyzw[TEMP_KILMASK_C].u[0] |= kilmask; } /** * Execute NVIDIA-style KIL which is predicated by a condition code. * Kill fragment if the condition code is TRUE. */ static void exec_kilp(struct tgsi_exec_machine *mach, const struct tgsi_full_instruction *inst) { uint kilmask; /* bit 0 = pixel 0, bit 1 = pixel 1, etc */ /* "unconditional" kil */ kilmask = mach->ExecMask; mach->Temps[TEMP_KILMASK_I].xyzw[TEMP_KILMASK_C].u[0] |= kilmask; } static void emit_vertex(struct tgsi_exec_machine *mach) { /* FIXME: check for exec mask correctly unsigned i; for (i = 0; i < QUAD_SIZE; ++i) { if ((mach->ExecMask & (1 << i))) */ if (mach->ExecMask) { mach->Temps[TEMP_OUTPUT_I].xyzw[TEMP_OUTPUT_C].u[0] += mach->NumOutputs; mach->Primitives[mach->Temps[TEMP_PRIMITIVE_I].xyzw[TEMP_PRIMITIVE_C].u[0]]++; } } static void emit_primitive(struct tgsi_exec_machine *mach) { unsigned *prim_count = &mach->Temps[TEMP_PRIMITIVE_I].xyzw[TEMP_PRIMITIVE_C].u[0]; /* FIXME: check for exec mask correctly unsigned i; for (i = 0; i < QUAD_SIZE; ++i) { if ((mach->ExecMask & (1 << i))) */ if (mach->ExecMask) { ++(*prim_count); debug_assert((*prim_count * mach->NumOutputs) < mach->MaxGeometryShaderOutputs); mach->Primitives[*prim_count] = 0; } } /* * Fetch a four texture samples using STR texture coordinates. */ static void fetch_texel( struct tgsi_sampler *sampler, const union tgsi_exec_channel *s, const union tgsi_exec_channel *t, const union tgsi_exec_channel *p, float lodbias, /* XXX should be float[4] */ union tgsi_exec_channel *r, union tgsi_exec_channel *g, union tgsi_exec_channel *b, union tgsi_exec_channel *a ) { uint j; float rgba[NUM_CHANNELS][QUAD_SIZE]; sampler->get_samples(sampler, s->f, t->f, p->f, lodbias, rgba); for (j = 0; j < 4; j++) { r->f[j] = rgba[0][j]; g->f[j] = rgba[1][j]; b->f[j] = rgba[2][j]; a->f[j] = rgba[3][j]; } } static void exec_tex(struct tgsi_exec_machine *mach, const struct tgsi_full_instruction *inst, boolean biasLod, boolean projected) { const uint unit = inst->Src[1].Register.Index; union tgsi_exec_channel r[4]; uint chan_index; float lodBias; /* debug_printf("Sampler %u unit %u\n", sampler, unit); */ switch (inst->Texture.Texture) { case TGSI_TEXTURE_1D: case TGSI_TEXTURE_SHADOW1D: FETCH(&r[0], 0, CHAN_X); if (projected) { FETCH(&r[1], 0, CHAN_W); micro_div( &r[0], &r[0], &r[1] ); } if (biasLod) { FETCH(&r[1], 0, CHAN_W); lodBias = r[2].f[0]; } else lodBias = 0.0; fetch_texel(mach->Samplers[unit], &r[0], &ZeroVec, &ZeroVec, lodBias, /* S, T, P, BIAS */ &r[0], &r[1], &r[2], &r[3]); /* R, G, B, A */ break; case TGSI_TEXTURE_2D: case TGSI_TEXTURE_RECT: case TGSI_TEXTURE_SHADOW2D: case TGSI_TEXTURE_SHADOWRECT: FETCH(&r[0], 0, CHAN_X); FETCH(&r[1], 0, CHAN_Y); FETCH(&r[2], 0, CHAN_Z); if (projected) { FETCH(&r[3], 0, CHAN_W); micro_div( &r[0], &r[0], &r[3] ); micro_div( &r[1], &r[1], &r[3] ); micro_div( &r[2], &r[2], &r[3] ); } if (biasLod) { FETCH(&r[3], 0, CHAN_W); lodBias = r[3].f[0]; } else lodBias = 0.0; fetch_texel(mach->Samplers[unit], &r[0], &r[1], &r[2], lodBias, /* inputs */ &r[0], &r[1], &r[2], &r[3]); /* outputs */ break; case TGSI_TEXTURE_3D: case TGSI_TEXTURE_CUBE: FETCH(&r[0], 0, CHAN_X); FETCH(&r[1], 0, CHAN_Y); FETCH(&r[2], 0, CHAN_Z); if (projected) { FETCH(&r[3], 0, CHAN_W); micro_div( &r[0], &r[0], &r[3] ); micro_div( &r[1], &r[1], &r[3] ); micro_div( &r[2], &r[2], &r[3] ); } if (biasLod) { FETCH(&r[3], 0, CHAN_W); lodBias = r[3].f[0]; } else lodBias = 0.0; fetch_texel(mach->Samplers[unit], &r[0], &r[1], &r[2], lodBias, &r[0], &r[1], &r[2], &r[3]); break; default: assert (0); } FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[chan_index], 0, chan_index ); } } static void exec_txd(struct tgsi_exec_machine *mach, const struct tgsi_full_instruction *inst) { const uint unit = inst->Src[3].Register.Index; union tgsi_exec_channel r[4]; uint chan_index; /* * XXX: This is fake TXD -- the derivatives are not taken into account, yet. */ switch (inst->Texture.Texture) { case TGSI_TEXTURE_1D: case TGSI_TEXTURE_SHADOW1D: FETCH(&r[0], 0, CHAN_X); fetch_texel(mach->Samplers[unit], &r[0], &ZeroVec, &ZeroVec, 0.0f, /* S, T, P, BIAS */ &r[0], &r[1], &r[2], &r[3]); /* R, G, B, A */ break; case TGSI_TEXTURE_2D: case TGSI_TEXTURE_RECT: case TGSI_TEXTURE_SHADOW2D: case TGSI_TEXTURE_SHADOWRECT: FETCH(&r[0], 0, CHAN_X); FETCH(&r[1], 0, CHAN_Y); FETCH(&r[2], 0, CHAN_Z); fetch_texel(mach->Samplers[unit], &r[0], &r[1], &r[2], 0.0f, /* inputs */ &r[0], &r[1], &r[2], &r[3]); /* outputs */ break; case TGSI_TEXTURE_3D: case TGSI_TEXTURE_CUBE: FETCH(&r[0], 0, CHAN_X); FETCH(&r[1], 0, CHAN_Y); FETCH(&r[2], 0, CHAN_Z); fetch_texel(mach->Samplers[unit], &r[0], &r[1], &r[2], 0.0f, &r[0], &r[1], &r[2], &r[3]); break; default: assert(0); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&r[chan_index], 0, chan_index); } } /** * Evaluate a constant-valued coefficient at the position of the * current quad. */ static void eval_constant_coef( struct tgsi_exec_machine *mach, unsigned attrib, unsigned chan ) { unsigned i; for( i = 0; i < QUAD_SIZE; i++ ) { mach->Inputs[attrib].xyzw[chan].f[i] = mach->InterpCoefs[attrib].a0[chan]; } } /** * Evaluate a linear-valued coefficient at the position of the * current quad. */ static void eval_linear_coef( struct tgsi_exec_machine *mach, unsigned attrib, unsigned chan ) { const float x = mach->QuadPos.xyzw[0].f[0]; const float y = mach->QuadPos.xyzw[1].f[0]; const float dadx = mach->InterpCoefs[attrib].dadx[chan]; const float dady = mach->InterpCoefs[attrib].dady[chan]; const float a0 = mach->InterpCoefs[attrib].a0[chan] + dadx * x + dady * y; mach->Inputs[attrib].xyzw[chan].f[0] = a0; mach->Inputs[attrib].xyzw[chan].f[1] = a0 + dadx; mach->Inputs[attrib].xyzw[chan].f[2] = a0 + dady; mach->Inputs[attrib].xyzw[chan].f[3] = a0 + dadx + dady; } /** * Evaluate a perspective-valued coefficient at the position of the * current quad. */ static void eval_perspective_coef( struct tgsi_exec_machine *mach, unsigned attrib, unsigned chan ) { const float x = mach->QuadPos.xyzw[0].f[0]; const float y = mach->QuadPos.xyzw[1].f[0]; const float dadx = mach->InterpCoefs[attrib].dadx[chan]; const float dady = mach->InterpCoefs[attrib].dady[chan]; const float a0 = mach->InterpCoefs[attrib].a0[chan] + dadx * x + dady * y; const float *w = mach->QuadPos.xyzw[3].f; /* divide by W here */ mach->Inputs[attrib].xyzw[chan].f[0] = a0 / w[0]; mach->Inputs[attrib].xyzw[chan].f[1] = (a0 + dadx) / w[1]; mach->Inputs[attrib].xyzw[chan].f[2] = (a0 + dady) / w[2]; mach->Inputs[attrib].xyzw[chan].f[3] = (a0 + dadx + dady) / w[3]; } typedef void (* eval_coef_func)( struct tgsi_exec_machine *mach, unsigned attrib, unsigned chan ); static void exec_declaration(struct tgsi_exec_machine *mach, const struct tgsi_full_declaration *decl) { if (mach->Processor == TGSI_PROCESSOR_FRAGMENT) { if (decl->Declaration.File == TGSI_FILE_INPUT || decl->Declaration.File == TGSI_FILE_SYSTEM_VALUE) { uint first, last, mask; first = decl->Range.First; last = decl->Range.Last; mask = decl->Declaration.UsageMask; if (decl->Semantic.Name == TGSI_SEMANTIC_POSITION) { assert(decl->Semantic.Index == 0); assert(first == last); assert(mask == TGSI_WRITEMASK_XYZW); mach->Inputs[first] = mach->QuadPos; } else if (decl->Semantic.Name == TGSI_SEMANTIC_FACE) { uint i; assert(decl->Semantic.Index == 0); assert(first == last); for (i = 0; i < QUAD_SIZE; i++) { mach->Inputs[first].xyzw[0].f[i] = mach->Face; } } else { eval_coef_func eval; uint i, j; switch (decl->Declaration.Interpolate) { case TGSI_INTERPOLATE_CONSTANT: eval = eval_constant_coef; break; case TGSI_INTERPOLATE_LINEAR: eval = eval_linear_coef; break; case TGSI_INTERPOLATE_PERSPECTIVE: eval = eval_perspective_coef; break; default: assert(0); return; } for (j = 0; j < NUM_CHANNELS; j++) { if (mask & (1 << j)) { for (i = first; i <= last; i++) { eval(mach, i, j); } } } } } } } typedef void (* micro_op)(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src); static void exec_vector_unary(struct tgsi_exec_machine *mach, const struct tgsi_full_instruction *inst, micro_op op, enum tgsi_exec_datatype dst_datatype, enum tgsi_exec_datatype src_datatype) { unsigned int chan; struct tgsi_exec_vector dst; for (chan = 0; chan < NUM_CHANNELS; chan++) { if (inst->Dst[0].Register.WriteMask & (1 << chan)) { union tgsi_exec_channel src; fetch_source(mach, &src, &inst->Src[0], chan, src_datatype); op(&dst.xyzw[chan], &src); } } for (chan = 0; chan < NUM_CHANNELS; chan++) { if (inst->Dst[0].Register.WriteMask & (1 << chan)) { store_dest(mach, &dst.xyzw[chan], &inst->Dst[0], inst, chan, dst_datatype); } } } static void exec_vector_binary(struct tgsi_exec_machine *mach, const struct tgsi_full_instruction *inst, micro_op op, enum tgsi_exec_datatype dst_datatype, enum tgsi_exec_datatype src_datatype) { unsigned int chan; struct tgsi_exec_vector dst; for (chan = 0; chan < NUM_CHANNELS; chan++) { if (inst->Dst[0].Register.WriteMask & (1 << chan)) { union tgsi_exec_channel src[2]; fetch_source(mach, &src[0], &inst->Src[0], chan, src_datatype); fetch_source(mach, &src[1], &inst->Src[1], chan, src_datatype); op(&dst.xyzw[chan], src); } } for (chan = 0; chan < NUM_CHANNELS; chan++) { if (inst->Dst[0].Register.WriteMask & (1 << chan)) { store_dest(mach, &dst.xyzw[chan], &inst->Dst[0], inst, chan, dst_datatype); } } } static void exec_vector_trinary(struct tgsi_exec_machine *mach, const struct tgsi_full_instruction *inst, micro_op op, enum tgsi_exec_datatype dst_datatype, enum tgsi_exec_datatype src_datatype) { unsigned int chan; struct tgsi_exec_vector dst; for (chan = 0; chan < NUM_CHANNELS; chan++) { if (inst->Dst[0].Register.WriteMask & (1 << chan)) { union tgsi_exec_channel src[3]; fetch_source(mach, &src[0], &inst->Src[0], chan, src_datatype); fetch_source(mach, &src[1], &inst->Src[1], chan, src_datatype); fetch_source(mach, &src[2], &inst->Src[2], chan, src_datatype); op(&dst.xyzw[chan], src); } } for (chan = 0; chan < NUM_CHANNELS; chan++) { if (inst->Dst[0].Register.WriteMask & (1 << chan)) { store_dest(mach, &dst.xyzw[chan], &inst->Dst[0], inst, chan, dst_datatype); } } } static void exec_break(struct tgsi_exec_machine *mach) { if (mach->BreakType == TGSI_EXEC_BREAK_INSIDE_LOOP) { /* turn off loop channels for each enabled exec channel */ mach->LoopMask &= ~mach->ExecMask; /* Todo: if mach->LoopMask == 0, jump to end of loop */ UPDATE_EXEC_MASK(mach); } else { assert(mach->BreakType == TGSI_EXEC_BREAK_INSIDE_SWITCH); mach->Switch.mask = 0x0; UPDATE_EXEC_MASK(mach); } } static void exec_switch(struct tgsi_exec_machine *mach, const struct tgsi_full_instruction *inst) { assert(mach->SwitchStackTop < TGSI_EXEC_MAX_SWITCH_NESTING); assert(mach->BreakStackTop < TGSI_EXEC_MAX_BREAK_STACK); mach->SwitchStack[mach->SwitchStackTop++] = mach->Switch; fetch_source(mach, &mach->Switch.selector, &inst->Src[0], CHAN_X, TGSI_EXEC_DATA_UINT); mach->Switch.mask = 0x0; mach->Switch.defaultMask = 0x0; mach->BreakStack[mach->BreakStackTop++] = mach->BreakType; mach->BreakType = TGSI_EXEC_BREAK_INSIDE_SWITCH; UPDATE_EXEC_MASK(mach); } static void exec_case(struct tgsi_exec_machine *mach, const struct tgsi_full_instruction *inst) { uint prevMask = mach->SwitchStack[mach->SwitchStackTop - 1].mask; union tgsi_exec_channel src; uint mask = 0; fetch_source(mach, &src, &inst->Src[0], CHAN_X, TGSI_EXEC_DATA_UINT); if (mach->Switch.selector.u[0] == src.u[0]) { mask |= 0x1; } if (mach->Switch.selector.u[1] == src.u[1]) { mask |= 0x2; } if (mach->Switch.selector.u[2] == src.u[2]) { mask |= 0x4; } if (mach->Switch.selector.u[3] == src.u[3]) { mask |= 0x8; } mach->Switch.defaultMask |= mask; mach->Switch.mask |= mask & prevMask; UPDATE_EXEC_MASK(mach); } static void exec_default(struct tgsi_exec_machine *mach) { uint prevMask = mach->SwitchStack[mach->SwitchStackTop - 1].mask; mach->Switch.mask |= ~mach->Switch.defaultMask & prevMask; UPDATE_EXEC_MASK(mach); } static void exec_endswitch(struct tgsi_exec_machine *mach) { mach->Switch = mach->SwitchStack[--mach->SwitchStackTop]; mach->BreakType = mach->BreakStack[--mach->BreakStackTop]; UPDATE_EXEC_MASK(mach); } static void micro_i2f(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->f[0] = (float)src->i[0]; dst->f[1] = (float)src->i[1]; dst->f[2] = (float)src->i[2]; dst->f[3] = (float)src->i[3]; } static void micro_not(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = ~src->u[0]; dst->u[1] = ~src->u[1]; dst->u[2] = ~src->u[2]; dst->u[3] = ~src->u[3]; } static void micro_shl(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] << src[1].u[0]; dst->u[1] = src[0].u[1] << src[1].u[1]; dst->u[2] = src[0].u[2] << src[1].u[2]; dst->u[3] = src[0].u[3] << src[1].u[3]; } static void micro_and(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] & src[1].u[0]; dst->u[1] = src[0].u[1] & src[1].u[1]; dst->u[2] = src[0].u[2] & src[1].u[2]; dst->u[3] = src[0].u[3] & src[1].u[3]; } static void micro_or(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] | src[1].u[0]; dst->u[1] = src[0].u[1] | src[1].u[1]; dst->u[2] = src[0].u[2] | src[1].u[2]; dst->u[3] = src[0].u[3] | src[1].u[3]; } static void micro_xor(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] ^ src[1].u[0]; dst->u[1] = src[0].u[1] ^ src[1].u[1]; dst->u[2] = src[0].u[2] ^ src[1].u[2]; dst->u[3] = src[0].u[3] ^ src[1].u[3]; } static void micro_f2i(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->i[0] = (int)src->f[0]; dst->i[1] = (int)src->f[1]; dst->i[2] = (int)src->f[2]; dst->i[3] = (int)src->f[3]; } static void micro_idiv(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->i[0] = src[0].i[0] / src[1].i[0]; dst->i[1] = src[0].i[1] / src[1].i[1]; dst->i[2] = src[0].i[2] / src[1].i[2]; dst->i[3] = src[0].i[3] / src[1].i[3]; } static void micro_imax(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->i[0] = src[0].i[0] > src[1].i[0] ? src[0].i[0] : src[1].i[0]; dst->i[1] = src[0].i[1] > src[1].i[1] ? src[0].i[1] : src[1].i[1]; dst->i[2] = src[0].i[2] > src[1].i[2] ? src[0].i[2] : src[1].i[2]; dst->i[3] = src[0].i[3] > src[1].i[3] ? src[0].i[3] : src[1].i[3]; } static void micro_imin(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->i[0] = src[0].i[0] < src[1].i[0] ? src[0].i[0] : src[1].i[0]; dst->i[1] = src[0].i[1] < src[1].i[1] ? src[0].i[1] : src[1].i[1]; dst->i[2] = src[0].i[2] < src[1].i[2] ? src[0].i[2] : src[1].i[2]; dst->i[3] = src[0].i[3] < src[1].i[3] ? src[0].i[3] : src[1].i[3]; } static void micro_isge(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->i[0] = src[0].i[0] >= src[1].i[0] ? -1 : 0; dst->i[1] = src[0].i[1] >= src[1].i[1] ? -1 : 0; dst->i[2] = src[0].i[2] >= src[1].i[2] ? -1 : 0; dst->i[3] = src[0].i[3] >= src[1].i[3] ? -1 : 0; } static void micro_ishr(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->i[0] = src[0].i[0] >> src[1].i[0]; dst->i[1] = src[0].i[1] >> src[1].i[1]; dst->i[2] = src[0].i[2] >> src[1].i[2]; dst->i[3] = src[0].i[3] >> src[1].i[3]; } static void micro_islt(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->i[0] = src[0].i[0] < src[1].i[0] ? -1 : 0; dst->i[1] = src[0].i[1] < src[1].i[1] ? -1 : 0; dst->i[2] = src[0].i[2] < src[1].i[2] ? -1 : 0; dst->i[3] = src[0].i[3] < src[1].i[3] ? -1 : 0; } static void micro_f2u(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = (uint)src->f[0]; dst->u[1] = (uint)src->f[1]; dst->u[2] = (uint)src->f[2]; dst->u[3] = (uint)src->f[3]; } static void micro_u2f(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->f[0] = (float)src->u[0]; dst->f[1] = (float)src->u[1]; dst->f[2] = (float)src->u[2]; dst->f[3] = (float)src->u[3]; } static void micro_uadd(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] + src[1].u[0]; dst->u[1] = src[0].u[1] + src[1].u[1]; dst->u[2] = src[0].u[2] + src[1].u[2]; dst->u[3] = src[0].u[3] + src[1].u[3]; } static void micro_udiv(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] / src[1].u[0]; dst->u[1] = src[0].u[1] / src[1].u[1]; dst->u[2] = src[0].u[2] / src[1].u[2]; dst->u[3] = src[0].u[3] / src[1].u[3]; } static void micro_umad(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] * src[1].u[0] + src[2].u[0]; dst->u[1] = src[0].u[1] * src[1].u[1] + src[2].u[1]; dst->u[2] = src[0].u[2] * src[1].u[2] + src[2].u[2]; dst->u[3] = src[0].u[3] * src[1].u[3] + src[2].u[3]; } static void micro_umax(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] > src[1].u[0] ? src[0].u[0] : src[1].u[0]; dst->u[1] = src[0].u[1] > src[1].u[1] ? src[0].u[1] : src[1].u[1]; dst->u[2] = src[0].u[2] > src[1].u[2] ? src[0].u[2] : src[1].u[2]; dst->u[3] = src[0].u[3] > src[1].u[3] ? src[0].u[3] : src[1].u[3]; } static void micro_umin(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] < src[1].u[0] ? src[0].u[0] : src[1].u[0]; dst->u[1] = src[0].u[1] < src[1].u[1] ? src[0].u[1] : src[1].u[1]; dst->u[2] = src[0].u[2] < src[1].u[2] ? src[0].u[2] : src[1].u[2]; dst->u[3] = src[0].u[3] < src[1].u[3] ? src[0].u[3] : src[1].u[3]; } static void micro_umod(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] % src[1].u[0]; dst->u[1] = src[0].u[1] % src[1].u[1]; dst->u[2] = src[0].u[2] % src[1].u[2]; dst->u[3] = src[0].u[3] % src[1].u[3]; } static void micro_umul(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] * src[1].u[0]; dst->u[1] = src[0].u[1] * src[1].u[1]; dst->u[2] = src[0].u[2] * src[1].u[2]; dst->u[3] = src[0].u[3] * src[1].u[3]; } static void micro_useq(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] == src[1].u[0] ? ~0 : 0; dst->u[1] = src[0].u[1] == src[1].u[1] ? ~0 : 0; dst->u[2] = src[0].u[2] == src[1].u[2] ? ~0 : 0; dst->u[3] = src[0].u[3] == src[1].u[3] ? ~0 : 0; } static void micro_usge(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] >= src[1].u[0] ? ~0 : 0; dst->u[1] = src[0].u[1] >= src[1].u[1] ? ~0 : 0; dst->u[2] = src[0].u[2] >= src[1].u[2] ? ~0 : 0; dst->u[3] = src[0].u[3] >= src[1].u[3] ? ~0 : 0; } static void micro_ushr(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] >> src[1].u[0]; dst->u[1] = src[0].u[1] >> src[1].u[1]; dst->u[2] = src[0].u[2] >> src[1].u[2]; dst->u[3] = src[0].u[3] >> src[1].u[3]; } static void micro_uslt(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] < src[1].u[0] ? ~0 : 0; dst->u[1] = src[0].u[1] < src[1].u[1] ? ~0 : 0; dst->u[2] = src[0].u[2] < src[1].u[2] ? ~0 : 0; dst->u[3] = src[0].u[3] < src[1].u[3] ? ~0 : 0; } static void micro_usne(union tgsi_exec_channel *dst, const union tgsi_exec_channel *src) { dst->u[0] = src[0].u[0] != src[1].u[0] ? ~0 : 0; dst->u[1] = src[0].u[1] != src[1].u[1] ? ~0 : 0; dst->u[2] = src[0].u[2] != src[1].u[2] ? ~0 : 0; dst->u[3] = src[0].u[3] != src[1].u[3] ? ~0 : 0; } static void exec_instruction( struct tgsi_exec_machine *mach, const struct tgsi_full_instruction *inst, int *pc ) { uint chan_index; union tgsi_exec_channel r[10]; union tgsi_exec_channel d[8]; (*pc)++; switch (inst->Instruction.Opcode) { case TGSI_OPCODE_ARL: case TGSI_OPCODE_FLR: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); micro_flr(&d[chan_index], &r[0]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_MOV: exec_vector_unary(mach, inst, micro_mov, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_LIT: if (IS_CHANNEL_ENABLED( *inst, CHAN_Y ) || IS_CHANNEL_ENABLED( *inst, CHAN_Z )) { FETCH( &r[0], 0, CHAN_X ); if (IS_CHANNEL_ENABLED( *inst, CHAN_Y )) { micro_max(&d[CHAN_Y], &r[0], &mach->Temps[TEMP_0_I].xyzw[TEMP_0_C]); } if (IS_CHANNEL_ENABLED( *inst, CHAN_Z )) { FETCH( &r[1], 0, CHAN_Y ); micro_max( &r[1], &r[1], &mach->Temps[TEMP_0_I].xyzw[TEMP_0_C] ); FETCH( &r[2], 0, CHAN_W ); micro_min( &r[2], &r[2], &mach->Temps[TEMP_128_I].xyzw[TEMP_128_C] ); micro_max( &r[2], &r[2], &mach->Temps[TEMP_M128_I].xyzw[TEMP_M128_C] ); micro_pow( &r[1], &r[1], &r[2] ); micro_lt(&d[CHAN_Z], &mach->Temps[TEMP_0_I].xyzw[TEMP_0_C], &r[0], &r[1], &mach->Temps[TEMP_0_I].xyzw[TEMP_0_C]); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Y)) { STORE(&d[CHAN_Y], 0, CHAN_Y); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Z)) { STORE(&d[CHAN_Z], 0, CHAN_Z); } } if (IS_CHANNEL_ENABLED( *inst, CHAN_X )) { STORE( &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], 0, CHAN_X ); } if (IS_CHANNEL_ENABLED( *inst, CHAN_W )) { STORE( &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], 0, CHAN_W ); } break; case TGSI_OPCODE_RCP: /* TGSI_OPCODE_RECIP */ FETCH( &r[0], 0, CHAN_X ); micro_div( &r[0], &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], &r[0] ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[0], 0, chan_index ); } break; case TGSI_OPCODE_RSQ: /* TGSI_OPCODE_RECIPSQRT */ FETCH( &r[0], 0, CHAN_X ); micro_abs( &r[0], &r[0] ); micro_sqrt( &r[0], &r[0] ); micro_div( &r[0], &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], &r[0] ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[0], 0, chan_index ); } break; case TGSI_OPCODE_EXP: FETCH( &r[0], 0, CHAN_X ); micro_flr( &r[1], &r[0] ); /* r1 = floor(r0) */ if (IS_CHANNEL_ENABLED( *inst, CHAN_X )) { micro_exp2( &r[2], &r[1] ); /* r2 = 2 ^ r1 */ STORE( &r[2], 0, CHAN_X ); /* store r2 */ } if (IS_CHANNEL_ENABLED( *inst, CHAN_Y )) { micro_sub( &r[2], &r[0], &r[1] ); /* r2 = r0 - r1 */ STORE( &r[2], 0, CHAN_Y ); /* store r2 */ } if (IS_CHANNEL_ENABLED( *inst, CHAN_Z )) { micro_exp2( &r[2], &r[0] ); /* r2 = 2 ^ r0 */ STORE( &r[2], 0, CHAN_Z ); /* store r2 */ } if (IS_CHANNEL_ENABLED( *inst, CHAN_W )) { STORE( &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], 0, CHAN_W ); } break; case TGSI_OPCODE_LOG: FETCH( &r[0], 0, CHAN_X ); micro_abs( &r[2], &r[0] ); /* r2 = abs(r0) */ micro_lg2( &r[1], &r[2] ); /* r1 = lg2(r2) */ micro_flr( &r[0], &r[1] ); /* r0 = floor(r1) */ if (IS_CHANNEL_ENABLED( *inst, CHAN_X )) { STORE( &r[0], 0, CHAN_X ); } if (IS_CHANNEL_ENABLED( *inst, CHAN_Y )) { micro_exp2( &r[0], &r[0] ); /* r0 = 2 ^ r0 */ micro_div( &r[0], &r[2], &r[0] ); /* r0 = r2 / r0 */ STORE( &r[0], 0, CHAN_Y ); } if (IS_CHANNEL_ENABLED( *inst, CHAN_Z )) { STORE( &r[1], 0, CHAN_Z ); } if (IS_CHANNEL_ENABLED( *inst, CHAN_W )) { STORE( &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], 0, CHAN_W ); } break; case TGSI_OPCODE_MUL: FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { FETCH(&r[0], 0, chan_index); FETCH(&r[1], 1, chan_index); micro_mul(&d[chan_index], &r[0], &r[1]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_ADD: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); FETCH( &r[1], 1, chan_index ); micro_add(&d[chan_index], &r[0], &r[1]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_DP3: /* TGSI_OPCODE_DOT3 */ FETCH( &r[0], 0, CHAN_X ); FETCH( &r[1], 1, CHAN_X ); micro_mul( &r[0], &r[0], &r[1] ); FETCH( &r[1], 0, CHAN_Y ); FETCH( &r[2], 1, CHAN_Y ); micro_mul( &r[1], &r[1], &r[2] ); micro_add( &r[0], &r[0], &r[1] ); FETCH( &r[1], 0, CHAN_Z ); FETCH( &r[2], 1, CHAN_Z ); micro_mul( &r[1], &r[1], &r[2] ); micro_add( &r[0], &r[0], &r[1] ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[0], 0, chan_index ); } break; case TGSI_OPCODE_DP4: /* TGSI_OPCODE_DOT4 */ FETCH(&r[0], 0, CHAN_X); FETCH(&r[1], 1, CHAN_X); micro_mul( &r[0], &r[0], &r[1] ); FETCH(&r[1], 0, CHAN_Y); FETCH(&r[2], 1, CHAN_Y); micro_mul( &r[1], &r[1], &r[2] ); micro_add( &r[0], &r[0], &r[1] ); FETCH(&r[1], 0, CHAN_Z); FETCH(&r[2], 1, CHAN_Z); micro_mul( &r[1], &r[1], &r[2] ); micro_add( &r[0], &r[0], &r[1] ); FETCH(&r[1], 0, CHAN_W); FETCH(&r[2], 1, CHAN_W); micro_mul( &r[1], &r[1], &r[2] ); micro_add( &r[0], &r[0], &r[1] ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[0], 0, chan_index ); } break; case TGSI_OPCODE_DST: if (IS_CHANNEL_ENABLED( *inst, CHAN_Y )) { FETCH( &r[0], 0, CHAN_Y ); FETCH( &r[1], 1, CHAN_Y); micro_mul(&d[CHAN_Y], &r[0], &r[1]); } if (IS_CHANNEL_ENABLED( *inst, CHAN_Z )) { FETCH(&d[CHAN_Z], 0, CHAN_Z); } if (IS_CHANNEL_ENABLED( *inst, CHAN_W )) { FETCH(&d[CHAN_W], 1, CHAN_W); } if (IS_CHANNEL_ENABLED(*inst, CHAN_X)) { STORE(&mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], 0, CHAN_X); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Y)) { STORE(&d[CHAN_Y], 0, CHAN_Y); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Z)) { STORE(&d[CHAN_Z], 0, CHAN_Z); } if (IS_CHANNEL_ENABLED(*inst, CHAN_W)) { STORE(&d[CHAN_W], 0, CHAN_W); } break; case TGSI_OPCODE_MIN: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH(&r[0], 0, chan_index); FETCH(&r[1], 1, chan_index); /* XXX use micro_min()?? */ micro_lt(&d[chan_index], &r[0], &r[1], &r[0], &r[1]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_MAX: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH(&r[0], 0, chan_index); FETCH(&r[1], 1, chan_index); /* XXX use micro_max()?? */ micro_lt(&d[chan_index], &r[0], &r[1], &r[1], &r[0] ); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_SLT: /* TGSI_OPCODE_SETLT */ FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); FETCH( &r[1], 1, chan_index ); micro_lt(&d[chan_index], &r[0], &r[1], &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], &mach->Temps[TEMP_0_I].xyzw[TEMP_0_C]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_SGE: /* TGSI_OPCODE_SETGE */ FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); FETCH( &r[1], 1, chan_index ); micro_le(&d[chan_index], &r[1], &r[0], &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], &mach->Temps[TEMP_0_I].xyzw[TEMP_0_C]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_MAD: /* TGSI_OPCODE_MADD */ FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); FETCH( &r[1], 1, chan_index ); micro_mul( &r[0], &r[0], &r[1] ); FETCH( &r[1], 2, chan_index ); micro_add(&d[chan_index], &r[0], &r[1]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_SUB: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH(&r[0], 0, chan_index); FETCH(&r[1], 1, chan_index); micro_sub(&d[chan_index], &r[0], &r[1]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_LRP: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH(&r[0], 0, chan_index); FETCH(&r[1], 1, chan_index); FETCH(&r[2], 2, chan_index); micro_sub( &r[1], &r[1], &r[2] ); micro_mul( &r[0], &r[0], &r[1] ); micro_add(&d[chan_index], &r[0], &r[2]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_CND: FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { FETCH(&r[0], 0, chan_index); FETCH(&r[1], 1, chan_index); FETCH(&r[2], 2, chan_index); micro_lt(&d[chan_index], &mach->Temps[TEMP_HALF_I].xyzw[TEMP_HALF_C], &r[2], &r[0], &r[1]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_DP2A: FETCH( &r[0], 0, CHAN_X ); FETCH( &r[1], 1, CHAN_X ); micro_mul( &r[0], &r[0], &r[1] ); FETCH( &r[1], 0, CHAN_Y ); FETCH( &r[2], 1, CHAN_Y ); micro_mul( &r[1], &r[1], &r[2] ); micro_add( &r[0], &r[0], &r[1] ); FETCH( &r[2], 2, CHAN_X ); micro_add( &r[0], &r[0], &r[2] ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[0], 0, chan_index ); } break; case TGSI_OPCODE_FRC: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); micro_frc(&d[chan_index], &r[0]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_CLAMP: FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { FETCH(&r[0], 0, chan_index); FETCH(&r[1], 1, chan_index); micro_max(&r[0], &r[0], &r[1]); FETCH(&r[1], 2, chan_index); micro_min(&d[chan_index], &r[0], &r[1]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_ROUND: case TGSI_OPCODE_ARR: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); micro_rnd(&d[chan_index], &r[0]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_EX2: FETCH(&r[0], 0, CHAN_X); micro_exp2( &r[0], &r[0] ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[0], 0, chan_index ); } break; case TGSI_OPCODE_LG2: FETCH( &r[0], 0, CHAN_X ); micro_lg2( &r[0], &r[0] ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[0], 0, chan_index ); } break; case TGSI_OPCODE_POW: FETCH(&r[0], 0, CHAN_X); FETCH(&r[1], 1, CHAN_X); micro_pow( &r[0], &r[0], &r[1] ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[0], 0, chan_index ); } break; case TGSI_OPCODE_XPD: FETCH(&r[0], 0, CHAN_Y); FETCH(&r[1], 1, CHAN_Z); micro_mul( &r[2], &r[0], &r[1] ); FETCH(&r[3], 0, CHAN_Z); FETCH(&r[4], 1, CHAN_Y); micro_mul( &r[5], &r[3], &r[4] ); micro_sub(&d[CHAN_X], &r[2], &r[5]); FETCH(&r[2], 1, CHAN_X); micro_mul( &r[3], &r[3], &r[2] ); FETCH(&r[5], 0, CHAN_X); micro_mul( &r[1], &r[1], &r[5] ); micro_sub(&d[CHAN_Y], &r[3], &r[1]); micro_mul( &r[5], &r[5], &r[4] ); micro_mul( &r[0], &r[0], &r[2] ); micro_sub(&d[CHAN_Z], &r[5], &r[0]); if (IS_CHANNEL_ENABLED(*inst, CHAN_X)) { STORE(&d[CHAN_X], 0, CHAN_X); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Y)) { STORE(&d[CHAN_Y], 0, CHAN_Y); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Z)) { STORE(&d[CHAN_Z], 0, CHAN_Z); } if (IS_CHANNEL_ENABLED( *inst, CHAN_W )) { STORE( &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], 0, CHAN_W ); } break; case TGSI_OPCODE_ABS: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH(&r[0], 0, chan_index); micro_abs(&d[chan_index], &r[0]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_RCC: FETCH(&r[0], 0, CHAN_X); micro_div(&r[0], &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], &r[0]); micro_float_clamp(&r[0], &r[0]); FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&r[0], 0, chan_index); } break; case TGSI_OPCODE_DPH: FETCH(&r[0], 0, CHAN_X); FETCH(&r[1], 1, CHAN_X); micro_mul( &r[0], &r[0], &r[1] ); FETCH(&r[1], 0, CHAN_Y); FETCH(&r[2], 1, CHAN_Y); micro_mul( &r[1], &r[1], &r[2] ); micro_add( &r[0], &r[0], &r[1] ); FETCH(&r[1], 0, CHAN_Z); FETCH(&r[2], 1, CHAN_Z); micro_mul( &r[1], &r[1], &r[2] ); micro_add( &r[0], &r[0], &r[1] ); FETCH(&r[1], 1, CHAN_W); micro_add( &r[0], &r[0], &r[1] ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[0], 0, chan_index ); } break; case TGSI_OPCODE_COS: FETCH(&r[0], 0, CHAN_X); micro_cos( &r[0], &r[0] ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[0], 0, chan_index ); } break; case TGSI_OPCODE_DDX: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); micro_ddx(&d[chan_index], &r[0]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_DDY: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); micro_ddy(&d[chan_index], &r[0]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_KILP: exec_kilp (mach, inst); break; case TGSI_OPCODE_KIL: exec_kil (mach, inst); break; case TGSI_OPCODE_PK2H: assert (0); break; case TGSI_OPCODE_PK2US: assert (0); break; case TGSI_OPCODE_PK4B: assert (0); break; case TGSI_OPCODE_PK4UB: assert (0); break; case TGSI_OPCODE_RFL: if (IS_CHANNEL_ENABLED(*inst, CHAN_X) || IS_CHANNEL_ENABLED(*inst, CHAN_Y) || IS_CHANNEL_ENABLED(*inst, CHAN_Z)) { /* r0 = dp3(src0, src0) */ FETCH(&r[2], 0, CHAN_X); micro_mul(&r[0], &r[2], &r[2]); FETCH(&r[4], 0, CHAN_Y); micro_mul(&r[8], &r[4], &r[4]); micro_add(&r[0], &r[0], &r[8]); FETCH(&r[6], 0, CHAN_Z); micro_mul(&r[8], &r[6], &r[6]); micro_add(&r[0], &r[0], &r[8]); /* r1 = dp3(src0, src1) */ FETCH(&r[3], 1, CHAN_X); micro_mul(&r[1], &r[2], &r[3]); FETCH(&r[5], 1, CHAN_Y); micro_mul(&r[8], &r[4], &r[5]); micro_add(&r[1], &r[1], &r[8]); FETCH(&r[7], 1, CHAN_Z); micro_mul(&r[8], &r[6], &r[7]); micro_add(&r[1], &r[1], &r[8]); /* r1 = 2 * r1 / r0 */ micro_add(&r[1], &r[1], &r[1]); micro_div(&r[1], &r[1], &r[0]); if (IS_CHANNEL_ENABLED(*inst, CHAN_X)) { micro_mul(&r[2], &r[2], &r[1]); micro_sub(&r[2], &r[2], &r[3]); STORE(&r[2], 0, CHAN_X); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Y)) { micro_mul(&r[4], &r[4], &r[1]); micro_sub(&r[4], &r[4], &r[5]); STORE(&r[4], 0, CHAN_Y); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Z)) { micro_mul(&r[6], &r[6], &r[1]); micro_sub(&r[6], &r[6], &r[7]); STORE(&r[6], 0, CHAN_Z); } } if (IS_CHANNEL_ENABLED(*inst, CHAN_W)) { STORE(&mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], 0, CHAN_W); } break; case TGSI_OPCODE_SEQ: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); FETCH( &r[1], 1, chan_index ); micro_eq(&d[chan_index], &r[0], &r[1], &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], &mach->Temps[TEMP_0_I].xyzw[TEMP_0_C]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_SFL: FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&mach->Temps[TEMP_0_I].xyzw[TEMP_0_C], 0, chan_index); } break; case TGSI_OPCODE_SGT: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); FETCH( &r[1], 1, chan_index ); micro_le(&d[chan_index], &r[0], &r[1], &mach->Temps[TEMP_0_I].xyzw[TEMP_0_C], &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_SIN: FETCH( &r[0], 0, CHAN_X ); micro_sin( &r[0], &r[0] ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[0], 0, chan_index ); } break; case TGSI_OPCODE_SLE: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); FETCH( &r[1], 1, chan_index ); micro_le(&d[chan_index], &r[0], &r[1], &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], &mach->Temps[TEMP_0_I].xyzw[TEMP_0_C]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_SNE: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); FETCH( &r[1], 1, chan_index ); micro_eq(&d[chan_index], &r[0], &r[1], &mach->Temps[TEMP_0_I].xyzw[TEMP_0_C], &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_STR: FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], 0, chan_index); } break; case TGSI_OPCODE_TEX: /* simple texture lookup */ /* src[0] = texcoord */ /* src[1] = sampler unit */ exec_tex(mach, inst, FALSE, FALSE); break; case TGSI_OPCODE_TXB: /* Texture lookup with lod bias */ /* src[0] = texcoord (src[0].w = LOD bias) */ /* src[1] = sampler unit */ exec_tex(mach, inst, TRUE, FALSE); break; case TGSI_OPCODE_TXD: /* Texture lookup with explict partial derivatives */ /* src[0] = texcoord */ /* src[1] = d[strq]/dx */ /* src[2] = d[strq]/dy */ /* src[3] = sampler unit */ exec_txd(mach, inst); break; case TGSI_OPCODE_TXL: /* Texture lookup with explit LOD */ /* src[0] = texcoord (src[0].w = LOD) */ /* src[1] = sampler unit */ exec_tex(mach, inst, TRUE, FALSE); break; case TGSI_OPCODE_TXP: /* Texture lookup with projection */ /* src[0] = texcoord (src[0].w = projection) */ /* src[1] = sampler unit */ exec_tex(mach, inst, FALSE, TRUE); break; case TGSI_OPCODE_UP2H: assert (0); break; case TGSI_OPCODE_UP2US: assert (0); break; case TGSI_OPCODE_UP4B: assert (0); break; case TGSI_OPCODE_UP4UB: assert (0); break; case TGSI_OPCODE_X2D: FETCH(&r[0], 1, CHAN_X); FETCH(&r[1], 1, CHAN_Y); if (IS_CHANNEL_ENABLED(*inst, CHAN_X) || IS_CHANNEL_ENABLED(*inst, CHAN_Z)) { FETCH(&r[2], 2, CHAN_X); micro_mul(&r[2], &r[2], &r[0]); FETCH(&r[3], 2, CHAN_Y); micro_mul(&r[3], &r[3], &r[1]); micro_add(&r[2], &r[2], &r[3]); FETCH(&r[3], 0, CHAN_X); micro_add(&d[CHAN_X], &r[2], &r[3]); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Y) || IS_CHANNEL_ENABLED(*inst, CHAN_W)) { FETCH(&r[2], 2, CHAN_Z); micro_mul(&r[2], &r[2], &r[0]); FETCH(&r[3], 2, CHAN_W); micro_mul(&r[3], &r[3], &r[1]); micro_add(&r[2], &r[2], &r[3]); FETCH(&r[3], 0, CHAN_Y); micro_add(&d[CHAN_Y], &r[2], &r[3]); } if (IS_CHANNEL_ENABLED(*inst, CHAN_X)) { STORE(&d[CHAN_X], 0, CHAN_X); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Y)) { STORE(&d[CHAN_Y], 0, CHAN_Y); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Z)) { STORE(&d[CHAN_X], 0, CHAN_Z); } if (IS_CHANNEL_ENABLED(*inst, CHAN_W)) { STORE(&d[CHAN_Y], 0, CHAN_W); } break; case TGSI_OPCODE_ARA: assert (0); break; case TGSI_OPCODE_BRA: assert (0); break; case TGSI_OPCODE_CAL: /* skip the call if no execution channels are enabled */ if (mach->ExecMask) { /* do the call */ /* First, record the depths of the execution stacks. * This is important for deeply nested/looped return statements. * We have to unwind the stacks by the correct amount. For a * real code generator, we could determine the number of entries * to pop off each stack with simple static analysis and avoid * implementing this data structure at run time. */ mach->CallStack[mach->CallStackTop].CondStackTop = mach->CondStackTop; mach->CallStack[mach->CallStackTop].LoopStackTop = mach->LoopStackTop; mach->CallStack[mach->CallStackTop].ContStackTop = mach->ContStackTop; mach->CallStack[mach->CallStackTop].SwitchStackTop = mach->SwitchStackTop; mach->CallStack[mach->CallStackTop].BreakStackTop = mach->BreakStackTop; /* note that PC was already incremented above */ mach->CallStack[mach->CallStackTop].ReturnAddr = *pc; mach->CallStackTop++; /* Second, push the Cond, Loop, Cont, Func stacks */ assert(mach->CondStackTop < TGSI_EXEC_MAX_COND_NESTING); assert(mach->LoopStackTop < TGSI_EXEC_MAX_LOOP_NESTING); assert(mach->ContStackTop < TGSI_EXEC_MAX_LOOP_NESTING); assert(mach->SwitchStackTop < TGSI_EXEC_MAX_SWITCH_NESTING); assert(mach->BreakStackTop < TGSI_EXEC_MAX_BREAK_STACK); assert(mach->FuncStackTop < TGSI_EXEC_MAX_CALL_NESTING); mach->CondStack[mach->CondStackTop++] = mach->CondMask; mach->LoopStack[mach->LoopStackTop++] = mach->LoopMask; mach->ContStack[mach->ContStackTop++] = mach->ContMask; mach->SwitchStack[mach->SwitchStackTop++] = mach->Switch; mach->BreakStack[mach->BreakStackTop++] = mach->BreakType; mach->FuncStack[mach->FuncStackTop++] = mach->FuncMask; /* Finally, jump to the subroutine */ *pc = inst->Label.Label; } break; case TGSI_OPCODE_RET: mach->FuncMask &= ~mach->ExecMask; UPDATE_EXEC_MASK(mach); if (mach->FuncMask == 0x0) { /* really return now (otherwise, keep executing */ if (mach->CallStackTop == 0) { /* returning from main() */ *pc = -1; return; } assert(mach->CallStackTop > 0); mach->CallStackTop--; mach->CondStackTop = mach->CallStack[mach->CallStackTop].CondStackTop; mach->CondMask = mach->CondStack[mach->CondStackTop]; mach->LoopStackTop = mach->CallStack[mach->CallStackTop].LoopStackTop; mach->LoopMask = mach->LoopStack[mach->LoopStackTop]; mach->ContStackTop = mach->CallStack[mach->CallStackTop].ContStackTop; mach->ContMask = mach->ContStack[mach->ContStackTop]; mach->SwitchStackTop = mach->CallStack[mach->CallStackTop].SwitchStackTop; mach->Switch = mach->SwitchStack[mach->SwitchStackTop]; mach->BreakStackTop = mach->CallStack[mach->CallStackTop].BreakStackTop; mach->BreakType = mach->BreakStack[mach->BreakStackTop]; assert(mach->FuncStackTop > 0); mach->FuncMask = mach->FuncStack[--mach->FuncStackTop]; *pc = mach->CallStack[mach->CallStackTop].ReturnAddr; UPDATE_EXEC_MASK(mach); } break; case TGSI_OPCODE_SSG: /* TGSI_OPCODE_SGN */ FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); micro_sgn(&d[chan_index], &r[0]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_CMP: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH(&r[0], 0, chan_index); FETCH(&r[1], 1, chan_index); FETCH(&r[2], 2, chan_index); micro_lt(&d[chan_index], &r[0], &mach->Temps[TEMP_0_I].xyzw[TEMP_0_C], &r[1], &r[2]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_SCS: if( IS_CHANNEL_ENABLED( *inst, CHAN_X ) || IS_CHANNEL_ENABLED( *inst, CHAN_Y ) ) { FETCH( &r[0], 0, CHAN_X ); if (IS_CHANNEL_ENABLED(*inst, CHAN_X)) { micro_cos(&r[1], &r[0]); STORE(&r[1], 0, CHAN_X); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Y)) { micro_sin(&r[1], &r[0]); STORE(&r[1], 0, CHAN_Y); } } if( IS_CHANNEL_ENABLED( *inst, CHAN_Z ) ) { STORE( &mach->Temps[TEMP_0_I].xyzw[TEMP_0_C], 0, CHAN_Z ); } if( IS_CHANNEL_ENABLED( *inst, CHAN_W ) ) { STORE( &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], 0, CHAN_W ); } break; case TGSI_OPCODE_NRM: /* 3-component vector normalize */ if(IS_CHANNEL_ENABLED(*inst, CHAN_X) || IS_CHANNEL_ENABLED(*inst, CHAN_Y) || IS_CHANNEL_ENABLED(*inst, CHAN_Z)) { /* r3 = sqrt(dp3(src0, src0)) */ FETCH(&r[0], 0, CHAN_X); micro_mul(&r[3], &r[0], &r[0]); FETCH(&r[1], 0, CHAN_Y); micro_mul(&r[4], &r[1], &r[1]); micro_add(&r[3], &r[3], &r[4]); FETCH(&r[2], 0, CHAN_Z); micro_mul(&r[4], &r[2], &r[2]); micro_add(&r[3], &r[3], &r[4]); micro_sqrt(&r[3], &r[3]); if (IS_CHANNEL_ENABLED(*inst, CHAN_X)) { micro_div(&r[0], &r[0], &r[3]); STORE(&r[0], 0, CHAN_X); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Y)) { micro_div(&r[1], &r[1], &r[3]); STORE(&r[1], 0, CHAN_Y); } if (IS_CHANNEL_ENABLED(*inst, CHAN_Z)) { micro_div(&r[2], &r[2], &r[3]); STORE(&r[2], 0, CHAN_Z); } } if (IS_CHANNEL_ENABLED(*inst, CHAN_W)) { STORE(&mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], 0, CHAN_W); } break; case TGSI_OPCODE_NRM4: /* 4-component vector normalize */ { union tgsi_exec_channel tmp, dot; /* tmp = dp4(src0, src0): */ FETCH( &r[0], 0, CHAN_X ); micro_mul( &tmp, &r[0], &r[0] ); FETCH( &r[1], 0, CHAN_Y ); micro_mul( &dot, &r[1], &r[1] ); micro_add( &tmp, &tmp, &dot ); FETCH( &r[2], 0, CHAN_Z ); micro_mul( &dot, &r[2], &r[2] ); micro_add( &tmp, &tmp, &dot ); FETCH( &r[3], 0, CHAN_W ); micro_mul( &dot, &r[3], &r[3] ); micro_add( &tmp, &tmp, &dot ); /* tmp = 1 / sqrt(tmp) */ micro_sqrt( &tmp, &tmp ); micro_div( &tmp, &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C], &tmp ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { /* chan = chan * tmp */ micro_mul( &r[chan_index], &tmp, &r[chan_index] ); STORE( &r[chan_index], 0, chan_index ); } } break; case TGSI_OPCODE_DIV: assert( 0 ); break; case TGSI_OPCODE_DP2: FETCH( &r[0], 0, CHAN_X ); FETCH( &r[1], 1, CHAN_X ); micro_mul( &r[0], &r[0], &r[1] ); FETCH( &r[1], 0, CHAN_Y ); FETCH( &r[2], 1, CHAN_Y ); micro_mul( &r[1], &r[1], &r[2] ); micro_add( &r[0], &r[0], &r[1] ); FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { STORE( &r[0], 0, chan_index ); } break; case TGSI_OPCODE_IF: /* push CondMask */ assert(mach->CondStackTop < TGSI_EXEC_MAX_COND_NESTING); mach->CondStack[mach->CondStackTop++] = mach->CondMask; FETCH( &r[0], 0, CHAN_X ); /* update CondMask */ if( ! r[0].u[0] ) { mach->CondMask &= ~0x1; } if( ! r[0].u[1] ) { mach->CondMask &= ~0x2; } if( ! r[0].u[2] ) { mach->CondMask &= ~0x4; } if( ! r[0].u[3] ) { mach->CondMask &= ~0x8; } UPDATE_EXEC_MASK(mach); /* Todo: If CondMask==0, jump to ELSE */ break; case TGSI_OPCODE_ELSE: /* invert CondMask wrt previous mask */ { uint prevMask; assert(mach->CondStackTop > 0); prevMask = mach->CondStack[mach->CondStackTop - 1]; mach->CondMask = ~mach->CondMask & prevMask; UPDATE_EXEC_MASK(mach); /* Todo: If CondMask==0, jump to ENDIF */ } break; case TGSI_OPCODE_ENDIF: /* pop CondMask */ assert(mach->CondStackTop > 0); mach->CondMask = mach->CondStack[--mach->CondStackTop]; UPDATE_EXEC_MASK(mach); break; case TGSI_OPCODE_END: /* halt execution */ *pc = -1; break; case TGSI_OPCODE_REP: assert (0); break; case TGSI_OPCODE_ENDREP: assert (0); break; case TGSI_OPCODE_PUSHA: assert (0); break; case TGSI_OPCODE_POPA: assert (0); break; case TGSI_OPCODE_CEIL: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); micro_ceil(&d[chan_index], &r[0]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_I2F: exec_vector_unary(mach, inst, micro_i2f, TGSI_EXEC_DATA_FLOAT, TGSI_EXEC_DATA_INT); break; case TGSI_OPCODE_NOT: exec_vector_unary(mach, inst, micro_not, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_TRUNC: FOR_EACH_ENABLED_CHANNEL( *inst, chan_index ) { FETCH( &r[0], 0, chan_index ); micro_trunc(&d[chan_index], &r[0]); } FOR_EACH_ENABLED_CHANNEL(*inst, chan_index) { STORE(&d[chan_index], 0, chan_index); } break; case TGSI_OPCODE_SHL: exec_vector_binary(mach, inst, micro_shl, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_AND: exec_vector_binary(mach, inst, micro_and, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_OR: exec_vector_binary(mach, inst, micro_or, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_MOD: assert (0); break; case TGSI_OPCODE_XOR: exec_vector_binary(mach, inst, micro_xor, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_SAD: assert (0); break; case TGSI_OPCODE_TXF: assert (0); break; case TGSI_OPCODE_TXQ: assert (0); break; case TGSI_OPCODE_EMIT: emit_vertex(mach); break; case TGSI_OPCODE_ENDPRIM: emit_primitive(mach); break; case TGSI_OPCODE_BGNFOR: assert(mach->LoopCounterStackTop < TGSI_EXEC_MAX_LOOP_NESTING); for (chan_index = 0; chan_index < 3; chan_index++) { FETCH( &mach->LoopCounterStack[mach->LoopCounterStackTop].xyzw[chan_index], 0, chan_index ); } ++mach->LoopCounterStackTop; STORE(&mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_X], 0, CHAN_X); /* update LoopMask */ if (mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_Y].f[0] <= 0.0f) { mach->LoopMask &= ~0x1; } if (mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_Y].f[1] <= 0.0f) { mach->LoopMask &= ~0x2; } if (mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_Y].f[2] <= 0.0f) { mach->LoopMask &= ~0x4; } if (mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_Y].f[3] <= 0.0f) { mach->LoopMask &= ~0x8; } /* TODO: if mach->LoopMask == 0, jump to end of loop */ UPDATE_EXEC_MASK(mach); /* fall-through (for now) */ case TGSI_OPCODE_BGNLOOP: /* push LoopMask and ContMasks */ assert(mach->LoopStackTop < TGSI_EXEC_MAX_LOOP_NESTING); assert(mach->ContStackTop < TGSI_EXEC_MAX_LOOP_NESTING); assert(mach->LoopLabelStackTop < TGSI_EXEC_MAX_LOOP_NESTING); assert(mach->BreakStackTop < TGSI_EXEC_MAX_BREAK_STACK); mach->LoopStack[mach->LoopStackTop++] = mach->LoopMask; mach->ContStack[mach->ContStackTop++] = mach->ContMask; mach->LoopLabelStack[mach->LoopLabelStackTop++] = *pc - 1; mach->BreakStack[mach->BreakStackTop++] = mach->BreakType; mach->BreakType = TGSI_EXEC_BREAK_INSIDE_LOOP; break; case TGSI_OPCODE_ENDFOR: assert(mach->LoopCounterStackTop > 0); micro_sub(&mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_Y], &mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_Y], &mach->Temps[TEMP_1_I].xyzw[TEMP_1_C]); /* update LoopMask */ if (mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_Y].f[0] <= 0.0f) { mach->LoopMask &= ~0x1; } if (mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_Y].f[1] <= 0.0f) { mach->LoopMask &= ~0x2; } if (mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_Y].f[2] <= 0.0f) { mach->LoopMask &= ~0x4; } if (mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_Y].f[3] <= 0.0f) { mach->LoopMask &= ~0x8; } micro_add(&mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_X], &mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_X], &mach->LoopCounterStack[mach->LoopCounterStackTop - 1].xyzw[CHAN_Z]); assert(mach->LoopLabelStackTop > 0); inst = mach->Instructions + mach->LoopLabelStack[mach->LoopLabelStackTop - 1]; STORE(&mach->LoopCounterStack[mach->LoopCounterStackTop].xyzw[CHAN_X], 0, CHAN_X); /* Restore ContMask, but don't pop */ assert(mach->ContStackTop > 0); mach->ContMask = mach->ContStack[mach->ContStackTop - 1]; UPDATE_EXEC_MASK(mach); if (mach->ExecMask) { /* repeat loop: jump to instruction just past BGNLOOP */ assert(mach->LoopLabelStackTop > 0); *pc = mach->LoopLabelStack[mach->LoopLabelStackTop - 1] + 1; } else { /* exit loop: pop LoopMask */ assert(mach->LoopStackTop > 0); mach->LoopMask = mach->LoopStack[--mach->LoopStackTop]; /* pop ContMask */ assert(mach->ContStackTop > 0); mach->ContMask = mach->ContStack[--mach->ContStackTop]; assert(mach->LoopLabelStackTop > 0); --mach->LoopLabelStackTop; assert(mach->LoopCounterStackTop > 0); --mach->LoopCounterStackTop; mach->BreakType = mach->BreakStack[--mach->BreakStackTop]; } UPDATE_EXEC_MASK(mach); break; case TGSI_OPCODE_ENDLOOP: /* Restore ContMask, but don't pop */ assert(mach->ContStackTop > 0); mach->ContMask = mach->ContStack[mach->ContStackTop - 1]; UPDATE_EXEC_MASK(mach); if (mach->ExecMask) { /* repeat loop: jump to instruction just past BGNLOOP */ assert(mach->LoopLabelStackTop > 0); *pc = mach->LoopLabelStack[mach->LoopLabelStackTop - 1] + 1; } else { /* exit loop: pop LoopMask */ assert(mach->LoopStackTop > 0); mach->LoopMask = mach->LoopStack[--mach->LoopStackTop]; /* pop ContMask */ assert(mach->ContStackTop > 0); mach->ContMask = mach->ContStack[--mach->ContStackTop]; assert(mach->LoopLabelStackTop > 0); --mach->LoopLabelStackTop; mach->BreakType = mach->BreakStack[--mach->BreakStackTop]; } UPDATE_EXEC_MASK(mach); break; case TGSI_OPCODE_BRK: exec_break(mach); break; case TGSI_OPCODE_CONT: /* turn off cont channels for each enabled exec channel */ mach->ContMask &= ~mach->ExecMask; /* Todo: if mach->LoopMask == 0, jump to end of loop */ UPDATE_EXEC_MASK(mach); break; case TGSI_OPCODE_BGNSUB: /* no-op */ break; case TGSI_OPCODE_ENDSUB: /* * XXX: This really should be a no-op. We should never reach this opcode. */ assert(mach->CallStackTop > 0); mach->CallStackTop--; mach->CondStackTop = mach->CallStack[mach->CallStackTop].CondStackTop; mach->CondMask = mach->CondStack[mach->CondStackTop]; mach->LoopStackTop = mach->CallStack[mach->CallStackTop].LoopStackTop; mach->LoopMask = mach->LoopStack[mach->LoopStackTop]; mach->ContStackTop = mach->CallStack[mach->CallStackTop].ContStackTop; mach->ContMask = mach->ContStack[mach->ContStackTop]; mach->SwitchStackTop = mach->CallStack[mach->CallStackTop].SwitchStackTop; mach->Switch = mach->SwitchStack[mach->SwitchStackTop]; mach->BreakStackTop = mach->CallStack[mach->CallStackTop].BreakStackTop; mach->BreakType = mach->BreakStack[mach->BreakStackTop]; assert(mach->FuncStackTop > 0); mach->FuncMask = mach->FuncStack[--mach->FuncStackTop]; *pc = mach->CallStack[mach->CallStackTop].ReturnAddr; UPDATE_EXEC_MASK(mach); break; case TGSI_OPCODE_NOP: break; case TGSI_OPCODE_BREAKC: FETCH(&r[0], 0, CHAN_X); /* update CondMask */ if (r[0].u[0] && (mach->ExecMask & 0x1)) { mach->LoopMask &= ~0x1; } if (r[0].u[1] && (mach->ExecMask & 0x2)) { mach->LoopMask &= ~0x2; } if (r[0].u[2] && (mach->ExecMask & 0x4)) { mach->LoopMask &= ~0x4; } if (r[0].u[3] && (mach->ExecMask & 0x8)) { mach->LoopMask &= ~0x8; } /* Todo: if mach->LoopMask == 0, jump to end of loop */ UPDATE_EXEC_MASK(mach); break; case TGSI_OPCODE_F2I: exec_vector_unary(mach, inst, micro_f2i, TGSI_EXEC_DATA_INT, TGSI_EXEC_DATA_FLOAT); break; case TGSI_OPCODE_IDIV: exec_vector_binary(mach, inst, micro_idiv, TGSI_EXEC_DATA_INT, TGSI_EXEC_DATA_INT); break; case TGSI_OPCODE_IMAX: exec_vector_binary(mach, inst, micro_imax, TGSI_EXEC_DATA_INT, TGSI_EXEC_DATA_INT); break; case TGSI_OPCODE_IMIN: exec_vector_binary(mach, inst, micro_imin, TGSI_EXEC_DATA_INT, TGSI_EXEC_DATA_INT); break; case TGSI_OPCODE_INEG: exec_vector_unary(mach, inst, micro_ineg, TGSI_EXEC_DATA_INT, TGSI_EXEC_DATA_INT); break; case TGSI_OPCODE_ISGE: exec_vector_binary(mach, inst, micro_isge, TGSI_EXEC_DATA_INT, TGSI_EXEC_DATA_INT); break; case TGSI_OPCODE_ISHR: exec_vector_binary(mach, inst, micro_ishr, TGSI_EXEC_DATA_INT, TGSI_EXEC_DATA_INT); break; case TGSI_OPCODE_ISLT: exec_vector_binary(mach, inst, micro_islt, TGSI_EXEC_DATA_INT, TGSI_EXEC_DATA_INT); break; case TGSI_OPCODE_F2U: exec_vector_unary(mach, inst, micro_f2u, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_FLOAT); break; case TGSI_OPCODE_U2F: exec_vector_unary(mach, inst, micro_u2f, TGSI_EXEC_DATA_FLOAT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_UADD: exec_vector_binary(mach, inst, micro_uadd, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_UDIV: exec_vector_binary(mach, inst, micro_udiv, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_UMAD: exec_vector_trinary(mach, inst, micro_umad, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_UMAX: exec_vector_binary(mach, inst, micro_umax, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_UMIN: exec_vector_binary(mach, inst, micro_umin, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_UMOD: exec_vector_binary(mach, inst, micro_umod, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_UMUL: exec_vector_binary(mach, inst, micro_umul, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_USEQ: exec_vector_binary(mach, inst, micro_useq, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_USGE: exec_vector_binary(mach, inst, micro_usge, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_USHR: exec_vector_binary(mach, inst, micro_ushr, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_USLT: exec_vector_binary(mach, inst, micro_uslt, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_USNE: exec_vector_binary(mach, inst, micro_usne, TGSI_EXEC_DATA_UINT, TGSI_EXEC_DATA_UINT); break; case TGSI_OPCODE_SWITCH: exec_switch(mach, inst); break; case TGSI_OPCODE_CASE: exec_case(mach, inst); break; case TGSI_OPCODE_DEFAULT: exec_default(mach); break; case TGSI_OPCODE_ENDSWITCH: exec_endswitch(mach); break; default: assert( 0 ); } } #define DEBUG_EXECUTION 0 /** * Run TGSI interpreter. * \return bitmask of "alive" quad components */ uint tgsi_exec_machine_run( struct tgsi_exec_machine *mach ) { uint i; int pc = 0; mach->CondMask = 0xf; mach->LoopMask = 0xf; mach->ContMask = 0xf; mach->FuncMask = 0xf; mach->ExecMask = 0xf; mach->Switch.mask = 0xf; assert(mach->CondStackTop == 0); assert(mach->LoopStackTop == 0); assert(mach->ContStackTop == 0); assert(mach->SwitchStackTop == 0); assert(mach->BreakStackTop == 0); assert(mach->CallStackTop == 0); mach->Temps[TEMP_KILMASK_I].xyzw[TEMP_KILMASK_C].u[0] = 0; mach->Temps[TEMP_OUTPUT_I].xyzw[TEMP_OUTPUT_C].u[0] = 0; if( mach->Processor == TGSI_PROCESSOR_GEOMETRY ) { mach->Temps[TEMP_PRIMITIVE_I].xyzw[TEMP_PRIMITIVE_C].u[0] = 0; mach->Primitives[0] = 0; } for (i = 0; i < QUAD_SIZE; i++) { mach->Temps[TEMP_CC_I].xyzw[TEMP_CC_C].u[i] = (TGSI_EXEC_CC_EQ << TGSI_EXEC_CC_X_SHIFT) | (TGSI_EXEC_CC_EQ << TGSI_EXEC_CC_Y_SHIFT) | (TGSI_EXEC_CC_EQ << TGSI_EXEC_CC_Z_SHIFT) | (TGSI_EXEC_CC_EQ << TGSI_EXEC_CC_W_SHIFT); } /* execute declarations (interpolants) */ for (i = 0; i < mach->NumDeclarations; i++) { exec_declaration( mach, mach->Declarations+i ); } { #if DEBUG_EXECUTION struct tgsi_exec_vector temps[TGSI_EXEC_NUM_TEMPS + TGSI_EXEC_NUM_TEMP_EXTRAS]; struct tgsi_exec_vector outputs[PIPE_MAX_ATTRIBS]; uint inst = 1; memcpy(temps, mach->Temps, sizeof(temps)); memcpy(outputs, mach->Outputs, sizeof(outputs)); #endif /* execute instructions, until pc is set to -1 */ while (pc != -1) { #if DEBUG_EXECUTION uint i; tgsi_dump_instruction(&mach->Instructions[pc], inst++); #endif assert(pc < (int) mach->NumInstructions); exec_instruction(mach, mach->Instructions + pc, &pc); #if DEBUG_EXECUTION for (i = 0; i < TGSI_EXEC_NUM_TEMPS + TGSI_EXEC_NUM_TEMP_EXTRAS; i++) { if (memcmp(&temps[i], &mach->Temps[i], sizeof(temps[i]))) { uint j; memcpy(&temps[i], &mach->Temps[i], sizeof(temps[i])); debug_printf("TEMP[%2u] = ", i); for (j = 0; j < 4; j++) { if (j > 0) { debug_printf(" "); } debug_printf("(%6f %u, %6f %u, %6f %u, %6f %u)\n", temps[i].xyzw[0].f[j], temps[i].xyzw[0].u[j], temps[i].xyzw[1].f[j], temps[i].xyzw[1].u[j], temps[i].xyzw[2].f[j], temps[i].xyzw[2].u[j], temps[i].xyzw[3].f[j], temps[i].xyzw[3].u[j]); } } } for (i = 0; i < PIPE_MAX_ATTRIBS; i++) { if (memcmp(&outputs[i], &mach->Outputs[i], sizeof(outputs[i]))) { uint j; memcpy(&outputs[i], &mach->Outputs[i], sizeof(outputs[i])); debug_printf("OUT[%2u] = ", i); for (j = 0; j < 4; j++) { if (j > 0) { debug_printf(" "); } debug_printf("(%6f %u, %6f %u, %6f %u, %6f %u)\n", outputs[i].xyzw[0].f[j], outputs[i].xyzw[0].u[j], outputs[i].xyzw[1].f[j], outputs[i].xyzw[1].u[j], outputs[i].xyzw[2].f[j], outputs[i].xyzw[2].u[j], outputs[i].xyzw[3].f[j], outputs[i].xyzw[3].u[j]); } } } #endif } } #if 0 /* we scale from floats in [0,1] to Zbuffer ints in sp_quad_depth_test.c */ if (mach->Processor == TGSI_PROCESSOR_FRAGMENT) { /* * Scale back depth component. */ for (i = 0; i < 4; i++) mach->Outputs[0].xyzw[2].f[i] *= ctx->DrawBuffer->_DepthMaxF; } #endif assert(mach->CondStackTop == 0); assert(mach->LoopStackTop == 0); assert(mach->ContStackTop == 0); assert(mach->SwitchStackTop == 0); assert(mach->BreakStackTop == 0); assert(mach->CallStackTop == 0); return ~mach->Temps[TEMP_KILMASK_I].xyzw[TEMP_KILMASK_C].u[0]; }