/* $Id: DevVGA-SVGA3d-dx-shader.cpp 103416 2024-02-19 08:06:54Z vboxsync $ */ /** @file * DevVMWare - VMWare SVGA device - VGPU10+ (DX) shader utilities. */ /* * Copyright (C) 2020-2023 Oracle and/or its affiliates. * * This file is part of VirtualBox base platform packages, as * available from https://www.virtualbox.org. * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License * as published by the Free Software Foundation, in version 3 of the * License. * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, see . * * SPDX-License-Identifier: GPL-3.0-only */ /********************************************************************************************************************************* * Header Files * *********************************************************************************************************************************/ #define LOG_GROUP LOG_GROUP_DEV_VMSVGA #include #include #include #include #include #include #include #include "DevVGA-SVGA3d-dx-shader.h" #ifdef RT_OS_WINDOWS #include #else #define D3D11_SB_EXTENDED_OPCODE_RESOURCE_DIM 2 #define D3D11_SB_EXTENDED_OPCODE_RESOURCE_RETURN_TYPE 3 #endif /* * * DXBC shader binary format definitions. * */ /* DXBC container header. */ typedef struct DXBCHeader { uint32_t u32DXBC; /* 0x43425844 = 'D', 'X', 'B', 'C' */ uint8_t au8Hash[16]; /* Modified MD5 hash. See dxbcHash. */ uint32_t u32Version; /* 1 */ uint32_t cbTotal; /* Total size in bytes. Including the header. */ uint32_t cBlob; /* Number of entries in aBlobOffset array. */ uint32_t aBlobOffset[1]; /* Offsets of blobs from the start of DXBC header. */ } DXBCHeader; #define DXBC_MAGIC RT_MAKE_U32_FROM_U8('D', 'X', 'B', 'C') /* DXBC blob header. */ typedef struct DXBCBlobHeader { uint32_t u32BlobType; /* FourCC code. DXBC_BLOB_TYPE_* */ uint32_t cbBlob; /* Size of the blob excluding the blob header. 4 bytes aligned. */ /* Followed by the blob's data. */ } DXBCBlobHeader; /* DXBC blob types. */ #define DXBC_BLOB_TYPE_ISGN RT_MAKE_U32_FROM_U8('I', 'S', 'G', 'N') #define DXBC_BLOB_TYPE_OSGN RT_MAKE_U32_FROM_U8('O', 'S', 'G', 'N') #define DXBC_BLOB_TYPE_PCSG RT_MAKE_U32_FROM_U8('P', 'C', 'S', 'G') #define DXBC_BLOB_TYPE_SHDR RT_MAKE_U32_FROM_U8('S', 'H', 'D', 'R') /** @todo More... */ /* 'SHDR' blob data format. */ typedef struct DXBCBlobSHDR { VGPU10ProgramToken programToken; uint32_t cToken; /* Number of 32 bit tokens including programToken and cToken. */ uint32_t au32Token[1]; /* cToken - 2 number of tokens. */ } DXBCBlobSHDR; /* Element of an input or output signature. */ typedef struct DXBCBlobIOSGNElement { uint32_t offElementName; /* Offset of the semantic's name relative to the start of the blob data. */ uint32_t idxSemantic; /* Semantic index. */ uint32_t enmSystemValue; /* SVGA3dDXSignatureSemanticName */ uint32_t enmComponentType; /* 1 - unsigned, 2 - integer, 3 - float. */ uint32_t idxRegister; /* Shader register index. Elements must be sorted by register index. */ union { struct { uint32_t mask : 8; /* Component mask. Lower 4 bits represent X, Y, Z, W channels. */ uint32_t mask2 : 8; /* Which components are used in the shader. */ uint32_t pad : 16; } m; uint32_t mask; } u; } DXBCBlobIOSGNElement; /* 'ISGN' and 'OSGN' blob data format. */ typedef struct DXBCBlobIOSGN { uint32_t cElement; /* Number of signature elements. */ uint32_t offElement; /* Offset of the first element from the start of the blob. Equals to 8. */ DXBCBlobIOSGNElement aElement[1]; /* Signature elements. Size is cElement. */ /* Followed by ASCIIZ semantic names. */ } DXBCBlobIOSGN; /* * VGPU10 shader parser definitions. */ /* Parsed info about an operand index. */ typedef struct VGPUOperandIndex { uint32_t indexRepresentation; /* VGPU10_OPERAND_INDEX_REPRESENTATION */ uint64_t iOperandImmediate; /* Needs up to a qword. */ struct VGPUOperand *pOperandRelative; /* For VGPU10_OPERAND_INDEX_*RELATIVE */ } VGPUOperandIndex; /* Parsed info about an operand. */ typedef struct VGPUOperand { uint32_t numComponents : 2; /* VGPU10_OPERAND_NUM_COMPONENTS */ uint32_t selectionMode : 2; /* VGPU10_OPERAND_4_COMPONENT_SELECTION_MODE */ uint32_t mask : 4; /* 4-bits X, Y, Z, W mask for VGPU10_OPERAND_4_COMPONENT_MASK_MODE. */ uint32_t operandType : 8; /* VGPU10_OPERAND_TYPE */ uint32_t indexDimension : 2; /* VGPU10_OPERAND_INDEX_DIMENSION */ VGPUOperandIndex aOperandIndex[VGPU10_OPERAND_INDEX_3D]; /* Up to 3. */ uint32_t aImm[4]; /* Immediate values for VGPU10_OPERAND_TYPE_IMMEDIATE* */ uint32_t cOperandToken; /* Number of tokens in this operand. */ uint32_t const *paOperandToken; /* Pointer to operand tokens in the input buffer. */ } VGPUOperand; /* Parsed info about an opcode. */ typedef struct VGPUOpcode { uint32_t cOpcodeToken; /* Number of tokens for this operation. */ uint32_t opcodeType; /* VGPU10_OPCODE_* */ uint32_t opcodeSubtype; /* For example VGPU10_VMWARE_OPCODE_* */ uint32_t semanticName; /* SVGA3dDXSignatureSemanticName for system value declarations. */ uint32_t cOperand; /* Number of operands for this instruction. */ uint32_t aIdxOperand[8]; /* Indices of the instruction operands in the aValOperand array. */ /* 8 should be enough for everyone. */ VGPUOperand aValOperand[16]; /* Operands including VGPU10_OPERAND_INDEX_*RELATIVE if they are used: */ /* Operand1, VGPU10_OPERAND_INDEX_*RELATIVE for Operand1, ... */ /* ... */ /* OperandN, VGPU10_OPERAND_INDEX_*RELATIVE for OperandN, ... */ /* 16 probably should be enough for everyone. */ uint32_t const *paOpcodeToken; /* Pointer to opcode tokens in the input buffer. */ } VGPUOpcode; typedef struct VGPUOpcodeInfo { uint32_t cOperand; /* Number of operands for this opcode. */ } VGPUOpcodeInfo; static VGPUOpcodeInfo const g_aOpcodeInfo[] = { { 3 }, /* VGPU10_OPCODE_ADD */ { 3 }, /* VGPU10_OPCODE_AND */ { 0 }, /* VGPU10_OPCODE_BREAK */ { 1 }, /* VGPU10_OPCODE_BREAKC */ { 1 }, /* VGPU10_OPCODE_CALL */ { 2 }, /* VGPU10_OPCODE_CALLC */ { 1 }, /* VGPU10_OPCODE_CASE */ { 0 }, /* VGPU10_OPCODE_CONTINUE */ { 1 }, /* VGPU10_OPCODE_CONTINUEC */ { 0 }, /* VGPU10_OPCODE_CUT */ { 0 }, /* VGPU10_OPCODE_DEFAULT */ { 2 }, /* VGPU10_OPCODE_DERIV_RTX */ { 2 }, /* VGPU10_OPCODE_DERIV_RTY */ { 1 }, /* VGPU10_OPCODE_DISCARD */ { 3 }, /* VGPU10_OPCODE_DIV */ { 3 }, /* VGPU10_OPCODE_DP2 */ { 3 }, /* VGPU10_OPCODE_DP3 */ { 3 }, /* VGPU10_OPCODE_DP4 */ { 0 }, /* VGPU10_OPCODE_ELSE */ { 0 }, /* VGPU10_OPCODE_EMIT */ { 0 }, /* VGPU10_OPCODE_EMITTHENCUT */ { 0 }, /* VGPU10_OPCODE_ENDIF */ { 0 }, /* VGPU10_OPCODE_ENDLOOP */ { 0 }, /* VGPU10_OPCODE_ENDSWITCH */ { 3 }, /* VGPU10_OPCODE_EQ */ { 2 }, /* VGPU10_OPCODE_EXP */ { 2 }, /* VGPU10_OPCODE_FRC */ { 2 }, /* VGPU10_OPCODE_FTOI */ { 2 }, /* VGPU10_OPCODE_FTOU */ { 3 }, /* VGPU10_OPCODE_GE */ { 3 }, /* VGPU10_OPCODE_IADD */ { 1 }, /* VGPU10_OPCODE_IF */ { 3 }, /* VGPU10_OPCODE_IEQ */ { 3 }, /* VGPU10_OPCODE_IGE */ { 3 }, /* VGPU10_OPCODE_ILT */ { 4 }, /* VGPU10_OPCODE_IMAD */ { 3 }, /* VGPU10_OPCODE_IMAX */ { 3 }, /* VGPU10_OPCODE_IMIN */ { 4 }, /* VGPU10_OPCODE_IMUL */ { 3 }, /* VGPU10_OPCODE_INE */ { 2 }, /* VGPU10_OPCODE_INEG */ { 3 }, /* VGPU10_OPCODE_ISHL */ { 3 }, /* VGPU10_OPCODE_ISHR */ { 2 }, /* VGPU10_OPCODE_ITOF */ { 1 }, /* VGPU10_OPCODE_LABEL */ { 3 }, /* VGPU10_OPCODE_LD */ { 4 }, /* VGPU10_OPCODE_LD_MS */ { 2 }, /* VGPU10_OPCODE_LOG */ { 0 }, /* VGPU10_OPCODE_LOOP */ { 3 }, /* VGPU10_OPCODE_LT */ { 4 }, /* VGPU10_OPCODE_MAD */ { 3 }, /* VGPU10_OPCODE_MIN */ { 3 }, /* VGPU10_OPCODE_MAX */ { UINT32_MAX }, /* VGPU10_OPCODE_CUSTOMDATA: special opcode */ { 2 }, /* VGPU10_OPCODE_MOV */ { 4 }, /* VGPU10_OPCODE_MOVC */ { 3 }, /* VGPU10_OPCODE_MUL */ { 3 }, /* VGPU10_OPCODE_NE */ { 0 }, /* VGPU10_OPCODE_NOP */ { 2 }, /* VGPU10_OPCODE_NOT */ { 3 }, /* VGPU10_OPCODE_OR */ { 3 }, /* VGPU10_OPCODE_RESINFO */ { 0 }, /* VGPU10_OPCODE_RET */ { 1 }, /* VGPU10_OPCODE_RETC */ { 2 }, /* VGPU10_OPCODE_ROUND_NE */ { 2 }, /* VGPU10_OPCODE_ROUND_NI */ { 2 }, /* VGPU10_OPCODE_ROUND_PI */ { 2 }, /* VGPU10_OPCODE_ROUND_Z */ { 2 }, /* VGPU10_OPCODE_RSQ */ { 4 }, /* VGPU10_OPCODE_SAMPLE */ { 5 }, /* VGPU10_OPCODE_SAMPLE_C */ { 5 }, /* VGPU10_OPCODE_SAMPLE_C_LZ */ { 5 }, /* VGPU10_OPCODE_SAMPLE_L */ { 6 }, /* VGPU10_OPCODE_SAMPLE_D */ { 5 }, /* VGPU10_OPCODE_SAMPLE_B */ { 2 }, /* VGPU10_OPCODE_SQRT */ { 1 }, /* VGPU10_OPCODE_SWITCH */ { 3 }, /* VGPU10_OPCODE_SINCOS */ { 4 }, /* VGPU10_OPCODE_UDIV */ { 3 }, /* VGPU10_OPCODE_ULT */ { 3 }, /* VGPU10_OPCODE_UGE */ { 4 }, /* VGPU10_OPCODE_UMUL */ { 4 }, /* VGPU10_OPCODE_UMAD */ { 3 }, /* VGPU10_OPCODE_UMAX */ { 3 }, /* VGPU10_OPCODE_UMIN */ { 3 }, /* VGPU10_OPCODE_USHR */ { 2 }, /* VGPU10_OPCODE_UTOF */ { 3 }, /* VGPU10_OPCODE_XOR */ { 1 }, /* VGPU10_OPCODE_DCL_RESOURCE */ { 1 }, /* VGPU10_OPCODE_DCL_CONSTANT_BUFFER */ { 1 }, /* VGPU10_OPCODE_DCL_SAMPLER */ { 1 }, /* VGPU10_OPCODE_DCL_INDEX_RANGE */ { 0 }, /* VGPU10_OPCODE_DCL_GS_OUTPUT_PRIMITIVE_TOPOLOGY */ { 0 }, /* VGPU10_OPCODE_DCL_GS_INPUT_PRIMITIVE */ { 0 }, /* VGPU10_OPCODE_DCL_MAX_OUTPUT_VERTEX_COUNT */ { 1 }, /* VGPU10_OPCODE_DCL_INPUT */ { 1 }, /* VGPU10_OPCODE_DCL_INPUT_SGV */ { 1 }, /* VGPU10_OPCODE_DCL_INPUT_SIV */ { 1 }, /* VGPU10_OPCODE_DCL_INPUT_PS */ { 1 }, /* VGPU10_OPCODE_DCL_INPUT_PS_SGV */ { 1 }, /* VGPU10_OPCODE_DCL_INPUT_PS_SIV */ { 1 }, /* VGPU10_OPCODE_DCL_OUTPUT */ { 1 }, /* VGPU10_OPCODE_DCL_OUTPUT_SGV */ { 1 }, /* VGPU10_OPCODE_DCL_OUTPUT_SIV */ { 0 }, /* VGPU10_OPCODE_DCL_TEMPS */ { 0 }, /* VGPU10_OPCODE_DCL_INDEXABLE_TEMP */ { 0 }, /* VGPU10_OPCODE_DCL_GLOBAL_FLAGS */ { UINT32_MAX }, /* VGPU10_OPCODE_VMWARE: special opcode */ { 4 }, /* VGPU10_OPCODE_LOD */ { 4 }, /* VGPU10_OPCODE_GATHER4 */ { 3 }, /* VGPU10_OPCODE_SAMPLE_POS */ { 2 }, /* VGPU10_OPCODE_SAMPLE_INFO */ { UINT32_MAX }, /* VGPU10_OPCODE_RESERVED1: special opcode */ { 0 }, /* VGPU10_OPCODE_HS_DECLS */ { 0 }, /* VGPU10_OPCODE_HS_CONTROL_POINT_PHASE */ { 0 }, /* VGPU10_OPCODE_HS_FORK_PHASE */ { 0 }, /* VGPU10_OPCODE_HS_JOIN_PHASE */ { 1 }, /* VGPU10_OPCODE_EMIT_STREAM */ { 1 }, /* VGPU10_OPCODE_CUT_STREAM */ { 1 }, /* VGPU10_OPCODE_EMITTHENCUT_STREAM */ { 1 }, /* VGPU10_OPCODE_INTERFACE_CALL */ { 2 }, /* VGPU10_OPCODE_BUFINFO */ { 2 }, /* VGPU10_OPCODE_DERIV_RTX_COARSE */ { 2 }, /* VGPU10_OPCODE_DERIV_RTX_FINE */ { 2 }, /* VGPU10_OPCODE_DERIV_RTY_COARSE */ { 2 }, /* VGPU10_OPCODE_DERIV_RTY_FINE */ { 5 }, /* VGPU10_OPCODE_GATHER4_C */ { 5 }, /* VGPU10_OPCODE_GATHER4_PO */ { 6 }, /* VGPU10_OPCODE_GATHER4_PO_C */ { 2 }, /* VGPU10_OPCODE_RCP */ { 2 }, /* VGPU10_OPCODE_F32TOF16 */ { 2 }, /* VGPU10_OPCODE_F16TOF32 */ { 4 }, /* VGPU10_OPCODE_UADDC */ { 4 }, /* VGPU10_OPCODE_USUBB */ { 2 }, /* VGPU10_OPCODE_COUNTBITS */ { 2 }, /* VGPU10_OPCODE_FIRSTBIT_HI */ { 2 }, /* VGPU10_OPCODE_FIRSTBIT_LO */ { 2 }, /* VGPU10_OPCODE_FIRSTBIT_SHI */ { 4 }, /* VGPU10_OPCODE_UBFE */ { 4 }, /* VGPU10_OPCODE_IBFE */ { 5 }, /* VGPU10_OPCODE_BFI */ { 2 }, /* VGPU10_OPCODE_BFREV */ { 5 }, /* VGPU10_OPCODE_SWAPC */ { 1 }, /* VGPU10_OPCODE_DCL_STREAM */ { 0 }, /* VGPU10_OPCODE_DCL_FUNCTION_BODY */ { 0 }, /* VGPU10_OPCODE_DCL_FUNCTION_TABLE */ { 0 }, /* VGPU10_OPCODE_DCL_INTERFACE */ { 0 }, /* VGPU10_OPCODE_DCL_INPUT_CONTROL_POINT_COUNT */ { 0 }, /* VGPU10_OPCODE_DCL_OUTPUT_CONTROL_POINT_COUNT */ { 0 }, /* VGPU10_OPCODE_DCL_TESS_DOMAIN */ { 0 }, /* VGPU10_OPCODE_DCL_TESS_PARTITIONING */ { 0 }, /* VGPU10_OPCODE_DCL_TESS_OUTPUT_PRIMITIVE */ { 0 }, /* VGPU10_OPCODE_DCL_HS_MAX_TESSFACTOR */ { 0 }, /* VGPU10_OPCODE_DCL_HS_FORK_PHASE_INSTANCE_COUNT */ { 0 }, /* VGPU10_OPCODE_DCL_HS_JOIN_PHASE_INSTANCE_COUNT */ { 0 }, /* VGPU10_OPCODE_DCL_THREAD_GROUP */ { 1 }, /* VGPU10_OPCODE_DCL_UAV_TYPED */ { 1 }, /* VGPU10_OPCODE_DCL_UAV_RAW */ { 1 }, /* VGPU10_OPCODE_DCL_UAV_STRUCTURED */ { 1 }, /* VGPU10_OPCODE_DCL_TGSM_RAW */ { 1 }, /* VGPU10_OPCODE_DCL_TGSM_STRUCTURED */ { 1 }, /* VGPU10_OPCODE_DCL_RESOURCE_RAW */ { 1 }, /* VGPU10_OPCODE_DCL_RESOURCE_STRUCTURED */ { 3 }, /* VGPU10_OPCODE_LD_UAV_TYPED */ { 3 }, /* VGPU10_OPCODE_STORE_UAV_TYPED */ { 3 }, /* VGPU10_OPCODE_LD_RAW */ { 3 }, /* VGPU10_OPCODE_STORE_RAW */ { 4 }, /* VGPU10_OPCODE_LD_STRUCTURED */ { 4 }, /* VGPU10_OPCODE_STORE_STRUCTURED */ { 3 }, /* VGPU10_OPCODE_ATOMIC_AND */ { 3 }, /* VGPU10_OPCODE_ATOMIC_OR */ { 3 }, /* VGPU10_OPCODE_ATOMIC_XOR */ { 4 }, /* VGPU10_OPCODE_ATOMIC_CMP_STORE */ { 3 }, /* VGPU10_OPCODE_ATOMIC_IADD */ { 3 }, /* VGPU10_OPCODE_ATOMIC_IMAX */ { 3 }, /* VGPU10_OPCODE_ATOMIC_IMIN */ { 3 }, /* VGPU10_OPCODE_ATOMIC_UMAX */ { 3 }, /* VGPU10_OPCODE_ATOMIC_UMIN */ { 2 }, /* VGPU10_OPCODE_IMM_ATOMIC_ALLOC */ { 2 }, /* VGPU10_OPCODE_IMM_ATOMIC_CONSUME */ { 4 }, /* VGPU10_OPCODE_IMM_ATOMIC_IADD */ { 4 }, /* VGPU10_OPCODE_IMM_ATOMIC_AND */ { 4 }, /* VGPU10_OPCODE_IMM_ATOMIC_OR */ { 4 }, /* VGPU10_OPCODE_IMM_ATOMIC_XOR */ { 4 }, /* VGPU10_OPCODE_IMM_ATOMIC_EXCH */ { 5 }, /* VGPU10_OPCODE_IMM_ATOMIC_CMP_EXCH */ { 4 }, /* VGPU10_OPCODE_IMM_ATOMIC_IMAX */ { 4 }, /* VGPU10_OPCODE_IMM_ATOMIC_IMIN */ { 4 }, /* VGPU10_OPCODE_IMM_ATOMIC_UMAX */ { 4 }, /* VGPU10_OPCODE_IMM_ATOMIC_UMIN */ { 0 }, /* VGPU10_OPCODE_SYNC */ { 3 }, /* VGPU10_OPCODE_DADD */ { 3 }, /* VGPU10_OPCODE_DMAX */ { 3 }, /* VGPU10_OPCODE_DMIN */ { 3 }, /* VGPU10_OPCODE_DMUL */ { 3 }, /* VGPU10_OPCODE_DEQ */ { 3 }, /* VGPU10_OPCODE_DGE */ { 3 }, /* VGPU10_OPCODE_DLT */ { 3 }, /* VGPU10_OPCODE_DNE */ { 2 }, /* VGPU10_OPCODE_DMOV */ { 4 }, /* VGPU10_OPCODE_DMOVC */ { 2 }, /* VGPU10_OPCODE_DTOF */ { 2 }, /* VGPU10_OPCODE_FTOD */ { 3 }, /* VGPU10_OPCODE_EVAL_SNAPPED */ { 3 }, /* VGPU10_OPCODE_EVAL_SAMPLE_INDEX */ { 2 }, /* VGPU10_OPCODE_EVAL_CENTROID */ { 0 }, /* VGPU10_OPCODE_DCL_GS_INSTANCE_COUNT */ { 0 }, /* VGPU10_OPCODE_ABORT */ { 0 }, /* VGPU10_OPCODE_DEBUG_BREAK */ { 0 }, /* VGPU10_OPCODE_RESERVED0 */ { 3 }, /* VGPU10_OPCODE_DDIV */ { 4 }, /* VGPU10_OPCODE_DFMA */ { 2 }, /* VGPU10_OPCODE_DRCP */ { 4 }, /* VGPU10_OPCODE_MSAD */ { 2 }, /* VGPU10_OPCODE_DTOI */ { 2 }, /* VGPU10_OPCODE_DTOU */ { 2 }, /* VGPU10_OPCODE_ITOD */ { 2 }, /* VGPU10_OPCODE_UTOD */ }; AssertCompile(RT_ELEMENTS(g_aOpcodeInfo) == VGPU10_NUM_OPCODES); #ifdef LOG_ENABLED /* * * Helpers to translate a VGPU10 shader constant to a string. * */ #define SVGA_CASE_ID2STR(idx) case idx: return #idx static const char *dxbcOpcodeToString(uint32_t opcodeType) { VGPU10_OPCODE_TYPE enm = (VGPU10_OPCODE_TYPE)opcodeType; switch (enm) { SVGA_CASE_ID2STR(VGPU10_OPCODE_ADD); SVGA_CASE_ID2STR(VGPU10_OPCODE_AND); SVGA_CASE_ID2STR(VGPU10_OPCODE_BREAK); SVGA_CASE_ID2STR(VGPU10_OPCODE_BREAKC); SVGA_CASE_ID2STR(VGPU10_OPCODE_CALL); SVGA_CASE_ID2STR(VGPU10_OPCODE_CALLC); SVGA_CASE_ID2STR(VGPU10_OPCODE_CASE); SVGA_CASE_ID2STR(VGPU10_OPCODE_CONTINUE); SVGA_CASE_ID2STR(VGPU10_OPCODE_CONTINUEC); SVGA_CASE_ID2STR(VGPU10_OPCODE_CUT); SVGA_CASE_ID2STR(VGPU10_OPCODE_DEFAULT); SVGA_CASE_ID2STR(VGPU10_OPCODE_DERIV_RTX); SVGA_CASE_ID2STR(VGPU10_OPCODE_DERIV_RTY); SVGA_CASE_ID2STR(VGPU10_OPCODE_DISCARD); SVGA_CASE_ID2STR(VGPU10_OPCODE_DIV); SVGA_CASE_ID2STR(VGPU10_OPCODE_DP2); SVGA_CASE_ID2STR(VGPU10_OPCODE_DP3); SVGA_CASE_ID2STR(VGPU10_OPCODE_DP4); SVGA_CASE_ID2STR(VGPU10_OPCODE_ELSE); SVGA_CASE_ID2STR(VGPU10_OPCODE_EMIT); SVGA_CASE_ID2STR(VGPU10_OPCODE_EMITTHENCUT); SVGA_CASE_ID2STR(VGPU10_OPCODE_ENDIF); SVGA_CASE_ID2STR(VGPU10_OPCODE_ENDLOOP); SVGA_CASE_ID2STR(VGPU10_OPCODE_ENDSWITCH); SVGA_CASE_ID2STR(VGPU10_OPCODE_EQ); SVGA_CASE_ID2STR(VGPU10_OPCODE_EXP); SVGA_CASE_ID2STR(VGPU10_OPCODE_FRC); SVGA_CASE_ID2STR(VGPU10_OPCODE_FTOI); SVGA_CASE_ID2STR(VGPU10_OPCODE_FTOU); SVGA_CASE_ID2STR(VGPU10_OPCODE_GE); SVGA_CASE_ID2STR(VGPU10_OPCODE_IADD); SVGA_CASE_ID2STR(VGPU10_OPCODE_IF); SVGA_CASE_ID2STR(VGPU10_OPCODE_IEQ); SVGA_CASE_ID2STR(VGPU10_OPCODE_IGE); SVGA_CASE_ID2STR(VGPU10_OPCODE_ILT); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMAD); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMAX); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMIN); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMUL); SVGA_CASE_ID2STR(VGPU10_OPCODE_INE); SVGA_CASE_ID2STR(VGPU10_OPCODE_INEG); SVGA_CASE_ID2STR(VGPU10_OPCODE_ISHL); SVGA_CASE_ID2STR(VGPU10_OPCODE_ISHR); SVGA_CASE_ID2STR(VGPU10_OPCODE_ITOF); SVGA_CASE_ID2STR(VGPU10_OPCODE_LABEL); SVGA_CASE_ID2STR(VGPU10_OPCODE_LD); SVGA_CASE_ID2STR(VGPU10_OPCODE_LD_MS); SVGA_CASE_ID2STR(VGPU10_OPCODE_LOG); SVGA_CASE_ID2STR(VGPU10_OPCODE_LOOP); SVGA_CASE_ID2STR(VGPU10_OPCODE_LT); SVGA_CASE_ID2STR(VGPU10_OPCODE_MAD); SVGA_CASE_ID2STR(VGPU10_OPCODE_MIN); SVGA_CASE_ID2STR(VGPU10_OPCODE_MAX); SVGA_CASE_ID2STR(VGPU10_OPCODE_CUSTOMDATA); SVGA_CASE_ID2STR(VGPU10_OPCODE_MOV); SVGA_CASE_ID2STR(VGPU10_OPCODE_MOVC); SVGA_CASE_ID2STR(VGPU10_OPCODE_MUL); SVGA_CASE_ID2STR(VGPU10_OPCODE_NE); SVGA_CASE_ID2STR(VGPU10_OPCODE_NOP); SVGA_CASE_ID2STR(VGPU10_OPCODE_NOT); SVGA_CASE_ID2STR(VGPU10_OPCODE_OR); SVGA_CASE_ID2STR(VGPU10_OPCODE_RESINFO); SVGA_CASE_ID2STR(VGPU10_OPCODE_RET); SVGA_CASE_ID2STR(VGPU10_OPCODE_RETC); SVGA_CASE_ID2STR(VGPU10_OPCODE_ROUND_NE); SVGA_CASE_ID2STR(VGPU10_OPCODE_ROUND_NI); SVGA_CASE_ID2STR(VGPU10_OPCODE_ROUND_PI); SVGA_CASE_ID2STR(VGPU10_OPCODE_ROUND_Z); SVGA_CASE_ID2STR(VGPU10_OPCODE_RSQ); SVGA_CASE_ID2STR(VGPU10_OPCODE_SAMPLE); SVGA_CASE_ID2STR(VGPU10_OPCODE_SAMPLE_C); SVGA_CASE_ID2STR(VGPU10_OPCODE_SAMPLE_C_LZ); SVGA_CASE_ID2STR(VGPU10_OPCODE_SAMPLE_L); SVGA_CASE_ID2STR(VGPU10_OPCODE_SAMPLE_D); SVGA_CASE_ID2STR(VGPU10_OPCODE_SAMPLE_B); SVGA_CASE_ID2STR(VGPU10_OPCODE_SQRT); SVGA_CASE_ID2STR(VGPU10_OPCODE_SWITCH); SVGA_CASE_ID2STR(VGPU10_OPCODE_SINCOS); SVGA_CASE_ID2STR(VGPU10_OPCODE_UDIV); SVGA_CASE_ID2STR(VGPU10_OPCODE_ULT); SVGA_CASE_ID2STR(VGPU10_OPCODE_UGE); SVGA_CASE_ID2STR(VGPU10_OPCODE_UMUL); SVGA_CASE_ID2STR(VGPU10_OPCODE_UMAD); SVGA_CASE_ID2STR(VGPU10_OPCODE_UMAX); SVGA_CASE_ID2STR(VGPU10_OPCODE_UMIN); SVGA_CASE_ID2STR(VGPU10_OPCODE_USHR); SVGA_CASE_ID2STR(VGPU10_OPCODE_UTOF); SVGA_CASE_ID2STR(VGPU10_OPCODE_XOR); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_RESOURCE); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_CONSTANT_BUFFER); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_SAMPLER); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_INDEX_RANGE); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_GS_OUTPUT_PRIMITIVE_TOPOLOGY); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_GS_INPUT_PRIMITIVE); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_MAX_OUTPUT_VERTEX_COUNT); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_INPUT); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_INPUT_SGV); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_INPUT_SIV); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_INPUT_PS); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_INPUT_PS_SGV); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_INPUT_PS_SIV); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_OUTPUT); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_OUTPUT_SGV); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_OUTPUT_SIV); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_TEMPS); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_INDEXABLE_TEMP); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_GLOBAL_FLAGS); SVGA_CASE_ID2STR(VGPU10_OPCODE_VMWARE); SVGA_CASE_ID2STR(VGPU10_OPCODE_LOD); SVGA_CASE_ID2STR(VGPU10_OPCODE_GATHER4); SVGA_CASE_ID2STR(VGPU10_OPCODE_SAMPLE_POS); SVGA_CASE_ID2STR(VGPU10_OPCODE_SAMPLE_INFO); SVGA_CASE_ID2STR(VGPU10_OPCODE_RESERVED1); SVGA_CASE_ID2STR(VGPU10_OPCODE_HS_DECLS); SVGA_CASE_ID2STR(VGPU10_OPCODE_HS_CONTROL_POINT_PHASE); SVGA_CASE_ID2STR(VGPU10_OPCODE_HS_FORK_PHASE); SVGA_CASE_ID2STR(VGPU10_OPCODE_HS_JOIN_PHASE); SVGA_CASE_ID2STR(VGPU10_OPCODE_EMIT_STREAM); SVGA_CASE_ID2STR(VGPU10_OPCODE_CUT_STREAM); SVGA_CASE_ID2STR(VGPU10_OPCODE_EMITTHENCUT_STREAM); SVGA_CASE_ID2STR(VGPU10_OPCODE_INTERFACE_CALL); SVGA_CASE_ID2STR(VGPU10_OPCODE_BUFINFO); SVGA_CASE_ID2STR(VGPU10_OPCODE_DERIV_RTX_COARSE); SVGA_CASE_ID2STR(VGPU10_OPCODE_DERIV_RTX_FINE); SVGA_CASE_ID2STR(VGPU10_OPCODE_DERIV_RTY_COARSE); SVGA_CASE_ID2STR(VGPU10_OPCODE_DERIV_RTY_FINE); SVGA_CASE_ID2STR(VGPU10_OPCODE_GATHER4_C); SVGA_CASE_ID2STR(VGPU10_OPCODE_GATHER4_PO); SVGA_CASE_ID2STR(VGPU10_OPCODE_GATHER4_PO_C); SVGA_CASE_ID2STR(VGPU10_OPCODE_RCP); SVGA_CASE_ID2STR(VGPU10_OPCODE_F32TOF16); SVGA_CASE_ID2STR(VGPU10_OPCODE_F16TOF32); SVGA_CASE_ID2STR(VGPU10_OPCODE_UADDC); SVGA_CASE_ID2STR(VGPU10_OPCODE_USUBB); SVGA_CASE_ID2STR(VGPU10_OPCODE_COUNTBITS); SVGA_CASE_ID2STR(VGPU10_OPCODE_FIRSTBIT_HI); SVGA_CASE_ID2STR(VGPU10_OPCODE_FIRSTBIT_LO); SVGA_CASE_ID2STR(VGPU10_OPCODE_FIRSTBIT_SHI); SVGA_CASE_ID2STR(VGPU10_OPCODE_UBFE); SVGA_CASE_ID2STR(VGPU10_OPCODE_IBFE); SVGA_CASE_ID2STR(VGPU10_OPCODE_BFI); SVGA_CASE_ID2STR(VGPU10_OPCODE_BFREV); SVGA_CASE_ID2STR(VGPU10_OPCODE_SWAPC); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_STREAM); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_FUNCTION_BODY); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_FUNCTION_TABLE); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_INTERFACE); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_INPUT_CONTROL_POINT_COUNT); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_OUTPUT_CONTROL_POINT_COUNT); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_TESS_DOMAIN); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_TESS_PARTITIONING); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_TESS_OUTPUT_PRIMITIVE); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_HS_MAX_TESSFACTOR); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_HS_FORK_PHASE_INSTANCE_COUNT); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_HS_JOIN_PHASE_INSTANCE_COUNT); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_THREAD_GROUP); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_UAV_TYPED); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_UAV_RAW); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_UAV_STRUCTURED); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_TGSM_RAW); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_TGSM_STRUCTURED); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_RESOURCE_RAW); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_RESOURCE_STRUCTURED); SVGA_CASE_ID2STR(VGPU10_OPCODE_LD_UAV_TYPED); SVGA_CASE_ID2STR(VGPU10_OPCODE_STORE_UAV_TYPED); SVGA_CASE_ID2STR(VGPU10_OPCODE_LD_RAW); SVGA_CASE_ID2STR(VGPU10_OPCODE_STORE_RAW); SVGA_CASE_ID2STR(VGPU10_OPCODE_LD_STRUCTURED); SVGA_CASE_ID2STR(VGPU10_OPCODE_STORE_STRUCTURED); SVGA_CASE_ID2STR(VGPU10_OPCODE_ATOMIC_AND); SVGA_CASE_ID2STR(VGPU10_OPCODE_ATOMIC_OR); SVGA_CASE_ID2STR(VGPU10_OPCODE_ATOMIC_XOR); SVGA_CASE_ID2STR(VGPU10_OPCODE_ATOMIC_CMP_STORE); SVGA_CASE_ID2STR(VGPU10_OPCODE_ATOMIC_IADD); SVGA_CASE_ID2STR(VGPU10_OPCODE_ATOMIC_IMAX); SVGA_CASE_ID2STR(VGPU10_OPCODE_ATOMIC_IMIN); SVGA_CASE_ID2STR(VGPU10_OPCODE_ATOMIC_UMAX); SVGA_CASE_ID2STR(VGPU10_OPCODE_ATOMIC_UMIN); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMM_ATOMIC_ALLOC); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMM_ATOMIC_CONSUME); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMM_ATOMIC_IADD); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMM_ATOMIC_AND); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMM_ATOMIC_OR); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMM_ATOMIC_XOR); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMM_ATOMIC_EXCH); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMM_ATOMIC_CMP_EXCH); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMM_ATOMIC_IMAX); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMM_ATOMIC_IMIN); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMM_ATOMIC_UMAX); SVGA_CASE_ID2STR(VGPU10_OPCODE_IMM_ATOMIC_UMIN); SVGA_CASE_ID2STR(VGPU10_OPCODE_SYNC); SVGA_CASE_ID2STR(VGPU10_OPCODE_DADD); SVGA_CASE_ID2STR(VGPU10_OPCODE_DMAX); SVGA_CASE_ID2STR(VGPU10_OPCODE_DMIN); SVGA_CASE_ID2STR(VGPU10_OPCODE_DMUL); SVGA_CASE_ID2STR(VGPU10_OPCODE_DEQ); SVGA_CASE_ID2STR(VGPU10_OPCODE_DGE); SVGA_CASE_ID2STR(VGPU10_OPCODE_DLT); SVGA_CASE_ID2STR(VGPU10_OPCODE_DNE); SVGA_CASE_ID2STR(VGPU10_OPCODE_DMOV); SVGA_CASE_ID2STR(VGPU10_OPCODE_DMOVC); SVGA_CASE_ID2STR(VGPU10_OPCODE_DTOF); SVGA_CASE_ID2STR(VGPU10_OPCODE_FTOD); SVGA_CASE_ID2STR(VGPU10_OPCODE_EVAL_SNAPPED); SVGA_CASE_ID2STR(VGPU10_OPCODE_EVAL_SAMPLE_INDEX); SVGA_CASE_ID2STR(VGPU10_OPCODE_EVAL_CENTROID); SVGA_CASE_ID2STR(VGPU10_OPCODE_DCL_GS_INSTANCE_COUNT); SVGA_CASE_ID2STR(VGPU10_OPCODE_ABORT); SVGA_CASE_ID2STR(VGPU10_OPCODE_DEBUG_BREAK); SVGA_CASE_ID2STR(VGPU10_OPCODE_RESERVED0); SVGA_CASE_ID2STR(VGPU10_OPCODE_DDIV); SVGA_CASE_ID2STR(VGPU10_OPCODE_DFMA); SVGA_CASE_ID2STR(VGPU10_OPCODE_DRCP); SVGA_CASE_ID2STR(VGPU10_OPCODE_MSAD); SVGA_CASE_ID2STR(VGPU10_OPCODE_DTOI); SVGA_CASE_ID2STR(VGPU10_OPCODE_DTOU); SVGA_CASE_ID2STR(VGPU10_OPCODE_ITOD); SVGA_CASE_ID2STR(VGPU10_OPCODE_UTOD); SVGA_CASE_ID2STR(VGPU10_NUM_OPCODES); } return NULL; } static const char *dxbcShaderTypeToString(uint32_t value) { VGPU10_PROGRAM_TYPE enm = (VGPU10_PROGRAM_TYPE)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_PIXEL_SHADER); SVGA_CASE_ID2STR(VGPU10_VERTEX_SHADER); SVGA_CASE_ID2STR(VGPU10_GEOMETRY_SHADER); SVGA_CASE_ID2STR(VGPU10_HULL_SHADER); SVGA_CASE_ID2STR(VGPU10_DOMAIN_SHADER); SVGA_CASE_ID2STR(VGPU10_COMPUTE_SHADER); } return NULL; } static const char *dxbcCustomDataClassToString(uint32_t value) { VGPU10_CUSTOMDATA_CLASS enm = (VGPU10_CUSTOMDATA_CLASS)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_CUSTOMDATA_COMMENT); SVGA_CASE_ID2STR(VGPU10_CUSTOMDATA_DEBUGINFO); SVGA_CASE_ID2STR(VGPU10_CUSTOMDATA_OPAQUE); SVGA_CASE_ID2STR(VGPU10_CUSTOMDATA_DCL_IMMEDIATE_CONSTANT_BUFFER); } return NULL; } static const char *dxbcSystemNameToString(uint32_t value) { VGPU10_SYSTEM_NAME enm = (VGPU10_SYSTEM_NAME)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_NAME_UNDEFINED); SVGA_CASE_ID2STR(VGPU10_NAME_POSITION); SVGA_CASE_ID2STR(VGPU10_NAME_CLIP_DISTANCE); SVGA_CASE_ID2STR(VGPU10_NAME_CULL_DISTANCE); SVGA_CASE_ID2STR(VGPU10_NAME_RENDER_TARGET_ARRAY_INDEX); SVGA_CASE_ID2STR(VGPU10_NAME_VIEWPORT_ARRAY_INDEX); SVGA_CASE_ID2STR(VGPU10_NAME_VERTEX_ID); SVGA_CASE_ID2STR(VGPU10_NAME_PRIMITIVE_ID); SVGA_CASE_ID2STR(VGPU10_NAME_INSTANCE_ID); SVGA_CASE_ID2STR(VGPU10_NAME_IS_FRONT_FACE); SVGA_CASE_ID2STR(VGPU10_NAME_SAMPLE_INDEX); SVGA_CASE_ID2STR(VGPU10_NAME_FINAL_QUAD_U_EQ_0_EDGE_TESSFACTOR); SVGA_CASE_ID2STR(VGPU10_NAME_FINAL_QUAD_V_EQ_0_EDGE_TESSFACTOR); SVGA_CASE_ID2STR(VGPU10_NAME_FINAL_QUAD_U_EQ_1_EDGE_TESSFACTOR); SVGA_CASE_ID2STR(VGPU10_NAME_FINAL_QUAD_V_EQ_1_EDGE_TESSFACTOR); SVGA_CASE_ID2STR(VGPU10_NAME_FINAL_QUAD_U_INSIDE_TESSFACTOR); SVGA_CASE_ID2STR(VGPU10_NAME_FINAL_QUAD_V_INSIDE_TESSFACTOR); SVGA_CASE_ID2STR(VGPU10_NAME_FINAL_TRI_U_EQ_0_EDGE_TESSFACTOR); SVGA_CASE_ID2STR(VGPU10_NAME_FINAL_TRI_V_EQ_0_EDGE_TESSFACTOR); SVGA_CASE_ID2STR(VGPU10_NAME_FINAL_TRI_W_EQ_0_EDGE_TESSFACTOR); SVGA_CASE_ID2STR(VGPU10_NAME_FINAL_TRI_INSIDE_TESSFACTOR); SVGA_CASE_ID2STR(VGPU10_NAME_FINAL_LINE_DETAIL_TESSFACTOR); SVGA_CASE_ID2STR(VGPU10_NAME_FINAL_LINE_DENSITY_TESSFACTOR); } return NULL; } static const char *dxbcOperandTypeToString(uint32_t value) { VGPU10_OPERAND_TYPE enm = (VGPU10_OPERAND_TYPE)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_TEMP); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_OUTPUT); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INDEXABLE_TEMP); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_IMMEDIATE32); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_IMMEDIATE64); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_SAMPLER); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_RESOURCE); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_CONSTANT_BUFFER); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_IMMEDIATE_CONSTANT_BUFFER); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_LABEL); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT_PRIMITIVEID); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_OUTPUT_DEPTH); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_NULL); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_RASTERIZER); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_OUTPUT_COVERAGE_MASK); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_STREAM); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_FUNCTION_BODY); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_FUNCTION_TABLE); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INTERFACE); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_FUNCTION_INPUT); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_FUNCTION_OUTPUT); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_OUTPUT_CONTROL_POINT_ID); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT_FORK_INSTANCE_ID); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT_JOIN_INSTANCE_ID); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT_CONTROL_POINT); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_OUTPUT_CONTROL_POINT); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT_PATCH_CONSTANT); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT_DOMAIN_POINT); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_THIS_POINTER); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_UAV); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_THREAD_GROUP_SHARED_MEMORY); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT_THREAD_ID); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT_THREAD_GROUP_ID); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT_THREAD_ID_IN_GROUP); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT_COVERAGE_MASK); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT_THREAD_ID_IN_GROUP_FLATTENED); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_INPUT_GS_INSTANCE_ID); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_OUTPUT_DEPTH_GREATER_EQUAL); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_OUTPUT_DEPTH_LESS_EQUAL); SVGA_CASE_ID2STR(VGPU10_OPERAND_TYPE_CYCLE_COUNTER); SVGA_CASE_ID2STR(VGPU10_NUM_OPERANDS); } return NULL; } static const char *dxbcExtendedOperandTypeToString(uint32_t value) { VGPU10_EXTENDED_OPERAND_TYPE enm = (VGPU10_EXTENDED_OPERAND_TYPE)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_EXTENDED_OPERAND_EMPTY); SVGA_CASE_ID2STR(VGPU10_EXTENDED_OPERAND_MODIFIER); } return NULL; } static const char *dxbcOperandModifierToString(uint32_t value) { VGPU10_OPERAND_MODIFIER enm = (VGPU10_OPERAND_MODIFIER)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_OPERAND_MODIFIER_NONE); SVGA_CASE_ID2STR(VGPU10_OPERAND_MODIFIER_NEG); SVGA_CASE_ID2STR(VGPU10_OPERAND_MODIFIER_ABS); SVGA_CASE_ID2STR(VGPU10_OPERAND_MODIFIER_ABSNEG); } return NULL; } static const char *dxbcOperandNumComponentsToString(uint32_t value) { VGPU10_OPERAND_NUM_COMPONENTS enm = (VGPU10_OPERAND_NUM_COMPONENTS)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_OPERAND_0_COMPONENT); SVGA_CASE_ID2STR(VGPU10_OPERAND_1_COMPONENT); SVGA_CASE_ID2STR(VGPU10_OPERAND_4_COMPONENT); SVGA_CASE_ID2STR(VGPU10_OPERAND_N_COMPONENT); } return NULL; } static const char *dxbcOperandComponentModeToString(uint32_t value) { VGPU10_OPERAND_4_COMPONENT_SELECTION_MODE enm = (VGPU10_OPERAND_4_COMPONENT_SELECTION_MODE)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_OPERAND_4_COMPONENT_MASK_MODE); SVGA_CASE_ID2STR(VGPU10_OPERAND_4_COMPONENT_SWIZZLE_MODE); SVGA_CASE_ID2STR(VGPU10_OPERAND_4_COMPONENT_SELECT_1_MODE); } return NULL; } static const char *dxbcOperandComponentNameToString(uint32_t value) { VGPU10_COMPONENT_NAME enm = (VGPU10_COMPONENT_NAME)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_COMPONENT_X); SVGA_CASE_ID2STR(VGPU10_COMPONENT_Y); SVGA_CASE_ID2STR(VGPU10_COMPONENT_Z); SVGA_CASE_ID2STR(VGPU10_COMPONENT_W); } return NULL; } static const char *dxbcOperandIndexDimensionToString(uint32_t value) { VGPU10_OPERAND_INDEX_DIMENSION enm = (VGPU10_OPERAND_INDEX_DIMENSION)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_OPERAND_INDEX_0D); SVGA_CASE_ID2STR(VGPU10_OPERAND_INDEX_1D); SVGA_CASE_ID2STR(VGPU10_OPERAND_INDEX_2D); SVGA_CASE_ID2STR(VGPU10_OPERAND_INDEX_3D); } return NULL; } static const char *dxbcOperandIndexRepresentationToString(uint32_t value) { VGPU10_OPERAND_INDEX_REPRESENTATION enm = (VGPU10_OPERAND_INDEX_REPRESENTATION)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_OPERAND_INDEX_IMMEDIATE32); SVGA_CASE_ID2STR(VGPU10_OPERAND_INDEX_IMMEDIATE64); SVGA_CASE_ID2STR(VGPU10_OPERAND_INDEX_RELATIVE); SVGA_CASE_ID2STR(VGPU10_OPERAND_INDEX_IMMEDIATE32_PLUS_RELATIVE); SVGA_CASE_ID2STR(VGPU10_OPERAND_INDEX_IMMEDIATE64_PLUS_RELATIVE); } return NULL; } static const char *dxbcInterpolationModeToString(uint32_t value) { VGPU10_INTERPOLATION_MODE enm = (VGPU10_INTERPOLATION_MODE)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_INTERPOLATION_UNDEFINED); SVGA_CASE_ID2STR(VGPU10_INTERPOLATION_CONSTANT); SVGA_CASE_ID2STR(VGPU10_INTERPOLATION_LINEAR); SVGA_CASE_ID2STR(VGPU10_INTERPOLATION_LINEAR_CENTROID); SVGA_CASE_ID2STR(VGPU10_INTERPOLATION_LINEAR_NOPERSPECTIVE); SVGA_CASE_ID2STR(VGPU10_INTERPOLATION_LINEAR_NOPERSPECTIVE_CENTROID); SVGA_CASE_ID2STR(VGPU10_INTERPOLATION_LINEAR_SAMPLE); SVGA_CASE_ID2STR(VGPU10_INTERPOLATION_LINEAR_NOPERSPECTIVE_SAMPLE); } return NULL; } static const char *dxbcResourceDimensionToString(uint32_t value) { VGPU10_RESOURCE_DIMENSION enm = (VGPU10_RESOURCE_DIMENSION)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_RESOURCE_DIMENSION_UNKNOWN); SVGA_CASE_ID2STR(VGPU10_RESOURCE_DIMENSION_BUFFER); SVGA_CASE_ID2STR(VGPU10_RESOURCE_DIMENSION_TEXTURE1D); SVGA_CASE_ID2STR(VGPU10_RESOURCE_DIMENSION_TEXTURE2D); SVGA_CASE_ID2STR(VGPU10_RESOURCE_DIMENSION_TEXTURE2DMS); SVGA_CASE_ID2STR(VGPU10_RESOURCE_DIMENSION_TEXTURE3D); SVGA_CASE_ID2STR(VGPU10_RESOURCE_DIMENSION_TEXTURECUBE); SVGA_CASE_ID2STR(VGPU10_RESOURCE_DIMENSION_TEXTURE1DARRAY); SVGA_CASE_ID2STR(VGPU10_RESOURCE_DIMENSION_TEXTURE2DARRAY); SVGA_CASE_ID2STR(VGPU10_RESOURCE_DIMENSION_TEXTURE2DMSARRAY); SVGA_CASE_ID2STR(VGPU10_RESOURCE_DIMENSION_TEXTURECUBEARRAY); } return NULL; } static const char *dxbcVmwareOpcodeTypeToString(uint32_t value) { VGPU10_VMWARE_OPCODE_TYPE enm = (VGPU10_VMWARE_OPCODE_TYPE)value; switch (enm) { SVGA_CASE_ID2STR(VGPU10_VMWARE_OPCODE_IDIV); SVGA_CASE_ID2STR(VGPU10_VMWARE_OPCODE_DFRC); SVGA_CASE_ID2STR(VGPU10_VMWARE_OPCODE_DRSQ); SVGA_CASE_ID2STR(VGPU10_VMWARE_NUM_OPCODES); } return NULL; } #endif /* LOG_ENABLED */ /* * MD5 from IPRT (alt-md5.cpp) for DXBC hash calculation. * DXBC hash function uses a different padding for the data, see dxbcHash. * Therefore RTMd5Final is not needed. Two functions have been renamed: dxbcRTMd5Update dxbcRTMd5Init. */ /* The four core functions - F1 is optimized somewhat */ /* #define F1(x, y, z) (x & y | ~x & z) */ #define F1(x, y, z) (z ^ (x & (y ^ z))) #define F2(x, y, z) F1(z, x, y) #define F3(x, y, z) (x ^ y ^ z) #define F4(x, y, z) (y ^ (x | ~z)) /* This is the central step in the MD5 algorithm. */ #define MD5STEP(f, w, x, y, z, data, s) \ ( w += f(x, y, z) + (data), w = w<>(32-s), w += x ) /** * The core of the MD5 algorithm, this alters an existing MD5 hash to reflect * the addition of 16 longwords of new data. RTMd5Update blocks the data and * converts bytes into longwords for this routine. */ static void rtMd5Transform(uint32_t buf[4], uint32_t const in[16]) { uint32_t a, b, c, d; a = buf[0]; b = buf[1]; c = buf[2]; d = buf[3]; /* fn, w, x, y, z, data, s) */ MD5STEP(F1, a, b, c, d, in[ 0] + 0xd76aa478, 7); MD5STEP(F1, d, a, b, c, in[ 1] + 0xe8c7b756, 12); MD5STEP(F1, c, d, a, b, in[ 2] + 0x242070db, 17); MD5STEP(F1, b, c, d, a, in[ 3] + 0xc1bdceee, 22); MD5STEP(F1, a, b, c, d, in[ 4] + 0xf57c0faf, 7); MD5STEP(F1, d, a, b, c, in[ 5] + 0x4787c62a, 12); MD5STEP(F1, c, d, a, b, in[ 6] + 0xa8304613, 17); MD5STEP(F1, b, c, d, a, in[ 7] + 0xfd469501, 22); MD5STEP(F1, a, b, c, d, in[ 8] + 0x698098d8, 7); MD5STEP(F1, d, a, b, c, in[ 9] + 0x8b44f7af, 12); MD5STEP(F1, c, d, a, b, in[10] + 0xffff5bb1, 17); MD5STEP(F1, b, c, d, a, in[11] + 0x895cd7be, 22); MD5STEP(F1, a, b, c, d, in[12] + 0x6b901122, 7); MD5STEP(F1, d, a, b, c, in[13] + 0xfd987193, 12); MD5STEP(F1, c, d, a, b, in[14] + 0xa679438e, 17); MD5STEP(F1, b, c, d, a, in[15] + 0x49b40821, 22); MD5STEP(F2, a, b, c, d, in[ 1] + 0xf61e2562, 5); MD5STEP(F2, d, a, b, c, in[ 6] + 0xc040b340, 9); MD5STEP(F2, c, d, a, b, in[11] + 0x265e5a51, 14); MD5STEP(F2, b, c, d, a, in[ 0] + 0xe9b6c7aa, 20); MD5STEP(F2, a, b, c, d, in[ 5] + 0xd62f105d, 5); MD5STEP(F2, d, a, b, c, in[10] + 0x02441453, 9); MD5STEP(F2, c, d, a, b, in[15] + 0xd8a1e681, 14); MD5STEP(F2, b, c, d, a, in[ 4] + 0xe7d3fbc8, 20); MD5STEP(F2, a, b, c, d, in[ 9] + 0x21e1cde6, 5); MD5STEP(F2, d, a, b, c, in[14] + 0xc33707d6, 9); MD5STEP(F2, c, d, a, b, in[ 3] + 0xf4d50d87, 14); MD5STEP(F2, b, c, d, a, in[ 8] + 0x455a14ed, 20); MD5STEP(F2, a, b, c, d, in[13] + 0xa9e3e905, 5); MD5STEP(F2, d, a, b, c, in[ 2] + 0xfcefa3f8, 9); MD5STEP(F2, c, d, a, b, in[ 7] + 0x676f02d9, 14); MD5STEP(F2, b, c, d, a, in[12] + 0x8d2a4c8a, 20); MD5STEP(F3, a, b, c, d, in[ 5] + 0xfffa3942, 4); MD5STEP(F3, d, a, b, c, in[ 8] + 0x8771f681, 11); MD5STEP(F3, c, d, a, b, in[11] + 0x6d9d6122, 16); MD5STEP(F3, b, c, d, a, in[14] + 0xfde5380c, 23); MD5STEP(F3, a, b, c, d, in[ 1] + 0xa4beea44, 4); MD5STEP(F3, d, a, b, c, in[ 4] + 0x4bdecfa9, 11); MD5STEP(F3, c, d, a, b, in[ 7] + 0xf6bb4b60, 16); MD5STEP(F3, b, c, d, a, in[10] + 0xbebfbc70, 23); MD5STEP(F3, a, b, c, d, in[13] + 0x289b7ec6, 4); MD5STEP(F3, d, a, b, c, in[ 0] + 0xeaa127fa, 11); MD5STEP(F3, c, d, a, b, in[ 3] + 0xd4ef3085, 16); MD5STEP(F3, b, c, d, a, in[ 6] + 0x04881d05, 23); MD5STEP(F3, a, b, c, d, in[ 9] + 0xd9d4d039, 4); MD5STEP(F3, d, a, b, c, in[12] + 0xe6db99e5, 11); MD5STEP(F3, c, d, a, b, in[15] + 0x1fa27cf8, 16); MD5STEP(F3, b, c, d, a, in[ 2] + 0xc4ac5665, 23); MD5STEP(F4, a, b, c, d, in[ 0] + 0xf4292244, 6); MD5STEP(F4, d, a, b, c, in[ 7] + 0x432aff97, 10); MD5STEP(F4, c, d, a, b, in[14] + 0xab9423a7, 15); MD5STEP(F4, b, c, d, a, in[ 5] + 0xfc93a039, 21); MD5STEP(F4, a, b, c, d, in[12] + 0x655b59c3, 6); MD5STEP(F4, d, a, b, c, in[ 3] + 0x8f0ccc92, 10); MD5STEP(F4, c, d, a, b, in[10] + 0xffeff47d, 15); MD5STEP(F4, b, c, d, a, in[ 1] + 0x85845dd1, 21); MD5STEP(F4, a, b, c, d, in[ 8] + 0x6fa87e4f, 6); MD5STEP(F4, d, a, b, c, in[15] + 0xfe2ce6e0, 10); MD5STEP(F4, c, d, a, b, in[ 6] + 0xa3014314, 15); MD5STEP(F4, b, c, d, a, in[13] + 0x4e0811a1, 21); MD5STEP(F4, a, b, c, d, in[ 4] + 0xf7537e82, 6); MD5STEP(F4, d, a, b, c, in[11] + 0xbd3af235, 10); MD5STEP(F4, c, d, a, b, in[ 2] + 0x2ad7d2bb, 15); MD5STEP(F4, b, c, d, a, in[ 9] + 0xeb86d391, 21); buf[0] += a; buf[1] += b; buf[2] += c; buf[3] += d; } #ifdef RT_BIG_ENDIAN /* * Note: this code is harmless on little-endian machines. */ static void rtMd5ByteReverse(uint32_t *buf, unsigned int longs) { uint32_t t; do { t = *buf; t = RT_LE2H_U32(t); *buf = t; buf++; } while (--longs); } #else /* little endian - do nothing */ # define rtMd5ByteReverse(buf, len) do { /* Nothing */ } while (0) #endif /* * Start MD5 accumulation. Set bit count to 0 and buffer to mysterious * initialization constants. */ static void dxbcRTMd5Init(PRTMD5CONTEXT pCtx) { pCtx->AltPrivate.buf[0] = 0x67452301; pCtx->AltPrivate.buf[1] = 0xefcdab89; pCtx->AltPrivate.buf[2] = 0x98badcfe; pCtx->AltPrivate.buf[3] = 0x10325476; pCtx->AltPrivate.bits[0] = 0; pCtx->AltPrivate.bits[1] = 0; } /* * Update context to reflect the concatenation of another buffer full * of bytes. */ /** @todo Optimize this, because len is always a multiple of 64. */ static void dxbcRTMd5Update(PRTMD5CONTEXT pCtx, const void *pvBuf, size_t len) { const uint8_t *buf = (const uint8_t *)pvBuf; uint32_t t; /* Update bitcount */ t = pCtx->AltPrivate.bits[0]; if ((pCtx->AltPrivate.bits[0] = t + ((uint32_t) len << 3)) < t) pCtx->AltPrivate.bits[1]++; /* Carry from low to high */ pCtx->AltPrivate.bits[1] += (uint32_t)(len >> 29); t = (t >> 3) & 0x3f; /* Bytes already in shsInfo->data */ /* Handle any leading odd-sized chunks */ if (t) { uint8_t *p = (uint8_t *) pCtx->AltPrivate.in + t; t = 64 - t; if (len < t) { memcpy(p, buf, len); return; } memcpy(p, buf, t); rtMd5ByteReverse(pCtx->AltPrivate.in, 16); rtMd5Transform(pCtx->AltPrivate.buf, pCtx->AltPrivate.in); buf += t; len -= t; } /* Process data in 64-byte chunks */ #ifndef RT_BIG_ENDIAN if (!((uintptr_t)buf & 0x3)) { while (len >= 64) { rtMd5Transform(pCtx->AltPrivate.buf, (uint32_t const *)buf); buf += 64; len -= 64; } } else #endif { while (len >= 64) { memcpy(pCtx->AltPrivate.in, buf, 64); rtMd5ByteReverse(pCtx->AltPrivate.in, 16); rtMd5Transform(pCtx->AltPrivate.buf, pCtx->AltPrivate.in); buf += 64; len -= 64; } } /* Handle any remaining bytes of data */ memcpy(pCtx->AltPrivate.in, buf, len); } static void dxbcHash(void const *pvData, uint32_t cbData, uint8_t pabDigest[RTMD5HASHSIZE]) { size_t const kBlockSize = 64; uint8_t au8BlockBuffer[kBlockSize]; static uint8_t const s_au8Padding[kBlockSize] = { 0x80, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, }; RTMD5CONTEXT Ctx; PRTMD5CONTEXT const pCtx = &Ctx; dxbcRTMd5Init(pCtx); uint8_t const *pu8Data = (uint8_t *)pvData; size_t cbRemaining = cbData; size_t const cbCompleteBlocks = cbData & ~ (kBlockSize - 1); dxbcRTMd5Update(pCtx, pu8Data, cbCompleteBlocks); pu8Data += cbCompleteBlocks; cbRemaining -= cbCompleteBlocks; /* Custom padding. */ if (cbRemaining >= kBlockSize - 2 * sizeof(uint32_t)) { /* Two additional blocks. */ memcpy(&au8BlockBuffer[0], pu8Data, cbRemaining); memcpy(&au8BlockBuffer[cbRemaining], s_au8Padding, kBlockSize - cbRemaining); dxbcRTMd5Update(pCtx, au8BlockBuffer, kBlockSize); memset(&au8BlockBuffer[sizeof(uint32_t)], 0, kBlockSize - 2 * sizeof(uint32_t)); } else { /* One additional block. */ memcpy(&au8BlockBuffer[sizeof(uint32_t)], pu8Data, cbRemaining); memcpy(&au8BlockBuffer[sizeof(uint32_t) + cbRemaining], s_au8Padding, kBlockSize - cbRemaining - 2 * sizeof(uint32_t)); } /* Set the first and last dwords of the last block. */ *(uint32_t *)&au8BlockBuffer[0] = cbData << 3; *(uint32_t *)&au8BlockBuffer[kBlockSize - sizeof(uint32_t)] = (cbData << 1) | 1; dxbcRTMd5Update(pCtx, au8BlockBuffer, kBlockSize); AssertCompile(sizeof(pCtx->AltPrivate.buf) == RTMD5HASHSIZE); memcpy(pabDigest, pCtx->AltPrivate.buf, RTMD5HASHSIZE); } /* * * Shader token reader. * */ typedef struct DXBCTokenReader { uint32_t const *pToken; /* Next token to read. */ uint32_t cToken; /* How many tokens total. */ uint32_t cRemainingToken; /* How many tokens remain. */ } DXBCTokenReader; #ifdef LOG_ENABLED DECLINLINE(uint32_t) dxbcTokenReaderByteOffset(DXBCTokenReader *r) { return (r->cToken - r->cRemainingToken) * 4; } #endif #if 0 // Unused for now DECLINLINE(uint32_t) dxbcTokenReaderRemaining(DXBCTokenReader *r) { return r->cRemainingToken; } #endif DECLINLINE(uint32_t const *) dxbcTokenReaderPtr(DXBCTokenReader *r) { return r->pToken; } DECLINLINE(bool) dxbcTokenReaderCanRead(DXBCTokenReader *r, uint32_t cToken) { return cToken <= r->cRemainingToken; } DECLINLINE(void) dxbcTokenReaderSkip(DXBCTokenReader *r, uint32_t cToken) { AssertReturnVoid(r->cRemainingToken >= cToken); r->cRemainingToken -= cToken; r->pToken += cToken; } DECLINLINE(uint32_t) dxbcTokenReaderRead32(DXBCTokenReader *r) { AssertReturn(r->cRemainingToken, 0); --r->cRemainingToken; return *(r->pToken++); } DECLINLINE(uint64_t) dxbcTokenReaderRead64(DXBCTokenReader *r) { uint64_t const u64Low = dxbcTokenReaderRead32(r); uint64_t const u64High = dxbcTokenReaderRead32(r); return u64Low + (u64High << 32); } /* * * Byte writer. * */ typedef struct DXBCByteWriter { uint8_t *pu8ByteCodeBegin; /* First byte of the buffer. */ uint8_t *pu8ByteCodePtr; /* Next byte to be written. */ uint32_t cbAllocated; /* How many bytes allocated in the buffer. */ uint32_t cbRemaining; /* How many bytes remain in the buffer. */ uint32_t cbWritten; /* Offset of first never written byte. * Since the writer allows to jump in the buffer, this field tracks * the upper boundary of the written data. */ int32_t rc; } DXBCByteWriter; typedef struct DXBCByteWriterState { uint32_t off; /* Offset of the next free byte. */ } DXBCByteWriterState; DECLINLINE(void *) dxbcByteWriterPtr(DXBCByteWriter *w) { return w->pu8ByteCodePtr; } DECLINLINE(uint32_t) dxbcByteWriterSize(DXBCByteWriter *w) { return (uint32_t)(w->pu8ByteCodePtr - w->pu8ByteCodeBegin); } static bool dxbcByteWriterRealloc(DXBCByteWriter *w, uint32_t cbNew) { void *pvNew = RTMemAllocZ(cbNew); if (!pvNew) { w->rc = VERR_NO_MEMORY; return false; } uint32_t const cbCurrent = dxbcByteWriterSize(w); if (cbCurrent) { memcpy(pvNew, w->pu8ByteCodeBegin, cbCurrent); RTMemFree(w->pu8ByteCodeBegin); } else Assert(w->pu8ByteCodeBegin == NULL); w->pu8ByteCodeBegin = (uint8_t *)pvNew; w->pu8ByteCodePtr = w->pu8ByteCodeBegin + cbCurrent; w->cbAllocated = cbNew; w->cbRemaining = cbNew - cbCurrent; return true; } DECLINLINE(bool) dxbcByteWriterSetOffset(DXBCByteWriter *w, uint32_t off, DXBCByteWriterState *pSavedWriterState) { if (RT_FAILURE(w->rc)) return false; uint32_t const cbNew = RT_ALIGN_32(off, 1024); uint32_t const cbMax = 2 * SVGA3D_MAX_SHADER_MEMORY_BYTES; AssertReturnStmt(off < cbMax && cbNew < cbMax, w->rc = VERR_INVALID_PARAMETER, false); if (cbNew > w->cbAllocated) { if (!dxbcByteWriterRealloc(w, cbNew)) return false; } pSavedWriterState->off = dxbcByteWriterSize(w); w->pu8ByteCodePtr = w->pu8ByteCodeBegin + off; w->cbRemaining = w->cbAllocated - off; return true; } DECLINLINE(void) dxbcByteWriterRestore(DXBCByteWriter *w, DXBCByteWriterState *pSavedWriterState) { w->pu8ByteCodePtr = w->pu8ByteCodeBegin + pSavedWriterState->off; w->cbRemaining = w->cbAllocated - pSavedWriterState->off; } DECLINLINE(void) dxbcByteWriterCommit(DXBCByteWriter *w, uint32_t cbCommit) { if (RT_FAILURE(w->rc)) return; Assert(cbCommit < w->cbRemaining); cbCommit = RT_MIN(cbCommit, w->cbRemaining); w->pu8ByteCodePtr += cbCommit; w->cbRemaining -= cbCommit; w->cbWritten = RT_MAX(w->cbWritten, w->cbAllocated - w->cbRemaining); } DECLINLINE(bool) dxbcByteWriterCanWrite(DXBCByteWriter *w, uint32_t cbMore) { if (RT_FAILURE(w->rc)) return false; if (cbMore <= w->cbRemaining) return true; /* Do not allow to allocate more than 2 * SVGA3D_MAX_SHADER_MEMORY_BYTES */ uint32_t const cbMax = 2 * SVGA3D_MAX_SHADER_MEMORY_BYTES; AssertReturnStmt(cbMore < cbMax && RT_ALIGN_32(cbMore, 4096) <= cbMax - w->cbAllocated, w->rc = VERR_INVALID_PARAMETER, false); uint32_t cbNew = w->cbAllocated + RT_ALIGN_32(cbMore, 4096); return dxbcByteWriterRealloc(w, cbNew); } DECLINLINE(bool) dxbcByteWriterAddTokens(DXBCByteWriter *w, uint32_t const *paToken, uint32_t cToken) { uint32_t const cbWrite = cToken * sizeof(uint32_t); if (dxbcByteWriterCanWrite(w, cbWrite)) { memcpy(dxbcByteWriterPtr(w), paToken, cbWrite); dxbcByteWriterCommit(w, cbWrite); return true; } AssertFailed(); return false; } DECLINLINE(bool) dxbcByteWriterInit(DXBCByteWriter *w, uint32_t cbInitial) { RT_ZERO(*w); return dxbcByteWriterCanWrite(w, cbInitial); } DECLINLINE(void) dxbcByteWriterReset(DXBCByteWriter *w) { RTMemFree(w->pu8ByteCodeBegin); RT_ZERO(*w); } DECLINLINE(void) dxbcByteWriterFetchData(DXBCByteWriter *w, void **ppv, uint32_t *pcb) { *ppv = w->pu8ByteCodeBegin; *pcb = w->cbWritten; w->pu8ByteCodeBegin = NULL; dxbcByteWriterReset(w); } /* * * VGPU10 shader parser. * */ /* Parse an instruction operand. */ static int dxbcParseOperand(DXBCTokenReader *r, VGPUOperand *paOperand, uint32_t *pcOperandRemain) { ASSERT_GUEST_RETURN(*pcOperandRemain > 0, VERR_NOT_SUPPORTED); ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); paOperand->paOperandToken = dxbcTokenReaderPtr(r); paOperand->cOperandToken = 0; VGPU10OperandToken0 operand0; operand0.value = dxbcTokenReaderRead32(r); Log6((" %s(%d) %s(%d) %s(%d) %s(%d)\n", dxbcOperandNumComponentsToString(operand0.numComponents), operand0.numComponents, dxbcOperandComponentModeToString(operand0.selectionMode), operand0.selectionMode, dxbcOperandTypeToString(operand0.operandType), operand0.operandType, dxbcOperandIndexDimensionToString(operand0.indexDimension), operand0.indexDimension)); ASSERT_GUEST_RETURN(operand0.numComponents <= VGPU10_OPERAND_4_COMPONENT, VERR_INVALID_PARAMETER); if ( operand0.operandType != VGPU10_OPERAND_TYPE_IMMEDIATE32 && operand0.operandType != VGPU10_OPERAND_TYPE_IMMEDIATE64) { if (operand0.numComponents == VGPU10_OPERAND_4_COMPONENT) { ASSERT_GUEST_RETURN(operand0.selectionMode <= VGPU10_OPERAND_4_COMPONENT_SELECT_1_MODE, VERR_INVALID_PARAMETER); switch (operand0.selectionMode) { case VGPU10_OPERAND_4_COMPONENT_MASK_MODE: Log6((" Mask %#x\n", operand0.mask)); break; case VGPU10_OPERAND_4_COMPONENT_SWIZZLE_MODE: Log6((" Swizzle %s(%d) %s(%d) %s(%d) %s(%d)\n", dxbcOperandComponentNameToString(operand0.swizzleX), operand0.swizzleX, dxbcOperandComponentNameToString(operand0.swizzleY), operand0.swizzleY, dxbcOperandComponentNameToString(operand0.swizzleZ), operand0.swizzleZ, dxbcOperandComponentNameToString(operand0.swizzleW), operand0.swizzleW)); break; case VGPU10_OPERAND_4_COMPONENT_SELECT_1_MODE: Log6((" Select %s(%d)\n", dxbcOperandComponentNameToString(operand0.selectMask), operand0.selectMask)); break; default: /* Never happens. */ break; } } } if (operand0.extended) { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); VGPU10OperandToken1 operand1; operand1.value = dxbcTokenReaderRead32(r); Log6((" %s(%d) %s(%d)\n", dxbcExtendedOperandTypeToString(operand1.extendedOperandType), operand1.extendedOperandType, dxbcOperandModifierToString(operand1.operandModifier), operand1.operandModifier)); } ASSERT_GUEST_RETURN(operand0.operandType < VGPU10_NUM_OPERANDS, VERR_INVALID_PARAMETER); if ( operand0.operandType == VGPU10_OPERAND_TYPE_IMMEDIATE32 || operand0.operandType == VGPU10_OPERAND_TYPE_IMMEDIATE64) { uint32_t cComponent = 0; if (operand0.numComponents == VGPU10_OPERAND_4_COMPONENT) cComponent = 4; else if (operand0.numComponents == VGPU10_OPERAND_1_COMPONENT) cComponent = 1; for (uint32_t i = 0; i < cComponent; ++i) { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); paOperand->aImm[i] = dxbcTokenReaderRead32(r); } } paOperand->numComponents = operand0.numComponents; paOperand->selectionMode = operand0.selectionMode; paOperand->mask = operand0.mask; paOperand->operandType = operand0.operandType; paOperand->indexDimension = operand0.indexDimension; int rc = VINF_SUCCESS; /* 'indexDimension' tells the number of indices. 'i' is the array index, i.e. i = 0 for 1D, etc. */ for (uint32_t i = 0; i < operand0.indexDimension; ++i) { if (i == 0) /* VGPU10_OPERAND_INDEX_1D */ paOperand->aOperandIndex[i].indexRepresentation = operand0.index0Representation; else if (i == 1) /* VGPU10_OPERAND_INDEX_2D */ paOperand->aOperandIndex[i].indexRepresentation = operand0.index1Representation; else /* VGPU10_OPERAND_INDEX_3D */ continue; /* Skip because it is "rarely if ever used" and is not supported by VGPU10. */ uint32_t const indexRepresentation = paOperand->aOperandIndex[i].indexRepresentation; switch (indexRepresentation) { case VGPU10_OPERAND_INDEX_IMMEDIATE32: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); paOperand->aOperandIndex[i].iOperandImmediate = dxbcTokenReaderRead32(r); break; } case VGPU10_OPERAND_INDEX_IMMEDIATE64: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 2), VERR_INVALID_PARAMETER); paOperand->aOperandIndex[i].iOperandImmediate = dxbcTokenReaderRead64(r); break; } case VGPU10_OPERAND_INDEX_RELATIVE: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); paOperand->aOperandIndex[i].pOperandRelative = &paOperand[1]; Log6((" [operand index %d] parsing relative\n", i)); rc = dxbcParseOperand(r, &paOperand[1], pcOperandRemain); break; } case VGPU10_OPERAND_INDEX_IMMEDIATE32_PLUS_RELATIVE: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 2), VERR_INVALID_PARAMETER); paOperand->aOperandIndex[i].iOperandImmediate = dxbcTokenReaderRead32(r); paOperand->aOperandIndex[i].pOperandRelative = &paOperand[1]; Log6((" [operand index %d] parsing relative\n", i)); rc = dxbcParseOperand(r, &paOperand[1], pcOperandRemain); break; } case VGPU10_OPERAND_INDEX_IMMEDIATE64_PLUS_RELATIVE: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 3), VERR_INVALID_PARAMETER); paOperand->aOperandIndex[i].iOperandImmediate = dxbcTokenReaderRead64(r); paOperand->aOperandIndex[i].pOperandRelative = &paOperand[1]; Log6((" [operand index %d] parsing relative\n", i)); rc = dxbcParseOperand(r, &paOperand[1], pcOperandRemain); break; } default: ASSERT_GUEST_FAILED_RETURN(VERR_INVALID_PARAMETER); } Log6((" [operand index %d] %s(%d): %#llx%s\n", i, dxbcOperandIndexRepresentationToString(indexRepresentation), indexRepresentation, paOperand->aOperandIndex[i].iOperandImmediate, paOperand->aOperandIndex[i].pOperandRelative ? " + relative" : "")); if (RT_FAILURE(rc)) break; } paOperand->cOperandToken = dxbcTokenReaderPtr(r) - paOperand->paOperandToken; *pcOperandRemain -= 1; return VINF_SUCCESS; } /* Parse an instruction. */ static int dxbcParseOpcode(DXBCTokenReader *r, VGPUOpcode *pOpcode) { RT_ZERO(*pOpcode); ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); pOpcode->paOpcodeToken = dxbcTokenReaderPtr(r); VGPU10OpcodeToken0 opcode; opcode.value = dxbcTokenReaderRead32(r); pOpcode->opcodeType = opcode.opcodeType; ASSERT_GUEST_RETURN(pOpcode->opcodeType < VGPU10_NUM_OPCODES, VERR_INVALID_PARAMETER); Log6(("[%#x] %s length %d\n", dxbcTokenReaderByteOffset(r) - 4, dxbcOpcodeToString(pOpcode->opcodeType), opcode.instructionLength)); uint32_t const cOperand = g_aOpcodeInfo[pOpcode->opcodeType].cOperand; if (cOperand != UINT32_MAX) { ASSERT_GUEST_RETURN(cOperand < RT_ELEMENTS(pOpcode->aIdxOperand), VERR_INVALID_PARAMETER); pOpcode->cOpcodeToken = opcode.instructionLength; uint32_t cOpcode = 1; /* Opcode token + extended opcode tokens. */ if (opcode.extended) { if ( pOpcode->opcodeType == VGPU10_OPCODE_DCL_FUNCTION_BODY || pOpcode->opcodeType == VGPU10_OPCODE_DCL_FUNCTION_TABLE || pOpcode->opcodeType == VGPU10_OPCODE_DCL_INTERFACE || pOpcode->opcodeType == VGPU10_OPCODE_INTERFACE_CALL || pOpcode->opcodeType == VGPU10_OPCODE_DCL_THREAD_GROUP) { /* "next DWORD contains ... the actual instruction length in DWORD since it may not fit into 7 bits" */ ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); pOpcode->cOpcodeToken = dxbcTokenReaderRead32(r); ++cOpcode; } else { VGPU10OpcodeToken1 opcode1; do { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); opcode1.value = dxbcTokenReaderRead32(r); ++cOpcode; ASSERT_GUEST( opcode1.opcodeType == VGPU10_EXTENDED_OPCODE_SAMPLE_CONTROLS || opcode1.opcodeType == D3D11_SB_EXTENDED_OPCODE_RESOURCE_DIM || opcode1.opcodeType == D3D11_SB_EXTENDED_OPCODE_RESOURCE_RETURN_TYPE); } while(opcode1.extended); } } ASSERT_GUEST_RETURN(pOpcode->cOpcodeToken >= 1 && pOpcode->cOpcodeToken < 256, VERR_INVALID_PARAMETER); ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, pOpcode->cOpcodeToken - cOpcode), VERR_INVALID_PARAMETER); #ifdef LOG_ENABLED Log6((" %08X", opcode.value)); for (uint32_t i = 1; i < pOpcode->cOpcodeToken; ++i) Log6((" %08X", r->pToken[i - 1])); Log6(("\n")); if (pOpcode->opcodeType == VGPU10_OPCODE_DCL_RESOURCE) Log6((" %s\n", dxbcResourceDimensionToString(opcode.resourceDimension))); else Log6((" %s\n", dxbcInterpolationModeToString(opcode.interpolationMode))); #endif /* Additional tokens before operands. */ switch (pOpcode->opcodeType) { case VGPU10_OPCODE_INTERFACE_CALL: ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 1); /* Function index */ break; default: break; } /* Operands. */ uint32_t cOperandRemain = RT_ELEMENTS(pOpcode->aValOperand); for (uint32_t i = 0; i < cOperand; ++i) { Log6((" [operand %d]\n", i)); uint32_t const idxOperand = RT_ELEMENTS(pOpcode->aValOperand) - cOperandRemain; pOpcode->aIdxOperand[i] = idxOperand; int rc = dxbcParseOperand(r, &pOpcode->aValOperand[idxOperand], &cOperandRemain); ASSERT_GUEST_RETURN(RT_SUCCESS(rc), VERR_INVALID_PARAMETER); } pOpcode->cOperand = cOperand; /* Additional tokens after operands. */ switch (pOpcode->opcodeType) { case VGPU10_OPCODE_DCL_INPUT_SIV: case VGPU10_OPCODE_DCL_INPUT_SGV: case VGPU10_OPCODE_DCL_INPUT_PS_SIV: case VGPU10_OPCODE_DCL_INPUT_PS_SGV: case VGPU10_OPCODE_DCL_OUTPUT_SIV: case VGPU10_OPCODE_DCL_OUTPUT_SGV: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); VGPU10NameToken name; name.value = dxbcTokenReaderRead32(r); Log6((" %s(%d)\n", dxbcSystemNameToString(name.name), name.name)); pOpcode->semanticName = name.name; break; } case VGPU10_OPCODE_DCL_RESOURCE: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 1); /* ResourceReturnTypeToken */ break; } case VGPU10_OPCODE_DCL_TEMPS: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 1); /* number of temps */ break; } case VGPU10_OPCODE_DCL_INDEXABLE_TEMP: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 3), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 3); /* register index; number of registers; number of components */ break; } case VGPU10_OPCODE_DCL_INDEX_RANGE: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 1); /* count of registers */ break; } case VGPU10_OPCODE_DCL_MAX_OUTPUT_VERTEX_COUNT: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 1); /* maximum number of primitives */ break; } case VGPU10_OPCODE_DCL_GS_INSTANCE_COUNT: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 1); /* number of instances */ break; } case VGPU10_OPCODE_DCL_HS_MAX_TESSFACTOR: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 1); /* maximum TessFactor */ break; } case VGPU10_OPCODE_DCL_HS_FORK_PHASE_INSTANCE_COUNT: case VGPU10_OPCODE_DCL_HS_JOIN_PHASE_INSTANCE_COUNT: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 1); /* number of instances of the current fork/join phase program to execute */ break; } case VGPU10_OPCODE_DCL_THREAD_GROUP: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 3), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 3); /* Thread Group dimensions as UINT32: x, y, z */ break; } case VGPU10_OPCODE_DCL_UAV_TYPED: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 1); /* ResourceReturnTypeToken */ break; } case VGPU10_OPCODE_DCL_UAV_STRUCTURED: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 1); /* byte stride */ break; } case VGPU10_OPCODE_DCL_TGSM_RAW: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 1); /* element count */ break; } case VGPU10_OPCODE_DCL_TGSM_STRUCTURED: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 2), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 2); /* struct byte stride; struct count */ break; } case VGPU10_OPCODE_DCL_RESOURCE_STRUCTURED: { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); dxbcTokenReaderSkip(r, 1); /* struct byte stride */ break; } default: break; } } else { /* Special opcodes. */ if (pOpcode->opcodeType == VGPU10_OPCODE_CUSTOMDATA) { ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, 1), VERR_INVALID_PARAMETER); pOpcode->cOpcodeToken = dxbcTokenReaderRead32(r); if (pOpcode->cOpcodeToken < 2) pOpcode->cOpcodeToken = 2; ASSERT_GUEST_RETURN(dxbcTokenReaderCanRead(r, pOpcode->cOpcodeToken - 2), VERR_INVALID_PARAMETER); #ifdef LOG_ENABLED Log6((" %08X", opcode.value)); for (uint32_t i = 1; i < pOpcode->cOpcodeToken; ++i) Log6((" %08X", r->pToken[i - 1])); Log6(("\n")); Log6((" %s\n", dxbcCustomDataClassToString(opcode.customDataClass))); #endif dxbcTokenReaderSkip(r, pOpcode->cOpcodeToken - 2); } else if (pOpcode->opcodeType == VGPU10_OPCODE_VMWARE) { pOpcode->cOpcodeToken = opcode.instructionLength; pOpcode->opcodeSubtype = opcode.vmwareOpcodeType; #ifdef LOG_ENABLED Log6((" %08X", opcode.value)); for (uint32_t i = 1; i < pOpcode->cOpcodeToken; ++i) Log6((" %08X", r->pToken[i - 1])); Log6(("\n")); Log6((" %s(%d)\n", dxbcVmwareOpcodeTypeToString(opcode.vmwareOpcodeType), opcode.vmwareOpcodeType)); #endif if (opcode.vmwareOpcodeType == VGPU10_VMWARE_OPCODE_IDIV) { /* Integer divide. */ pOpcode->cOperand = 4; /* dstQuit, dstRem, src0, src1. */ } else if (opcode.vmwareOpcodeType == VGPU10_VMWARE_OPCODE_DFRC) { /* Double precision fraction. */ pOpcode->cOperand = 2; /* dst, src. */ } else if (opcode.vmwareOpcodeType == VGPU10_VMWARE_OPCODE_DRSQ) { /* Double precision reciprocal square root. */ pOpcode->cOperand = 2; /* dst, src. */ } else { ASSERT_GUEST_FAILED_RETURN(VERR_INVALID_PARAMETER); } /* Operands. */ uint32_t cOperandRemain = RT_ELEMENTS(pOpcode->aValOperand); for (uint32_t i = 0; i < pOpcode->cOperand; ++i) { Log6((" [operand %d]\n", i)); uint32_t const idxOperand = RT_ELEMENTS(pOpcode->aValOperand) - cOperandRemain; pOpcode->aIdxOperand[i] = idxOperand; int rc = dxbcParseOperand(r, &pOpcode->aValOperand[idxOperand], &cOperandRemain); ASSERT_GUEST_RETURN(RT_SUCCESS(rc), VERR_INVALID_PARAMETER); } } else ASSERT_GUEST_FAILED_RETURN(VERR_INVALID_PARAMETER); // pOpcode->cOperand = 0; } return VINF_SUCCESS; } typedef struct DXBCOUTPUTCTX { VGPU10ProgramToken programToken; uint32_t cToken; /* Number of tokens in the original shader code. */ uint32_t offSubroutine; /* Current offset where to write subroutines. */ } DXBCOUTPUTCTX; static void dxbcOutputInit(DXBCOUTPUTCTX *pOutctx, VGPU10ProgramToken const *pProgramToken, uint32_t cToken) { RT_ZERO(*pOutctx); pOutctx->programToken = *pProgramToken; pOutctx->cToken = cToken; pOutctx->offSubroutine = cToken * 4; } static void dxbcEmitCall(DXBCByteWriter *w, VGPUOpcode const *pOpcode, uint32_t label) { VGPU10OpcodeToken0 opcode; VGPU10OperandToken0 operand; opcode.value = 0; opcode.opcodeType = VGPU10_OPCODE_CALL; opcode.instructionLength = 3; dxbcByteWriterAddTokens(w, &opcode.value, 1); operand.value = 0; operand.numComponents = VGPU10_OPERAND_1_COMPONENT; operand.operandType = VGPU10_OPERAND_TYPE_LABEL; operand.indexDimension = VGPU10_OPERAND_INDEX_1D; operand.index0Representation = VGPU10_OPERAND_INDEX_IMMEDIATE32; dxbcByteWriterAddTokens(w, &operand.value, 1); dxbcByteWriterAddTokens(w, &label, 1); opcode.value = 0; opcode.opcodeType = VGPU10_OPCODE_NOP; opcode.instructionLength = 1; for (unsigned i = 0; i < pOpcode->cOpcodeToken - 3; ++i) dxbcByteWriterAddTokens(w, &opcode.value, 1); } static void dxbcEmitLabel(DXBCByteWriter *w, uint32_t label) { VGPU10OpcodeToken0 opcode; VGPU10OperandToken0 operand; opcode.value = 0; opcode.opcodeType = VGPU10_OPCODE_LABEL; opcode.instructionLength = 3; dxbcByteWriterAddTokens(w, &opcode.value, 1); operand.value = 0; operand.numComponents = VGPU10_OPERAND_1_COMPONENT; operand.operandType = VGPU10_OPERAND_TYPE_LABEL; operand.indexDimension = VGPU10_OPERAND_INDEX_1D; operand.index0Representation = VGPU10_OPERAND_INDEX_IMMEDIATE32; dxbcByteWriterAddTokens(w, &operand.value, 1); dxbcByteWriterAddTokens(w, &label, 1); } static void dxbcEmitRet(DXBCByteWriter *w) { VGPU10OpcodeToken0 opcode; opcode.value = 0; opcode.opcodeType = VGPU10_OPCODE_RET; opcode.instructionLength = 1; dxbcByteWriterAddTokens(w, &opcode.value, 1); } static int dxbcEmitVmwareIDIV(DXBCOUTPUTCTX *pOutctx, DXBCByteWriter *w, VGPUOpcode *pOpcode) { /* Insert a call and append a subroutne. */ VGPU10OpcodeToken0 opcode; uint32_t const label = (pOutctx->offSubroutine - dxbcByteWriterSize(w)) / 4; dxbcEmitCall(w, pOpcode, label); /* * Subroutine. */ DXBCByteWriterState savedWriterState; if (!dxbcByteWriterSetOffset(w, pOutctx->offSubroutine, &savedWriterState)) return w->rc; dxbcEmitLabel(w, label); /* Just output UDIV for now. */ opcode.value = 0; opcode.opcodeType = VGPU10_OPCODE_UDIV; opcode.instructionLength = pOpcode->cOpcodeToken; dxbcByteWriterAddTokens(w, &opcode.value, 1); dxbcByteWriterAddTokens(w, &pOpcode->paOpcodeToken[1], pOpcode->cOpcodeToken - 1); dxbcEmitRet(w); pOutctx->offSubroutine = dxbcByteWriterSize(w); dxbcByteWriterRestore(w, &savedWriterState); return w->rc; } static int dxbcEmitVmwareDFRC(DXBCOUTPUTCTX *pOutctx, DXBCByteWriter *w, VGPUOpcode *pOpcode) { /* Insert a call and append a subroutine. */ VGPU10OpcodeToken0 opcode; uint32_t const label = (pOutctx->offSubroutine - dxbcByteWriterSize(w)) / 4; dxbcEmitCall(w, pOpcode, label); /* * Subroutine. */ DXBCByteWriterState savedWriterState; if (!dxbcByteWriterSetOffset(w, pOutctx->offSubroutine, &savedWriterState)) return w->rc; dxbcEmitLabel(w, label); /* Just output a MOV for now. */ opcode.value = 0; opcode.opcodeType = VGPU10_OPCODE_MOV; opcode.instructionLength = pOpcode->cOpcodeToken; dxbcByteWriterAddTokens(w, &opcode.value, 1); dxbcByteWriterAddTokens(w, &pOpcode->paOpcodeToken[1], pOpcode->cOpcodeToken - 1); dxbcEmitRet(w); pOutctx->offSubroutine = dxbcByteWriterSize(w); dxbcByteWriterRestore(w, &savedWriterState); return w->rc; } static int dxbcEmitVmwareDRSQ(DXBCOUTPUTCTX *pOutctx, DXBCByteWriter *w, VGPUOpcode *pOpcode) { /* Insert a call and append a subroutine. */ VGPU10OpcodeToken0 opcode; uint32_t const label = (pOutctx->offSubroutine - dxbcByteWriterSize(w)) / 4; dxbcEmitCall(w, pOpcode, label); /* * Subroutine. */ DXBCByteWriterState savedWriterState; if (!dxbcByteWriterSetOffset(w, pOutctx->offSubroutine, &savedWriterState)) return w->rc; dxbcEmitLabel(w, label); /* Just output a MOV for now. */ opcode.value = 0; opcode.opcodeType = VGPU10_OPCODE_MOV; opcode.instructionLength = pOpcode->cOpcodeToken; dxbcByteWriterAddTokens(w, &opcode.value, 1); dxbcByteWriterAddTokens(w, &pOpcode->paOpcodeToken[1], pOpcode->cOpcodeToken - 1); dxbcEmitRet(w); pOutctx->offSubroutine = dxbcByteWriterSize(w); dxbcByteWriterRestore(w, &savedWriterState); return w->rc; } static int dxbcOutputOpcode(DXBCOUTPUTCTX *pOutctx, DXBCByteWriter *w, VGPUOpcode *pOpcode) { #ifdef DEBUG void *pvBegin = dxbcByteWriterPtr(w); #endif if ( pOutctx->programToken.programType == VGPU10_PIXEL_SHADER && pOpcode->opcodeType == VGPU10_OPCODE_DCL_RESOURCE) { /** @todo This is a workaround. */ /* Sometimes the guest (Mesa) created a shader with uninitialized resource dimension. * Use texture 2d because it is what a pixel shader normally uses. */ ASSERT_GUEST_RETURN(pOpcode->cOpcodeToken == 4, VERR_INVALID_PARAMETER); VGPU10OpcodeToken0 opcode; opcode.value = pOpcode->paOpcodeToken[0]; if (opcode.resourceDimension == VGPU10_RESOURCE_DIMENSION_BUFFER) { opcode.resourceDimension = VGPU10_RESOURCE_DIMENSION_TEXTURE2D; dxbcByteWriterAddTokens(w, &opcode.value, 1); dxbcByteWriterAddTokens(w, &pOpcode->paOpcodeToken[1], 2); uint32_t const returnType = 0x5555; /* float */ dxbcByteWriterAddTokens(w, &returnType, 1); return VINF_SUCCESS; } } else if (pOpcode->opcodeType == VGPU10_OPCODE_VMWARE) { if (pOpcode->opcodeSubtype == VGPU10_VMWARE_OPCODE_IDIV) return dxbcEmitVmwareIDIV(pOutctx, w, pOpcode); if (pOpcode->opcodeSubtype == VGPU10_VMWARE_OPCODE_DFRC) return dxbcEmitVmwareDFRC(pOutctx, w, pOpcode); if (pOpcode->opcodeSubtype == VGPU10_VMWARE_OPCODE_DRSQ) return dxbcEmitVmwareDRSQ(pOutctx, w, pOpcode); ASSERT_GUEST_FAILED_RETURN(VERR_NOT_SUPPORTED); } #ifdef DEBUG /* The code above must emit either nothing or everything. */ Assert((uintptr_t)pvBegin == (uintptr_t)dxbcByteWriterPtr(w)); #endif /* Just emit the unmodified instruction. */ dxbcByteWriterAddTokens(w, pOpcode->paOpcodeToken, pOpcode->cOpcodeToken); return VINF_SUCCESS; } static int dxbcOutputFinalize(DXBCOUTPUTCTX *pOutctx, DXBCByteWriter *w) { RT_NOREF(pOutctx, w); return VINF_SUCCESS; } static DECLCALLBACK(int) signatureEntryCmp(void const *pvElement1, void const *pvElement2, void *pvUser) { SVGA3dDXSignatureEntry const *e1 = (SVGA3dDXSignatureEntry *)pvElement1; SVGA3dDXSignatureEntry const *e2 = (SVGA3dDXSignatureEntry *)pvElement2; RT_NOREF(pvUser); if (e1->registerIndex < e2->registerIndex) return -1; if (e1->registerIndex > e2->registerIndex) return 1; if ((e1->mask & 0xf) < (e2->mask & 0xf)) return -1; if ((e1->mask & 0xf) > (e2->mask & 0xf)) return 1; return 0; } static void dxbcGenerateSemantics(DXShaderInfo *pInfo, uint32_t cSignature, SVGA3dDXSignatureEntry *paSignature, DXShaderAttributeSemantic *paSemantic, uint32_t u32BlobType); /* * Parse and verify the shader byte code. Extract input and output signatures into pInfo. */ int DXShaderParse(void const *pvShaderCode, uint32_t cbShaderCode, DXShaderInfo *pInfo) { if (pInfo) RT_ZERO(*pInfo); ASSERT_GUEST_RETURN(cbShaderCode <= SVGA3D_MAX_SHADER_MEMORY_BYTES, VERR_INVALID_PARAMETER); ASSERT_GUEST_RETURN((cbShaderCode & 0x3) == 0, VERR_INVALID_PARAMETER); /* Aligned to the token size. */ ASSERT_GUEST_RETURN(cbShaderCode >= 8, VERR_INVALID_PARAMETER); /* At least program and length tokens. */ uint32_t const *paToken = (uint32_t *)pvShaderCode; VGPU10ProgramToken const *pProgramToken = (VGPU10ProgramToken *)&paToken[0]; ASSERT_GUEST_RETURN( pProgramToken->majorVersion >= 4 && pProgramToken->programType <= VGPU10_COMPUTE_SHADER, VERR_INVALID_PARAMETER); if (pInfo) pInfo->enmProgramType = (VGPU10_PROGRAM_TYPE)pProgramToken->programType; uint32_t const cToken = paToken[1]; Log6(("Shader version %d.%d type %s(%d) Length %d\n", pProgramToken->majorVersion, pProgramToken->minorVersion, dxbcShaderTypeToString(pProgramToken->programType), pProgramToken->programType, cToken)); ASSERT_GUEST_RETURN(cbShaderCode / 4 >= cToken, VERR_INVALID_PARAMETER); /* Declared length should be less or equal to the actual. */ /* Write the parsed (and possibly modified) shader to a memory buffer. */ DXBCByteWriter dxbcByteWriter; DXBCByteWriter *w = &dxbcByteWriter; if (!dxbcByteWriterInit(w, 4096 + cbShaderCode)) return VERR_NO_MEMORY; dxbcByteWriterAddTokens(w, paToken, 2); DXBCTokenReader parser; RT_ZERO(parser); DXBCTokenReader *r = &parser; r->pToken = &paToken[2]; r->cToken = r->cRemainingToken = cToken - 2; DXBCOUTPUTCTX outctx; dxbcOutputInit(&outctx, pProgramToken, cToken); int rc = VINF_SUCCESS; while (dxbcTokenReaderCanRead(r, 1)) { uint32_t const offOpcode = dxbcByteWriterSize(w); VGPUOpcode opcode; rc = dxbcParseOpcode(r, &opcode); ASSERT_GUEST_STMT_BREAK(RT_SUCCESS(rc), rc = VERR_INVALID_PARAMETER); rc = dxbcOutputOpcode(&outctx, w, &opcode); AssertRCBreak(rc); if (pInfo) { /* Remember offsets of DCL_RESOURCE instructions. */ if ( outctx.programToken.programType == VGPU10_PIXEL_SHADER && opcode.opcodeType == VGPU10_OPCODE_DCL_RESOURCE) { if ( opcode.cOperand == 1 && opcode.aValOperand[0].indexDimension == VGPU10_OPERAND_INDEX_1D && opcode.aValOperand[0].aOperandIndex[0].indexRepresentation == VGPU10_OPERAND_INDEX_IMMEDIATE32) { uint32_t const indexResource = opcode.aValOperand[0].aOperandIndex[0].iOperandImmediate; if (indexResource < SVGA3D_DX_MAX_SRVIEWS) { ASSERT_GUEST(pInfo->aOffDclResource[indexResource] == 0); pInfo->aOffDclResource[indexResource] = offOpcode; pInfo->cDclResource = RT_MAX(pInfo->cDclResource, indexResource + 1); } else ASSERT_GUEST_FAILED(); } else ASSERT_GUEST_FAILED(); } /* Fetch signatures. */ SVGA3dDXSignatureEntry *pSignatureEntry = NULL; switch (opcode.opcodeType) { case VGPU10_OPCODE_DCL_INPUT: case VGPU10_OPCODE_DCL_INPUT_SIV: //case VGPU10_OPCODE_DCL_INPUT_SGV: case VGPU10_OPCODE_DCL_INPUT_PS: //case VGPU10_OPCODE_DCL_INPUT_PS_SIV: //case VGPU10_OPCODE_DCL_INPUT_PS_SGV: //case VGPU10_OPCODE_DCL_INPUT_CONTROL_POINT_COUNT: ASSERT_GUEST_STMT_BREAK(pInfo->cInputSignature < RT_ELEMENTS(pInfo->aInputSignature), rc = VERR_INVALID_PARAMETER); pSignatureEntry = &pInfo->aInputSignature[pInfo->cInputSignature++]; break; case VGPU10_OPCODE_DCL_OUTPUT: case VGPU10_OPCODE_DCL_OUTPUT_SIV: case VGPU10_OPCODE_DCL_OUTPUT_SGV: //case VGPU10_OPCODE_DCL_OUTPUT_CONTROL_POINT_COUNT: ASSERT_GUEST_STMT_BREAK(pInfo->cOutputSignature < RT_ELEMENTS(pInfo->aOutputSignature), rc = VERR_INVALID_PARAMETER); pSignatureEntry = &pInfo->aOutputSignature[pInfo->cOutputSignature++]; break; default: break; } if (RT_FAILURE(rc)) break; if (pSignatureEntry) { ASSERT_GUEST_STMT_BREAK( opcode.aValOperand[0].aOperandIndex[0].indexRepresentation == VGPU10_OPERAND_INDEX_IMMEDIATE32 || opcode.aValOperand[0].aOperandIndex[0].indexRepresentation == VGPU10_OPERAND_INDEX_IMMEDIATE64, rc = VERR_NOT_SUPPORTED); uint32_t const indexDimension = opcode.aValOperand[0].indexDimension; if (indexDimension == VGPU10_OPERAND_INDEX_0D) { if (opcode.aValOperand[0].operandType == VGPU10_OPERAND_TYPE_INPUT_PRIMITIVEID) { pSignatureEntry->registerIndex = 0; pSignatureEntry->semanticName = SVGADX_SIGNATURE_SEMANTIC_NAME_PRIMITIVE_ID; } else if (opcode.aValOperand[0].operandType == VGPU10_OPERAND_TYPE_OUTPUT_DEPTH) { /* oDepth is always last in the signature. Register index is equal to 0xFFFFFFFF. */ pSignatureEntry->registerIndex = 0xFFFFFFFF; pSignatureEntry->semanticName = SVGADX_SIGNATURE_SEMANTIC_NAME_UNDEFINED; } else if (opcode.aValOperand[0].operandType <= VGPU10_OPERAND_TYPE_SM50_MAX) { pSignatureEntry->registerIndex = 0; pSignatureEntry->semanticName = opcode.semanticName; } else ASSERT_GUEST_FAILED_STMT_BREAK(rc = VERR_NOT_SUPPORTED); } else { ASSERT_GUEST_STMT_BREAK( indexDimension == VGPU10_OPERAND_INDEX_1D || indexDimension == VGPU10_OPERAND_INDEX_2D || indexDimension == VGPU10_OPERAND_INDEX_3D, rc = VERR_NOT_SUPPORTED); /* The register index seems to be in the highest dimension. */ pSignatureEntry->registerIndex = opcode.aValOperand[0].aOperandIndex[indexDimension - VGPU10_OPERAND_INDEX_1D].iOperandImmediate; pSignatureEntry->semanticName = opcode.semanticName; } pSignatureEntry->mask = opcode.aValOperand[0].mask; pSignatureEntry->componentType = SVGADX_SIGNATURE_REGISTER_COMPONENT_UNKNOWN; // Will be updated by vboxDXUpdateVSInputSignature pSignatureEntry->minPrecision = SVGADX_SIGNATURE_MIN_PRECISION_DEFAULT; } } } if (RT_FAILURE(rc)) { return rc; } rc = dxbcOutputFinalize(&outctx, w); if (RT_FAILURE(rc)) { return rc; } dxbcByteWriterFetchData(w, &pInfo->pvBytecode, &pInfo->cbBytecode); uint32_t *pcOutputToken = (uint32_t *)pInfo->pvBytecode + 1; *pcOutputToken = pInfo->cbBytecode / 4; /* Sort signatures by register index and mask because the host API need them to be sorted. */ if (pInfo->cInputSignature) { RTSortShell(pInfo->aInputSignature, pInfo->cInputSignature, sizeof(pInfo->aInputSignature[0]), signatureEntryCmp, NULL); dxbcGenerateSemantics(pInfo, pInfo->cInputSignature, pInfo->aInputSignature, pInfo->aInputSemantic, DXBC_BLOB_TYPE_ISGN); } if (pInfo->cOutputSignature) { RTSortShell(pInfo->aOutputSignature, pInfo->cOutputSignature, sizeof(pInfo->aOutputSignature[0]), signatureEntryCmp, NULL); dxbcGenerateSemantics(pInfo, pInfo->cOutputSignature, pInfo->aOutputSignature, pInfo->aOutputSemantic, DXBC_BLOB_TYPE_OSGN); } if (pInfo->cPatchConstantSignature) { RTSortShell(pInfo->aPatchConstantSignature, pInfo->cPatchConstantSignature, sizeof(pInfo->aPatchConstantSignature[0]), signatureEntryCmp, NULL); dxbcGenerateSemantics(pInfo, pInfo->cPatchConstantSignature, pInfo->aPatchConstantSignature, pInfo->aPatchConstantSemantic, DXBC_BLOB_TYPE_PCSG); } #ifdef LOG_ENABLED if (pInfo->cInputSignature) { Log6(("Input signatures:\n")); for (uint32_t i = 0; i < pInfo->cInputSignature; ++i) Log6((" [%u]: %u %u 0x%X, %s %d\n", i, pInfo->aInputSignature[i].registerIndex, pInfo->aInputSignature[i].semanticName, pInfo->aInputSignature[i].mask, pInfo->aInputSemantic[i].pcszSemanticName, pInfo->aInputSemantic[i].SemanticIndex)); } if (pInfo->cOutputSignature) { Log6(("Output signatures:\n")); for (uint32_t i = 0; i < pInfo->cOutputSignature; ++i) Log6((" [%u]: %u %u 0x%X, %s %d\n", i, pInfo->aOutputSignature[i].registerIndex, pInfo->aOutputSignature[i].semanticName, pInfo->aOutputSignature[i].mask, pInfo->aOutputSemantic[i].pcszSemanticName, pInfo->aOutputSemantic[i].SemanticIndex)); } if (pInfo->cPatchConstantSignature) { Log6(("Patch constant signatures:\n")); for (uint32_t i = 0; i < pInfo->cPatchConstantSignature; ++i) Log6((" [%u]: %u %u 0x%X, %s %d\n", i, pInfo->aPatchConstantSignature[i].registerIndex, pInfo->aPatchConstantSignature[i].semanticName, pInfo->aPatchConstantSignature[i].mask, pInfo->aPatchConstantSemantic[i].pcszSemanticName, pInfo->aPatchConstantSemantic[i].SemanticIndex)); } #endif return VINF_SUCCESS; } void DXShaderGenerateSemantics(DXShaderInfo *pInfo) { if (pInfo->cInputSignature) dxbcGenerateSemantics(pInfo, pInfo->cInputSignature, pInfo->aInputSignature, pInfo->aInputSemantic, DXBC_BLOB_TYPE_ISGN); if (pInfo->cOutputSignature) dxbcGenerateSemantics(pInfo, pInfo->cOutputSignature, pInfo->aOutputSignature, pInfo->aOutputSemantic, DXBC_BLOB_TYPE_OSGN); if (pInfo->cPatchConstantSignature) dxbcGenerateSemantics(pInfo, pInfo->cPatchConstantSignature, pInfo->aPatchConstantSignature, pInfo->aPatchConstantSemantic, DXBC_BLOB_TYPE_PCSG); } void DXShaderSortSignatures(DXShaderInfo *pInfo) { /* Sort signatures by register index and mask because the host API need them to be sorted. */ if (pInfo->cInputSignature) { RTSortShell(pInfo->aInputSignature, pInfo->cInputSignature, sizeof(pInfo->aInputSignature[0]), signatureEntryCmp, NULL); } if (pInfo->cOutputSignature) { RTSortShell(pInfo->aOutputSignature, pInfo->cOutputSignature, sizeof(pInfo->aOutputSignature[0]), signatureEntryCmp, NULL); } if (pInfo->cPatchConstantSignature) { RTSortShell(pInfo->aPatchConstantSignature, pInfo->cPatchConstantSignature, sizeof(pInfo->aPatchConstantSignature[0]), signatureEntryCmp, NULL); } } void DXShaderFree(DXShaderInfo *pInfo) { RTMemFree(pInfo->pvBytecode); RT_ZERO(*pInfo); } #if 0 // Unused. Replaced with dxbcSemanticInfo. static char const *dxbcSemanticName(SVGA3dDXSignatureSemanticName enmSemanticName) { /* https://docs.microsoft.com/en-us/windows/win32/direct3dhlsl/dx-graphics-hlsl-semantics#system-value-semantics */ switch (enmSemanticName) { case SVGADX_SIGNATURE_SEMANTIC_NAME_POSITION: return "SV_Position"; case SVGADX_SIGNATURE_SEMANTIC_NAME_CLIP_DISTANCE: return "SV_ClipDistance"; case SVGADX_SIGNATURE_SEMANTIC_NAME_CULL_DISTANCE: return "SV_CullDistance"; case SVGADX_SIGNATURE_SEMANTIC_NAME_RENDER_TARGET_ARRAY_INDEX: return "SV_RenderTargetArrayIndex"; case SVGADX_SIGNATURE_SEMANTIC_NAME_VIEWPORT_ARRAY_INDEX: return "SV_ViewportArrayIndex"; case SVGADX_SIGNATURE_SEMANTIC_NAME_VERTEX_ID: return "SV_VertexID"; case SVGADX_SIGNATURE_SEMANTIC_NAME_PRIMITIVE_ID: return "SV_PrimitiveID"; case SVGADX_SIGNATURE_SEMANTIC_NAME_INSTANCE_ID: return "SV_InstanceID"; case SVGADX_SIGNATURE_SEMANTIC_NAME_IS_FRONT_FACE: return "SV_IsFrontFace"; case SVGADX_SIGNATURE_SEMANTIC_NAME_SAMPLE_INDEX: return "SV_SampleIndex"; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_U_EQ_0_EDGE_TESSFACTOR: return "SV_FinalQuadUeq0EdgeTessFactor"; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_V_EQ_0_EDGE_TESSFACTOR: return "SV_FinalQuadVeq0EdgeTessFactor"; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_U_EQ_1_EDGE_TESSFACTOR: return "SV_FinalQuadUeq1EdgeTessFactor"; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_V_EQ_1_EDGE_TESSFACTOR: return "SV_FinalQuadVeq1EdgeTessFactor"; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_U_INSIDE_TESSFACTOR: return "SV_FinalQuadUInsideTessFactor"; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_V_INSIDE_TESSFACTOR: return "SV_FinalQuadVInsideTessFactor"; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_TRI_U_EQ_0_EDGE_TESSFACTOR: return "SV_FinalTriUeq0EdgeTessFactor"; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_TRI_V_EQ_0_EDGE_TESSFACTOR: return "SV_FinalTriVeq0EdgeTessFactor"; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_TRI_W_EQ_0_EDGE_TESSFACTOR: return "SV_FinalTriWeq0EdgeTessFactor"; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_TRI_INSIDE_TESSFACTOR: return "SV_FinalTriInsideTessFactor"; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_LINE_DETAIL_TESSFACTOR: return "SV_FinalLineDetailTessFactor"; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_LINE_DENSITY_TESSFACTOR: return "SV_FinalLineDensityTessFactor"; default: Assert(enmSemanticName == SVGADX_SIGNATURE_SEMANTIC_NAME_UNDEFINED); break; } /* Generic. Arbitrary name. It does not have any meaning. */ return "ATTRIB"; } #endif /* https://docs.microsoft.com/en-us/windows/win32/direct3dhlsl/dx-graphics-hlsl-semantics#system-value-semantics * Type: * 0 - undefined * 1 - unsigned int * 2 - signed int * 3 - float */ typedef struct VGPUSemanticInfo { char const *pszName; uint32_t u32Type; } VGPUSemanticInfo; static VGPUSemanticInfo const g_aSemanticInfo[SVGADX_SIGNATURE_SEMANTIC_NAME_MAX] = { { "ATTRIB", 0 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_UNDEFINED 0 { "SV_Position", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_POSITION 1 { "SV_ClipDistance", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_CLIP_DISTANCE 2 { "SV_CullDistance", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_CULL_DISTANCE 3 { "SV_RenderTargetArrayIndex", 1 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_RENDER_TARGET_ARRAY_INDEX 4 { "SV_ViewportArrayIndex", 1 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_VIEWPORT_ARRAY_INDEX 5 { "SV_VertexID", 1 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_VERTEX_ID 6 { "SV_PrimitiveID", 1 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_PRIMITIVE_ID 7 { "SV_InstanceID", 1 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_INSTANCE_ID 8 { "SV_IsFrontFace", 1 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_IS_FRONT_FACE 9 { "SV_SampleIndex", 1 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_SAMPLE_INDEX 10 { "SV_TessFactor", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_U_EQ_0_EDGE_TESSFACTOR 11 { "SV_TessFactor", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_V_EQ_0_EDGE_TESSFACTOR 12 { "SV_TessFactor", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_U_EQ_1_EDGE_TESSFACTOR 13 { "SV_TessFactor", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_V_EQ_1_EDGE_TESSFACTOR 14 { "SV_InsideTessFactor", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_U_INSIDE_TESSFACTOR 15 { "SV_InsideTessFactor", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_V_INSIDE_TESSFACTOR 16 { "SV_TessFactor", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_TRI_U_EQ_0_EDGE_TESSFACTOR 17 { "SV_TessFactor", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_TRI_V_EQ_0_EDGE_TESSFACTOR 18 { "SV_TessFactor", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_TRI_W_EQ_0_EDGE_TESSFACTOR 19 { "SV_InsideTessFactor", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_TRI_INSIDE_TESSFACTOR 20 { "SV_TessFactor", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_LINE_DETAIL_TESSFACTOR 21 { "SV_TessFactor", 3 }, // SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_LINE_DENSITY_TESSFACTOR 22 }; static VGPUSemanticInfo const g_SemanticPSOutput = { "SV_TARGET", 3 }; // SVGADX_SIGNATURE_SEMANTIC_NAME_UNDEFINED 0 /* A clone of D3D_NAME */ typedef enum { D3D_SV_UNDEFINED = 0, D3D_SV_POSITION = 1, D3D_SV_CLIP_DISTANCE = 2, D3D_SV_CULL_DISTANCE = 3, D3D_SV_RENDER_TARGET_ARRAY_INDEX = 4, D3D_SV_VIEWPORT_ARRAY_INDEX = 5, D3D_SV_VERTEX_ID = 6, D3D_SV_PRIMITIVE_ID = 7, D3D_SV_INSTANCE_ID = 8, D3D_SV_IS_FRONT_FACE = 9, D3D_SV_SAMPLE_INDEX = 10, D3D_SV_FINAL_QUAD_EDGE_TESSFACTOR = 11, D3D_SV_FINAL_QUAD_INSIDE_TESSFACTOR = 12, D3D_SV_FINAL_TRI_EDGE_TESSFACTOR = 13, D3D_SV_FINAL_TRI_INSIDE_TESSFACTOR = 14, D3D_SV_FINAL_LINE_DETAIL_TESSFACTOR = 15, D3D_SV_FINAL_LINE_DENSITY_TESSFACTOR = 16 } D3DSYSTEMVALUE; static uint32_t svga2dxSystemValue(SVGA3dDXSignatureSemanticName semanticName) { switch (semanticName) { case SVGADX_SIGNATURE_SEMANTIC_NAME_UNDEFINED: return D3D_SV_UNDEFINED; case SVGADX_SIGNATURE_SEMANTIC_NAME_POSITION: return D3D_SV_POSITION; case SVGADX_SIGNATURE_SEMANTIC_NAME_CLIP_DISTANCE: return D3D_SV_CLIP_DISTANCE; case SVGADX_SIGNATURE_SEMANTIC_NAME_CULL_DISTANCE: return D3D_SV_CULL_DISTANCE; case SVGADX_SIGNATURE_SEMANTIC_NAME_RENDER_TARGET_ARRAY_INDEX: return D3D_SV_RENDER_TARGET_ARRAY_INDEX; case SVGADX_SIGNATURE_SEMANTIC_NAME_VIEWPORT_ARRAY_INDEX: return D3D_SV_VIEWPORT_ARRAY_INDEX; case SVGADX_SIGNATURE_SEMANTIC_NAME_VERTEX_ID: return D3D_SV_VERTEX_ID; case SVGADX_SIGNATURE_SEMANTIC_NAME_PRIMITIVE_ID: return D3D_SV_PRIMITIVE_ID; case SVGADX_SIGNATURE_SEMANTIC_NAME_INSTANCE_ID: return D3D_SV_INSTANCE_ID; case SVGADX_SIGNATURE_SEMANTIC_NAME_IS_FRONT_FACE: return D3D_SV_IS_FRONT_FACE; case SVGADX_SIGNATURE_SEMANTIC_NAME_SAMPLE_INDEX: return D3D_SV_SAMPLE_INDEX; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_U_EQ_0_EDGE_TESSFACTOR: return D3D_SV_FINAL_QUAD_EDGE_TESSFACTOR; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_V_EQ_0_EDGE_TESSFACTOR: return D3D_SV_FINAL_QUAD_EDGE_TESSFACTOR; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_U_EQ_1_EDGE_TESSFACTOR: return D3D_SV_FINAL_QUAD_EDGE_TESSFACTOR; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_V_EQ_1_EDGE_TESSFACTOR: return D3D_SV_FINAL_QUAD_EDGE_TESSFACTOR; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_U_INSIDE_TESSFACTOR: return D3D_SV_FINAL_QUAD_INSIDE_TESSFACTOR; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_QUAD_V_INSIDE_TESSFACTOR: return D3D_SV_FINAL_QUAD_INSIDE_TESSFACTOR; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_TRI_U_EQ_0_EDGE_TESSFACTOR: return D3D_SV_FINAL_TRI_EDGE_TESSFACTOR; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_TRI_V_EQ_0_EDGE_TESSFACTOR: return D3D_SV_FINAL_TRI_EDGE_TESSFACTOR; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_TRI_W_EQ_0_EDGE_TESSFACTOR: return D3D_SV_FINAL_TRI_EDGE_TESSFACTOR; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_TRI_INSIDE_TESSFACTOR: return D3D_SV_FINAL_TRI_INSIDE_TESSFACTOR; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_LINE_DETAIL_TESSFACTOR: return D3D_SV_FINAL_LINE_DETAIL_TESSFACTOR; case SVGADX_SIGNATURE_SEMANTIC_NAME_FINAL_LINE_DENSITY_TESSFACTOR: return D3D_SV_FINAL_LINE_DENSITY_TESSFACTOR; } AssertFailedReturn(D3D_SV_UNDEFINED); } static VGPUSemanticInfo const *dxbcSemanticInfo(DXShaderInfo const *pInfo, SVGA3dDXSignatureSemanticName enmSemanticName, uint32_t u32BlobType) { if (enmSemanticName < RT_ELEMENTS(g_aSemanticInfo)) { if ( enmSemanticName == 0 && pInfo->enmProgramType == VGPU10_PIXEL_SHADER && u32BlobType == DXBC_BLOB_TYPE_OSGN) return &g_SemanticPSOutput; return &g_aSemanticInfo[enmSemanticName]; } return &g_aSemanticInfo[0]; } static void dxbcGenerateSemantics(DXShaderInfo *pInfo, uint32_t cSignature, SVGA3dDXSignatureEntry *paSignature, DXShaderAttributeSemantic *paSemantic, uint32_t u32BlobType) { for (uint32_t iSignatureEntry = 0; iSignatureEntry < cSignature; ++iSignatureEntry) { SVGA3dDXSignatureEntry *pSignatureEntry = &paSignature[iSignatureEntry]; DXShaderAttributeSemantic *pSemantic = &paSemantic[iSignatureEntry]; ASSERT_GUEST_RETURN_VOID(pSignatureEntry->semanticName < SVGADX_SIGNATURE_SEMANTIC_NAME_MAX); VGPUSemanticInfo const *pSemanticInfo = dxbcSemanticInfo(pInfo, pSignatureEntry->semanticName, u32BlobType); pSemantic->pcszSemanticName = pSemanticInfo->pszName; pSemantic->SemanticIndex = 0; if (pSignatureEntry->componentType == SVGADX_SIGNATURE_REGISTER_COMPONENT_UNKNOWN) pSignatureEntry->componentType = pSemanticInfo->u32Type; for (uint32_t i = 0; i < iSignatureEntry; ++i) { DXShaderAttributeSemantic const *pPriorSemantic = &paSemantic[i]; if (RTStrCmp(pPriorSemantic->pcszSemanticName, pSemantic->pcszSemanticName) == 0) ++pSemantic->SemanticIndex; } } } static int dxbcCreateIOSGNBlob(DXShaderInfo const *pInfo, DXBCHeader *pHdr, uint32_t u32BlobType, uint32_t cSignature, SVGA3dDXSignatureEntry const *paSignature, DXShaderAttributeSemantic const *paSemantic, DXBCByteWriter *w) { RT_NOREF(pInfo); AssertReturn(cSignature <= SVGA3D_DX_SM41_MAX_VERTEXINPUTREGISTERS, VERR_INVALID_PARAMETER); uint32_t cbBlob = RT_UOFFSETOF_DYN(DXBCBlobIOSGN, aElement[cSignature]); if (!dxbcByteWriterCanWrite(w, sizeof(DXBCBlobHeader) + cbBlob)) return VERR_NO_MEMORY; Log6(("Create signature type %c%c%c%c (0x%RX32)\n", RT_BYTE1(u32BlobType), RT_BYTE2(u32BlobType), RT_BYTE3(u32BlobType), RT_BYTE4(u32BlobType), u32BlobType)); DXBCBlobHeader *pHdrBlob = (DXBCBlobHeader *)dxbcByteWriterPtr(w); pHdrBlob->u32BlobType = u32BlobType; // pHdrBlob->cbBlob = 0; DXBCBlobIOSGN *pHdrISGN = (DXBCBlobIOSGN *)&pHdrBlob[1]; pHdrISGN->cElement = cSignature; pHdrISGN->offElement = RT_UOFFSETOF(DXBCBlobIOSGN, aElement[0]); #ifdef DEBUG /* Check that signatures are sorted by register index because the host API need them to be sorted. */ uint32_t idxRegisterLast = 0; #endif for (uint32_t iSignatureEntry = 0; iSignatureEntry < cSignature; ++iSignatureEntry) { SVGA3dDXSignatureEntry const *srcEntry = &paSignature[iSignatureEntry]; DXShaderAttributeSemantic const *srcSemantic = &paSemantic[iSignatureEntry]; DXBCBlobIOSGNElement *dst = &pHdrISGN->aElement[iSignatureEntry]; dst->offElementName = 0; for (uint32_t i = 0; i < iSignatureEntry; ++i) { DXBCBlobIOSGNElement const *pElement = &pHdrISGN->aElement[i]; char const *pszElementName = (char *)pHdrISGN + pElement->offElementName; if (RTStrCmp(pszElementName, srcSemantic->pcszSemanticName) == 0) { dst->offElementName = pElement->offElementName; break; } } dst->idxSemantic = srcSemantic->SemanticIndex; dst->enmSystemValue = svga2dxSystemValue(srcEntry->semanticName); dst->enmComponentType = srcEntry->componentType; dst->idxRegister = srcEntry->registerIndex; dst->u.mask = srcEntry->mask; Log6((" [%u]: %s[%u] sv %u type %u reg %u mask %X\n", iSignatureEntry, srcSemantic->pcszSemanticName, dst->idxSemantic, dst->enmSystemValue, dst->enmComponentType, dst->idxRegister, dst->u.mask)); #ifdef DEBUG Assert(idxRegisterLast <= dst->idxRegister); idxRegisterLast = dst->idxRegister; #endif if (dst->offElementName == 0) { /* Store the semantic name for this element. */ dst->offElementName = cbBlob; /* Offset of the semantic's name relative to the start of the blob (without DXBCBlobHeader). */ uint32_t const cbElementName = (uint32_t)strlen(srcSemantic->pcszSemanticName) + 1; if (!dxbcByteWriterCanWrite(w, sizeof(DXBCBlobHeader) + cbBlob + cbElementName)) return VERR_NO_MEMORY; memcpy((char *)pHdrISGN + dst->offElementName, srcSemantic->pcszSemanticName, cbElementName); cbBlob += cbElementName; } } /* Blobs are 4 bytes aligned. Commit the blob data. */ cbBlob = RT_ALIGN_32(cbBlob, 4); pHdrBlob->cbBlob = cbBlob; pHdr->cbTotal += cbBlob + sizeof(DXBCBlobHeader); dxbcByteWriterCommit(w, cbBlob + sizeof(DXBCBlobHeader)); return VINF_SUCCESS; } static int dxbcCreateSHDRBlob(DXBCHeader *pHdr, uint32_t u32BlobType, void const *pvShader, uint32_t cbShader, DXBCByteWriter *w) { uint32_t cbBlob = cbShader; if (!dxbcByteWriterCanWrite(w, sizeof(DXBCBlobHeader) + cbBlob)) return VERR_NO_MEMORY; DXBCBlobHeader *pHdrBlob = (DXBCBlobHeader *)dxbcByteWriterPtr(w); pHdrBlob->u32BlobType = u32BlobType; // pHdrBlob->cbBlob = 0; memcpy(&pHdrBlob[1], pvShader, cbShader); /* Blobs are 4 bytes aligned. Commit the blob data. */ cbBlob = RT_ALIGN_32(cbBlob, 4); pHdrBlob->cbBlob = cbBlob; pHdr->cbTotal += cbBlob + sizeof(DXBCBlobHeader); dxbcByteWriterCommit(w, cbBlob + sizeof(DXBCBlobHeader)); return VINF_SUCCESS; } /* * Create a DXBC container with signature and shader code data blobs. */ static int dxbcCreateFromInfo(DXShaderInfo const *pInfo, void const *pvShader, uint32_t cbShader, DXBCByteWriter *w) { int rc; /* Create a DXBC container with ISGN, OSGN and SHDR blobs. */ uint32_t cBlob = 3; if ( pInfo->enmProgramType == VGPU10_HULL_SHADER || pInfo->enmProgramType == VGPU10_DOMAIN_SHADER) ++cBlob; uint32_t const cbHdr = RT_UOFFSETOF_DYN(DXBCHeader, aBlobOffset[cBlob]); /* Header with blob offsets. */ if (!dxbcByteWriterCanWrite(w, cbHdr)) return VERR_NO_MEMORY; /* Container header. */ DXBCHeader *pHdr = (DXBCHeader *)dxbcByteWriterPtr(w); pHdr->u32DXBC = DXBC_MAGIC; // RT_ZERO(pHdr->au8Hash); pHdr->u32Version = 1; pHdr->cbTotal = cbHdr; pHdr->cBlob = cBlob; //RT_ZERO(pHdr->aBlobOffset); dxbcByteWriterCommit(w, cbHdr); #ifdef LOG_ENABLED if (pInfo->cInputSignature) { Log6(("Input signatures:\n")); for (uint32_t i = 0; i < pInfo->cInputSignature; ++i) Log6((" [%u]: %u %u 0x%X, %s %d\n", i, pInfo->aInputSignature[i].registerIndex, pInfo->aInputSignature[i].semanticName, pInfo->aInputSignature[i].mask, pInfo->aInputSemantic[i].pcszSemanticName, pInfo->aInputSemantic[i].SemanticIndex)); } if (pInfo->cOutputSignature) { Log6(("Output signatures:\n")); for (uint32_t i = 0; i < pInfo->cOutputSignature; ++i) Log6((" [%u]: %u %u 0x%X, %s %d\n", i, pInfo->aOutputSignature[i].registerIndex, pInfo->aOutputSignature[i].semanticName, pInfo->aOutputSignature[i].mask, pInfo->aOutputSemantic[i].pcszSemanticName, pInfo->aOutputSemantic[i].SemanticIndex)); } if (pInfo->cPatchConstantSignature) { Log6(("Patch constant signatures:\n")); for (uint32_t i = 0; i < pInfo->cPatchConstantSignature; ++i) Log6((" [%u]: %u %u 0x%X, %s %d\n", i, pInfo->aPatchConstantSignature[i].registerIndex, pInfo->aPatchConstantSignature[i].semanticName, pInfo->aPatchConstantSignature[i].mask, pInfo->aPatchConstantSemantic[i].pcszSemanticName, pInfo->aPatchConstantSemantic[i].SemanticIndex)); } #endif /* Blobs. */ uint32_t iBlob = 0; pHdr->aBlobOffset[iBlob++] = dxbcByteWriterSize(w); rc = dxbcCreateIOSGNBlob(pInfo, pHdr, DXBC_BLOB_TYPE_ISGN, pInfo->cInputSignature, &pInfo->aInputSignature[0], pInfo->aInputSemantic, w); AssertRCReturn(rc, rc); pHdr->aBlobOffset[iBlob++] = dxbcByteWriterSize(w); rc = dxbcCreateIOSGNBlob(pInfo, pHdr, DXBC_BLOB_TYPE_OSGN, pInfo->cOutputSignature, &pInfo->aOutputSignature[0], pInfo->aOutputSemantic, w); AssertRCReturn(rc, rc); if ( pInfo->enmProgramType == VGPU10_HULL_SHADER || pInfo->enmProgramType == VGPU10_DOMAIN_SHADER) { pHdr->aBlobOffset[iBlob++] = dxbcByteWriterSize(w); rc = dxbcCreateIOSGNBlob(pInfo, pHdr, DXBC_BLOB_TYPE_PCSG, pInfo->cPatchConstantSignature, &pInfo->aPatchConstantSignature[0], pInfo->aPatchConstantSemantic, w); AssertRCReturn(rc, rc); } pHdr->aBlobOffset[iBlob++] = dxbcByteWriterSize(w); rc = dxbcCreateSHDRBlob(pHdr, DXBC_BLOB_TYPE_SHDR, pvShader, cbShader, w); AssertRCReturn(rc, rc); Assert(iBlob == cBlob); AssertCompile(RT_UOFFSETOF(DXBCHeader, u32Version) == 0x14); dxbcHash(&pHdr->u32Version, pHdr->cbTotal - RT_UOFFSETOF(DXBCHeader, u32Version), pHdr->au8Hash); return VINF_SUCCESS; } int DXShaderCreateDXBC(DXShaderInfo const *pInfo, void **ppvDXBC, uint32_t *pcbDXBC) { /* Build DXBC container. */ int rc; DXBCByteWriter dxbcByteWriter; DXBCByteWriter *w = &dxbcByteWriter; if (dxbcByteWriterInit(w, 4096 + pInfo->cbBytecode)) { rc = dxbcCreateFromInfo(pInfo, pInfo->pvBytecode, pInfo->cbBytecode, w); if (RT_SUCCESS(rc)) dxbcByteWriterFetchData(w, ppvDXBC, pcbDXBC); } else rc = VERR_NO_MEMORY; return rc; } static char const *dxbcGetOutputSemanticName(DXShaderInfo const *pInfo, uint32_t idxRegister, uint32_t u32BlobType, uint32_t cSignature, SVGA3dDXSignatureEntry const *paSignature, SVGA3dDXSignatureSemanticName *pSemanticName) { for (uint32_t i = 0; i < cSignature; ++i) { SVGA3dDXSignatureEntry const *p = &paSignature[i]; if (p->registerIndex == idxRegister) { AssertReturn(p->semanticName < SVGADX_SIGNATURE_SEMANTIC_NAME_MAX, NULL); VGPUSemanticInfo const *pSemanticInfo = dxbcSemanticInfo(pInfo, p->semanticName, u32BlobType); *pSemanticName = p->semanticName; return pSemanticInfo->pszName; } } return NULL; } char const *DXShaderGetOutputSemanticName(DXShaderInfo const *pInfo, uint32_t idxRegister, SVGA3dDXSignatureSemanticName *pSemanticName) { return dxbcGetOutputSemanticName(pInfo, idxRegister, DXBC_BLOB_TYPE_OSGN, pInfo->cOutputSignature, &pInfo->aOutputSignature[0], pSemanticName); } VGPU10_RESOURCE_RETURN_TYPE DXShaderResourceReturnTypeFromFormat(SVGA3dSurfaceFormat format) { /** @todo This is auto-generated from format names and needs a review. */ switch (format) { case SVGA3D_R32G32B32A32_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R32G32B32A32_SINT: return VGPU10_RETURN_TYPE_SINT; case SVGA3D_R32G32B32_FLOAT: return VGPU10_RETURN_TYPE_FLOAT; case SVGA3D_R32G32B32_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R32G32B32_SINT: return VGPU10_RETURN_TYPE_SINT; case SVGA3D_R16G16B16A16_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R16G16B16A16_SNORM: return VGPU10_RETURN_TYPE_SNORM; case SVGA3D_R16G16B16A16_SINT: return VGPU10_RETURN_TYPE_SINT; case SVGA3D_R32G32_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R32G32_SINT: return VGPU10_RETURN_TYPE_SINT; case SVGA3D_D32_FLOAT_S8X24_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R32_FLOAT_X8X24: return VGPU10_RETURN_TYPE_FLOAT; case SVGA3D_X32_G8X24_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R10G10B10A2_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R11G11B10_FLOAT: return VGPU10_RETURN_TYPE_FLOAT; case SVGA3D_R8G8B8A8_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_R8G8B8A8_UNORM_SRGB: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_R8G8B8A8_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R8G8B8A8_SINT: return VGPU10_RETURN_TYPE_SINT; case SVGA3D_R16G16_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R16G16_SINT: return VGPU10_RETURN_TYPE_SINT; case SVGA3D_D32_FLOAT: return VGPU10_RETURN_TYPE_FLOAT; case SVGA3D_R32_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R32_SINT: return VGPU10_RETURN_TYPE_SINT; case SVGA3D_D24_UNORM_S8_UINT: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_R24_UNORM_X8: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_X24_G8_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R8G8_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_R8G8_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R8G8_SINT: return VGPU10_RETURN_TYPE_SINT; case SVGA3D_R16_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_R16_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R16_SNORM: return VGPU10_RETURN_TYPE_SNORM; case SVGA3D_R16_SINT: return VGPU10_RETURN_TYPE_SINT; case SVGA3D_R8_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_R8_UINT: return VGPU10_RETURN_TYPE_UINT; case SVGA3D_R8_SNORM: return VGPU10_RETURN_TYPE_SNORM; case SVGA3D_R8_SINT: return VGPU10_RETURN_TYPE_SINT; case SVGA3D_R8G8_B8G8_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_G8R8_G8B8_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_BC1_UNORM_SRGB: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_BC2_UNORM_SRGB: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_BC3_UNORM_SRGB: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_BC4_SNORM: return VGPU10_RETURN_TYPE_SNORM; case SVGA3D_BC5_SNORM: return VGPU10_RETURN_TYPE_SNORM; case SVGA3D_R10G10B10_XR_BIAS_A2_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_B8G8R8A8_UNORM_SRGB: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_B8G8R8X8_UNORM_SRGB: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_R32G32B32A32_FLOAT: return VGPU10_RETURN_TYPE_FLOAT; case SVGA3D_R16G16B16A16_FLOAT: return VGPU10_RETURN_TYPE_FLOAT; case SVGA3D_R16G16B16A16_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_R32G32_FLOAT: return VGPU10_RETURN_TYPE_FLOAT; case SVGA3D_R10G10B10A2_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_R8G8B8A8_SNORM: return VGPU10_RETURN_TYPE_SNORM; case SVGA3D_R16G16_FLOAT: return VGPU10_RETURN_TYPE_FLOAT; case SVGA3D_R16G16_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_R16G16_SNORM: return VGPU10_RETURN_TYPE_SNORM; case SVGA3D_R32_FLOAT: return VGPU10_RETURN_TYPE_FLOAT; case SVGA3D_R8G8_SNORM: return VGPU10_RETURN_TYPE_SNORM; case SVGA3D_R16_FLOAT: return VGPU10_RETURN_TYPE_FLOAT; case SVGA3D_D16_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_A8_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_BC1_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_BC2_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_BC3_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_B5G6R5_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_B5G5R5A1_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_B8G8R8A8_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_B8G8R8X8_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_BC4_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_BC5_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_B4G4R4A4_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_BC7_UNORM: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_BC7_UNORM_SRGB: return VGPU10_RETURN_TYPE_UNORM; case SVGA3D_R9G9B9E5_SHAREDEXP: return VGPU10_RETURN_TYPE_FLOAT; default: break; } return VGPU10_RETURN_TYPE_UNORM; } SVGA3dDXSignatureRegisterComponentType DXShaderComponentTypeFromFormat(SVGA3dSurfaceFormat format) { /** @todo This is auto-generated from format names and needs a review. */ switch (format) { case SVGA3D_R32G32B32A32_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R32G32B32A32_SINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_SINT32; case SVGA3D_R32G32B32_FLOAT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R32G32B32_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R32G32B32_SINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_SINT32; case SVGA3D_R16G16B16A16_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R16G16B16A16_SNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R16G16B16A16_SINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_SINT32; case SVGA3D_R32G32_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R32G32_SINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_SINT32; case SVGA3D_D32_FLOAT_S8X24_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R32_FLOAT_X8X24: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_X32_G8X24_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R10G10B10A2_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R11G11B10_FLOAT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R8G8B8A8_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R8G8B8A8_UNORM_SRGB: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R8G8B8A8_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R8G8B8A8_SINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_SINT32; case SVGA3D_R16G16_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R16G16_SINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_SINT32; case SVGA3D_D32_FLOAT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R32_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R32_SINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_SINT32; case SVGA3D_D24_UNORM_S8_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R24_UNORM_X8: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_X24_G8_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R8G8_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R8G8_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R8G8_SINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_SINT32; case SVGA3D_R16_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R16_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R16_SNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R16_SINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_SINT32; case SVGA3D_R8_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R8_UINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_UINT32; case SVGA3D_R8_SNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R8_SINT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_SINT32; case SVGA3D_R8G8_B8G8_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_G8R8_G8B8_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_BC1_UNORM_SRGB: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_BC2_UNORM_SRGB: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_BC3_UNORM_SRGB: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_BC4_SNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_BC5_SNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R10G10B10_XR_BIAS_A2_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_B8G8R8A8_UNORM_SRGB: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_B8G8R8X8_UNORM_SRGB: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R32G32B32A32_FLOAT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R16G16B16A16_FLOAT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R16G16B16A16_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R32G32_FLOAT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R10G10B10A2_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R8G8B8A8_SNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R16G16_FLOAT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R16G16_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R16G16_SNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R32_FLOAT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R8G8_SNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R16_FLOAT: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_D16_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_A8_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_BC1_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_BC2_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_BC3_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_B5G6R5_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_B5G5R5A1_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_B8G8R8A8_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_B8G8R8X8_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_BC4_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_BC5_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_B4G4R4A4_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_BC7_UNORM: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_BC7_UNORM_SRGB: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; case SVGA3D_R9G9B9E5_SHAREDEXP: return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; default: break; } return SVGADX_SIGNATURE_REGISTER_COMPONENT_FLOAT32; } int DXShaderUpdateResources(DXShaderInfo const *pInfo, VGPU10_RESOURCE_DIMENSION *paResourceDimension, VGPU10_RESOURCE_RETURN_TYPE *paResourceReturnType, uint32_t cResources) { for (uint32_t i = 0; i < pInfo->cDclResource; ++i) { VGPU10_RESOURCE_DIMENSION const resourceDimension = i < cResources ? paResourceDimension[i] : VGPU10_RESOURCE_DIMENSION_TEXTURE2D; AssertContinue(resourceDimension <= VGPU10_RESOURCE_DIMENSION_TEXTURECUBEARRAY); VGPU10_RESOURCE_RETURN_TYPE const resourceReturnType = i < cResources ? paResourceReturnType[i] : VGPU10_RETURN_TYPE_FLOAT; AssertContinue(resourceReturnType <= VGPU10_RETURN_TYPE_MIXED); uint32_t const offToken = pInfo->aOffDclResource[i]; AssertContinue(offToken < pInfo->cbBytecode); if (offToken == 0) /* nothing at this index */ continue; uint32_t *paToken = (uint32_t *)((uintptr_t)pInfo->pvBytecode + offToken); VGPU10OpcodeToken0 *pOpcode = (VGPU10OpcodeToken0 *)&paToken[0]; if (resourceDimension != VGPU10_RESOURCE_DIMENSION_UNKNOWN) pOpcode->resourceDimension = resourceDimension; // paToken[1] unmodified // paToken[2] unmodified VGPU10ResourceReturnTypeToken *pReturnTypeToken = (VGPU10ResourceReturnTypeToken *)&paToken[3]; if ((uint8_t)resourceReturnType != 0) { pReturnTypeToken->component0 = (uint8_t)resourceReturnType; pReturnTypeToken->component1 = (uint8_t)resourceReturnType; pReturnTypeToken->component2 = (uint8_t)resourceReturnType; pReturnTypeToken->component3 = (uint8_t)resourceReturnType; } } return VINF_SUCCESS; } #ifdef DXBC_STANDALONE_TEST static int dxbcCreateFromBytecode(void const *pvShaderCode, uint32_t cbShaderCode, void **ppvDXBC, uint32_t *pcbDXBC) { /* Parse the shader bytecode and create DXBC container with resource, signature and shader bytecode blobs. */ DXShaderInfo info; RT_ZERO(info); int rc = DXShaderParse(pvShaderCode, cbShaderCode, &info); if (RT_SUCCESS(rc)) rc = DXShaderCreateDXBC(&info, ppvDXBC, pcbDXBC); return rc; } static int parseShaderVM(void const *pvShaderCode, uint32_t cbShaderCode) { void *pv = NULL; uint32_t cb = 0; int rc = dxbcCreateFromBytecode(pvShaderCode, cbShaderCode, &pv, &cb); if (RT_SUCCESS(rc)) { /* Hexdump DXBC */ printf("{\n"); uint8_t *pu8 = (uint8_t *)pv; for (uint32_t i = 0; i < cb; ++i) { if ((i % 16) == 0) { if (i > 0) printf(",\n"); printf(" 0x%02x", pu8[i]); } else { printf(", 0x%02x", pu8[i]); } } printf("\n"); printf("};\n"); RTMemFree(pv); } return rc; } static DXBCBlobHeader *dxbcFindBlob(DXBCHeader *pDXBCHeader, uint32_t u32BlobType) { uint8_t const *pu8DXBCBegin = (uint8_t *)pDXBCHeader; for (uint32_t i = 0; i < pDXBCHeader->cBlob; ++i) { DXBCBlobHeader *pCurrentBlob = (DXBCBlobHeader *)&pu8DXBCBegin[pDXBCHeader->aBlobOffset[i]]; if (pCurrentBlob->u32BlobType == u32BlobType) return pCurrentBlob; } return NULL; } static int dxbcExtractShaderCode(DXBCHeader *pDXBCHeader, void **ppvCode, uint32_t *pcbCode) { DXBCBlobHeader *pBlob = dxbcFindBlob(pDXBCHeader, DXBC_BLOB_TYPE_SHDR); AssertReturn(pBlob, VERR_NOT_IMPLEMENTED); DXBCBlobSHDR *pSHDR = (DXBCBlobSHDR *)&pBlob[1]; *pcbCode = pSHDR->cToken * 4; *ppvCode = RTMemAlloc(*pcbCode); AssertReturn(*ppvCode, VERR_NO_MEMORY); memcpy(*ppvCode, pSHDR, *pcbCode); return VINF_SUCCESS; } static int parseShaderDXBC(void const *pvDXBC) { DXBCHeader *pDXBCHeader = (DXBCHeader *)pvDXBC; void *pvShaderCode = NULL; uint32_t cbShaderCode = 0; int rc = dxbcExtractShaderCode(pDXBCHeader, &pvShaderCode, &cbShaderCode); if (RT_SUCCESS(rc)) { rc = parseShaderVM(pvShaderCode, cbShaderCode); RTMemFree(pvShaderCode); } return rc; } #endif /* DXBC_STANDALONE_TEST */