mirror of
https://github.com/Yours3lf/rpi-vk-driver.git
synced 2025-01-05 23:46:14 +01:00
1451 lines
51 KiB
C
1451 lines
51 KiB
C
/*
|
|
* Copyright © 2016 Broadcom
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
|
* IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include <string.h>
|
|
#include <strings.h>
|
|
#include "../common/macros.h"
|
|
|
|
#include "../common/v3d_device_info.h"
|
|
#include "qpu_instr.h"
|
|
|
|
#ifndef QPU_MASK
|
|
#define QPU_MASK(high, low) ((((uint64_t)1<<((high)-(low)+1))-1)<<(low))
|
|
/* Using the GNU statement expression extension */
|
|
#define QPU_SET_FIELD(value, field) \
|
|
({ \
|
|
uint64_t fieldval = (uint64_t)(value) << field ## _SHIFT; \
|
|
assert((fieldval & ~ field ## _MASK) == 0); \
|
|
fieldval & field ## _MASK; \
|
|
})
|
|
|
|
#define QPU_GET_FIELD(word, field) ((uint32_t)(((word) & field ## _MASK) >> field ## _SHIFT))
|
|
|
|
#define QPU_UPDATE_FIELD(inst, value, field) \
|
|
(((inst) & ~(field ## _MASK)) | QPU_SET_FIELD(value, field))
|
|
#endif /* QPU_MASK */
|
|
|
|
#define VC5_QPU_OP_MUL_SHIFT 58
|
|
#define VC5_QPU_OP_MUL_MASK QPU_MASK(63, 58)
|
|
|
|
#define VC5_QPU_SIG_SHIFT 53
|
|
#define VC5_QPU_SIG_MASK QPU_MASK(57, 53)
|
|
|
|
#define VC5_QPU_COND_SHIFT 46
|
|
#define VC5_QPU_COND_MASK QPU_MASK(52, 46)
|
|
#define VC5_QPU_COND_SIG_MAGIC_ADDR (1 << 6)
|
|
|
|
#define VC5_QPU_MM QPU_MASK(45, 45)
|
|
#define VC5_QPU_MA QPU_MASK(44, 44)
|
|
|
|
#define V3D_QPU_WADDR_M_SHIFT 38
|
|
#define V3D_QPU_WADDR_M_MASK QPU_MASK(43, 38)
|
|
|
|
#define VC5_QPU_BRANCH_ADDR_LOW_SHIFT 35
|
|
#define VC5_QPU_BRANCH_ADDR_LOW_MASK QPU_MASK(55, 35)
|
|
|
|
#define V3D_QPU_WADDR_A_SHIFT 32
|
|
#define V3D_QPU_WADDR_A_MASK QPU_MASK(37, 32)
|
|
|
|
#define VC5_QPU_BRANCH_COND_SHIFT 32
|
|
#define VC5_QPU_BRANCH_COND_MASK QPU_MASK(34, 32)
|
|
|
|
#define VC5_QPU_BRANCH_ADDR_HIGH_SHIFT 24
|
|
#define VC5_QPU_BRANCH_ADDR_HIGH_MASK QPU_MASK(31, 24)
|
|
|
|
#define VC5_QPU_OP_ADD_SHIFT 24
|
|
#define VC5_QPU_OP_ADD_MASK QPU_MASK(31, 24)
|
|
|
|
#define VC5_QPU_MUL_B_SHIFT 21
|
|
#define VC5_QPU_MUL_B_MASK QPU_MASK(23, 21)
|
|
|
|
#define VC5_QPU_BRANCH_MSFIGN_SHIFT 21
|
|
#define VC5_QPU_BRANCH_MSFIGN_MASK QPU_MASK(22, 21)
|
|
|
|
#define VC5_QPU_MUL_A_SHIFT 18
|
|
#define VC5_QPU_MUL_A_MASK QPU_MASK(20, 18)
|
|
|
|
#define VC5_QPU_ADD_B_SHIFT 15
|
|
#define VC5_QPU_ADD_B_MASK QPU_MASK(17, 15)
|
|
|
|
#define VC5_QPU_BRANCH_BDU_SHIFT 15
|
|
#define VC5_QPU_BRANCH_BDU_MASK QPU_MASK(17, 15)
|
|
|
|
#define VC5_QPU_BRANCH_UB QPU_MASK(14, 14)
|
|
|
|
#define VC5_QPU_ADD_A_SHIFT 12
|
|
#define VC5_QPU_ADD_A_MASK QPU_MASK(14, 12)
|
|
|
|
#define VC5_QPU_BRANCH_BDI_SHIFT 12
|
|
#define VC5_QPU_BRANCH_BDI_MASK QPU_MASK(13, 12)
|
|
|
|
#define VC5_QPU_RADDR_A_SHIFT 6
|
|
#define VC5_QPU_RADDR_A_MASK QPU_MASK(11, 6)
|
|
|
|
#define VC5_QPU_RADDR_B_SHIFT 0
|
|
#define VC5_QPU_RADDR_B_MASK QPU_MASK(5, 0)
|
|
|
|
#define THRSW .thrsw = true
|
|
#define LDUNIF .ldunif = true
|
|
#define LDUNIFRF .ldunifrf = true
|
|
#define LDUNIFA .ldunifa = true
|
|
#define LDUNIFARF .ldunifarf = true
|
|
#define LDTMU .ldtmu = true
|
|
#define LDVARY .ldvary = true
|
|
#define LDVPM .ldvpm = true
|
|
#define SMIMM .small_imm = true
|
|
#define LDTLB .ldtlb = true
|
|
#define LDTLBU .ldtlbu = true
|
|
#define UCB .ucb = true
|
|
#define ROT .rotate = true
|
|
#define WRTMUC .wrtmuc = true
|
|
|
|
static const struct v3d_qpu_sig v33_sig_map[] = {
|
|
/* MISC R3 R4 R5 */
|
|
[0] = { },
|
|
[1] = { THRSW, },
|
|
[2] = { LDUNIF },
|
|
[3] = { THRSW, LDUNIF },
|
|
[4] = { LDTMU, },
|
|
[5] = { THRSW, LDTMU, },
|
|
[6] = { LDTMU, LDUNIF },
|
|
[7] = { THRSW, LDTMU, LDUNIF },
|
|
[8] = { LDVARY, },
|
|
[9] = { THRSW, LDVARY, },
|
|
[10] = { LDVARY, LDUNIF },
|
|
[11] = { THRSW, LDVARY, LDUNIF },
|
|
[12] = { LDVARY, LDTMU, },
|
|
[13] = { THRSW, LDVARY, LDTMU, },
|
|
[14] = { SMIMM, LDVARY, },
|
|
[15] = { SMIMM, },
|
|
[16] = { LDTLB, },
|
|
[17] = { LDTLBU, },
|
|
/* 18-21 reserved */
|
|
[22] = { UCB, },
|
|
[23] = { ROT, },
|
|
[24] = { LDVPM, },
|
|
[25] = { THRSW, LDVPM, },
|
|
[26] = { LDVPM, LDUNIF },
|
|
[27] = { THRSW, LDVPM, LDUNIF },
|
|
[28] = { LDVPM, LDTMU, },
|
|
[29] = { THRSW, LDVPM, LDTMU, },
|
|
[30] = { SMIMM, LDVPM, },
|
|
[31] = { SMIMM, },
|
|
};
|
|
|
|
static const struct v3d_qpu_sig v40_sig_map[] = {
|
|
/* MISC R3 R4 R5 */
|
|
[0] = { },
|
|
[1] = { THRSW, },
|
|
[2] = { LDUNIF },
|
|
[3] = { THRSW, LDUNIF },
|
|
[4] = { LDTMU, },
|
|
[5] = { THRSW, LDTMU, },
|
|
[6] = { LDTMU, LDUNIF },
|
|
[7] = { THRSW, LDTMU, LDUNIF },
|
|
[8] = { LDVARY, },
|
|
[9] = { THRSW, LDVARY, },
|
|
[10] = { LDVARY, LDUNIF },
|
|
[11] = { THRSW, LDVARY, LDUNIF },
|
|
/* 12-13 reserved */
|
|
[14] = { SMIMM, LDVARY, },
|
|
[15] = { SMIMM, },
|
|
[16] = { LDTLB, },
|
|
[17] = { LDTLBU, },
|
|
[18] = { WRTMUC },
|
|
[19] = { THRSW, WRTMUC },
|
|
[20] = { LDVARY, WRTMUC },
|
|
[21] = { THRSW, LDVARY, WRTMUC },
|
|
[22] = { UCB, },
|
|
[23] = { ROT, },
|
|
/* 24-30 reserved */
|
|
[31] = { SMIMM, LDTMU, },
|
|
};
|
|
|
|
static const struct v3d_qpu_sig v41_sig_map[] = {
|
|
/* MISC phys R5 */
|
|
[0] = { },
|
|
[1] = { THRSW, },
|
|
[2] = { LDUNIF },
|
|
[3] = { THRSW, LDUNIF },
|
|
[4] = { LDTMU, },
|
|
[5] = { THRSW, LDTMU, },
|
|
[6] = { LDTMU, LDUNIF },
|
|
[7] = { THRSW, LDTMU, LDUNIF },
|
|
[8] = { LDVARY, },
|
|
[9] = { THRSW, LDVARY, },
|
|
[10] = { LDVARY, LDUNIF },
|
|
[11] = { THRSW, LDVARY, LDUNIF },
|
|
[12] = { LDUNIFRF },
|
|
[13] = { THRSW, LDUNIFRF },
|
|
[14] = { SMIMM, LDVARY, },
|
|
[15] = { SMIMM, },
|
|
[16] = { LDTLB, },
|
|
[17] = { LDTLBU, },
|
|
[18] = { WRTMUC },
|
|
[19] = { THRSW, WRTMUC },
|
|
[20] = { LDVARY, WRTMUC },
|
|
[21] = { THRSW, LDVARY, WRTMUC },
|
|
[22] = { UCB, },
|
|
[23] = { ROT, },
|
|
/* 24-30 reserved */
|
|
[24] = { LDUNIFA},
|
|
[25] = { LDUNIFARF },
|
|
[31] = { SMIMM, LDTMU, },
|
|
};
|
|
|
|
bool
|
|
v3d_qpu_sig_unpack(const struct v3d_device_info *devinfo,
|
|
uint32_t packed_sig,
|
|
struct v3d_qpu_sig *sig)
|
|
{
|
|
if (packed_sig >= ARRAY_SIZE(v33_sig_map))
|
|
return false;
|
|
|
|
if (devinfo->ver >= 41)
|
|
*sig = v41_sig_map[packed_sig];
|
|
else if (devinfo->ver == 40)
|
|
*sig = v40_sig_map[packed_sig];
|
|
else
|
|
*sig = v33_sig_map[packed_sig];
|
|
|
|
/* Signals with zeroed unpacked contents after element 0 are reserved. */
|
|
return (packed_sig == 0 ||
|
|
memcmp(sig, &v33_sig_map[0], sizeof(*sig)) != 0);
|
|
}
|
|
|
|
bool
|
|
v3d_qpu_sig_pack(const struct v3d_device_info *devinfo,
|
|
const struct v3d_qpu_sig *sig,
|
|
uint32_t *packed_sig)
|
|
{
|
|
static const struct v3d_qpu_sig *map;
|
|
|
|
if (devinfo->ver >= 41)
|
|
map = v41_sig_map;
|
|
else if (devinfo->ver == 40)
|
|
map = v40_sig_map;
|
|
else
|
|
map = v33_sig_map;
|
|
|
|
for (int i = 0; i < ARRAY_SIZE(v33_sig_map); i++) {
|
|
if (memcmp(&map[i], sig, sizeof(*sig)) == 0) {
|
|
*packed_sig = i;
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
static inline unsigned
|
|
fui( float f )
|
|
{
|
|
union {float f; unsigned ui;} fi;
|
|
fi.f = f;
|
|
return fi.ui;
|
|
}
|
|
|
|
static const uint32_t small_immediates[] = {
|
|
0, 1, 2, 3,
|
|
4, 5, 6, 7,
|
|
8, 9, 10, 11,
|
|
12, 13, 14, 15,
|
|
-16, -15, -14, -13,
|
|
-12, -11, -10, -9,
|
|
-8, -7, -6, -5,
|
|
-4, -3, -2, -1,
|
|
0x3b800000, /* 2.0^-8 */
|
|
0x3c000000, /* 2.0^-7 */
|
|
0x3c800000, /* 2.0^-6 */
|
|
0x3d000000, /* 2.0^-5 */
|
|
0x3d800000, /* 2.0^-4 */
|
|
0x3e000000, /* 2.0^-3 */
|
|
0x3e800000, /* 2.0^-2 */
|
|
0x3f000000, /* 2.0^-1 */
|
|
0x3f800000, /* 2.0^0 */
|
|
0x40000000, /* 2.0^1 */
|
|
0x40800000, /* 2.0^2 */
|
|
0x41000000, /* 2.0^3 */
|
|
0x41800000, /* 2.0^4 */
|
|
0x42000000, /* 2.0^5 */
|
|
0x42800000, /* 2.0^6 */
|
|
0x43000000, /* 2.0^7 */
|
|
};
|
|
|
|
bool
|
|
v3d_qpu_small_imm_unpack(const struct v3d_device_info *devinfo,
|
|
uint32_t packed_small_immediate,
|
|
uint32_t *small_immediate)
|
|
{
|
|
if (packed_small_immediate >= ARRAY_SIZE(small_immediates))
|
|
return false;
|
|
|
|
*small_immediate = small_immediates[packed_small_immediate];
|
|
return true;
|
|
}
|
|
|
|
bool
|
|
v3d_qpu_small_imm_pack(const struct v3d_device_info *devinfo,
|
|
uint32_t value,
|
|
uint32_t *packed_small_immediate)
|
|
{
|
|
STATIC_ASSERT(ARRAY_SIZE(small_immediates) == 48);
|
|
|
|
for (int i = 0; i < ARRAY_SIZE(small_immediates); i++) {
|
|
if (small_immediates[i] == value) {
|
|
*packed_small_immediate = i;
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool
|
|
v3d_qpu_flags_unpack(const struct v3d_device_info *devinfo,
|
|
uint32_t packed_cond,
|
|
struct v3d_qpu_flags *cond)
|
|
{
|
|
static const enum v3d_qpu_cond cond_map[4] = {
|
|
[0] = V3D_QPU_COND_IFA,
|
|
[1] = V3D_QPU_COND_IFB,
|
|
[2] = V3D_QPU_COND_IFNA,
|
|
[3] = V3D_QPU_COND_IFNB,
|
|
};
|
|
|
|
cond->ac = V3D_QPU_COND_NONE;
|
|
cond->mc = V3D_QPU_COND_NONE;
|
|
cond->apf = V3D_QPU_PF_NONE;
|
|
cond->mpf = V3D_QPU_PF_NONE;
|
|
cond->auf = V3D_QPU_UF_NONE;
|
|
cond->muf = V3D_QPU_UF_NONE;
|
|
|
|
if (packed_cond == 0) {
|
|
return true;
|
|
} else if (packed_cond >> 2 == 0) {
|
|
cond->apf = packed_cond & 0x3;
|
|
} else if (packed_cond >> 4 == 0) {
|
|
cond->auf = (packed_cond & 0xf) - 4 + V3D_QPU_UF_ANDZ;
|
|
} else if (packed_cond == 0x10) {
|
|
return false;
|
|
} else if (packed_cond >> 2 == 0x4) {
|
|
cond->mpf = packed_cond & 0x3;
|
|
} else if (packed_cond >> 4 == 0x1) {
|
|
cond->muf = (packed_cond & 0xf) - 4 + V3D_QPU_UF_ANDZ;
|
|
} else if (packed_cond >> 4 == 0x2) {
|
|
cond->ac = ((packed_cond >> 2) & 0x3) + V3D_QPU_COND_IFA;
|
|
cond->mpf = packed_cond & 0x3;
|
|
} else if (packed_cond >> 4 == 0x3) {
|
|
cond->mc = ((packed_cond >> 2) & 0x3) + V3D_QPU_COND_IFA;
|
|
cond->apf = packed_cond & 0x3;
|
|
} else if (packed_cond >> 6) {
|
|
cond->mc = cond_map[(packed_cond >> 4) & 0x3];
|
|
if (((packed_cond >> 2) & 0x3) == 0) {
|
|
cond->ac = cond_map[packed_cond & 0x3];
|
|
} else {
|
|
cond->auf = (packed_cond & 0xf) - 4 + V3D_QPU_UF_ANDZ;
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool
|
|
v3d_qpu_flags_pack(const struct v3d_device_info *devinfo,
|
|
const struct v3d_qpu_flags *cond,
|
|
uint32_t *packed_cond)
|
|
{
|
|
#define AC (1 << 0)
|
|
#define MC (1 << 1)
|
|
#define APF (1 << 2)
|
|
#define MPF (1 << 3)
|
|
#define AUF (1 << 4)
|
|
#define MUF (1 << 5)
|
|
static const struct {
|
|
uint8_t flags_present;
|
|
uint8_t bits;
|
|
} flags_table[] = {
|
|
{ 0, 0 },
|
|
{ APF, 0 },
|
|
{ AUF, 0 },
|
|
{ MPF, (1 << 4) },
|
|
{ MUF, (1 << 4) },
|
|
{ AC, (1 << 5) },
|
|
{ AC | MPF, (1 << 5) },
|
|
{ MC, (1 << 5) | (1 << 4) },
|
|
{ MC | APF, (1 << 5) | (1 << 4) },
|
|
{ MC | AC, (1 << 6) },
|
|
{ MC | AUF, (1 << 6) },
|
|
};
|
|
|
|
uint8_t flags_present = 0;
|
|
if (cond->ac != V3D_QPU_COND_NONE)
|
|
flags_present |= AC;
|
|
if (cond->mc != V3D_QPU_COND_NONE)
|
|
flags_present |= MC;
|
|
if (cond->apf != V3D_QPU_PF_NONE)
|
|
flags_present |= APF;
|
|
if (cond->mpf != V3D_QPU_PF_NONE)
|
|
flags_present |= MPF;
|
|
if (cond->auf != V3D_QPU_UF_NONE)
|
|
flags_present |= AUF;
|
|
if (cond->muf != V3D_QPU_UF_NONE)
|
|
flags_present |= MUF;
|
|
|
|
for (int i = 0; i < ARRAY_SIZE(flags_table); i++) {
|
|
if (flags_table[i].flags_present != flags_present)
|
|
continue;
|
|
|
|
*packed_cond = flags_table[i].bits;
|
|
|
|
*packed_cond |= cond->apf;
|
|
*packed_cond |= cond->mpf;
|
|
|
|
if (flags_present & AUF)
|
|
*packed_cond |= cond->auf - V3D_QPU_UF_ANDZ + 4;
|
|
if (flags_present & MUF)
|
|
*packed_cond |= cond->muf - V3D_QPU_UF_ANDZ + 4;
|
|
|
|
if (flags_present & AC)
|
|
*packed_cond |= (cond->ac - V3D_QPU_COND_IFA) << 2;
|
|
|
|
if (flags_present & MC) {
|
|
if (*packed_cond & (1 << 6))
|
|
*packed_cond |= (cond->mc -
|
|
V3D_QPU_COND_IFA) << 4;
|
|
else
|
|
*packed_cond |= (cond->mc -
|
|
V3D_QPU_COND_IFA) << 2;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
/* Make a mapping of the table of opcodes in the spec. The opcode is
|
|
* determined by a combination of the opcode field, and in the case of 0 or
|
|
* 1-arg opcodes, the mux_b field as well.
|
|
*/
|
|
#define MUX_MASK(bot, top) (((1 << (top + 1)) - 1) - ((1 << (bot)) - 1))
|
|
#define ANYMUX MUX_MASK(0, 7)
|
|
|
|
struct opcode_desc {
|
|
uint8_t opcode_first;
|
|
uint8_t opcode_last;
|
|
uint8_t mux_b_mask;
|
|
uint8_t mux_a_mask;
|
|
uint8_t op;
|
|
/* 0 if it's the same across V3D versions, or a specific V3D version. */
|
|
uint8_t ver;
|
|
};
|
|
|
|
static const struct opcode_desc add_ops[] = {
|
|
/* FADD is FADDNF depending on the order of the mux_a/mux_b. */
|
|
{ 0, 47, ANYMUX, ANYMUX, V3D_QPU_A_FADD },
|
|
{ 0, 47, ANYMUX, ANYMUX, V3D_QPU_A_FADDNF },
|
|
{ 53, 55, ANYMUX, ANYMUX, V3D_QPU_A_VFPACK },
|
|
{ 56, 56, ANYMUX, ANYMUX, V3D_QPU_A_ADD },
|
|
{ 57, 59, ANYMUX, ANYMUX, V3D_QPU_A_VFPACK },
|
|
{ 60, 60, ANYMUX, ANYMUX, V3D_QPU_A_SUB },
|
|
{ 61, 63, ANYMUX, ANYMUX, V3D_QPU_A_VFPACK },
|
|
{ 64, 111, ANYMUX, ANYMUX, V3D_QPU_A_FSUB },
|
|
{ 120, 120, ANYMUX, ANYMUX, V3D_QPU_A_MIN },
|
|
{ 121, 121, ANYMUX, ANYMUX, V3D_QPU_A_MAX },
|
|
{ 122, 122, ANYMUX, ANYMUX, V3D_QPU_A_UMIN },
|
|
{ 123, 123, ANYMUX, ANYMUX, V3D_QPU_A_UMAX },
|
|
{ 124, 124, ANYMUX, ANYMUX, V3D_QPU_A_SHL },
|
|
{ 125, 125, ANYMUX, ANYMUX, V3D_QPU_A_SHR },
|
|
{ 126, 126, ANYMUX, ANYMUX, V3D_QPU_A_ASR },
|
|
{ 127, 127, ANYMUX, ANYMUX, V3D_QPU_A_ROR },
|
|
/* FMIN is instead FMAX depending on the order of the mux_a/mux_b. */
|
|
{ 128, 175, ANYMUX, ANYMUX, V3D_QPU_A_FMIN },
|
|
{ 128, 175, ANYMUX, ANYMUX, V3D_QPU_A_FMAX },
|
|
{ 176, 180, ANYMUX, ANYMUX, V3D_QPU_A_VFMIN },
|
|
|
|
{ 181, 181, ANYMUX, ANYMUX, V3D_QPU_A_AND },
|
|
{ 182, 182, ANYMUX, ANYMUX, V3D_QPU_A_OR },
|
|
{ 183, 183, ANYMUX, ANYMUX, V3D_QPU_A_XOR },
|
|
|
|
{ 184, 184, ANYMUX, ANYMUX, V3D_QPU_A_VADD },
|
|
{ 185, 185, ANYMUX, ANYMUX, V3D_QPU_A_VSUB },
|
|
{ 186, 186, 1 << 0, ANYMUX, V3D_QPU_A_NOT },
|
|
{ 186, 186, 1 << 1, ANYMUX, V3D_QPU_A_NEG },
|
|
{ 186, 186, 1 << 2, ANYMUX, V3D_QPU_A_FLAPUSH },
|
|
{ 186, 186, 1 << 3, ANYMUX, V3D_QPU_A_FLBPUSH },
|
|
{ 186, 186, 1 << 4, ANYMUX, V3D_QPU_A_FLPOP },
|
|
{ 186, 186, 1 << 5, ANYMUX, V3D_QPU_A_RECIP },
|
|
{ 186, 186, 1 << 6, ANYMUX, V3D_QPU_A_SETMSF },
|
|
{ 186, 186, 1 << 7, ANYMUX, V3D_QPU_A_SETREVF },
|
|
{ 187, 187, 1 << 0, 1 << 0, V3D_QPU_A_NOP, 0 },
|
|
{ 187, 187, 1 << 0, 1 << 1, V3D_QPU_A_TIDX },
|
|
{ 187, 187, 1 << 0, 1 << 2, V3D_QPU_A_EIDX },
|
|
{ 187, 187, 1 << 0, 1 << 3, V3D_QPU_A_LR },
|
|
{ 187, 187, 1 << 0, 1 << 4, V3D_QPU_A_VFLA },
|
|
{ 187, 187, 1 << 0, 1 << 5, V3D_QPU_A_VFLNA },
|
|
{ 187, 187, 1 << 0, 1 << 6, V3D_QPU_A_VFLB },
|
|
{ 187, 187, 1 << 0, 1 << 7, V3D_QPU_A_VFLNB },
|
|
|
|
{ 187, 187, 1 << 1, MUX_MASK(0, 2), V3D_QPU_A_FXCD },
|
|
{ 187, 187, 1 << 1, 1 << 3, V3D_QPU_A_XCD },
|
|
{ 187, 187, 1 << 1, MUX_MASK(4, 6), V3D_QPU_A_FYCD },
|
|
{ 187, 187, 1 << 1, 1 << 7, V3D_QPU_A_YCD },
|
|
|
|
{ 187, 187, 1 << 2, 1 << 0, V3D_QPU_A_MSF },
|
|
{ 187, 187, 1 << 2, 1 << 1, V3D_QPU_A_REVF },
|
|
{ 187, 187, 1 << 2, 1 << 2, V3D_QPU_A_VDWWT, 33 },
|
|
{ 187, 187, 1 << 2, 1 << 2, V3D_QPU_A_IID, 40 },
|
|
{ 187, 187, 1 << 2, 1 << 3, V3D_QPU_A_SAMPID, 40 },
|
|
{ 187, 187, 1 << 2, 1 << 4, V3D_QPU_A_BARRIERID, 40 },
|
|
{ 187, 187, 1 << 2, 1 << 5, V3D_QPU_A_TMUWT },
|
|
{ 187, 187, 1 << 2, 1 << 6, V3D_QPU_A_VPMWT },
|
|
|
|
{ 187, 187, 1 << 3, ANYMUX, V3D_QPU_A_VPMSETUP, 33 },
|
|
{ 188, 188, 1 << 0, ANYMUX, V3D_QPU_A_LDVPMV_IN, 40 },
|
|
{ 188, 188, 1 << 1, ANYMUX, V3D_QPU_A_LDVPMD_IN, 40 },
|
|
{ 188, 188, 1 << 2, ANYMUX, V3D_QPU_A_LDVPMP, 40 },
|
|
{ 188, 188, 1 << 3, ANYMUX, V3D_QPU_A_RSQRT, 41 },
|
|
{ 188, 188, 1 << 4, ANYMUX, V3D_QPU_A_EXP, 41 },
|
|
{ 188, 188, 1 << 5, ANYMUX, V3D_QPU_A_LOG, 41 },
|
|
{ 188, 188, 1 << 6, ANYMUX, V3D_QPU_A_SIN, 41 },
|
|
{ 188, 188, 1 << 7, ANYMUX, V3D_QPU_A_RSQRT2, 41 },
|
|
{ 189, 189, ANYMUX, ANYMUX, V3D_QPU_A_LDVPMG_IN, 40 },
|
|
|
|
/* FIXME: MORE COMPLICATED */
|
|
/* { 190, 191, ANYMUX, ANYMUX, V3D_QPU_A_VFMOVABSNEGNAB }, */
|
|
|
|
{ 192, 239, ANYMUX, ANYMUX, V3D_QPU_A_FCMP },
|
|
{ 240, 244, ANYMUX, ANYMUX, V3D_QPU_A_VFMAX },
|
|
|
|
{ 245, 245, MUX_MASK(0, 2), ANYMUX, V3D_QPU_A_FROUND },
|
|
{ 245, 245, 1 << 3, ANYMUX, V3D_QPU_A_FTOIN },
|
|
{ 245, 245, MUX_MASK(4, 6), ANYMUX, V3D_QPU_A_FTRUNC },
|
|
{ 245, 245, 1 << 7, ANYMUX, V3D_QPU_A_FTOIZ },
|
|
{ 246, 246, MUX_MASK(0, 2), ANYMUX, V3D_QPU_A_FFLOOR },
|
|
{ 246, 246, 1 << 3, ANYMUX, V3D_QPU_A_FTOUZ },
|
|
{ 246, 246, MUX_MASK(4, 6), ANYMUX, V3D_QPU_A_FCEIL },
|
|
{ 246, 246, 1 << 7, ANYMUX, V3D_QPU_A_FTOC },
|
|
|
|
{ 247, 247, MUX_MASK(0, 2), ANYMUX, V3D_QPU_A_FDX },
|
|
{ 247, 247, MUX_MASK(4, 6), ANYMUX, V3D_QPU_A_FDY },
|
|
|
|
/* The stvpms are distinguished by the waddr field. */
|
|
{ 248, 248, ANYMUX, ANYMUX, V3D_QPU_A_STVPMV },
|
|
{ 248, 248, ANYMUX, ANYMUX, V3D_QPU_A_STVPMD },
|
|
{ 248, 248, ANYMUX, ANYMUX, V3D_QPU_A_STVPMP },
|
|
|
|
{ 252, 252, MUX_MASK(0, 2), ANYMUX, V3D_QPU_A_ITOF },
|
|
{ 252, 252, 1 << 3, ANYMUX, V3D_QPU_A_CLZ },
|
|
{ 252, 252, MUX_MASK(4, 6), ANYMUX, V3D_QPU_A_UTOF },
|
|
};
|
|
|
|
static const struct opcode_desc mul_ops[] = {
|
|
{ 1, 1, ANYMUX, ANYMUX, V3D_QPU_M_ADD },
|
|
{ 2, 2, ANYMUX, ANYMUX, V3D_QPU_M_SUB },
|
|
{ 3, 3, ANYMUX, ANYMUX, V3D_QPU_M_UMUL24 },
|
|
{ 4, 8, ANYMUX, ANYMUX, V3D_QPU_M_VFMUL },
|
|
{ 9, 9, ANYMUX, ANYMUX, V3D_QPU_M_SMUL24 },
|
|
{ 10, 10, ANYMUX, ANYMUX, V3D_QPU_M_MULTOP },
|
|
{ 14, 14, ANYMUX, ANYMUX, V3D_QPU_M_FMOV },
|
|
{ 15, 15, MUX_MASK(0, 3), ANYMUX, V3D_QPU_M_FMOV },
|
|
{ 15, 15, 1 << 4, 1 << 0, V3D_QPU_M_NOP, 0 },
|
|
{ 15, 15, 1 << 7, ANYMUX, V3D_QPU_M_MOV },
|
|
{ 16, 63, ANYMUX, ANYMUX, V3D_QPU_M_FMUL },
|
|
};
|
|
|
|
static const struct opcode_desc *
|
|
lookup_opcode(const struct opcode_desc *opcodes, size_t num_opcodes,
|
|
uint32_t opcode, uint32_t mux_a, uint32_t mux_b)
|
|
{
|
|
for (int i = 0; i < num_opcodes; i++) {
|
|
const struct opcode_desc *op_desc = &opcodes[i];
|
|
|
|
if (opcode < op_desc->opcode_first ||
|
|
opcode > op_desc->opcode_last)
|
|
continue;
|
|
|
|
if (!(op_desc->mux_b_mask & (1 << mux_b)))
|
|
continue;
|
|
|
|
if (!(op_desc->mux_a_mask & (1 << mux_a)))
|
|
continue;
|
|
|
|
return op_desc;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_float32_unpack_unpack(uint32_t packed,
|
|
enum v3d_qpu_input_unpack *unpacked)
|
|
{
|
|
switch (packed) {
|
|
case 0:
|
|
*unpacked = V3D_QPU_UNPACK_ABS;
|
|
return true;
|
|
case 1:
|
|
*unpacked = V3D_QPU_UNPACK_NONE;
|
|
return true;
|
|
case 2:
|
|
*unpacked = V3D_QPU_UNPACK_L;
|
|
return true;
|
|
case 3:
|
|
*unpacked = V3D_QPU_UNPACK_H;
|
|
return true;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_float32_unpack_pack(enum v3d_qpu_input_unpack unpacked,
|
|
uint32_t *packed)
|
|
{
|
|
switch (unpacked) {
|
|
case V3D_QPU_UNPACK_ABS:
|
|
*packed = 0;
|
|
return true;
|
|
case V3D_QPU_UNPACK_NONE:
|
|
*packed = 1;
|
|
return true;
|
|
case V3D_QPU_UNPACK_L:
|
|
*packed = 2;
|
|
return true;
|
|
case V3D_QPU_UNPACK_H:
|
|
*packed = 3;
|
|
return true;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_float16_unpack_unpack(uint32_t packed,
|
|
enum v3d_qpu_input_unpack *unpacked)
|
|
{
|
|
switch (packed) {
|
|
case 0:
|
|
*unpacked = V3D_QPU_UNPACK_NONE;
|
|
return true;
|
|
case 1:
|
|
*unpacked = V3D_QPU_UNPACK_REPLICATE_32F_16;
|
|
return true;
|
|
case 2:
|
|
*unpacked = V3D_QPU_UNPACK_REPLICATE_L_16;
|
|
return true;
|
|
case 3:
|
|
*unpacked = V3D_QPU_UNPACK_REPLICATE_H_16;
|
|
return true;
|
|
case 4:
|
|
*unpacked = V3D_QPU_UNPACK_SWAP_16;
|
|
return true;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_float16_unpack_pack(enum v3d_qpu_input_unpack unpacked,
|
|
uint32_t *packed)
|
|
{
|
|
switch (unpacked) {
|
|
case V3D_QPU_UNPACK_NONE:
|
|
*packed = 0;
|
|
return true;
|
|
case V3D_QPU_UNPACK_REPLICATE_32F_16:
|
|
*packed = 1;
|
|
return true;
|
|
case V3D_QPU_UNPACK_REPLICATE_L_16:
|
|
*packed = 2;
|
|
return true;
|
|
case V3D_QPU_UNPACK_REPLICATE_H_16:
|
|
*packed = 3;
|
|
return true;
|
|
case V3D_QPU_UNPACK_SWAP_16:
|
|
*packed = 4;
|
|
return true;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_float32_pack_pack(enum v3d_qpu_input_unpack unpacked,
|
|
uint32_t *packed)
|
|
{
|
|
switch (unpacked) {
|
|
case V3D_QPU_PACK_NONE:
|
|
*packed = 0;
|
|
return true;
|
|
case V3D_QPU_PACK_L:
|
|
*packed = 1;
|
|
return true;
|
|
case V3D_QPU_PACK_H:
|
|
*packed = 2;
|
|
return true;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_add_unpack(const struct v3d_device_info *devinfo, uint64_t packed_inst,
|
|
struct v3d_qpu_instr *instr)
|
|
{
|
|
uint32_t op = QPU_GET_FIELD(packed_inst, VC5_QPU_OP_ADD);
|
|
uint32_t mux_a = QPU_GET_FIELD(packed_inst, VC5_QPU_ADD_A);
|
|
uint32_t mux_b = QPU_GET_FIELD(packed_inst, VC5_QPU_ADD_B);
|
|
uint32_t waddr = QPU_GET_FIELD(packed_inst, V3D_QPU_WADDR_A);
|
|
|
|
uint32_t map_op = op;
|
|
/* Some big clusters of opcodes are replicated with unpack
|
|
* flags
|
|
*/
|
|
if (map_op >= 249 && map_op <= 251)
|
|
map_op = (map_op - 249 + 245);
|
|
if (map_op >= 253 && map_op <= 255)
|
|
map_op = (map_op - 253 + 245);
|
|
|
|
const struct opcode_desc *desc =
|
|
lookup_opcode(add_ops, ARRAY_SIZE(add_ops),
|
|
map_op, mux_a, mux_b);
|
|
if (!desc)
|
|
return false;
|
|
|
|
instr->alu.add.op = desc->op;
|
|
|
|
/* FADD/FADDNF and FMIN/FMAX are determined by the orders of the
|
|
* operands.
|
|
*/
|
|
if (((op >> 2) & 3) * 8 + mux_a > (op & 3) * 8 + mux_b) {
|
|
if (instr->alu.add.op == V3D_QPU_A_FMIN)
|
|
instr->alu.add.op = V3D_QPU_A_FMAX;
|
|
if (instr->alu.add.op == V3D_QPU_A_FADD)
|
|
instr->alu.add.op = V3D_QPU_A_FADDNF;
|
|
}
|
|
|
|
/* Some QPU ops require a bit more than just basic opcode and mux a/b
|
|
* comparisons to distinguish them.
|
|
*/
|
|
switch (instr->alu.add.op) {
|
|
case V3D_QPU_A_STVPMV:
|
|
case V3D_QPU_A_STVPMD:
|
|
case V3D_QPU_A_STVPMP:
|
|
switch (waddr) {
|
|
case 0:
|
|
instr->alu.add.op = V3D_QPU_A_STVPMV;
|
|
break;
|
|
case 1:
|
|
instr->alu.add.op = V3D_QPU_A_STVPMD;
|
|
break;
|
|
case 2:
|
|
instr->alu.add.op = V3D_QPU_A_STVPMP;
|
|
break;
|
|
default:
|
|
return false;
|
|
}
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
switch (instr->alu.add.op) {
|
|
case V3D_QPU_A_FADD:
|
|
case V3D_QPU_A_FADDNF:
|
|
case V3D_QPU_A_FSUB:
|
|
case V3D_QPU_A_FMIN:
|
|
case V3D_QPU_A_FMAX:
|
|
case V3D_QPU_A_FCMP:
|
|
instr->alu.add.output_pack = (op >> 4) & 0x3;
|
|
|
|
if (!v3d_qpu_float32_unpack_unpack((op >> 2) & 0x3,
|
|
&instr->alu.add.a_unpack)) {
|
|
return false;
|
|
}
|
|
|
|
if (!v3d_qpu_float32_unpack_unpack((op >> 0) & 0x3,
|
|
&instr->alu.add.b_unpack)) {
|
|
return false;
|
|
}
|
|
break;
|
|
|
|
case V3D_QPU_A_FFLOOR:
|
|
case V3D_QPU_A_FROUND:
|
|
case V3D_QPU_A_FTRUNC:
|
|
case V3D_QPU_A_FCEIL:
|
|
case V3D_QPU_A_FDX:
|
|
case V3D_QPU_A_FDY:
|
|
instr->alu.add.output_pack = mux_b & 0x3;
|
|
|
|
if (!v3d_qpu_float32_unpack_unpack((op >> 2) & 0x3,
|
|
&instr->alu.add.a_unpack)) {
|
|
return false;
|
|
}
|
|
break;
|
|
|
|
case V3D_QPU_A_FTOIN:
|
|
case V3D_QPU_A_FTOIZ:
|
|
case V3D_QPU_A_FTOUZ:
|
|
case V3D_QPU_A_FTOC:
|
|
instr->alu.add.output_pack = V3D_QPU_PACK_NONE;
|
|
|
|
if (!v3d_qpu_float32_unpack_unpack((op >> 2) & 0x3,
|
|
&instr->alu.add.a_unpack)) {
|
|
return false;
|
|
}
|
|
break;
|
|
|
|
case V3D_QPU_A_VFMIN:
|
|
case V3D_QPU_A_VFMAX:
|
|
if (!v3d_qpu_float16_unpack_unpack(op & 0x7,
|
|
&instr->alu.add.a_unpack)) {
|
|
return false;
|
|
}
|
|
|
|
instr->alu.add.output_pack = V3D_QPU_PACK_NONE;
|
|
instr->alu.add.b_unpack = V3D_QPU_UNPACK_NONE;
|
|
break;
|
|
|
|
default:
|
|
instr->alu.add.output_pack = V3D_QPU_PACK_NONE;
|
|
instr->alu.add.a_unpack = V3D_QPU_UNPACK_NONE;
|
|
instr->alu.add.b_unpack = V3D_QPU_UNPACK_NONE;
|
|
break;
|
|
}
|
|
|
|
instr->alu.add.a = mux_a;
|
|
instr->alu.add.b = mux_b;
|
|
instr->alu.add.waddr = QPU_GET_FIELD(packed_inst, V3D_QPU_WADDR_A);
|
|
|
|
instr->alu.add.magic_write = false;
|
|
if (packed_inst & VC5_QPU_MA) {
|
|
switch (instr->alu.add.op) {
|
|
case V3D_QPU_A_LDVPMV_IN:
|
|
instr->alu.add.op = V3D_QPU_A_LDVPMV_OUT;
|
|
break;
|
|
case V3D_QPU_A_LDVPMD_IN:
|
|
instr->alu.add.op = V3D_QPU_A_LDVPMD_OUT;
|
|
break;
|
|
case V3D_QPU_A_LDVPMG_IN:
|
|
instr->alu.add.op = V3D_QPU_A_LDVPMG_OUT;
|
|
break;
|
|
default:
|
|
instr->alu.add.magic_write = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_mul_unpack(const struct v3d_device_info *devinfo, uint64_t packed_inst,
|
|
struct v3d_qpu_instr *instr)
|
|
{
|
|
uint32_t op = QPU_GET_FIELD(packed_inst, VC5_QPU_OP_MUL);
|
|
uint32_t mux_a = QPU_GET_FIELD(packed_inst, VC5_QPU_MUL_A);
|
|
uint32_t mux_b = QPU_GET_FIELD(packed_inst, VC5_QPU_MUL_B);
|
|
|
|
{
|
|
const struct opcode_desc *desc =
|
|
lookup_opcode(mul_ops, ARRAY_SIZE(mul_ops),
|
|
op, mux_a, mux_b);
|
|
if (!desc)
|
|
return false;
|
|
|
|
instr->alu.mul.op = desc->op;
|
|
}
|
|
|
|
switch (instr->alu.mul.op) {
|
|
case V3D_QPU_M_FMUL:
|
|
instr->alu.mul.output_pack = ((op >> 4) & 0x3) - 1;
|
|
|
|
if (!v3d_qpu_float32_unpack_unpack((op >> 2) & 0x3,
|
|
&instr->alu.mul.a_unpack)) {
|
|
return false;
|
|
}
|
|
|
|
if (!v3d_qpu_float32_unpack_unpack((op >> 0) & 0x3,
|
|
&instr->alu.mul.b_unpack)) {
|
|
return false;
|
|
}
|
|
|
|
break;
|
|
|
|
case V3D_QPU_M_FMOV:
|
|
instr->alu.mul.output_pack = (((op & 1) << 1) +
|
|
((mux_b >> 2) & 1));
|
|
|
|
if (!v3d_qpu_float32_unpack_unpack(mux_b & 0x3,
|
|
&instr->alu.mul.a_unpack)) {
|
|
return false;
|
|
}
|
|
|
|
break;
|
|
|
|
case V3D_QPU_M_VFMUL:
|
|
instr->alu.mul.output_pack = V3D_QPU_PACK_NONE;
|
|
|
|
if (!v3d_qpu_float16_unpack_unpack(((op & 0x7) - 4) & 7,
|
|
&instr->alu.mul.a_unpack)) {
|
|
return false;
|
|
}
|
|
|
|
instr->alu.mul.b_unpack = V3D_QPU_UNPACK_NONE;
|
|
|
|
break;
|
|
|
|
default:
|
|
instr->alu.mul.output_pack = V3D_QPU_PACK_NONE;
|
|
instr->alu.mul.a_unpack = V3D_QPU_UNPACK_NONE;
|
|
instr->alu.mul.b_unpack = V3D_QPU_UNPACK_NONE;
|
|
break;
|
|
}
|
|
|
|
instr->alu.mul.a = mux_a;
|
|
instr->alu.mul.b = mux_b;
|
|
instr->alu.mul.waddr = QPU_GET_FIELD(packed_inst, V3D_QPU_WADDR_M);
|
|
instr->alu.mul.magic_write = packed_inst & VC5_QPU_MM;
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_add_pack(const struct v3d_device_info *devinfo,
|
|
const struct v3d_qpu_instr *instr, uint64_t *packed_instr)
|
|
{
|
|
uint32_t waddr = instr->alu.add.waddr;
|
|
uint32_t mux_a = instr->alu.add.a;
|
|
uint32_t mux_b = instr->alu.add.b;
|
|
int nsrc = v3d_qpu_add_op_num_src(instr->alu.add.op);
|
|
const struct opcode_desc *desc;
|
|
|
|
int opcode;
|
|
for (desc = add_ops; desc != &add_ops[ARRAY_SIZE(add_ops)];
|
|
desc++) {
|
|
if (desc->op == instr->alu.add.op)
|
|
break;
|
|
}
|
|
if (desc == &add_ops[ARRAY_SIZE(add_ops)])
|
|
return false;
|
|
|
|
opcode = desc->opcode_first;
|
|
|
|
/* If an operation doesn't use an arg, its mux values may be used to
|
|
* identify the operation type.
|
|
*/
|
|
if (nsrc < 2)
|
|
mux_b = ffs(desc->mux_b_mask) - 1;
|
|
|
|
if (nsrc < 1)
|
|
mux_a = ffs(desc->mux_a_mask) - 1;
|
|
|
|
bool no_magic_write = false;
|
|
|
|
switch (instr->alu.add.op) {
|
|
case V3D_QPU_A_STVPMV:
|
|
waddr = 0;
|
|
no_magic_write = true;
|
|
break;
|
|
case V3D_QPU_A_STVPMD:
|
|
waddr = 1;
|
|
no_magic_write = true;
|
|
break;
|
|
case V3D_QPU_A_STVPMP:
|
|
waddr = 2;
|
|
no_magic_write = true;
|
|
break;
|
|
|
|
case V3D_QPU_A_LDVPMV_IN:
|
|
case V3D_QPU_A_LDVPMD_IN:
|
|
case V3D_QPU_A_LDVPMP:
|
|
case V3D_QPU_A_LDVPMG_IN:
|
|
assert(!instr->alu.add.magic_write);
|
|
break;
|
|
|
|
case V3D_QPU_A_LDVPMV_OUT:
|
|
case V3D_QPU_A_LDVPMD_OUT:
|
|
case V3D_QPU_A_LDVPMG_OUT:
|
|
assert(!instr->alu.add.magic_write);
|
|
*packed_instr |= VC5_QPU_MA;
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
switch (instr->alu.add.op) {
|
|
case V3D_QPU_A_FADD:
|
|
case V3D_QPU_A_FADDNF:
|
|
case V3D_QPU_A_FSUB:
|
|
case V3D_QPU_A_FMIN:
|
|
case V3D_QPU_A_FMAX:
|
|
case V3D_QPU_A_FCMP: {
|
|
uint32_t output_pack;
|
|
uint32_t a_unpack;
|
|
uint32_t b_unpack;
|
|
|
|
if (!v3d_qpu_float32_pack_pack(instr->alu.add.output_pack,
|
|
&output_pack)) {
|
|
return false;
|
|
}
|
|
opcode |= output_pack << 4;
|
|
|
|
if (!v3d_qpu_float32_unpack_pack(instr->alu.add.a_unpack,
|
|
&a_unpack)) {
|
|
return false;
|
|
}
|
|
|
|
if (!v3d_qpu_float32_unpack_pack(instr->alu.add.b_unpack,
|
|
&b_unpack)) {
|
|
return false;
|
|
}
|
|
|
|
/* These operations with commutative operands are
|
|
* distinguished by which order their operands come in.
|
|
*/
|
|
bool ordering = a_unpack * 8 + mux_a > b_unpack * 8 + mux_b;
|
|
if (((instr->alu.add.op == V3D_QPU_A_FMIN ||
|
|
instr->alu.add.op == V3D_QPU_A_FADD) && ordering) ||
|
|
((instr->alu.add.op == V3D_QPU_A_FMAX ||
|
|
instr->alu.add.op == V3D_QPU_A_FADDNF) && !ordering)) {
|
|
uint32_t temp;
|
|
|
|
temp = a_unpack;
|
|
a_unpack = b_unpack;
|
|
b_unpack = temp;
|
|
|
|
temp = mux_a;
|
|
mux_a = mux_b;
|
|
mux_b = temp;
|
|
}
|
|
|
|
opcode |= a_unpack << 2;
|
|
opcode |= b_unpack << 0;
|
|
break;
|
|
}
|
|
|
|
case V3D_QPU_A_FFLOOR:
|
|
case V3D_QPU_A_FROUND:
|
|
case V3D_QPU_A_FTRUNC:
|
|
case V3D_QPU_A_FCEIL:
|
|
case V3D_QPU_A_FDX:
|
|
case V3D_QPU_A_FDY: {
|
|
uint32_t packed;
|
|
|
|
if (!v3d_qpu_float32_pack_pack(instr->alu.add.output_pack,
|
|
&packed)) {
|
|
return false;
|
|
}
|
|
mux_b |= packed;
|
|
|
|
if (!v3d_qpu_float32_unpack_pack(instr->alu.add.a_unpack,
|
|
&packed)) {
|
|
return false;
|
|
}
|
|
if (packed == 0)
|
|
return false;
|
|
opcode |= packed << 2;
|
|
break;
|
|
}
|
|
|
|
case V3D_QPU_A_FTOIN:
|
|
case V3D_QPU_A_FTOIZ:
|
|
case V3D_QPU_A_FTOUZ:
|
|
case V3D_QPU_A_FTOC:
|
|
if (instr->alu.add.output_pack != V3D_QPU_PACK_NONE)
|
|
return false;
|
|
|
|
uint32_t packed;
|
|
if (!v3d_qpu_float32_unpack_pack(instr->alu.add.a_unpack,
|
|
&packed)) {
|
|
return false;
|
|
}
|
|
if (packed == 0)
|
|
return false;
|
|
opcode |= packed << 2;
|
|
|
|
break;
|
|
|
|
case V3D_QPU_A_VFMIN:
|
|
case V3D_QPU_A_VFMAX:
|
|
if (instr->alu.add.output_pack != V3D_QPU_PACK_NONE ||
|
|
instr->alu.add.b_unpack != V3D_QPU_UNPACK_NONE) {
|
|
return false;
|
|
}
|
|
|
|
if (!v3d_qpu_float16_unpack_pack(instr->alu.add.a_unpack,
|
|
&packed)) {
|
|
return false;
|
|
}
|
|
opcode |= packed;
|
|
break;
|
|
|
|
default:
|
|
if (instr->alu.add.op != V3D_QPU_A_NOP &&
|
|
(instr->alu.add.output_pack != V3D_QPU_PACK_NONE ||
|
|
instr->alu.add.a_unpack != V3D_QPU_UNPACK_NONE ||
|
|
instr->alu.add.b_unpack != V3D_QPU_UNPACK_NONE)) {
|
|
return false;
|
|
}
|
|
break;
|
|
}
|
|
|
|
*packed_instr |= QPU_SET_FIELD(mux_a, VC5_QPU_ADD_A);
|
|
*packed_instr |= QPU_SET_FIELD(mux_b, VC5_QPU_ADD_B);
|
|
*packed_instr |= QPU_SET_FIELD(opcode, VC5_QPU_OP_ADD);
|
|
*packed_instr |= QPU_SET_FIELD(waddr, V3D_QPU_WADDR_A);
|
|
if (instr->alu.add.magic_write && !no_magic_write)
|
|
*packed_instr |= VC5_QPU_MA;
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_mul_pack(const struct v3d_device_info *devinfo,
|
|
const struct v3d_qpu_instr *instr, uint64_t *packed_instr)
|
|
{
|
|
uint32_t mux_a = instr->alu.mul.a;
|
|
uint32_t mux_b = instr->alu.mul.b;
|
|
int nsrc = v3d_qpu_mul_op_num_src(instr->alu.mul.op);
|
|
const struct opcode_desc *desc;
|
|
|
|
for (desc = mul_ops; desc != &mul_ops[ARRAY_SIZE(mul_ops)];
|
|
desc++) {
|
|
if (desc->op == instr->alu.mul.op)
|
|
break;
|
|
}
|
|
if (desc == &mul_ops[ARRAY_SIZE(mul_ops)])
|
|
return false;
|
|
|
|
uint32_t opcode = desc->opcode_first;
|
|
|
|
/* Some opcodes have a single valid value for their mux a/b, so set
|
|
* that here. If mux a/b determine packing, it will be set below.
|
|
*/
|
|
if (nsrc < 2)
|
|
mux_b = ffs(desc->mux_b_mask) - 1;
|
|
|
|
if (nsrc < 1)
|
|
mux_a = ffs(desc->mux_a_mask) - 1;
|
|
|
|
switch (instr->alu.mul.op) {
|
|
case V3D_QPU_M_FMUL: {
|
|
uint32_t packed;
|
|
|
|
if (!v3d_qpu_float32_pack_pack(instr->alu.mul.output_pack,
|
|
&packed)) {
|
|
return false;
|
|
}
|
|
/* No need for a +1 because desc->opcode_first has a 1 in this
|
|
* field.
|
|
*/
|
|
opcode += packed << 4;
|
|
|
|
if (!v3d_qpu_float32_unpack_pack(instr->alu.mul.a_unpack,
|
|
&packed)) {
|
|
return false;
|
|
}
|
|
opcode |= packed << 2;
|
|
|
|
if (!v3d_qpu_float32_unpack_pack(instr->alu.mul.b_unpack,
|
|
&packed)) {
|
|
return false;
|
|
}
|
|
opcode |= packed << 0;
|
|
break;
|
|
}
|
|
|
|
case V3D_QPU_M_FMOV: {
|
|
uint32_t packed;
|
|
|
|
if (!v3d_qpu_float32_pack_pack(instr->alu.mul.output_pack,
|
|
&packed)) {
|
|
return false;
|
|
}
|
|
opcode |= (packed >> 1) & 1;
|
|
mux_b = (packed & 1) << 2;
|
|
|
|
if (!v3d_qpu_float32_unpack_pack(instr->alu.mul.a_unpack,
|
|
&packed)) {
|
|
return false;
|
|
}
|
|
mux_b |= packed;
|
|
break;
|
|
}
|
|
|
|
case V3D_QPU_M_VFMUL: {
|
|
uint32_t packed;
|
|
|
|
if (instr->alu.mul.output_pack != V3D_QPU_PACK_NONE)
|
|
return false;
|
|
|
|
if (!v3d_qpu_float16_unpack_pack(instr->alu.mul.a_unpack,
|
|
&packed)) {
|
|
return false;
|
|
}
|
|
if (instr->alu.mul.a_unpack == V3D_QPU_UNPACK_SWAP_16)
|
|
opcode = 8;
|
|
else
|
|
opcode |= (packed + 4) & 7;
|
|
|
|
if (instr->alu.mul.b_unpack != V3D_QPU_UNPACK_NONE)
|
|
return false;
|
|
|
|
break;
|
|
}
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
*packed_instr |= QPU_SET_FIELD(mux_a, VC5_QPU_MUL_A);
|
|
*packed_instr |= QPU_SET_FIELD(mux_b, VC5_QPU_MUL_B);
|
|
|
|
*packed_instr |= QPU_SET_FIELD(opcode, VC5_QPU_OP_MUL);
|
|
*packed_instr |= QPU_SET_FIELD(instr->alu.mul.waddr, V3D_QPU_WADDR_M);
|
|
if (instr->alu.mul.magic_write)
|
|
*packed_instr |= VC5_QPU_MM;
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_instr_unpack_alu(const struct v3d_device_info *devinfo,
|
|
uint64_t packed_instr,
|
|
struct v3d_qpu_instr *instr)
|
|
{
|
|
instr->type = V3D_QPU_INSTR_TYPE_ALU;
|
|
|
|
if (!v3d_qpu_sig_unpack(devinfo,
|
|
QPU_GET_FIELD(packed_instr, VC5_QPU_SIG),
|
|
&instr->sig))
|
|
return false;
|
|
|
|
uint32_t packed_cond = QPU_GET_FIELD(packed_instr, VC5_QPU_COND);
|
|
if (v3d_qpu_sig_writes_address(devinfo, &instr->sig)) {
|
|
instr->sig_addr = packed_cond & ~VC5_QPU_COND_SIG_MAGIC_ADDR;
|
|
instr->sig_magic = packed_cond & VC5_QPU_COND_SIG_MAGIC_ADDR;
|
|
|
|
instr->flags.ac = V3D_QPU_COND_NONE;
|
|
instr->flags.mc = V3D_QPU_COND_NONE;
|
|
instr->flags.apf = V3D_QPU_PF_NONE;
|
|
instr->flags.mpf = V3D_QPU_PF_NONE;
|
|
instr->flags.auf = V3D_QPU_UF_NONE;
|
|
instr->flags.muf = V3D_QPU_UF_NONE;
|
|
} else {
|
|
if (!v3d_qpu_flags_unpack(devinfo, packed_cond, &instr->flags))
|
|
return false;
|
|
}
|
|
|
|
instr->raddr_a = QPU_GET_FIELD(packed_instr, VC5_QPU_RADDR_A);
|
|
instr->raddr_b = QPU_GET_FIELD(packed_instr, VC5_QPU_RADDR_B);
|
|
|
|
if (!v3d_qpu_add_unpack(devinfo, packed_instr, instr))
|
|
return false;
|
|
|
|
if (!v3d_qpu_mul_unpack(devinfo, packed_instr, instr))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_instr_unpack_branch(const struct v3d_device_info *devinfo,
|
|
uint64_t packed_instr,
|
|
struct v3d_qpu_instr *instr)
|
|
{
|
|
instr->type = V3D_QPU_INSTR_TYPE_BRANCH;
|
|
|
|
uint32_t cond = QPU_GET_FIELD(packed_instr, VC5_QPU_BRANCH_COND);
|
|
if (cond == 0)
|
|
instr->branch.cond = V3D_QPU_BRANCH_COND_ALWAYS;
|
|
else if (V3D_QPU_BRANCH_COND_A0 + (cond - 2) <=
|
|
V3D_QPU_BRANCH_COND_ALLNA)
|
|
instr->branch.cond = V3D_QPU_BRANCH_COND_A0 + (cond - 2);
|
|
else
|
|
return false;
|
|
|
|
uint32_t msfign = QPU_GET_FIELD(packed_instr, VC5_QPU_BRANCH_MSFIGN);
|
|
if (msfign == 3)
|
|
return false;
|
|
instr->branch.msfign = msfign;
|
|
|
|
instr->branch.bdi = QPU_GET_FIELD(packed_instr, VC5_QPU_BRANCH_BDI);
|
|
|
|
instr->branch.ub = packed_instr & VC5_QPU_BRANCH_UB;
|
|
if (instr->branch.ub) {
|
|
instr->branch.bdu = QPU_GET_FIELD(packed_instr,
|
|
VC5_QPU_BRANCH_BDU);
|
|
}
|
|
|
|
instr->branch.raddr_a = QPU_GET_FIELD(packed_instr,
|
|
VC5_QPU_RADDR_A);
|
|
|
|
instr->branch.offset = 0;
|
|
|
|
instr->branch.offset +=
|
|
QPU_GET_FIELD(packed_instr,
|
|
VC5_QPU_BRANCH_ADDR_LOW) << 3;
|
|
|
|
instr->branch.offset +=
|
|
QPU_GET_FIELD(packed_instr,
|
|
VC5_QPU_BRANCH_ADDR_HIGH) << 24;
|
|
|
|
return true;
|
|
}
|
|
|
|
bool
|
|
v3d_qpu_instr_unpack(const struct v3d_device_info *devinfo,
|
|
uint64_t packed_instr,
|
|
struct v3d_qpu_instr *instr)
|
|
{
|
|
if (QPU_GET_FIELD(packed_instr, VC5_QPU_OP_MUL) != 0) {
|
|
return v3d_qpu_instr_unpack_alu(devinfo, packed_instr, instr);
|
|
} else {
|
|
uint32_t sig = QPU_GET_FIELD(packed_instr, VC5_QPU_SIG);
|
|
|
|
if ((sig & 24) == 16) {
|
|
return v3d_qpu_instr_unpack_branch(devinfo, packed_instr,
|
|
instr);
|
|
} else {
|
|
return false;
|
|
}
|
|
}
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_instr_pack_alu(const struct v3d_device_info *devinfo,
|
|
const struct v3d_qpu_instr *instr,
|
|
uint64_t *packed_instr)
|
|
{
|
|
uint32_t sig;
|
|
if (!v3d_qpu_sig_pack(devinfo, &instr->sig, &sig))
|
|
return false;
|
|
*packed_instr |= QPU_SET_FIELD(sig, VC5_QPU_SIG);
|
|
|
|
if (instr->type == V3D_QPU_INSTR_TYPE_ALU) {
|
|
*packed_instr |= QPU_SET_FIELD(instr->raddr_a, VC5_QPU_RADDR_A);
|
|
*packed_instr |= QPU_SET_FIELD(instr->raddr_b, VC5_QPU_RADDR_B);
|
|
|
|
if (!v3d_qpu_add_pack(devinfo, instr, packed_instr))
|
|
return false;
|
|
if (!v3d_qpu_mul_pack(devinfo, instr, packed_instr))
|
|
return false;
|
|
|
|
uint32_t flags;
|
|
if (v3d_qpu_sig_writes_address(devinfo, &instr->sig)) {
|
|
if (instr->flags.ac != V3D_QPU_COND_NONE ||
|
|
instr->flags.mc != V3D_QPU_COND_NONE ||
|
|
instr->flags.apf != V3D_QPU_PF_NONE ||
|
|
instr->flags.mpf != V3D_QPU_PF_NONE ||
|
|
instr->flags.auf != V3D_QPU_UF_NONE ||
|
|
instr->flags.muf != V3D_QPU_UF_NONE) {
|
|
return false;
|
|
}
|
|
|
|
flags = instr->sig_addr;
|
|
if (instr->sig_magic)
|
|
flags |= VC5_QPU_COND_SIG_MAGIC_ADDR;
|
|
} else {
|
|
if (!v3d_qpu_flags_pack(devinfo, &instr->flags, &flags))
|
|
return false;
|
|
}
|
|
|
|
*packed_instr |= QPU_SET_FIELD(flags, VC5_QPU_COND);
|
|
} else {
|
|
if (v3d_qpu_sig_writes_address(devinfo, &instr->sig))
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool
|
|
v3d_qpu_instr_pack_branch(const struct v3d_device_info *devinfo,
|
|
const struct v3d_qpu_instr *instr,
|
|
uint64_t *packed_instr)
|
|
{
|
|
*packed_instr |= QPU_SET_FIELD(16, VC5_QPU_SIG);
|
|
|
|
if (instr->branch.cond != V3D_QPU_BRANCH_COND_ALWAYS) {
|
|
*packed_instr |= QPU_SET_FIELD(2 + (instr->branch.cond -
|
|
V3D_QPU_BRANCH_COND_A0),
|
|
VC5_QPU_BRANCH_COND);
|
|
}
|
|
|
|
*packed_instr |= QPU_SET_FIELD(instr->branch.msfign,
|
|
VC5_QPU_BRANCH_MSFIGN);
|
|
|
|
*packed_instr |= QPU_SET_FIELD(instr->branch.bdi,
|
|
VC5_QPU_BRANCH_BDI);
|
|
|
|
if (instr->branch.ub) {
|
|
*packed_instr |= VC5_QPU_BRANCH_UB;
|
|
*packed_instr |= QPU_SET_FIELD(instr->branch.bdu,
|
|
VC5_QPU_BRANCH_BDU);
|
|
}
|
|
|
|
switch (instr->branch.bdi) {
|
|
case V3D_QPU_BRANCH_DEST_ABS:
|
|
case V3D_QPU_BRANCH_DEST_REL:
|
|
*packed_instr |= QPU_SET_FIELD(instr->branch.msfign,
|
|
VC5_QPU_BRANCH_MSFIGN);
|
|
|
|
*packed_instr |= QPU_SET_FIELD((instr->branch.offset &
|
|
~0xff000000) >> 3,
|
|
VC5_QPU_BRANCH_ADDR_LOW);
|
|
|
|
*packed_instr |= QPU_SET_FIELD(instr->branch.offset >> 24,
|
|
VC5_QPU_BRANCH_ADDR_HIGH);
|
|
|
|
case V3D_QPU_BRANCH_DEST_REGFILE:
|
|
*packed_instr |= QPU_SET_FIELD(instr->branch.raddr_a,
|
|
VC5_QPU_RADDR_A);
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool
|
|
v3d_qpu_instr_pack(const struct v3d_device_info *devinfo,
|
|
const struct v3d_qpu_instr *instr,
|
|
uint64_t *packed_instr)
|
|
{
|
|
*packed_instr = 0;
|
|
|
|
switch (instr->type) {
|
|
case V3D_QPU_INSTR_TYPE_ALU:
|
|
return v3d_qpu_instr_pack_alu(devinfo, instr, packed_instr);
|
|
case V3D_QPU_INSTR_TYPE_BRANCH:
|
|
return v3d_qpu_instr_pack_branch(devinfo, instr, packed_instr);
|
|
default:
|
|
return false;
|
|
}
|
|
}
|