mirror of
https://gitlab.com/qemu-project/qemu
synced 2024-11-02 21:32:52 +00:00
ad80e36744
We pass a ResetType argument to the Resettable class enter phase method, but we don't pass it to hold and exit, even though the callsites have it readily available. This means that if a device cared about the ResetType it would need to record it in the enter phase method to use later on. Pass the type to all three of the phase methods to avoid having to do that. Commit created with for dir in hw target include; do \ spatch --macro-file scripts/cocci-macro-file.h \ --sp-file scripts/coccinelle/reset-type.cocci \ --keep-comments --smpl-spacing --in-place \ --include-headers --dir $dir; done and no manual edits. Signed-off-by: Peter Maydell <peter.maydell@linaro.org> Reviewed-by: Edgar E. Iglesias <edgar.iglesias@amd.com> Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Reviewed-by: Luc Michel <luc.michel@amd.com> Message-id: 20240412160809.1260625-5-peter.maydell@linaro.org
803 lines
25 KiB
C
803 lines
25 KiB
C
/*
|
|
* QEMU model of the Versal eFuse controller
|
|
*
|
|
* Copyright (c) 2020 Xilinx Inc.
|
|
* Copyright (c) 2023 Advanced Micro Devices, Inc.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
* of this software and associated documentation files (the "Software"), to deal
|
|
* in the Software without restriction, including without limitation the rights
|
|
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
* copies of the Software, and to permit persons to whom the Software is
|
|
* furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
|
* THE SOFTWARE.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "hw/nvram/xlnx-versal-efuse.h"
|
|
|
|
#include "qemu/log.h"
|
|
#include "qapi/error.h"
|
|
#include "migration/vmstate.h"
|
|
#include "hw/qdev-properties.h"
|
|
|
|
#ifndef XLNX_VERSAL_EFUSE_CTRL_ERR_DEBUG
|
|
#define XLNX_VERSAL_EFUSE_CTRL_ERR_DEBUG 0
|
|
#endif
|
|
|
|
REG32(WR_LOCK, 0x0)
|
|
FIELD(WR_LOCK, LOCK, 0, 16)
|
|
REG32(CFG, 0x4)
|
|
FIELD(CFG, SLVERR_ENABLE, 5, 1)
|
|
FIELD(CFG, MARGIN_RD, 2, 1)
|
|
FIELD(CFG, PGM_EN, 1, 1)
|
|
REG32(STATUS, 0x8)
|
|
FIELD(STATUS, AES_USER_KEY_1_CRC_PASS, 11, 1)
|
|
FIELD(STATUS, AES_USER_KEY_1_CRC_DONE, 10, 1)
|
|
FIELD(STATUS, AES_USER_KEY_0_CRC_PASS, 9, 1)
|
|
FIELD(STATUS, AES_USER_KEY_0_CRC_DONE, 8, 1)
|
|
FIELD(STATUS, AES_CRC_PASS, 7, 1)
|
|
FIELD(STATUS, AES_CRC_DONE, 6, 1)
|
|
FIELD(STATUS, CACHE_DONE, 5, 1)
|
|
FIELD(STATUS, CACHE_LOAD, 4, 1)
|
|
FIELD(STATUS, EFUSE_2_TBIT, 2, 1)
|
|
FIELD(STATUS, EFUSE_1_TBIT, 1, 1)
|
|
FIELD(STATUS, EFUSE_0_TBIT, 0, 1)
|
|
REG32(EFUSE_PGM_ADDR, 0xc)
|
|
FIELD(EFUSE_PGM_ADDR, PAGE, 13, 4)
|
|
FIELD(EFUSE_PGM_ADDR, ROW, 5, 8)
|
|
FIELD(EFUSE_PGM_ADDR, COLUMN, 0, 5)
|
|
REG32(EFUSE_RD_ADDR, 0x10)
|
|
FIELD(EFUSE_RD_ADDR, PAGE, 13, 4)
|
|
FIELD(EFUSE_RD_ADDR, ROW, 5, 8)
|
|
REG32(EFUSE_RD_DATA, 0x14)
|
|
REG32(TPGM, 0x18)
|
|
FIELD(TPGM, VALUE, 0, 16)
|
|
REG32(TRD, 0x1c)
|
|
FIELD(TRD, VALUE, 0, 8)
|
|
REG32(TSU_H_PS, 0x20)
|
|
FIELD(TSU_H_PS, VALUE, 0, 8)
|
|
REG32(TSU_H_PS_CS, 0x24)
|
|
FIELD(TSU_H_PS_CS, VALUE, 0, 8)
|
|
REG32(TRDM, 0x28)
|
|
FIELD(TRDM, VALUE, 0, 8)
|
|
REG32(TSU_H_CS, 0x2c)
|
|
FIELD(TSU_H_CS, VALUE, 0, 8)
|
|
REG32(EFUSE_ISR, 0x30)
|
|
FIELD(EFUSE_ISR, APB_SLVERR, 31, 1)
|
|
FIELD(EFUSE_ISR, CACHE_PARITY_E2, 14, 1)
|
|
FIELD(EFUSE_ISR, CACHE_PARITY_E1, 13, 1)
|
|
FIELD(EFUSE_ISR, CACHE_PARITY_E0S, 12, 1)
|
|
FIELD(EFUSE_ISR, CACHE_PARITY_E0R, 11, 1)
|
|
FIELD(EFUSE_ISR, CACHE_APB_SLVERR, 10, 1)
|
|
FIELD(EFUSE_ISR, CACHE_REQ_ERROR, 9, 1)
|
|
FIELD(EFUSE_ISR, MAIN_REQ_ERROR, 8, 1)
|
|
FIELD(EFUSE_ISR, READ_ON_CACHE_LD, 7, 1)
|
|
FIELD(EFUSE_ISR, CACHE_FSM_ERROR, 6, 1)
|
|
FIELD(EFUSE_ISR, MAIN_FSM_ERROR, 5, 1)
|
|
FIELD(EFUSE_ISR, CACHE_ERROR, 4, 1)
|
|
FIELD(EFUSE_ISR, RD_ERROR, 3, 1)
|
|
FIELD(EFUSE_ISR, RD_DONE, 2, 1)
|
|
FIELD(EFUSE_ISR, PGM_ERROR, 1, 1)
|
|
FIELD(EFUSE_ISR, PGM_DONE, 0, 1)
|
|
REG32(EFUSE_IMR, 0x34)
|
|
FIELD(EFUSE_IMR, APB_SLVERR, 31, 1)
|
|
FIELD(EFUSE_IMR, CACHE_PARITY_E2, 14, 1)
|
|
FIELD(EFUSE_IMR, CACHE_PARITY_E1, 13, 1)
|
|
FIELD(EFUSE_IMR, CACHE_PARITY_E0S, 12, 1)
|
|
FIELD(EFUSE_IMR, CACHE_PARITY_E0R, 11, 1)
|
|
FIELD(EFUSE_IMR, CACHE_APB_SLVERR, 10, 1)
|
|
FIELD(EFUSE_IMR, CACHE_REQ_ERROR, 9, 1)
|
|
FIELD(EFUSE_IMR, MAIN_REQ_ERROR, 8, 1)
|
|
FIELD(EFUSE_IMR, READ_ON_CACHE_LD, 7, 1)
|
|
FIELD(EFUSE_IMR, CACHE_FSM_ERROR, 6, 1)
|
|
FIELD(EFUSE_IMR, MAIN_FSM_ERROR, 5, 1)
|
|
FIELD(EFUSE_IMR, CACHE_ERROR, 4, 1)
|
|
FIELD(EFUSE_IMR, RD_ERROR, 3, 1)
|
|
FIELD(EFUSE_IMR, RD_DONE, 2, 1)
|
|
FIELD(EFUSE_IMR, PGM_ERROR, 1, 1)
|
|
FIELD(EFUSE_IMR, PGM_DONE, 0, 1)
|
|
REG32(EFUSE_IER, 0x38)
|
|
FIELD(EFUSE_IER, APB_SLVERR, 31, 1)
|
|
FIELD(EFUSE_IER, CACHE_PARITY_E2, 14, 1)
|
|
FIELD(EFUSE_IER, CACHE_PARITY_E1, 13, 1)
|
|
FIELD(EFUSE_IER, CACHE_PARITY_E0S, 12, 1)
|
|
FIELD(EFUSE_IER, CACHE_PARITY_E0R, 11, 1)
|
|
FIELD(EFUSE_IER, CACHE_APB_SLVERR, 10, 1)
|
|
FIELD(EFUSE_IER, CACHE_REQ_ERROR, 9, 1)
|
|
FIELD(EFUSE_IER, MAIN_REQ_ERROR, 8, 1)
|
|
FIELD(EFUSE_IER, READ_ON_CACHE_LD, 7, 1)
|
|
FIELD(EFUSE_IER, CACHE_FSM_ERROR, 6, 1)
|
|
FIELD(EFUSE_IER, MAIN_FSM_ERROR, 5, 1)
|
|
FIELD(EFUSE_IER, CACHE_ERROR, 4, 1)
|
|
FIELD(EFUSE_IER, RD_ERROR, 3, 1)
|
|
FIELD(EFUSE_IER, RD_DONE, 2, 1)
|
|
FIELD(EFUSE_IER, PGM_ERROR, 1, 1)
|
|
FIELD(EFUSE_IER, PGM_DONE, 0, 1)
|
|
REG32(EFUSE_IDR, 0x3c)
|
|
FIELD(EFUSE_IDR, APB_SLVERR, 31, 1)
|
|
FIELD(EFUSE_IDR, CACHE_PARITY_E2, 14, 1)
|
|
FIELD(EFUSE_IDR, CACHE_PARITY_E1, 13, 1)
|
|
FIELD(EFUSE_IDR, CACHE_PARITY_E0S, 12, 1)
|
|
FIELD(EFUSE_IDR, CACHE_PARITY_E0R, 11, 1)
|
|
FIELD(EFUSE_IDR, CACHE_APB_SLVERR, 10, 1)
|
|
FIELD(EFUSE_IDR, CACHE_REQ_ERROR, 9, 1)
|
|
FIELD(EFUSE_IDR, MAIN_REQ_ERROR, 8, 1)
|
|
FIELD(EFUSE_IDR, READ_ON_CACHE_LD, 7, 1)
|
|
FIELD(EFUSE_IDR, CACHE_FSM_ERROR, 6, 1)
|
|
FIELD(EFUSE_IDR, MAIN_FSM_ERROR, 5, 1)
|
|
FIELD(EFUSE_IDR, CACHE_ERROR, 4, 1)
|
|
FIELD(EFUSE_IDR, RD_ERROR, 3, 1)
|
|
FIELD(EFUSE_IDR, RD_DONE, 2, 1)
|
|
FIELD(EFUSE_IDR, PGM_ERROR, 1, 1)
|
|
FIELD(EFUSE_IDR, PGM_DONE, 0, 1)
|
|
REG32(EFUSE_CACHE_LOAD, 0x40)
|
|
FIELD(EFUSE_CACHE_LOAD, LOAD, 0, 1)
|
|
REG32(EFUSE_PGM_LOCK, 0x44)
|
|
FIELD(EFUSE_PGM_LOCK, SPK_ID_LOCK, 0, 1)
|
|
REG32(EFUSE_AES_CRC, 0x48)
|
|
REG32(EFUSE_AES_USR_KEY0_CRC, 0x4c)
|
|
REG32(EFUSE_AES_USR_KEY1_CRC, 0x50)
|
|
REG32(EFUSE_PD, 0x54)
|
|
REG32(EFUSE_ANLG_OSC_SW_1LP, 0x60)
|
|
REG32(EFUSE_TEST_CTRL, 0x100)
|
|
|
|
#define R_MAX (R_EFUSE_TEST_CTRL + 1)
|
|
|
|
#define R_WR_LOCK_UNLOCK_PASSCODE (0xDF0D)
|
|
|
|
/*
|
|
* eFuse layout references:
|
|
* https://github.com/Xilinx/embeddedsw/blob/release-2019.2/lib/sw_services/xilnvm/src/xnvm_efuse_hw.h
|
|
*/
|
|
#define BIT_POS_OF(A_) \
|
|
((uint32_t)((A_) & (R_EFUSE_PGM_ADDR_ROW_MASK | \
|
|
R_EFUSE_PGM_ADDR_COLUMN_MASK)))
|
|
|
|
#define BIT_POS(R_, C_) \
|
|
((uint32_t)((R_EFUSE_PGM_ADDR_ROW_MASK \
|
|
& ((R_) << R_EFUSE_PGM_ADDR_ROW_SHIFT)) \
|
|
| \
|
|
(R_EFUSE_PGM_ADDR_COLUMN_MASK \
|
|
& ((C_) << R_EFUSE_PGM_ADDR_COLUMN_SHIFT))))
|
|
|
|
#define EFUSE_TBIT_POS(A_) (BIT_POS_OF(A_) >= BIT_POS(0, 28))
|
|
|
|
#define EFUSE_ANCHOR_ROW (0)
|
|
#define EFUSE_ANCHOR_3_COL (27)
|
|
#define EFUSE_ANCHOR_1_COL (1)
|
|
|
|
#define EFUSE_AES_KEY_START BIT_POS(12, 0)
|
|
#define EFUSE_AES_KEY_END BIT_POS(19, 31)
|
|
#define EFUSE_USER_KEY_0_START BIT_POS(20, 0)
|
|
#define EFUSE_USER_KEY_0_END BIT_POS(27, 31)
|
|
#define EFUSE_USER_KEY_1_START BIT_POS(28, 0)
|
|
#define EFUSE_USER_KEY_1_END BIT_POS(35, 31)
|
|
|
|
#define EFUSE_RD_BLOCKED_START EFUSE_AES_KEY_START
|
|
#define EFUSE_RD_BLOCKED_END EFUSE_USER_KEY_1_END
|
|
|
|
#define EFUSE_GLITCH_DET_WR_LK BIT_POS(4, 31)
|
|
#define EFUSE_PPK0_WR_LK BIT_POS(43, 6)
|
|
#define EFUSE_PPK1_WR_LK BIT_POS(43, 7)
|
|
#define EFUSE_PPK2_WR_LK BIT_POS(43, 8)
|
|
#define EFUSE_AES_WR_LK BIT_POS(43, 11)
|
|
#define EFUSE_USER_KEY_0_WR_LK BIT_POS(43, 13)
|
|
#define EFUSE_USER_KEY_1_WR_LK BIT_POS(43, 15)
|
|
#define EFUSE_PUF_SYN_LK BIT_POS(43, 16)
|
|
#define EFUSE_DNA_WR_LK BIT_POS(43, 27)
|
|
#define EFUSE_BOOT_ENV_WR_LK BIT_POS(43, 28)
|
|
|
|
#define EFUSE_PGM_LOCKED_START BIT_POS(44, 0)
|
|
#define EFUSE_PGM_LOCKED_END BIT_POS(51, 31)
|
|
|
|
#define EFUSE_PUF_PAGE (2)
|
|
#define EFUSE_PUF_SYN_START BIT_POS(129, 0)
|
|
#define EFUSE_PUF_SYN_END BIT_POS(255, 27)
|
|
|
|
#define EFUSE_KEY_CRC_LK_ROW (43)
|
|
#define EFUSE_AES_KEY_CRC_LK_MASK ((1U << 9) | (1U << 10))
|
|
#define EFUSE_USER_KEY_0_CRC_LK_MASK (1U << 12)
|
|
#define EFUSE_USER_KEY_1_CRC_LK_MASK (1U << 14)
|
|
|
|
/*
|
|
* A handy macro to return value of an array element,
|
|
* or a specific default if given index is out of bound.
|
|
*/
|
|
#define ARRAY_GET(A_, I_, D_) \
|
|
((unsigned int)(I_) < ARRAY_SIZE(A_) ? (A_)[I_] : (D_))
|
|
|
|
QEMU_BUILD_BUG_ON(R_MAX != ARRAY_SIZE(((XlnxVersalEFuseCtrl *)0)->regs));
|
|
|
|
typedef struct XlnxEFuseLkSpec {
|
|
uint16_t row;
|
|
uint16_t lk_bit;
|
|
} XlnxEFuseLkSpec;
|
|
|
|
static void efuse_imr_update_irq(XlnxVersalEFuseCtrl *s)
|
|
{
|
|
bool pending = s->regs[R_EFUSE_ISR] & ~s->regs[R_EFUSE_IMR];
|
|
qemu_set_irq(s->irq_efuse_imr, pending);
|
|
}
|
|
|
|
static void efuse_isr_postw(RegisterInfo *reg, uint64_t val64)
|
|
{
|
|
XlnxVersalEFuseCtrl *s = XLNX_VERSAL_EFUSE_CTRL(reg->opaque);
|
|
efuse_imr_update_irq(s);
|
|
}
|
|
|
|
static uint64_t efuse_ier_prew(RegisterInfo *reg, uint64_t val64)
|
|
{
|
|
XlnxVersalEFuseCtrl *s = XLNX_VERSAL_EFUSE_CTRL(reg->opaque);
|
|
uint32_t val = val64;
|
|
|
|
s->regs[R_EFUSE_IMR] &= ~val;
|
|
efuse_imr_update_irq(s);
|
|
return 0;
|
|
}
|
|
|
|
static uint64_t efuse_idr_prew(RegisterInfo *reg, uint64_t val64)
|
|
{
|
|
XlnxVersalEFuseCtrl *s = XLNX_VERSAL_EFUSE_CTRL(reg->opaque);
|
|
uint32_t val = val64;
|
|
|
|
s->regs[R_EFUSE_IMR] |= val;
|
|
efuse_imr_update_irq(s);
|
|
return 0;
|
|
}
|
|
|
|
static void efuse_status_tbits_sync(XlnxVersalEFuseCtrl *s)
|
|
{
|
|
uint32_t check = xlnx_efuse_tbits_check(s->efuse);
|
|
uint32_t val = s->regs[R_STATUS];
|
|
|
|
val = FIELD_DP32(val, STATUS, EFUSE_0_TBIT, !!(check & (1 << 0)));
|
|
val = FIELD_DP32(val, STATUS, EFUSE_1_TBIT, !!(check & (1 << 1)));
|
|
val = FIELD_DP32(val, STATUS, EFUSE_2_TBIT, !!(check & (1 << 2)));
|
|
|
|
s->regs[R_STATUS] = val;
|
|
}
|
|
|
|
static void efuse_anchor_bits_check(XlnxVersalEFuseCtrl *s)
|
|
{
|
|
unsigned page;
|
|
|
|
if (!s->efuse || !s->efuse->init_tbits) {
|
|
return;
|
|
}
|
|
|
|
for (page = 0; page < s->efuse->efuse_nr; page++) {
|
|
uint32_t row = 0, bit;
|
|
|
|
row = FIELD_DP32(row, EFUSE_PGM_ADDR, PAGE, page);
|
|
row = FIELD_DP32(row, EFUSE_PGM_ADDR, ROW, EFUSE_ANCHOR_ROW);
|
|
|
|
bit = FIELD_DP32(row, EFUSE_PGM_ADDR, COLUMN, EFUSE_ANCHOR_3_COL);
|
|
if (!xlnx_efuse_get_bit(s->efuse, bit)) {
|
|
xlnx_efuse_set_bit(s->efuse, bit);
|
|
}
|
|
|
|
bit = FIELD_DP32(row, EFUSE_PGM_ADDR, COLUMN, EFUSE_ANCHOR_1_COL);
|
|
if (!xlnx_efuse_get_bit(s->efuse, bit)) {
|
|
xlnx_efuse_set_bit(s->efuse, bit);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void efuse_key_crc_check(RegisterInfo *reg, uint32_t crc,
|
|
uint32_t pass_mask, uint32_t done_mask,
|
|
unsigned first, uint32_t lk_mask)
|
|
{
|
|
XlnxVersalEFuseCtrl *s = XLNX_VERSAL_EFUSE_CTRL(reg->opaque);
|
|
uint32_t r, lk_bits;
|
|
|
|
/*
|
|
* To start, assume both DONE and PASS, and clear PASS by xor
|
|
* if CRC-check fails or CRC-check disabled by lock fuse.
|
|
*/
|
|
r = s->regs[R_STATUS] | done_mask | pass_mask;
|
|
|
|
lk_bits = xlnx_efuse_get_row(s->efuse, EFUSE_KEY_CRC_LK_ROW) & lk_mask;
|
|
if (lk_bits == 0 && xlnx_efuse_k256_check(s->efuse, crc, first)) {
|
|
pass_mask = 0;
|
|
}
|
|
|
|
s->regs[R_STATUS] = r ^ pass_mask;
|
|
}
|
|
|
|
static void efuse_data_sync(XlnxVersalEFuseCtrl *s)
|
|
{
|
|
efuse_status_tbits_sync(s);
|
|
}
|
|
|
|
static int efuse_lk_spec_cmp(const void *a, const void *b)
|
|
{
|
|
uint16_t r1 = ((const XlnxEFuseLkSpec *)a)->row;
|
|
uint16_t r2 = ((const XlnxEFuseLkSpec *)b)->row;
|
|
|
|
return (r1 > r2) - (r1 < r2);
|
|
}
|
|
|
|
static void efuse_lk_spec_sort(XlnxVersalEFuseCtrl *s)
|
|
{
|
|
XlnxEFuseLkSpec *ary = s->extra_pg0_lock_spec;
|
|
const uint32_t n8 = s->extra_pg0_lock_n16 * 2;
|
|
const uint32_t sz = sizeof(ary[0]);
|
|
const uint32_t cnt = n8 / sz;
|
|
|
|
if (ary && cnt) {
|
|
qsort(ary, cnt, sz, efuse_lk_spec_cmp);
|
|
}
|
|
}
|
|
|
|
static uint32_t efuse_lk_spec_find(XlnxVersalEFuseCtrl *s, uint32_t row)
|
|
{
|
|
const XlnxEFuseLkSpec *ary = s->extra_pg0_lock_spec;
|
|
const uint32_t n8 = s->extra_pg0_lock_n16 * 2;
|
|
const uint32_t sz = sizeof(ary[0]);
|
|
const uint32_t cnt = n8 / sz;
|
|
const XlnxEFuseLkSpec *item = NULL;
|
|
|
|
if (ary && cnt) {
|
|
XlnxEFuseLkSpec k = { .row = row, };
|
|
|
|
item = bsearch(&k, ary, cnt, sz, efuse_lk_spec_cmp);
|
|
}
|
|
|
|
return item ? item->lk_bit : 0;
|
|
}
|
|
|
|
static uint32_t efuse_bit_locked(XlnxVersalEFuseCtrl *s, uint32_t bit)
|
|
{
|
|
/* Hard-coded locks */
|
|
static const uint16_t pg0_hard_lock[] = {
|
|
[4] = EFUSE_GLITCH_DET_WR_LK,
|
|
[37] = EFUSE_BOOT_ENV_WR_LK,
|
|
|
|
[8 ... 11] = EFUSE_DNA_WR_LK,
|
|
[12 ... 19] = EFUSE_AES_WR_LK,
|
|
[20 ... 27] = EFUSE_USER_KEY_0_WR_LK,
|
|
[28 ... 35] = EFUSE_USER_KEY_1_WR_LK,
|
|
[64 ... 71] = EFUSE_PPK0_WR_LK,
|
|
[72 ... 79] = EFUSE_PPK1_WR_LK,
|
|
[80 ... 87] = EFUSE_PPK2_WR_LK,
|
|
};
|
|
|
|
uint32_t row = FIELD_EX32(bit, EFUSE_PGM_ADDR, ROW);
|
|
uint32_t lk_bit = ARRAY_GET(pg0_hard_lock, row, 0);
|
|
|
|
return lk_bit ? lk_bit : efuse_lk_spec_find(s, row);
|
|
}
|
|
|
|
static bool efuse_pgm_locked(XlnxVersalEFuseCtrl *s, unsigned int bit)
|
|
{
|
|
|
|
unsigned int lock = 1;
|
|
|
|
/* Global lock */
|
|
if (!ARRAY_FIELD_EX32(s->regs, CFG, PGM_EN)) {
|
|
goto ret_lock;
|
|
}
|
|
|
|
/* Row lock */
|
|
switch (FIELD_EX32(bit, EFUSE_PGM_ADDR, PAGE)) {
|
|
case 0:
|
|
if (ARRAY_FIELD_EX32(s->regs, EFUSE_PGM_LOCK, SPK_ID_LOCK) &&
|
|
bit >= EFUSE_PGM_LOCKED_START && bit <= EFUSE_PGM_LOCKED_END) {
|
|
goto ret_lock;
|
|
}
|
|
|
|
lock = efuse_bit_locked(s, bit);
|
|
break;
|
|
case EFUSE_PUF_PAGE:
|
|
if (bit < EFUSE_PUF_SYN_START || bit > EFUSE_PUF_SYN_END) {
|
|
lock = 0;
|
|
goto ret_lock;
|
|
}
|
|
|
|
lock = EFUSE_PUF_SYN_LK;
|
|
break;
|
|
default:
|
|
lock = 0;
|
|
goto ret_lock;
|
|
}
|
|
|
|
/* Row lock by an efuse bit */
|
|
if (lock) {
|
|
lock = xlnx_efuse_get_bit(s->efuse, lock);
|
|
}
|
|
|
|
ret_lock:
|
|
return lock != 0;
|
|
}
|
|
|
|
static void efuse_pgm_addr_postw(RegisterInfo *reg, uint64_t val64)
|
|
{
|
|
XlnxVersalEFuseCtrl *s = XLNX_VERSAL_EFUSE_CTRL(reg->opaque);
|
|
unsigned bit = val64;
|
|
bool ok = false;
|
|
|
|
/* Always zero out PGM_ADDR because it is write-only */
|
|
s->regs[R_EFUSE_PGM_ADDR] = 0;
|
|
|
|
/*
|
|
* Indicate error if bit is write-protected (or read-only
|
|
* as guarded by efuse_set_bit()).
|
|
*
|
|
* Keep it simple by not modeling program timing.
|
|
*
|
|
* Note: model must NEVER clear the PGM_ERROR bit; it is
|
|
* up to guest to do so (or by reset).
|
|
*/
|
|
if (efuse_pgm_locked(s, bit)) {
|
|
g_autofree char *path = object_get_canonical_path(OBJECT(s));
|
|
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"%s: Denied setting of efuse<%u, %u, %u>\n",
|
|
path,
|
|
FIELD_EX32(bit, EFUSE_PGM_ADDR, PAGE),
|
|
FIELD_EX32(bit, EFUSE_PGM_ADDR, ROW),
|
|
FIELD_EX32(bit, EFUSE_PGM_ADDR, COLUMN));
|
|
} else if (xlnx_efuse_set_bit(s->efuse, bit)) {
|
|
ok = true;
|
|
if (EFUSE_TBIT_POS(bit)) {
|
|
efuse_status_tbits_sync(s);
|
|
}
|
|
}
|
|
|
|
if (!ok) {
|
|
ARRAY_FIELD_DP32(s->regs, EFUSE_ISR, PGM_ERROR, 1);
|
|
}
|
|
|
|
ARRAY_FIELD_DP32(s->regs, EFUSE_ISR, PGM_DONE, 1);
|
|
efuse_imr_update_irq(s);
|
|
}
|
|
|
|
static void efuse_rd_addr_postw(RegisterInfo *reg, uint64_t val64)
|
|
{
|
|
XlnxVersalEFuseCtrl *s = XLNX_VERSAL_EFUSE_CTRL(reg->opaque);
|
|
unsigned bit = val64;
|
|
bool denied;
|
|
|
|
/* Always zero out RD_ADDR because it is write-only */
|
|
s->regs[R_EFUSE_RD_ADDR] = 0;
|
|
|
|
/*
|
|
* Indicate error if row is read-blocked.
|
|
*
|
|
* Note: model must NEVER clear the RD_ERROR bit; it is
|
|
* up to guest to do so (or by reset).
|
|
*/
|
|
s->regs[R_EFUSE_RD_DATA] = xlnx_versal_efuse_read_row(s->efuse,
|
|
bit, &denied);
|
|
if (denied) {
|
|
g_autofree char *path = object_get_canonical_path(OBJECT(s));
|
|
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"%s: Denied reading of efuse<%u, %u>\n",
|
|
path,
|
|
FIELD_EX32(bit, EFUSE_RD_ADDR, PAGE),
|
|
FIELD_EX32(bit, EFUSE_RD_ADDR, ROW));
|
|
|
|
ARRAY_FIELD_DP32(s->regs, EFUSE_ISR, RD_ERROR, 1);
|
|
}
|
|
|
|
ARRAY_FIELD_DP32(s->regs, EFUSE_ISR, RD_DONE, 1);
|
|
efuse_imr_update_irq(s);
|
|
return;
|
|
}
|
|
|
|
static uint64_t efuse_cache_load_prew(RegisterInfo *reg, uint64_t val64)
|
|
{
|
|
XlnxVersalEFuseCtrl *s = XLNX_VERSAL_EFUSE_CTRL(reg->opaque);
|
|
|
|
if (val64 & R_EFUSE_CACHE_LOAD_LOAD_MASK) {
|
|
efuse_data_sync(s);
|
|
|
|
ARRAY_FIELD_DP32(s->regs, STATUS, CACHE_DONE, 1);
|
|
efuse_imr_update_irq(s);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static uint64_t efuse_pgm_lock_prew(RegisterInfo *reg, uint64_t val64)
|
|
{
|
|
XlnxVersalEFuseCtrl *s = XLNX_VERSAL_EFUSE_CTRL(reg->opaque);
|
|
|
|
/* Ignore all other bits */
|
|
val64 = FIELD_EX32(val64, EFUSE_PGM_LOCK, SPK_ID_LOCK);
|
|
|
|
/* Once the bit is written 1, only reset will clear it to 0 */
|
|
val64 |= ARRAY_FIELD_EX32(s->regs, EFUSE_PGM_LOCK, SPK_ID_LOCK);
|
|
|
|
return val64;
|
|
}
|
|
|
|
static void efuse_aes_crc_postw(RegisterInfo *reg, uint64_t val64)
|
|
{
|
|
efuse_key_crc_check(reg, val64,
|
|
R_STATUS_AES_CRC_PASS_MASK,
|
|
R_STATUS_AES_CRC_DONE_MASK,
|
|
EFUSE_AES_KEY_START,
|
|
EFUSE_AES_KEY_CRC_LK_MASK);
|
|
}
|
|
|
|
static void efuse_aes_u0_crc_postw(RegisterInfo *reg, uint64_t val64)
|
|
{
|
|
efuse_key_crc_check(reg, val64,
|
|
R_STATUS_AES_USER_KEY_0_CRC_PASS_MASK,
|
|
R_STATUS_AES_USER_KEY_0_CRC_DONE_MASK,
|
|
EFUSE_USER_KEY_0_START,
|
|
EFUSE_USER_KEY_0_CRC_LK_MASK);
|
|
}
|
|
|
|
static void efuse_aes_u1_crc_postw(RegisterInfo *reg, uint64_t val64)
|
|
{
|
|
efuse_key_crc_check(reg, val64,
|
|
R_STATUS_AES_USER_KEY_1_CRC_PASS_MASK,
|
|
R_STATUS_AES_USER_KEY_1_CRC_DONE_MASK,
|
|
EFUSE_USER_KEY_1_START,
|
|
EFUSE_USER_KEY_1_CRC_LK_MASK);
|
|
}
|
|
|
|
static uint64_t efuse_wr_lock_prew(RegisterInfo *reg, uint64_t val)
|
|
{
|
|
return val != R_WR_LOCK_UNLOCK_PASSCODE;
|
|
}
|
|
|
|
static const RegisterAccessInfo efuse_ctrl_regs_info[] = {
|
|
{ .name = "WR_LOCK", .addr = A_WR_LOCK,
|
|
.reset = 0x1,
|
|
.pre_write = efuse_wr_lock_prew,
|
|
},{ .name = "CFG", .addr = A_CFG,
|
|
.rsvd = 0x9,
|
|
},{ .name = "STATUS", .addr = A_STATUS,
|
|
.rsvd = 0x8,
|
|
.ro = 0xfff,
|
|
},{ .name = "EFUSE_PGM_ADDR", .addr = A_EFUSE_PGM_ADDR,
|
|
.post_write = efuse_pgm_addr_postw,
|
|
},{ .name = "EFUSE_RD_ADDR", .addr = A_EFUSE_RD_ADDR,
|
|
.rsvd = 0x1f,
|
|
.post_write = efuse_rd_addr_postw,
|
|
},{ .name = "EFUSE_RD_DATA", .addr = A_EFUSE_RD_DATA,
|
|
.ro = 0xffffffff,
|
|
},{ .name = "TPGM", .addr = A_TPGM,
|
|
},{ .name = "TRD", .addr = A_TRD,
|
|
.reset = 0x19,
|
|
},{ .name = "TSU_H_PS", .addr = A_TSU_H_PS,
|
|
.reset = 0xff,
|
|
},{ .name = "TSU_H_PS_CS", .addr = A_TSU_H_PS_CS,
|
|
.reset = 0x11,
|
|
},{ .name = "TRDM", .addr = A_TRDM,
|
|
.reset = 0x3a,
|
|
},{ .name = "TSU_H_CS", .addr = A_TSU_H_CS,
|
|
.reset = 0x16,
|
|
},{ .name = "EFUSE_ISR", .addr = A_EFUSE_ISR,
|
|
.rsvd = 0x7fff8000,
|
|
.w1c = 0x80007fff,
|
|
.post_write = efuse_isr_postw,
|
|
},{ .name = "EFUSE_IMR", .addr = A_EFUSE_IMR,
|
|
.reset = 0x80007fff,
|
|
.rsvd = 0x7fff8000,
|
|
.ro = 0xffffffff,
|
|
},{ .name = "EFUSE_IER", .addr = A_EFUSE_IER,
|
|
.rsvd = 0x7fff8000,
|
|
.pre_write = efuse_ier_prew,
|
|
},{ .name = "EFUSE_IDR", .addr = A_EFUSE_IDR,
|
|
.rsvd = 0x7fff8000,
|
|
.pre_write = efuse_idr_prew,
|
|
},{ .name = "EFUSE_CACHE_LOAD", .addr = A_EFUSE_CACHE_LOAD,
|
|
.pre_write = efuse_cache_load_prew,
|
|
},{ .name = "EFUSE_PGM_LOCK", .addr = A_EFUSE_PGM_LOCK,
|
|
.pre_write = efuse_pgm_lock_prew,
|
|
},{ .name = "EFUSE_AES_CRC", .addr = A_EFUSE_AES_CRC,
|
|
.post_write = efuse_aes_crc_postw,
|
|
},{ .name = "EFUSE_AES_USR_KEY0_CRC", .addr = A_EFUSE_AES_USR_KEY0_CRC,
|
|
.post_write = efuse_aes_u0_crc_postw,
|
|
},{ .name = "EFUSE_AES_USR_KEY1_CRC", .addr = A_EFUSE_AES_USR_KEY1_CRC,
|
|
.post_write = efuse_aes_u1_crc_postw,
|
|
},{ .name = "EFUSE_PD", .addr = A_EFUSE_PD,
|
|
.ro = 0xfffffffe,
|
|
},{ .name = "EFUSE_ANLG_OSC_SW_1LP", .addr = A_EFUSE_ANLG_OSC_SW_1LP,
|
|
},{ .name = "EFUSE_TEST_CTRL", .addr = A_EFUSE_TEST_CTRL,
|
|
.reset = 0x8,
|
|
}
|
|
};
|
|
|
|
static void efuse_ctrl_reg_write(void *opaque, hwaddr addr,
|
|
uint64_t data, unsigned size)
|
|
{
|
|
RegisterInfoArray *reg_array = opaque;
|
|
XlnxVersalEFuseCtrl *s;
|
|
Object *dev;
|
|
|
|
assert(reg_array != NULL);
|
|
|
|
dev = reg_array->mem.owner;
|
|
assert(dev);
|
|
|
|
s = XLNX_VERSAL_EFUSE_CTRL(dev);
|
|
|
|
if (addr != A_WR_LOCK && s->regs[R_WR_LOCK]) {
|
|
g_autofree char *path = object_get_canonical_path(OBJECT(s));
|
|
|
|
qemu_log_mask(LOG_GUEST_ERROR,
|
|
"%s[reg_0x%02lx]: Attempt to write locked register.\n",
|
|
path, (long)addr);
|
|
} else {
|
|
register_write_memory(opaque, addr, data, size);
|
|
}
|
|
}
|
|
|
|
static void efuse_ctrl_register_reset(RegisterInfo *reg)
|
|
{
|
|
if (!reg->data || !reg->access) {
|
|
return;
|
|
}
|
|
|
|
/* Reset must not trigger some registers' writers */
|
|
switch (reg->access->addr) {
|
|
case A_EFUSE_AES_CRC:
|
|
case A_EFUSE_AES_USR_KEY0_CRC:
|
|
case A_EFUSE_AES_USR_KEY1_CRC:
|
|
*(uint32_t *)reg->data = reg->access->reset;
|
|
return;
|
|
}
|
|
|
|
register_reset(reg);
|
|
}
|
|
|
|
static void efuse_ctrl_reset_hold(Object *obj, ResetType type)
|
|
{
|
|
XlnxVersalEFuseCtrl *s = XLNX_VERSAL_EFUSE_CTRL(obj);
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(s->regs_info); ++i) {
|
|
efuse_ctrl_register_reset(&s->regs_info[i]);
|
|
}
|
|
|
|
efuse_anchor_bits_check(s);
|
|
efuse_data_sync(s);
|
|
efuse_imr_update_irq(s);
|
|
}
|
|
|
|
static const MemoryRegionOps efuse_ctrl_ops = {
|
|
.read = register_read_memory,
|
|
.write = efuse_ctrl_reg_write,
|
|
.endianness = DEVICE_LITTLE_ENDIAN,
|
|
.valid = {
|
|
.min_access_size = 4,
|
|
.max_access_size = 4,
|
|
},
|
|
};
|
|
|
|
static void efuse_ctrl_realize(DeviceState *dev, Error **errp)
|
|
{
|
|
XlnxVersalEFuseCtrl *s = XLNX_VERSAL_EFUSE_CTRL(dev);
|
|
const uint32_t lks_sz = sizeof(XlnxEFuseLkSpec) / 2;
|
|
|
|
if (!s->efuse) {
|
|
g_autofree char *path = object_get_canonical_path(OBJECT(s));
|
|
|
|
error_setg(errp, "%s.efuse: link property not connected to XLNX-EFUSE",
|
|
path);
|
|
return;
|
|
}
|
|
|
|
/* Sort property-defined pgm-locks for bsearch lookup */
|
|
if ((s->extra_pg0_lock_n16 % lks_sz) != 0) {
|
|
g_autofree char *path = object_get_canonical_path(OBJECT(s));
|
|
|
|
error_setg(errp,
|
|
"%s.pg0-lock: array property item-count not multiple of %u",
|
|
path, lks_sz);
|
|
return;
|
|
}
|
|
|
|
efuse_lk_spec_sort(s);
|
|
}
|
|
|
|
static void efuse_ctrl_init(Object *obj)
|
|
{
|
|
XlnxVersalEFuseCtrl *s = XLNX_VERSAL_EFUSE_CTRL(obj);
|
|
SysBusDevice *sbd = SYS_BUS_DEVICE(obj);
|
|
RegisterInfoArray *reg_array;
|
|
|
|
reg_array =
|
|
register_init_block32(DEVICE(obj), efuse_ctrl_regs_info,
|
|
ARRAY_SIZE(efuse_ctrl_regs_info),
|
|
s->regs_info, s->regs,
|
|
&efuse_ctrl_ops,
|
|
XLNX_VERSAL_EFUSE_CTRL_ERR_DEBUG,
|
|
R_MAX * 4);
|
|
|
|
sysbus_init_mmio(sbd, ®_array->mem);
|
|
sysbus_init_irq(sbd, &s->irq_efuse_imr);
|
|
}
|
|
|
|
static void efuse_ctrl_finalize(Object *obj)
|
|
{
|
|
XlnxVersalEFuseCtrl *s = XLNX_VERSAL_EFUSE_CTRL(obj);
|
|
|
|
g_free(s->extra_pg0_lock_spec);
|
|
}
|
|
|
|
static const VMStateDescription vmstate_efuse_ctrl = {
|
|
.name = TYPE_XLNX_VERSAL_EFUSE_CTRL,
|
|
.version_id = 1,
|
|
.minimum_version_id = 1,
|
|
.fields = (const VMStateField[]) {
|
|
VMSTATE_UINT32_ARRAY(regs, XlnxVersalEFuseCtrl, R_MAX),
|
|
VMSTATE_END_OF_LIST(),
|
|
}
|
|
};
|
|
|
|
static Property efuse_ctrl_props[] = {
|
|
DEFINE_PROP_LINK("efuse",
|
|
XlnxVersalEFuseCtrl, efuse,
|
|
TYPE_XLNX_EFUSE, XlnxEFuse *),
|
|
DEFINE_PROP_ARRAY("pg0-lock",
|
|
XlnxVersalEFuseCtrl, extra_pg0_lock_n16,
|
|
extra_pg0_lock_spec, qdev_prop_uint16, uint16_t),
|
|
|
|
DEFINE_PROP_END_OF_LIST(),
|
|
};
|
|
|
|
static void efuse_ctrl_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
DeviceClass *dc = DEVICE_CLASS(klass);
|
|
ResettableClass *rc = RESETTABLE_CLASS(klass);
|
|
|
|
rc->phases.hold = efuse_ctrl_reset_hold;
|
|
dc->realize = efuse_ctrl_realize;
|
|
dc->vmsd = &vmstate_efuse_ctrl;
|
|
device_class_set_props(dc, efuse_ctrl_props);
|
|
}
|
|
|
|
static const TypeInfo efuse_ctrl_info = {
|
|
.name = TYPE_XLNX_VERSAL_EFUSE_CTRL,
|
|
.parent = TYPE_SYS_BUS_DEVICE,
|
|
.instance_size = sizeof(XlnxVersalEFuseCtrl),
|
|
.class_init = efuse_ctrl_class_init,
|
|
.instance_init = efuse_ctrl_init,
|
|
.instance_finalize = efuse_ctrl_finalize,
|
|
};
|
|
|
|
static void efuse_ctrl_register_types(void)
|
|
{
|
|
type_register_static(&efuse_ctrl_info);
|
|
}
|
|
|
|
type_init(efuse_ctrl_register_types)
|
|
|
|
/*
|
|
* Retrieve a row, with unreadable bits returned as 0.
|
|
*/
|
|
uint32_t xlnx_versal_efuse_read_row(XlnxEFuse *efuse,
|
|
uint32_t bit, bool *denied)
|
|
{
|
|
bool dummy;
|
|
|
|
if (!denied) {
|
|
denied = &dummy;
|
|
}
|
|
|
|
if (bit >= EFUSE_RD_BLOCKED_START && bit <= EFUSE_RD_BLOCKED_END) {
|
|
*denied = true;
|
|
return 0;
|
|
}
|
|
|
|
*denied = false;
|
|
return xlnx_efuse_get_row(efuse, bit);
|
|
}
|