mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2026-05-16 10:11:38 -04:00
net: mana: hardening: Validate doorbell ID from GDMA_REGISTER_DEVICE response
As a part of MANA hardening for CVM, add validation for the doorbell
ID (db_id) received from hardware in the GDMA_REGISTER_DEVICE response
to prevent out-of-bounds memory access when calculating the doorbell
page address.
In mana_gd_ring_doorbell(), the doorbell page address is calculated as:
addr = db_page_base + db_page_size * db_index
= (bar0_va + db_page_off) + db_page_size * db_index
A hardware could return values that cause this address to fall outside
the BAR0 MMIO region. In Confidential VM environments, hardware responses
cannot be fully trusted.
Add the following validations:
- Store the BAR0 size (bar0_size) in gdma_context during probe.
- Validate the doorbell page offset (db_page_off) read from device
registers does not exceed bar0_size during initialization, converting
mana_gd_init_registers() to return an error code.
- Validate db_id from GDMA_REGISTER_DEVICE response against the
maximum number of doorbell pages that fit within BAR0.
Signed-off-by: Erni Sri Satya Vennela <ernis@linux.microsoft.com>
Link: https://patch.msgid.link/20260306211212.543376-1-ernis@linux.microsoft.com
Signed-off-by: Paolo Abeni <pabeni@redhat.com>
This commit is contained in:
committed by
Paolo Abeni
parent
46097d011f
commit
89fe91c659
@@ -39,49 +39,66 @@ static u64 mana_gd_r64(struct gdma_context *g, u64 offset)
|
||||
return readq(g->bar0_va + offset);
|
||||
}
|
||||
|
||||
static void mana_gd_init_pf_regs(struct pci_dev *pdev)
|
||||
static int mana_gd_init_pf_regs(struct pci_dev *pdev)
|
||||
{
|
||||
struct gdma_context *gc = pci_get_drvdata(pdev);
|
||||
void __iomem *sriov_base_va;
|
||||
u64 sriov_base_off;
|
||||
|
||||
gc->db_page_size = mana_gd_r32(gc, GDMA_PF_REG_DB_PAGE_SIZE) & 0xFFFF;
|
||||
gc->db_page_base = gc->bar0_va +
|
||||
mana_gd_r64(gc, GDMA_PF_REG_DB_PAGE_OFF);
|
||||
gc->db_page_off = mana_gd_r64(gc, GDMA_PF_REG_DB_PAGE_OFF);
|
||||
|
||||
gc->phys_db_page_base = gc->bar0_pa +
|
||||
mana_gd_r64(gc, GDMA_PF_REG_DB_PAGE_OFF);
|
||||
/* Validate doorbell offset is within BAR0 */
|
||||
if (gc->db_page_off >= gc->bar0_size) {
|
||||
dev_err(gc->dev,
|
||||
"Doorbell offset 0x%llx exceeds BAR0 size 0x%llx\n",
|
||||
gc->db_page_off, (u64)gc->bar0_size);
|
||||
return -EPROTO;
|
||||
}
|
||||
|
||||
gc->db_page_base = gc->bar0_va + gc->db_page_off;
|
||||
gc->phys_db_page_base = gc->bar0_pa + gc->db_page_off;
|
||||
|
||||
sriov_base_off = mana_gd_r64(gc, GDMA_SRIOV_REG_CFG_BASE_OFF);
|
||||
|
||||
sriov_base_va = gc->bar0_va + sriov_base_off;
|
||||
gc->shm_base = sriov_base_va +
|
||||
mana_gd_r64(gc, sriov_base_off + GDMA_PF_REG_SHM_OFF);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void mana_gd_init_vf_regs(struct pci_dev *pdev)
|
||||
static int mana_gd_init_vf_regs(struct pci_dev *pdev)
|
||||
{
|
||||
struct gdma_context *gc = pci_get_drvdata(pdev);
|
||||
|
||||
gc->db_page_size = mana_gd_r32(gc, GDMA_REG_DB_PAGE_SIZE) & 0xFFFF;
|
||||
gc->db_page_off = mana_gd_r64(gc, GDMA_REG_DB_PAGE_OFFSET);
|
||||
|
||||
gc->db_page_base = gc->bar0_va +
|
||||
mana_gd_r64(gc, GDMA_REG_DB_PAGE_OFFSET);
|
||||
/* Validate doorbell offset is within BAR0 */
|
||||
if (gc->db_page_off >= gc->bar0_size) {
|
||||
dev_err(gc->dev,
|
||||
"Doorbell offset 0x%llx exceeds BAR0 size 0x%llx\n",
|
||||
gc->db_page_off, (u64)gc->bar0_size);
|
||||
return -EPROTO;
|
||||
}
|
||||
|
||||
gc->phys_db_page_base = gc->bar0_pa +
|
||||
mana_gd_r64(gc, GDMA_REG_DB_PAGE_OFFSET);
|
||||
gc->db_page_base = gc->bar0_va + gc->db_page_off;
|
||||
gc->phys_db_page_base = gc->bar0_pa + gc->db_page_off;
|
||||
|
||||
gc->shm_base = gc->bar0_va + mana_gd_r64(gc, GDMA_REG_SHM_OFFSET);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void mana_gd_init_registers(struct pci_dev *pdev)
|
||||
static int mana_gd_init_registers(struct pci_dev *pdev)
|
||||
{
|
||||
struct gdma_context *gc = pci_get_drvdata(pdev);
|
||||
|
||||
if (gc->is_pf)
|
||||
mana_gd_init_pf_regs(pdev);
|
||||
return mana_gd_init_pf_regs(pdev);
|
||||
else
|
||||
mana_gd_init_vf_regs(pdev);
|
||||
return mana_gd_init_vf_regs(pdev);
|
||||
}
|
||||
|
||||
/* Suppress logging when we set timeout to zero */
|
||||
@@ -1256,6 +1273,17 @@ int mana_gd_register_device(struct gdma_dev *gd)
|
||||
return err ? err : -EPROTO;
|
||||
}
|
||||
|
||||
/* Validate that doorbell page for db_id is within the BAR0 region.
|
||||
* In mana_gd_ring_doorbell(), the address is calculated as:
|
||||
* addr = db_page_base + db_page_size * db_id
|
||||
* = (bar0_va + db_page_off) + (db_page_size * db_id)
|
||||
* So we need: db_page_off + db_page_size * (db_id + 1) <= bar0_size
|
||||
*/
|
||||
if (gc->db_page_off + gc->db_page_size * ((u64)resp.db_id + 1) > gc->bar0_size) {
|
||||
dev_err(gc->dev, "Doorbell ID %u out of range\n", resp.db_id);
|
||||
return -EPROTO;
|
||||
}
|
||||
|
||||
gd->pdid = resp.pdid;
|
||||
gd->gpa_mkey = resp.gpa_mkey;
|
||||
gd->doorbell = resp.db_id;
|
||||
@@ -1890,7 +1918,10 @@ static int mana_gd_setup(struct pci_dev *pdev)
|
||||
struct gdma_context *gc = pci_get_drvdata(pdev);
|
||||
int err;
|
||||
|
||||
mana_gd_init_registers(pdev);
|
||||
err = mana_gd_init_registers(pdev);
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
mana_smc_init(&gc->shm_channel, gc->dev, gc->shm_base);
|
||||
|
||||
gc->service_wq = alloc_ordered_workqueue("gdma_service_wq", 0);
|
||||
@@ -1996,6 +2027,7 @@ static int mana_gd_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
|
||||
mutex_init(&gc->eq_test_event_mutex);
|
||||
pci_set_drvdata(pdev, gc);
|
||||
gc->bar0_pa = pci_resource_start(pdev, 0);
|
||||
gc->bar0_size = pci_resource_len(pdev, 0);
|
||||
|
||||
bar0_va = pci_iomap(pdev, bar, 0);
|
||||
if (!bar0_va)
|
||||
|
||||
@@ -421,10 +421,12 @@ struct gdma_context {
|
||||
|
||||
phys_addr_t bar0_pa;
|
||||
void __iomem *bar0_va;
|
||||
resource_size_t bar0_size;
|
||||
void __iomem *shm_base;
|
||||
void __iomem *db_page_base;
|
||||
phys_addr_t phys_db_page_base;
|
||||
u32 db_page_size;
|
||||
u64 db_page_off;
|
||||
u64 db_page_size;
|
||||
int numa_node;
|
||||
|
||||
/* Shared memory chanenl (used to bootstrap HWC) */
|
||||
|
||||
Reference in New Issue
Block a user