mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-11-01 00:48:50 +00:00
172dbac35e
The dumper is only a debugging aid so we don't want to invoke the OOM killer if buffer for the potentially large GPU state can't be vmalloced. Signed-off-by: Lucas Stach <l.stach@pengutronix.de>
232 lines
6.1 KiB
C
232 lines
6.1 KiB
C
/*
|
|
* Copyright (C) 2015 Etnaviv Project
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License version 2 as published by
|
|
* the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along with
|
|
* this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include <linux/devcoredump.h>
|
|
#include "etnaviv_dump.h"
|
|
#include "etnaviv_gem.h"
|
|
#include "etnaviv_gpu.h"
|
|
#include "etnaviv_mmu.h"
|
|
#include "state.xml.h"
|
|
#include "state_hi.xml.h"
|
|
|
|
struct core_dump_iterator {
|
|
void *start;
|
|
struct etnaviv_dump_object_header *hdr;
|
|
void *data;
|
|
};
|
|
|
|
static const unsigned short etnaviv_dump_registers[] = {
|
|
VIVS_HI_AXI_STATUS,
|
|
VIVS_HI_CLOCK_CONTROL,
|
|
VIVS_HI_IDLE_STATE,
|
|
VIVS_HI_AXI_CONFIG,
|
|
VIVS_HI_INTR_ENBL,
|
|
VIVS_HI_CHIP_IDENTITY,
|
|
VIVS_HI_CHIP_FEATURE,
|
|
VIVS_HI_CHIP_MODEL,
|
|
VIVS_HI_CHIP_REV,
|
|
VIVS_HI_CHIP_DATE,
|
|
VIVS_HI_CHIP_TIME,
|
|
VIVS_HI_CHIP_MINOR_FEATURE_0,
|
|
VIVS_HI_CACHE_CONTROL,
|
|
VIVS_HI_AXI_CONTROL,
|
|
VIVS_PM_POWER_CONTROLS,
|
|
VIVS_PM_MODULE_CONTROLS,
|
|
VIVS_PM_MODULE_STATUS,
|
|
VIVS_PM_PULSE_EATER,
|
|
VIVS_MC_MMU_FE_PAGE_TABLE,
|
|
VIVS_MC_MMU_TX_PAGE_TABLE,
|
|
VIVS_MC_MMU_PE_PAGE_TABLE,
|
|
VIVS_MC_MMU_PEZ_PAGE_TABLE,
|
|
VIVS_MC_MMU_RA_PAGE_TABLE,
|
|
VIVS_MC_DEBUG_MEMORY,
|
|
VIVS_MC_MEMORY_BASE_ADDR_RA,
|
|
VIVS_MC_MEMORY_BASE_ADDR_FE,
|
|
VIVS_MC_MEMORY_BASE_ADDR_TX,
|
|
VIVS_MC_MEMORY_BASE_ADDR_PEZ,
|
|
VIVS_MC_MEMORY_BASE_ADDR_PE,
|
|
VIVS_MC_MEMORY_TIMING_CONTROL,
|
|
VIVS_MC_BUS_CONFIG,
|
|
VIVS_FE_DMA_STATUS,
|
|
VIVS_FE_DMA_DEBUG_STATE,
|
|
VIVS_FE_DMA_ADDRESS,
|
|
VIVS_FE_DMA_LOW,
|
|
VIVS_FE_DMA_HIGH,
|
|
VIVS_FE_AUTO_FLUSH,
|
|
};
|
|
|
|
static void etnaviv_core_dump_header(struct core_dump_iterator *iter,
|
|
u32 type, void *data_end)
|
|
{
|
|
struct etnaviv_dump_object_header *hdr = iter->hdr;
|
|
|
|
hdr->magic = cpu_to_le32(ETDUMP_MAGIC);
|
|
hdr->type = cpu_to_le32(type);
|
|
hdr->file_offset = cpu_to_le32(iter->data - iter->start);
|
|
hdr->file_size = cpu_to_le32(data_end - iter->data);
|
|
|
|
iter->hdr++;
|
|
iter->data += hdr->file_size;
|
|
}
|
|
|
|
static void etnaviv_core_dump_registers(struct core_dump_iterator *iter,
|
|
struct etnaviv_gpu *gpu)
|
|
{
|
|
struct etnaviv_dump_registers *reg = iter->data;
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(etnaviv_dump_registers); i++, reg++) {
|
|
reg->reg = etnaviv_dump_registers[i];
|
|
reg->value = gpu_read(gpu, etnaviv_dump_registers[i]);
|
|
}
|
|
|
|
etnaviv_core_dump_header(iter, ETDUMP_BUF_REG, reg);
|
|
}
|
|
|
|
static void etnaviv_core_dump_mmu(struct core_dump_iterator *iter,
|
|
struct etnaviv_gpu *gpu, size_t mmu_size)
|
|
{
|
|
etnaviv_iommu_dump(gpu->mmu, iter->data);
|
|
|
|
etnaviv_core_dump_header(iter, ETDUMP_BUF_MMU, iter->data + mmu_size);
|
|
}
|
|
|
|
static void etnaviv_core_dump_mem(struct core_dump_iterator *iter, u32 type,
|
|
void *ptr, size_t size, u64 iova)
|
|
{
|
|
memcpy(iter->data, ptr, size);
|
|
|
|
iter->hdr->iova = cpu_to_le64(iova);
|
|
|
|
etnaviv_core_dump_header(iter, type, iter->data + size);
|
|
}
|
|
|
|
void etnaviv_core_dump(struct etnaviv_gpu *gpu)
|
|
{
|
|
struct core_dump_iterator iter;
|
|
struct etnaviv_vram_mapping *vram;
|
|
struct etnaviv_gem_object *obj;
|
|
struct etnaviv_cmdbuf *cmd;
|
|
unsigned int n_obj, n_bomap_pages;
|
|
size_t file_size, mmu_size;
|
|
__le64 *bomap, *bomap_start;
|
|
|
|
mmu_size = etnaviv_iommu_dump_size(gpu->mmu);
|
|
|
|
/* We always dump registers, mmu, ring and end marker */
|
|
n_obj = 4;
|
|
n_bomap_pages = 0;
|
|
file_size = ARRAY_SIZE(etnaviv_dump_registers) *
|
|
sizeof(struct etnaviv_dump_registers) +
|
|
mmu_size + gpu->buffer->size;
|
|
|
|
/* Add in the active command buffers */
|
|
list_for_each_entry(cmd, &gpu->active_cmd_list, node) {
|
|
file_size += cmd->size;
|
|
n_obj++;
|
|
}
|
|
|
|
/* Add in the active buffer objects */
|
|
list_for_each_entry(vram, &gpu->mmu->mappings, mmu_node) {
|
|
if (!vram->use)
|
|
continue;
|
|
|
|
obj = vram->object;
|
|
file_size += obj->base.size;
|
|
n_bomap_pages += obj->base.size >> PAGE_SHIFT;
|
|
n_obj++;
|
|
}
|
|
|
|
/* If we have any buffer objects, add a bomap object */
|
|
if (n_bomap_pages) {
|
|
file_size += n_bomap_pages * sizeof(__le64);
|
|
n_obj++;
|
|
}
|
|
|
|
/* Add the size of the headers */
|
|
file_size += sizeof(*iter.hdr) * n_obj;
|
|
|
|
/* Allocate the file in vmalloc memory, it's likely to be big */
|
|
iter.start = __vmalloc(file_size, GFP_KERNEL | __GFP_HIGHMEM |
|
|
__GFP_NOWARN | __GFP_NORETRY, PAGE_KERNEL);
|
|
if (!iter.start) {
|
|
dev_warn(gpu->dev, "failed to allocate devcoredump file\n");
|
|
return;
|
|
}
|
|
|
|
/* Point the data member after the headers */
|
|
iter.hdr = iter.start;
|
|
iter.data = &iter.hdr[n_obj];
|
|
|
|
memset(iter.hdr, 0, iter.data - iter.start);
|
|
|
|
etnaviv_core_dump_registers(&iter, gpu);
|
|
etnaviv_core_dump_mmu(&iter, gpu, mmu_size);
|
|
etnaviv_core_dump_mem(&iter, ETDUMP_BUF_RING, gpu->buffer->vaddr,
|
|
gpu->buffer->size,
|
|
etnaviv_iommu_get_cmdbuf_va(gpu, gpu->buffer));
|
|
|
|
list_for_each_entry(cmd, &gpu->active_cmd_list, node)
|
|
etnaviv_core_dump_mem(&iter, ETDUMP_BUF_CMD, cmd->vaddr,
|
|
cmd->size,
|
|
etnaviv_iommu_get_cmdbuf_va(gpu, cmd));
|
|
|
|
/* Reserve space for the bomap */
|
|
if (n_bomap_pages) {
|
|
bomap_start = bomap = iter.data;
|
|
memset(bomap, 0, sizeof(*bomap) * n_bomap_pages);
|
|
etnaviv_core_dump_header(&iter, ETDUMP_BUF_BOMAP,
|
|
bomap + n_bomap_pages);
|
|
} else {
|
|
/* Silence warning */
|
|
bomap_start = bomap = NULL;
|
|
}
|
|
|
|
list_for_each_entry(vram, &gpu->mmu->mappings, mmu_node) {
|
|
struct page **pages;
|
|
void *vaddr;
|
|
|
|
if (vram->use == 0)
|
|
continue;
|
|
|
|
obj = vram->object;
|
|
|
|
mutex_lock(&obj->lock);
|
|
pages = etnaviv_gem_get_pages(obj);
|
|
mutex_unlock(&obj->lock);
|
|
if (pages) {
|
|
int j;
|
|
|
|
iter.hdr->data[0] = bomap - bomap_start;
|
|
|
|
for (j = 0; j < obj->base.size >> PAGE_SHIFT; j++)
|
|
*bomap++ = cpu_to_le64(page_to_phys(*pages++));
|
|
}
|
|
|
|
iter.hdr->iova = cpu_to_le64(vram->iova);
|
|
|
|
vaddr = etnaviv_gem_vmap(&obj->base);
|
|
if (vaddr)
|
|
memcpy(iter.data, vaddr, obj->base.size);
|
|
|
|
etnaviv_core_dump_header(&iter, ETDUMP_BUF_BO, iter.data +
|
|
obj->base.size);
|
|
}
|
|
|
|
etnaviv_core_dump_header(&iter, ETDUMP_BUF_END, iter.data);
|
|
|
|
dev_coredumpv(gpu->dev, iter.start, iter.data - iter.start, GFP_KERNEL);
|
|
}
|