mirror of
https://github.com/torvalds/linux
synced 2024-11-05 18:23:50 +00:00
drm/etnaviv: add initial etnaviv DRM driver
This adds the etnaviv DRM driver and hooks it up in Makefiles and Kconfig. Signed-off-by: Christian Gmeiner <christian.gmeiner@gmail.com> Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk> Signed-off-by: Lucas Stach <l.stach@pengutronix.de> Acked-by: Daniel Vetter <daniel.vetter@ffwll.ch>
This commit is contained in:
parent
f04b205ac1
commit
a8c21a5451
27 changed files with 7238 additions and 0 deletions
|
@ -266,3 +266,5 @@ source "drivers/gpu/drm/amd/amdkfd/Kconfig"
|
|||
source "drivers/gpu/drm/imx/Kconfig"
|
||||
|
||||
source "drivers/gpu/drm/vc4/Kconfig"
|
||||
|
||||
source "drivers/gpu/drm/etnaviv/Kconfig"
|
||||
|
|
|
@ -75,3 +75,4 @@ obj-y += i2c/
|
|||
obj-y += panel/
|
||||
obj-y += bridge/
|
||||
obj-$(CONFIG_DRM_FSL_DCU) += fsl-dcu/
|
||||
obj-$(CONFIG_DRM_ETNAVIV) += etnaviv/
|
||||
|
|
20
drivers/gpu/drm/etnaviv/Kconfig
Normal file
20
drivers/gpu/drm/etnaviv/Kconfig
Normal file
|
@ -0,0 +1,20 @@
|
|||
|
||||
config DRM_ETNAVIV
|
||||
tristate "ETNAVIV (DRM support for Vivante GPU IP cores)"
|
||||
depends on DRM
|
||||
depends on ARCH_MXC || ARCH_DOVE
|
||||
select SHMEM
|
||||
select TMPFS
|
||||
select IOMMU_API
|
||||
select IOMMU_SUPPORT
|
||||
select WANT_DEV_COREDUMP
|
||||
help
|
||||
DRM driver for Vivante GPUs.
|
||||
|
||||
config DRM_ETNAVIV_REGISTER_LOGGING
|
||||
bool "enable ETNAVIV register logging"
|
||||
depends on DRM_ETNAVIV
|
||||
help
|
||||
Compile in support for logging register reads/writes in a format
|
||||
that can be parsed by envytools demsm tool. If enabled, register
|
||||
logging can be switched on via etnaviv.reglog=y module param.
|
14
drivers/gpu/drm/etnaviv/Makefile
Normal file
14
drivers/gpu/drm/etnaviv/Makefile
Normal file
|
@ -0,0 +1,14 @@
|
|||
etnaviv-y := \
|
||||
etnaviv_buffer.o \
|
||||
etnaviv_cmd_parser.o \
|
||||
etnaviv_drv.o \
|
||||
etnaviv_dump.o \
|
||||
etnaviv_gem_prime.o \
|
||||
etnaviv_gem_submit.o \
|
||||
etnaviv_gem.o \
|
||||
etnaviv_gpu.o \
|
||||
etnaviv_iommu_v2.o \
|
||||
etnaviv_iommu.o \
|
||||
etnaviv_mmu.o
|
||||
|
||||
obj-$(CONFIG_DRM_ETNAVIV) += etnaviv.o
|
218
drivers/gpu/drm/etnaviv/cmdstream.xml.h
Normal file
218
drivers/gpu/drm/etnaviv/cmdstream.xml.h
Normal file
|
@ -0,0 +1,218 @@
|
|||
#ifndef CMDSTREAM_XML
|
||||
#define CMDSTREAM_XML
|
||||
|
||||
/* Autogenerated file, DO NOT EDIT manually!
|
||||
|
||||
This file was generated by the rules-ng-ng headergen tool in this git repository:
|
||||
http://0x04.net/cgit/index.cgi/rules-ng-ng
|
||||
git clone git://0x04.net/rules-ng-ng
|
||||
|
||||
The rules-ng-ng source files this header was generated from are:
|
||||
- cmdstream.xml ( 12589 bytes, from 2014-02-17 14:57:56)
|
||||
- common.xml ( 18437 bytes, from 2015-03-25 11:27:41)
|
||||
|
||||
Copyright (C) 2014
|
||||
*/
|
||||
|
||||
|
||||
#define FE_OPCODE_LOAD_STATE 0x00000001
|
||||
#define FE_OPCODE_END 0x00000002
|
||||
#define FE_OPCODE_NOP 0x00000003
|
||||
#define FE_OPCODE_DRAW_2D 0x00000004
|
||||
#define FE_OPCODE_DRAW_PRIMITIVES 0x00000005
|
||||
#define FE_OPCODE_DRAW_INDEXED_PRIMITIVES 0x00000006
|
||||
#define FE_OPCODE_WAIT 0x00000007
|
||||
#define FE_OPCODE_LINK 0x00000008
|
||||
#define FE_OPCODE_STALL 0x00000009
|
||||
#define FE_OPCODE_CALL 0x0000000a
|
||||
#define FE_OPCODE_RETURN 0x0000000b
|
||||
#define FE_OPCODE_CHIP_SELECT 0x0000000d
|
||||
#define PRIMITIVE_TYPE_POINTS 0x00000001
|
||||
#define PRIMITIVE_TYPE_LINES 0x00000002
|
||||
#define PRIMITIVE_TYPE_LINE_STRIP 0x00000003
|
||||
#define PRIMITIVE_TYPE_TRIANGLES 0x00000004
|
||||
#define PRIMITIVE_TYPE_TRIANGLE_STRIP 0x00000005
|
||||
#define PRIMITIVE_TYPE_TRIANGLE_FAN 0x00000006
|
||||
#define PRIMITIVE_TYPE_LINE_LOOP 0x00000007
|
||||
#define PRIMITIVE_TYPE_QUADS 0x00000008
|
||||
#define VIV_FE_LOAD_STATE 0x00000000
|
||||
|
||||
#define VIV_FE_LOAD_STATE_HEADER 0x00000000
|
||||
#define VIV_FE_LOAD_STATE_HEADER_OP__MASK 0xf8000000
|
||||
#define VIV_FE_LOAD_STATE_HEADER_OP__SHIFT 27
|
||||
#define VIV_FE_LOAD_STATE_HEADER_OP_LOAD_STATE 0x08000000
|
||||
#define VIV_FE_LOAD_STATE_HEADER_FIXP 0x04000000
|
||||
#define VIV_FE_LOAD_STATE_HEADER_COUNT__MASK 0x03ff0000
|
||||
#define VIV_FE_LOAD_STATE_HEADER_COUNT__SHIFT 16
|
||||
#define VIV_FE_LOAD_STATE_HEADER_COUNT(x) (((x) << VIV_FE_LOAD_STATE_HEADER_COUNT__SHIFT) & VIV_FE_LOAD_STATE_HEADER_COUNT__MASK)
|
||||
#define VIV_FE_LOAD_STATE_HEADER_OFFSET__MASK 0x0000ffff
|
||||
#define VIV_FE_LOAD_STATE_HEADER_OFFSET__SHIFT 0
|
||||
#define VIV_FE_LOAD_STATE_HEADER_OFFSET(x) (((x) << VIV_FE_LOAD_STATE_HEADER_OFFSET__SHIFT) & VIV_FE_LOAD_STATE_HEADER_OFFSET__MASK)
|
||||
#define VIV_FE_LOAD_STATE_HEADER_OFFSET__SHR 2
|
||||
|
||||
#define VIV_FE_END 0x00000000
|
||||
|
||||
#define VIV_FE_END_HEADER 0x00000000
|
||||
#define VIV_FE_END_HEADER_EVENT_ID__MASK 0x0000001f
|
||||
#define VIV_FE_END_HEADER_EVENT_ID__SHIFT 0
|
||||
#define VIV_FE_END_HEADER_EVENT_ID(x) (((x) << VIV_FE_END_HEADER_EVENT_ID__SHIFT) & VIV_FE_END_HEADER_EVENT_ID__MASK)
|
||||
#define VIV_FE_END_HEADER_EVENT_ENABLE 0x00000100
|
||||
#define VIV_FE_END_HEADER_OP__MASK 0xf8000000
|
||||
#define VIV_FE_END_HEADER_OP__SHIFT 27
|
||||
#define VIV_FE_END_HEADER_OP_END 0x10000000
|
||||
|
||||
#define VIV_FE_NOP 0x00000000
|
||||
|
||||
#define VIV_FE_NOP_HEADER 0x00000000
|
||||
#define VIV_FE_NOP_HEADER_OP__MASK 0xf8000000
|
||||
#define VIV_FE_NOP_HEADER_OP__SHIFT 27
|
||||
#define VIV_FE_NOP_HEADER_OP_NOP 0x18000000
|
||||
|
||||
#define VIV_FE_DRAW_2D 0x00000000
|
||||
|
||||
#define VIV_FE_DRAW_2D_HEADER 0x00000000
|
||||
#define VIV_FE_DRAW_2D_HEADER_COUNT__MASK 0x0000ff00
|
||||
#define VIV_FE_DRAW_2D_HEADER_COUNT__SHIFT 8
|
||||
#define VIV_FE_DRAW_2D_HEADER_COUNT(x) (((x) << VIV_FE_DRAW_2D_HEADER_COUNT__SHIFT) & VIV_FE_DRAW_2D_HEADER_COUNT__MASK)
|
||||
#define VIV_FE_DRAW_2D_HEADER_DATA_COUNT__MASK 0x07ff0000
|
||||
#define VIV_FE_DRAW_2D_HEADER_DATA_COUNT__SHIFT 16
|
||||
#define VIV_FE_DRAW_2D_HEADER_DATA_COUNT(x) (((x) << VIV_FE_DRAW_2D_HEADER_DATA_COUNT__SHIFT) & VIV_FE_DRAW_2D_HEADER_DATA_COUNT__MASK)
|
||||
#define VIV_FE_DRAW_2D_HEADER_OP__MASK 0xf8000000
|
||||
#define VIV_FE_DRAW_2D_HEADER_OP__SHIFT 27
|
||||
#define VIV_FE_DRAW_2D_HEADER_OP_DRAW_2D 0x20000000
|
||||
|
||||
#define VIV_FE_DRAW_2D_TOP_LEFT 0x00000008
|
||||
#define VIV_FE_DRAW_2D_TOP_LEFT_X__MASK 0x0000ffff
|
||||
#define VIV_FE_DRAW_2D_TOP_LEFT_X__SHIFT 0
|
||||
#define VIV_FE_DRAW_2D_TOP_LEFT_X(x) (((x) << VIV_FE_DRAW_2D_TOP_LEFT_X__SHIFT) & VIV_FE_DRAW_2D_TOP_LEFT_X__MASK)
|
||||
#define VIV_FE_DRAW_2D_TOP_LEFT_Y__MASK 0xffff0000
|
||||
#define VIV_FE_DRAW_2D_TOP_LEFT_Y__SHIFT 16
|
||||
#define VIV_FE_DRAW_2D_TOP_LEFT_Y(x) (((x) << VIV_FE_DRAW_2D_TOP_LEFT_Y__SHIFT) & VIV_FE_DRAW_2D_TOP_LEFT_Y__MASK)
|
||||
|
||||
#define VIV_FE_DRAW_2D_BOTTOM_RIGHT 0x0000000c
|
||||
#define VIV_FE_DRAW_2D_BOTTOM_RIGHT_X__MASK 0x0000ffff
|
||||
#define VIV_FE_DRAW_2D_BOTTOM_RIGHT_X__SHIFT 0
|
||||
#define VIV_FE_DRAW_2D_BOTTOM_RIGHT_X(x) (((x) << VIV_FE_DRAW_2D_BOTTOM_RIGHT_X__SHIFT) & VIV_FE_DRAW_2D_BOTTOM_RIGHT_X__MASK)
|
||||
#define VIV_FE_DRAW_2D_BOTTOM_RIGHT_Y__MASK 0xffff0000
|
||||
#define VIV_FE_DRAW_2D_BOTTOM_RIGHT_Y__SHIFT 16
|
||||
#define VIV_FE_DRAW_2D_BOTTOM_RIGHT_Y(x) (((x) << VIV_FE_DRAW_2D_BOTTOM_RIGHT_Y__SHIFT) & VIV_FE_DRAW_2D_BOTTOM_RIGHT_Y__MASK)
|
||||
|
||||
#define VIV_FE_DRAW_PRIMITIVES 0x00000000
|
||||
|
||||
#define VIV_FE_DRAW_PRIMITIVES_HEADER 0x00000000
|
||||
#define VIV_FE_DRAW_PRIMITIVES_HEADER_OP__MASK 0xf8000000
|
||||
#define VIV_FE_DRAW_PRIMITIVES_HEADER_OP__SHIFT 27
|
||||
#define VIV_FE_DRAW_PRIMITIVES_HEADER_OP_DRAW_PRIMITIVES 0x28000000
|
||||
|
||||
#define VIV_FE_DRAW_PRIMITIVES_COMMAND 0x00000004
|
||||
#define VIV_FE_DRAW_PRIMITIVES_COMMAND_TYPE__MASK 0x000000ff
|
||||
#define VIV_FE_DRAW_PRIMITIVES_COMMAND_TYPE__SHIFT 0
|
||||
#define VIV_FE_DRAW_PRIMITIVES_COMMAND_TYPE(x) (((x) << VIV_FE_DRAW_PRIMITIVES_COMMAND_TYPE__SHIFT) & VIV_FE_DRAW_PRIMITIVES_COMMAND_TYPE__MASK)
|
||||
|
||||
#define VIV_FE_DRAW_PRIMITIVES_START 0x00000008
|
||||
|
||||
#define VIV_FE_DRAW_PRIMITIVES_COUNT 0x0000000c
|
||||
|
||||
#define VIV_FE_DRAW_INDEXED_PRIMITIVES 0x00000000
|
||||
|
||||
#define VIV_FE_DRAW_INDEXED_PRIMITIVES_HEADER 0x00000000
|
||||
#define VIV_FE_DRAW_INDEXED_PRIMITIVES_HEADER_OP__MASK 0xf8000000
|
||||
#define VIV_FE_DRAW_INDEXED_PRIMITIVES_HEADER_OP__SHIFT 27
|
||||
#define VIV_FE_DRAW_INDEXED_PRIMITIVES_HEADER_OP_DRAW_INDEXED_PRIMITIVES 0x30000000
|
||||
|
||||
#define VIV_FE_DRAW_INDEXED_PRIMITIVES_COMMAND 0x00000004
|
||||
#define VIV_FE_DRAW_INDEXED_PRIMITIVES_COMMAND_TYPE__MASK 0x000000ff
|
||||
#define VIV_FE_DRAW_INDEXED_PRIMITIVES_COMMAND_TYPE__SHIFT 0
|
||||
#define VIV_FE_DRAW_INDEXED_PRIMITIVES_COMMAND_TYPE(x) (((x) << VIV_FE_DRAW_INDEXED_PRIMITIVES_COMMAND_TYPE__SHIFT) & VIV_FE_DRAW_INDEXED_PRIMITIVES_COMMAND_TYPE__MASK)
|
||||
|
||||
#define VIV_FE_DRAW_INDEXED_PRIMITIVES_START 0x00000008
|
||||
|
||||
#define VIV_FE_DRAW_INDEXED_PRIMITIVES_COUNT 0x0000000c
|
||||
|
||||
#define VIV_FE_DRAW_INDEXED_PRIMITIVES_OFFSET 0x00000010
|
||||
|
||||
#define VIV_FE_WAIT 0x00000000
|
||||
|
||||
#define VIV_FE_WAIT_HEADER 0x00000000
|
||||
#define VIV_FE_WAIT_HEADER_DELAY__MASK 0x0000ffff
|
||||
#define VIV_FE_WAIT_HEADER_DELAY__SHIFT 0
|
||||
#define VIV_FE_WAIT_HEADER_DELAY(x) (((x) << VIV_FE_WAIT_HEADER_DELAY__SHIFT) & VIV_FE_WAIT_HEADER_DELAY__MASK)
|
||||
#define VIV_FE_WAIT_HEADER_OP__MASK 0xf8000000
|
||||
#define VIV_FE_WAIT_HEADER_OP__SHIFT 27
|
||||
#define VIV_FE_WAIT_HEADER_OP_WAIT 0x38000000
|
||||
|
||||
#define VIV_FE_LINK 0x00000000
|
||||
|
||||
#define VIV_FE_LINK_HEADER 0x00000000
|
||||
#define VIV_FE_LINK_HEADER_PREFETCH__MASK 0x0000ffff
|
||||
#define VIV_FE_LINK_HEADER_PREFETCH__SHIFT 0
|
||||
#define VIV_FE_LINK_HEADER_PREFETCH(x) (((x) << VIV_FE_LINK_HEADER_PREFETCH__SHIFT) & VIV_FE_LINK_HEADER_PREFETCH__MASK)
|
||||
#define VIV_FE_LINK_HEADER_OP__MASK 0xf8000000
|
||||
#define VIV_FE_LINK_HEADER_OP__SHIFT 27
|
||||
#define VIV_FE_LINK_HEADER_OP_LINK 0x40000000
|
||||
|
||||
#define VIV_FE_LINK_ADDRESS 0x00000004
|
||||
|
||||
#define VIV_FE_STALL 0x00000000
|
||||
|
||||
#define VIV_FE_STALL_HEADER 0x00000000
|
||||
#define VIV_FE_STALL_HEADER_OP__MASK 0xf8000000
|
||||
#define VIV_FE_STALL_HEADER_OP__SHIFT 27
|
||||
#define VIV_FE_STALL_HEADER_OP_STALL 0x48000000
|
||||
|
||||
#define VIV_FE_STALL_TOKEN 0x00000004
|
||||
#define VIV_FE_STALL_TOKEN_FROM__MASK 0x0000001f
|
||||
#define VIV_FE_STALL_TOKEN_FROM__SHIFT 0
|
||||
#define VIV_FE_STALL_TOKEN_FROM(x) (((x) << VIV_FE_STALL_TOKEN_FROM__SHIFT) & VIV_FE_STALL_TOKEN_FROM__MASK)
|
||||
#define VIV_FE_STALL_TOKEN_TO__MASK 0x00001f00
|
||||
#define VIV_FE_STALL_TOKEN_TO__SHIFT 8
|
||||
#define VIV_FE_STALL_TOKEN_TO(x) (((x) << VIV_FE_STALL_TOKEN_TO__SHIFT) & VIV_FE_STALL_TOKEN_TO__MASK)
|
||||
|
||||
#define VIV_FE_CALL 0x00000000
|
||||
|
||||
#define VIV_FE_CALL_HEADER 0x00000000
|
||||
#define VIV_FE_CALL_HEADER_PREFETCH__MASK 0x0000ffff
|
||||
#define VIV_FE_CALL_HEADER_PREFETCH__SHIFT 0
|
||||
#define VIV_FE_CALL_HEADER_PREFETCH(x) (((x) << VIV_FE_CALL_HEADER_PREFETCH__SHIFT) & VIV_FE_CALL_HEADER_PREFETCH__MASK)
|
||||
#define VIV_FE_CALL_HEADER_OP__MASK 0xf8000000
|
||||
#define VIV_FE_CALL_HEADER_OP__SHIFT 27
|
||||
#define VIV_FE_CALL_HEADER_OP_CALL 0x50000000
|
||||
|
||||
#define VIV_FE_CALL_ADDRESS 0x00000004
|
||||
|
||||
#define VIV_FE_CALL_RETURN_PREFETCH 0x00000008
|
||||
|
||||
#define VIV_FE_CALL_RETURN_ADDRESS 0x0000000c
|
||||
|
||||
#define VIV_FE_RETURN 0x00000000
|
||||
|
||||
#define VIV_FE_RETURN_HEADER 0x00000000
|
||||
#define VIV_FE_RETURN_HEADER_OP__MASK 0xf8000000
|
||||
#define VIV_FE_RETURN_HEADER_OP__SHIFT 27
|
||||
#define VIV_FE_RETURN_HEADER_OP_RETURN 0x58000000
|
||||
|
||||
#define VIV_FE_CHIP_SELECT 0x00000000
|
||||
|
||||
#define VIV_FE_CHIP_SELECT_HEADER 0x00000000
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_OP__MASK 0xf8000000
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_OP__SHIFT 27
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_OP_CHIP_SELECT 0x68000000
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP15 0x00008000
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP14 0x00004000
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP13 0x00002000
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP12 0x00001000
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP11 0x00000800
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP10 0x00000400
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP9 0x00000200
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP8 0x00000100
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP7 0x00000080
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP6 0x00000040
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP5 0x00000020
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP4 0x00000010
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP3 0x00000008
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP2 0x00000004
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP1 0x00000002
|
||||
#define VIV_FE_CHIP_SELECT_HEADER_ENABLE_CHIP0 0x00000001
|
||||
|
||||
|
||||
#endif /* CMDSTREAM_XML */
|
249
drivers/gpu/drm/etnaviv/common.xml.h
Normal file
249
drivers/gpu/drm/etnaviv/common.xml.h
Normal file
|
@ -0,0 +1,249 @@
|
|||
#ifndef COMMON_XML
|
||||
#define COMMON_XML
|
||||
|
||||
/* Autogenerated file, DO NOT EDIT manually!
|
||||
|
||||
This file was generated by the rules-ng-ng headergen tool in this git repository:
|
||||
http://0x04.net/cgit/index.cgi/rules-ng-ng
|
||||
git clone git://0x04.net/rules-ng-ng
|
||||
|
||||
The rules-ng-ng source files this header was generated from are:
|
||||
- state_vg.xml ( 5973 bytes, from 2015-03-25 11:26:01)
|
||||
- common.xml ( 18437 bytes, from 2015-03-25 11:27:41)
|
||||
|
||||
Copyright (C) 2015
|
||||
*/
|
||||
|
||||
|
||||
#define PIPE_ID_PIPE_3D 0x00000000
|
||||
#define PIPE_ID_PIPE_2D 0x00000001
|
||||
#define SYNC_RECIPIENT_FE 0x00000001
|
||||
#define SYNC_RECIPIENT_RA 0x00000005
|
||||
#define SYNC_RECIPIENT_PE 0x00000007
|
||||
#define SYNC_RECIPIENT_DE 0x0000000b
|
||||
#define SYNC_RECIPIENT_VG 0x0000000f
|
||||
#define SYNC_RECIPIENT_TESSELATOR 0x00000010
|
||||
#define SYNC_RECIPIENT_VG2 0x00000011
|
||||
#define SYNC_RECIPIENT_TESSELATOR2 0x00000012
|
||||
#define SYNC_RECIPIENT_VG3 0x00000013
|
||||
#define SYNC_RECIPIENT_TESSELATOR3 0x00000014
|
||||
#define ENDIAN_MODE_NO_SWAP 0x00000000
|
||||
#define ENDIAN_MODE_SWAP_16 0x00000001
|
||||
#define ENDIAN_MODE_SWAP_32 0x00000002
|
||||
#define chipModel_GC300 0x00000300
|
||||
#define chipModel_GC320 0x00000320
|
||||
#define chipModel_GC350 0x00000350
|
||||
#define chipModel_GC355 0x00000355
|
||||
#define chipModel_GC400 0x00000400
|
||||
#define chipModel_GC410 0x00000410
|
||||
#define chipModel_GC420 0x00000420
|
||||
#define chipModel_GC450 0x00000450
|
||||
#define chipModel_GC500 0x00000500
|
||||
#define chipModel_GC530 0x00000530
|
||||
#define chipModel_GC600 0x00000600
|
||||
#define chipModel_GC700 0x00000700
|
||||
#define chipModel_GC800 0x00000800
|
||||
#define chipModel_GC860 0x00000860
|
||||
#define chipModel_GC880 0x00000880
|
||||
#define chipModel_GC1000 0x00001000
|
||||
#define chipModel_GC2000 0x00002000
|
||||
#define chipModel_GC2100 0x00002100
|
||||
#define chipModel_GC4000 0x00004000
|
||||
#define RGBA_BITS_R 0x00000001
|
||||
#define RGBA_BITS_G 0x00000002
|
||||
#define RGBA_BITS_B 0x00000004
|
||||
#define RGBA_BITS_A 0x00000008
|
||||
#define chipFeatures_FAST_CLEAR 0x00000001
|
||||
#define chipFeatures_SPECIAL_ANTI_ALIASING 0x00000002
|
||||
#define chipFeatures_PIPE_3D 0x00000004
|
||||
#define chipFeatures_DXT_TEXTURE_COMPRESSION 0x00000008
|
||||
#define chipFeatures_DEBUG_MODE 0x00000010
|
||||
#define chipFeatures_Z_COMPRESSION 0x00000020
|
||||
#define chipFeatures_YUV420_SCALER 0x00000040
|
||||
#define chipFeatures_MSAA 0x00000080
|
||||
#define chipFeatures_DC 0x00000100
|
||||
#define chipFeatures_PIPE_2D 0x00000200
|
||||
#define chipFeatures_ETC1_TEXTURE_COMPRESSION 0x00000400
|
||||
#define chipFeatures_FAST_SCALER 0x00000800
|
||||
#define chipFeatures_HIGH_DYNAMIC_RANGE 0x00001000
|
||||
#define chipFeatures_YUV420_TILER 0x00002000
|
||||
#define chipFeatures_MODULE_CG 0x00004000
|
||||
#define chipFeatures_MIN_AREA 0x00008000
|
||||
#define chipFeatures_NO_EARLY_Z 0x00010000
|
||||
#define chipFeatures_NO_422_TEXTURE 0x00020000
|
||||
#define chipFeatures_BUFFER_INTERLEAVING 0x00040000
|
||||
#define chipFeatures_BYTE_WRITE_2D 0x00080000
|
||||
#define chipFeatures_NO_SCALER 0x00100000
|
||||
#define chipFeatures_YUY2_AVERAGING 0x00200000
|
||||
#define chipFeatures_HALF_PE_CACHE 0x00400000
|
||||
#define chipFeatures_HALF_TX_CACHE 0x00800000
|
||||
#define chipFeatures_YUY2_RENDER_TARGET 0x01000000
|
||||
#define chipFeatures_MEM32 0x02000000
|
||||
#define chipFeatures_PIPE_VG 0x04000000
|
||||
#define chipFeatures_VGTS 0x08000000
|
||||
#define chipFeatures_FE20 0x10000000
|
||||
#define chipFeatures_BYTE_WRITE_3D 0x20000000
|
||||
#define chipFeatures_RS_YUV_TARGET 0x40000000
|
||||
#define chipFeatures_32_BIT_INDICES 0x80000000
|
||||
#define chipMinorFeatures0_FLIP_Y 0x00000001
|
||||
#define chipMinorFeatures0_DUAL_RETURN_BUS 0x00000002
|
||||
#define chipMinorFeatures0_ENDIANNESS_CONFIG 0x00000004
|
||||
#define chipMinorFeatures0_TEXTURE_8K 0x00000008
|
||||
#define chipMinorFeatures0_CORRECT_TEXTURE_CONVERTER 0x00000010
|
||||
#define chipMinorFeatures0_SPECIAL_MSAA_LOD 0x00000020
|
||||
#define chipMinorFeatures0_FAST_CLEAR_FLUSH 0x00000040
|
||||
#define chipMinorFeatures0_2DPE20 0x00000080
|
||||
#define chipMinorFeatures0_CORRECT_AUTO_DISABLE 0x00000100
|
||||
#define chipMinorFeatures0_RENDERTARGET_8K 0x00000200
|
||||
#define chipMinorFeatures0_2BITPERTILE 0x00000400
|
||||
#define chipMinorFeatures0_SEPARATE_TILE_STATUS_WHEN_INTERLEAVED 0x00000800
|
||||
#define chipMinorFeatures0_SUPER_TILED 0x00001000
|
||||
#define chipMinorFeatures0_VG_20 0x00002000
|
||||
#define chipMinorFeatures0_TS_EXTENDED_COMMANDS 0x00004000
|
||||
#define chipMinorFeatures0_COMPRESSION_FIFO_FIXED 0x00008000
|
||||
#define chipMinorFeatures0_HAS_SIGN_FLOOR_CEIL 0x00010000
|
||||
#define chipMinorFeatures0_VG_FILTER 0x00020000
|
||||
#define chipMinorFeatures0_VG_21 0x00040000
|
||||
#define chipMinorFeatures0_SHADER_HAS_W 0x00080000
|
||||
#define chipMinorFeatures0_HAS_SQRT_TRIG 0x00100000
|
||||
#define chipMinorFeatures0_MORE_MINOR_FEATURES 0x00200000
|
||||
#define chipMinorFeatures0_MC20 0x00400000
|
||||
#define chipMinorFeatures0_MSAA_SIDEBAND 0x00800000
|
||||
#define chipMinorFeatures0_BUG_FIXES0 0x01000000
|
||||
#define chipMinorFeatures0_VAA 0x02000000
|
||||
#define chipMinorFeatures0_BYPASS_IN_MSAA 0x04000000
|
||||
#define chipMinorFeatures0_HZ 0x08000000
|
||||
#define chipMinorFeatures0_NEW_TEXTURE 0x10000000
|
||||
#define chipMinorFeatures0_2D_A8_TARGET 0x20000000
|
||||
#define chipMinorFeatures0_CORRECT_STENCIL 0x40000000
|
||||
#define chipMinorFeatures0_ENHANCE_VR 0x80000000
|
||||
#define chipMinorFeatures1_RSUV_SWIZZLE 0x00000001
|
||||
#define chipMinorFeatures1_V2_COMPRESSION 0x00000002
|
||||
#define chipMinorFeatures1_VG_DOUBLE_BUFFER 0x00000004
|
||||
#define chipMinorFeatures1_EXTRA_EVENT_STATES 0x00000008
|
||||
#define chipMinorFeatures1_NO_STRIPING_NEEDED 0x00000010
|
||||
#define chipMinorFeatures1_TEXTURE_STRIDE 0x00000020
|
||||
#define chipMinorFeatures1_BUG_FIXES3 0x00000040
|
||||
#define chipMinorFeatures1_AUTO_DISABLE 0x00000080
|
||||
#define chipMinorFeatures1_AUTO_RESTART_TS 0x00000100
|
||||
#define chipMinorFeatures1_DISABLE_PE_GATING 0x00000200
|
||||
#define chipMinorFeatures1_L2_WINDOWING 0x00000400
|
||||
#define chipMinorFeatures1_HALF_FLOAT 0x00000800
|
||||
#define chipMinorFeatures1_PIXEL_DITHER 0x00001000
|
||||
#define chipMinorFeatures1_TWO_STENCIL_REFERENCE 0x00002000
|
||||
#define chipMinorFeatures1_EXTENDED_PIXEL_FORMAT 0x00004000
|
||||
#define chipMinorFeatures1_CORRECT_MIN_MAX_DEPTH 0x00008000
|
||||
#define chipMinorFeatures1_2D_DITHER 0x00010000
|
||||
#define chipMinorFeatures1_BUG_FIXES5 0x00020000
|
||||
#define chipMinorFeatures1_NEW_2D 0x00040000
|
||||
#define chipMinorFeatures1_NEW_FP 0x00080000
|
||||
#define chipMinorFeatures1_TEXTURE_HALIGN 0x00100000
|
||||
#define chipMinorFeatures1_NON_POWER_OF_TWO 0x00200000
|
||||
#define chipMinorFeatures1_LINEAR_TEXTURE_SUPPORT 0x00400000
|
||||
#define chipMinorFeatures1_HALTI0 0x00800000
|
||||
#define chipMinorFeatures1_CORRECT_OVERFLOW_VG 0x01000000
|
||||
#define chipMinorFeatures1_NEGATIVE_LOG_FIX 0x02000000
|
||||
#define chipMinorFeatures1_RESOLVE_OFFSET 0x04000000
|
||||
#define chipMinorFeatures1_OK_TO_GATE_AXI_CLOCK 0x08000000
|
||||
#define chipMinorFeatures1_MMU_VERSION 0x10000000
|
||||
#define chipMinorFeatures1_WIDE_LINE 0x20000000
|
||||
#define chipMinorFeatures1_BUG_FIXES6 0x40000000
|
||||
#define chipMinorFeatures1_FC_FLUSH_STALL 0x80000000
|
||||
#define chipMinorFeatures2_LINE_LOOP 0x00000001
|
||||
#define chipMinorFeatures2_LOGIC_OP 0x00000002
|
||||
#define chipMinorFeatures2_UNK2 0x00000004
|
||||
#define chipMinorFeatures2_SUPERTILED_TEXTURE 0x00000008
|
||||
#define chipMinorFeatures2_UNK4 0x00000010
|
||||
#define chipMinorFeatures2_RECT_PRIMITIVE 0x00000020
|
||||
#define chipMinorFeatures2_COMPOSITION 0x00000040
|
||||
#define chipMinorFeatures2_CORRECT_AUTO_DISABLE_COUNT 0x00000080
|
||||
#define chipMinorFeatures2_UNK8 0x00000100
|
||||
#define chipMinorFeatures2_UNK9 0x00000200
|
||||
#define chipMinorFeatures2_UNK10 0x00000400
|
||||
#define chipMinorFeatures2_SAMPLERBASE_16 0x00000800
|
||||
#define chipMinorFeatures2_UNK12 0x00001000
|
||||
#define chipMinorFeatures2_UNK13 0x00002000
|
||||
#define chipMinorFeatures2_UNK14 0x00004000
|
||||
#define chipMinorFeatures2_EXTRA_TEXTURE_STATE 0x00008000
|
||||
#define chipMinorFeatures2_FULL_DIRECTFB 0x00010000
|
||||
#define chipMinorFeatures2_2D_TILING 0x00020000
|
||||
#define chipMinorFeatures2_THREAD_WALKER_IN_PS 0x00040000
|
||||
#define chipMinorFeatures2_TILE_FILLER 0x00080000
|
||||
#define chipMinorFeatures2_UNK20 0x00100000
|
||||
#define chipMinorFeatures2_2D_MULTI_SOURCE_BLIT 0x00200000
|
||||
#define chipMinorFeatures2_UNK22 0x00400000
|
||||
#define chipMinorFeatures2_UNK23 0x00800000
|
||||
#define chipMinorFeatures2_UNK24 0x01000000
|
||||
#define chipMinorFeatures2_MIXED_STREAMS 0x02000000
|
||||
#define chipMinorFeatures2_2D_420_L2CACHE 0x04000000
|
||||
#define chipMinorFeatures2_UNK27 0x08000000
|
||||
#define chipMinorFeatures2_2D_NO_INDEX8_BRUSH 0x10000000
|
||||
#define chipMinorFeatures2_TEXTURE_TILED_READ 0x20000000
|
||||
#define chipMinorFeatures2_UNK30 0x40000000
|
||||
#define chipMinorFeatures2_UNK31 0x80000000
|
||||
#define chipMinorFeatures3_ROTATION_STALL_FIX 0x00000001
|
||||
#define chipMinorFeatures3_UNK1 0x00000002
|
||||
#define chipMinorFeatures3_2D_MULTI_SOURCE_BLT_EX 0x00000004
|
||||
#define chipMinorFeatures3_UNK3 0x00000008
|
||||
#define chipMinorFeatures3_UNK4 0x00000010
|
||||
#define chipMinorFeatures3_UNK5 0x00000020
|
||||
#define chipMinorFeatures3_UNK6 0x00000040
|
||||
#define chipMinorFeatures3_UNK7 0x00000080
|
||||
#define chipMinorFeatures3_UNK8 0x00000100
|
||||
#define chipMinorFeatures3_UNK9 0x00000200
|
||||
#define chipMinorFeatures3_BUG_FIXES10 0x00000400
|
||||
#define chipMinorFeatures3_UNK11 0x00000800
|
||||
#define chipMinorFeatures3_BUG_FIXES11 0x00001000
|
||||
#define chipMinorFeatures3_UNK13 0x00002000
|
||||
#define chipMinorFeatures3_UNK14 0x00004000
|
||||
#define chipMinorFeatures3_UNK15 0x00008000
|
||||
#define chipMinorFeatures3_UNK16 0x00010000
|
||||
#define chipMinorFeatures3_UNK17 0x00020000
|
||||
#define chipMinorFeatures3_UNK18 0x00040000
|
||||
#define chipMinorFeatures3_UNK19 0x00080000
|
||||
#define chipMinorFeatures3_UNK20 0x00100000
|
||||
#define chipMinorFeatures3_UNK21 0x00200000
|
||||
#define chipMinorFeatures3_UNK22 0x00400000
|
||||
#define chipMinorFeatures3_UNK23 0x00800000
|
||||
#define chipMinorFeatures3_UNK24 0x01000000
|
||||
#define chipMinorFeatures3_UNK25 0x02000000
|
||||
#define chipMinorFeatures3_UNK26 0x04000000
|
||||
#define chipMinorFeatures3_UNK27 0x08000000
|
||||
#define chipMinorFeatures3_UNK28 0x10000000
|
||||
#define chipMinorFeatures3_UNK29 0x20000000
|
||||
#define chipMinorFeatures3_UNK30 0x40000000
|
||||
#define chipMinorFeatures3_UNK31 0x80000000
|
||||
#define chipMinorFeatures4_UNK0 0x00000001
|
||||
#define chipMinorFeatures4_UNK1 0x00000002
|
||||
#define chipMinorFeatures4_UNK2 0x00000004
|
||||
#define chipMinorFeatures4_UNK3 0x00000008
|
||||
#define chipMinorFeatures4_UNK4 0x00000010
|
||||
#define chipMinorFeatures4_UNK5 0x00000020
|
||||
#define chipMinorFeatures4_UNK6 0x00000040
|
||||
#define chipMinorFeatures4_UNK7 0x00000080
|
||||
#define chipMinorFeatures4_UNK8 0x00000100
|
||||
#define chipMinorFeatures4_UNK9 0x00000200
|
||||
#define chipMinorFeatures4_UNK10 0x00000400
|
||||
#define chipMinorFeatures4_UNK11 0x00000800
|
||||
#define chipMinorFeatures4_UNK12 0x00001000
|
||||
#define chipMinorFeatures4_UNK13 0x00002000
|
||||
#define chipMinorFeatures4_UNK14 0x00004000
|
||||
#define chipMinorFeatures4_UNK15 0x00008000
|
||||
#define chipMinorFeatures4_UNK16 0x00010000
|
||||
#define chipMinorFeatures4_UNK17 0x00020000
|
||||
#define chipMinorFeatures4_UNK18 0x00040000
|
||||
#define chipMinorFeatures4_UNK19 0x00080000
|
||||
#define chipMinorFeatures4_UNK20 0x00100000
|
||||
#define chipMinorFeatures4_UNK21 0x00200000
|
||||
#define chipMinorFeatures4_UNK22 0x00400000
|
||||
#define chipMinorFeatures4_UNK23 0x00800000
|
||||
#define chipMinorFeatures4_UNK24 0x01000000
|
||||
#define chipMinorFeatures4_UNK25 0x02000000
|
||||
#define chipMinorFeatures4_UNK26 0x04000000
|
||||
#define chipMinorFeatures4_UNK27 0x08000000
|
||||
#define chipMinorFeatures4_UNK28 0x10000000
|
||||
#define chipMinorFeatures4_UNK29 0x20000000
|
||||
#define chipMinorFeatures4_UNK30 0x40000000
|
||||
#define chipMinorFeatures4_UNK31 0x80000000
|
||||
|
||||
#endif /* COMMON_XML */
|
268
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
Normal file
268
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
Normal file
|
@ -0,0 +1,268 @@
|
|||
/*
|
||||
* Copyright (C) 2014 Etnaviv Project
|
||||
* Author: Christian Gmeiner <christian.gmeiner@gmail.com>
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#include "etnaviv_gpu.h"
|
||||
#include "etnaviv_gem.h"
|
||||
#include "etnaviv_mmu.h"
|
||||
|
||||
#include "common.xml.h"
|
||||
#include "state.xml.h"
|
||||
#include "cmdstream.xml.h"
|
||||
|
||||
/*
|
||||
* Command Buffer helper:
|
||||
*/
|
||||
|
||||
|
||||
static inline void OUT(struct etnaviv_cmdbuf *buffer, u32 data)
|
||||
{
|
||||
u32 *vaddr = (u32 *)buffer->vaddr;
|
||||
|
||||
BUG_ON(buffer->user_size >= buffer->size);
|
||||
|
||||
vaddr[buffer->user_size / 4] = data;
|
||||
buffer->user_size += 4;
|
||||
}
|
||||
|
||||
static inline void CMD_LOAD_STATE(struct etnaviv_cmdbuf *buffer,
|
||||
u32 reg, u32 value)
|
||||
{
|
||||
u32 index = reg >> VIV_FE_LOAD_STATE_HEADER_OFFSET__SHR;
|
||||
|
||||
buffer->user_size = ALIGN(buffer->user_size, 8);
|
||||
|
||||
/* write a register via cmd stream */
|
||||
OUT(buffer, VIV_FE_LOAD_STATE_HEADER_OP_LOAD_STATE |
|
||||
VIV_FE_LOAD_STATE_HEADER_COUNT(1) |
|
||||
VIV_FE_LOAD_STATE_HEADER_OFFSET(index));
|
||||
OUT(buffer, value);
|
||||
}
|
||||
|
||||
static inline void CMD_END(struct etnaviv_cmdbuf *buffer)
|
||||
{
|
||||
buffer->user_size = ALIGN(buffer->user_size, 8);
|
||||
|
||||
OUT(buffer, VIV_FE_END_HEADER_OP_END);
|
||||
}
|
||||
|
||||
static inline void CMD_WAIT(struct etnaviv_cmdbuf *buffer)
|
||||
{
|
||||
buffer->user_size = ALIGN(buffer->user_size, 8);
|
||||
|
||||
OUT(buffer, VIV_FE_WAIT_HEADER_OP_WAIT | 200);
|
||||
}
|
||||
|
||||
static inline void CMD_LINK(struct etnaviv_cmdbuf *buffer,
|
||||
u16 prefetch, u32 address)
|
||||
{
|
||||
buffer->user_size = ALIGN(buffer->user_size, 8);
|
||||
|
||||
OUT(buffer, VIV_FE_LINK_HEADER_OP_LINK |
|
||||
VIV_FE_LINK_HEADER_PREFETCH(prefetch));
|
||||
OUT(buffer, address);
|
||||
}
|
||||
|
||||
static inline void CMD_STALL(struct etnaviv_cmdbuf *buffer,
|
||||
u32 from, u32 to)
|
||||
{
|
||||
buffer->user_size = ALIGN(buffer->user_size, 8);
|
||||
|
||||
OUT(buffer, VIV_FE_STALL_HEADER_OP_STALL);
|
||||
OUT(buffer, VIV_FE_STALL_TOKEN_FROM(from) | VIV_FE_STALL_TOKEN_TO(to));
|
||||
}
|
||||
|
||||
static void etnaviv_cmd_select_pipe(struct etnaviv_cmdbuf *buffer, u8 pipe)
|
||||
{
|
||||
u32 flush;
|
||||
u32 stall;
|
||||
|
||||
/*
|
||||
* This assumes that if we're switching to 2D, we're switching
|
||||
* away from 3D, and vice versa. Hence, if we're switching to
|
||||
* the 2D core, we need to flush the 3D depth and color caches,
|
||||
* otherwise we need to flush the 2D pixel engine cache.
|
||||
*/
|
||||
if (pipe == ETNA_PIPE_2D)
|
||||
flush = VIVS_GL_FLUSH_CACHE_DEPTH | VIVS_GL_FLUSH_CACHE_COLOR;
|
||||
else
|
||||
flush = VIVS_GL_FLUSH_CACHE_PE2D;
|
||||
|
||||
stall = VIVS_GL_SEMAPHORE_TOKEN_FROM(SYNC_RECIPIENT_FE) |
|
||||
VIVS_GL_SEMAPHORE_TOKEN_TO(SYNC_RECIPIENT_PE);
|
||||
|
||||
CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush);
|
||||
CMD_LOAD_STATE(buffer, VIVS_GL_SEMAPHORE_TOKEN, stall);
|
||||
|
||||
CMD_STALL(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
||||
|
||||
CMD_LOAD_STATE(buffer, VIVS_GL_PIPE_SELECT,
|
||||
VIVS_GL_PIPE_SELECT_PIPE(pipe));
|
||||
}
|
||||
|
||||
static u32 gpu_va(struct etnaviv_gpu *gpu, struct etnaviv_cmdbuf *buf)
|
||||
{
|
||||
return buf->paddr - gpu->memory_base;
|
||||
}
|
||||
|
||||
static void etnaviv_buffer_dump(struct etnaviv_gpu *gpu,
|
||||
struct etnaviv_cmdbuf *buf, u32 off, u32 len)
|
||||
{
|
||||
u32 size = buf->size;
|
||||
u32 *ptr = buf->vaddr + off;
|
||||
|
||||
dev_info(gpu->dev, "virt %p phys 0x%08x free 0x%08x\n",
|
||||
ptr, gpu_va(gpu, buf) + off, size - len * 4 - off);
|
||||
|
||||
print_hex_dump(KERN_INFO, "cmd ", DUMP_PREFIX_OFFSET, 16, 4,
|
||||
ptr, len * 4, 0);
|
||||
}
|
||||
|
||||
u16 etnaviv_buffer_init(struct etnaviv_gpu *gpu)
|
||||
{
|
||||
struct etnaviv_cmdbuf *buffer = gpu->buffer;
|
||||
|
||||
/* initialize buffer */
|
||||
buffer->user_size = 0;
|
||||
|
||||
CMD_WAIT(buffer);
|
||||
CMD_LINK(buffer, 2, gpu_va(gpu, buffer) + buffer->user_size - 4);
|
||||
|
||||
return buffer->user_size / 8;
|
||||
}
|
||||
|
||||
void etnaviv_buffer_end(struct etnaviv_gpu *gpu)
|
||||
{
|
||||
struct etnaviv_cmdbuf *buffer = gpu->buffer;
|
||||
|
||||
/* Replace the last WAIT with an END */
|
||||
buffer->user_size -= 16;
|
||||
|
||||
CMD_END(buffer);
|
||||
mb();
|
||||
}
|
||||
|
||||
void etnaviv_buffer_queue(struct etnaviv_gpu *gpu, unsigned int event,
|
||||
struct etnaviv_cmdbuf *cmdbuf)
|
||||
{
|
||||
struct etnaviv_cmdbuf *buffer = gpu->buffer;
|
||||
u32 *lw = buffer->vaddr + buffer->user_size - 16;
|
||||
u32 back, link_target, link_size, reserve_size, extra_size = 0;
|
||||
|
||||
if (drm_debug & DRM_UT_DRIVER)
|
||||
etnaviv_buffer_dump(gpu, buffer, 0, 0x50);
|
||||
|
||||
/*
|
||||
* If we need to flush the MMU prior to submitting this buffer, we
|
||||
* will need to append a mmu flush load state, followed by a new
|
||||
* link to this buffer - a total of four additional words.
|
||||
*/
|
||||
if (gpu->mmu->need_flush || gpu->switch_context) {
|
||||
/* link command */
|
||||
extra_size += 2;
|
||||
/* flush command */
|
||||
if (gpu->mmu->need_flush)
|
||||
extra_size += 2;
|
||||
/* pipe switch commands */
|
||||
if (gpu->switch_context)
|
||||
extra_size += 8;
|
||||
}
|
||||
|
||||
reserve_size = (6 + extra_size) * 4;
|
||||
|
||||
/*
|
||||
* if we are going to completely overflow the buffer, we need to wrap.
|
||||
*/
|
||||
if (buffer->user_size + reserve_size > buffer->size)
|
||||
buffer->user_size = 0;
|
||||
|
||||
/* save offset back into main buffer */
|
||||
back = buffer->user_size + reserve_size - 6 * 4;
|
||||
link_target = gpu_va(gpu, buffer) + buffer->user_size;
|
||||
link_size = 6;
|
||||
|
||||
/* Skip over any extra instructions */
|
||||
link_target += extra_size * sizeof(u32);
|
||||
|
||||
if (drm_debug & DRM_UT_DRIVER)
|
||||
pr_info("stream link to 0x%08x @ 0x%08x %p\n",
|
||||
link_target, gpu_va(gpu, cmdbuf), cmdbuf->vaddr);
|
||||
|
||||
/* jump back from cmd to main buffer */
|
||||
CMD_LINK(cmdbuf, link_size, link_target);
|
||||
|
||||
link_target = gpu_va(gpu, cmdbuf);
|
||||
link_size = cmdbuf->size / 8;
|
||||
|
||||
|
||||
|
||||
if (drm_debug & DRM_UT_DRIVER) {
|
||||
print_hex_dump(KERN_INFO, "cmd ", DUMP_PREFIX_OFFSET, 16, 4,
|
||||
cmdbuf->vaddr, cmdbuf->size, 0);
|
||||
|
||||
pr_info("link op: %p\n", lw);
|
||||
pr_info("link addr: %p\n", lw + 1);
|
||||
pr_info("addr: 0x%08x\n", link_target);
|
||||
pr_info("back: 0x%08x\n", gpu_va(gpu, buffer) + back);
|
||||
pr_info("event: %d\n", event);
|
||||
}
|
||||
|
||||
if (gpu->mmu->need_flush || gpu->switch_context) {
|
||||
u32 new_target = gpu_va(gpu, buffer) + buffer->user_size;
|
||||
|
||||
if (gpu->mmu->need_flush) {
|
||||
/* Add the MMU flush */
|
||||
CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_MMU,
|
||||
VIVS_GL_FLUSH_MMU_FLUSH_FEMMU |
|
||||
VIVS_GL_FLUSH_MMU_FLUSH_UNK1 |
|
||||
VIVS_GL_FLUSH_MMU_FLUSH_UNK2 |
|
||||
VIVS_GL_FLUSH_MMU_FLUSH_PEMMU |
|
||||
VIVS_GL_FLUSH_MMU_FLUSH_UNK4);
|
||||
|
||||
gpu->mmu->need_flush = false;
|
||||
}
|
||||
|
||||
if (gpu->switch_context) {
|
||||
etnaviv_cmd_select_pipe(buffer, cmdbuf->exec_state);
|
||||
gpu->switch_context = false;
|
||||
}
|
||||
|
||||
/* And the link to the first buffer */
|
||||
CMD_LINK(buffer, link_size, link_target);
|
||||
|
||||
/* Update the link target to point to above instructions */
|
||||
link_target = new_target;
|
||||
link_size = extra_size;
|
||||
}
|
||||
|
||||
/* trigger event */
|
||||
CMD_LOAD_STATE(buffer, VIVS_GL_EVENT, VIVS_GL_EVENT_EVENT_ID(event) |
|
||||
VIVS_GL_EVENT_FROM_PE);
|
||||
|
||||
/* append WAIT/LINK to main buffer */
|
||||
CMD_WAIT(buffer);
|
||||
CMD_LINK(buffer, 2, gpu_va(gpu, buffer) + (buffer->user_size - 4));
|
||||
|
||||
/* Change WAIT into a LINK command; write the address first. */
|
||||
*(lw + 1) = link_target;
|
||||
mb();
|
||||
*(lw) = VIV_FE_LINK_HEADER_OP_LINK |
|
||||
VIV_FE_LINK_HEADER_PREFETCH(link_size);
|
||||
mb();
|
||||
|
||||
if (drm_debug & DRM_UT_DRIVER)
|
||||
etnaviv_buffer_dump(gpu, buffer, 0, 0x50);
|
||||
}
|
209
drivers/gpu/drm/etnaviv/etnaviv_cmd_parser.c
Normal file
209
drivers/gpu/drm/etnaviv/etnaviv_cmd_parser.c
Normal file
|
@ -0,0 +1,209 @@
|
|||
/*
|
||||
* Copyright (C) 2015 Etnaviv Project
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#include <linux/kernel.h>
|
||||
|
||||
#include "etnaviv_gem.h"
|
||||
#include "etnaviv_gpu.h"
|
||||
|
||||
#include "cmdstream.xml.h"
|
||||
|
||||
#define EXTRACT(val, field) (((val) & field##__MASK) >> field##__SHIFT)
|
||||
|
||||
struct etna_validation_state {
|
||||
struct etnaviv_gpu *gpu;
|
||||
const struct drm_etnaviv_gem_submit_reloc *relocs;
|
||||
unsigned int num_relocs;
|
||||
u32 *start;
|
||||
};
|
||||
|
||||
static const struct {
|
||||
u16 offset;
|
||||
u16 size;
|
||||
} etnaviv_sensitive_states[] __initconst = {
|
||||
#define ST(start, num) { (start) >> 2, (num) }
|
||||
/* 2D */
|
||||
ST(0x1200, 1),
|
||||
ST(0x1228, 1),
|
||||
ST(0x1238, 1),
|
||||
ST(0x1284, 1),
|
||||
ST(0x128c, 1),
|
||||
ST(0x1304, 1),
|
||||
ST(0x1310, 1),
|
||||
ST(0x1318, 1),
|
||||
ST(0x12800, 4),
|
||||
ST(0x128a0, 4),
|
||||
ST(0x128c0, 4),
|
||||
ST(0x12970, 4),
|
||||
ST(0x12a00, 8),
|
||||
ST(0x12b40, 8),
|
||||
ST(0x12b80, 8),
|
||||
ST(0x12ce0, 8),
|
||||
/* 3D */
|
||||
ST(0x0644, 1),
|
||||
ST(0x064c, 1),
|
||||
ST(0x0680, 8),
|
||||
ST(0x1410, 1),
|
||||
ST(0x1430, 1),
|
||||
ST(0x1458, 1),
|
||||
ST(0x1460, 8),
|
||||
ST(0x1480, 8),
|
||||
ST(0x1500, 8),
|
||||
ST(0x1520, 8),
|
||||
ST(0x1608, 1),
|
||||
ST(0x1610, 1),
|
||||
ST(0x1658, 1),
|
||||
ST(0x165c, 1),
|
||||
ST(0x1664, 1),
|
||||
ST(0x1668, 1),
|
||||
ST(0x16a4, 1),
|
||||
ST(0x16c0, 8),
|
||||
ST(0x16e0, 8),
|
||||
ST(0x1740, 8),
|
||||
ST(0x2400, 14 * 16),
|
||||
ST(0x10800, 32 * 16),
|
||||
#undef ST
|
||||
};
|
||||
|
||||
#define ETNAVIV_STATES_SIZE (VIV_FE_LOAD_STATE_HEADER_OFFSET__MASK + 1u)
|
||||
static DECLARE_BITMAP(etnaviv_states, ETNAVIV_STATES_SIZE);
|
||||
|
||||
void __init etnaviv_validate_init(void)
|
||||
{
|
||||
unsigned int i;
|
||||
|
||||
for (i = 0; i < ARRAY_SIZE(etnaviv_sensitive_states); i++)
|
||||
bitmap_set(etnaviv_states, etnaviv_sensitive_states[i].offset,
|
||||
etnaviv_sensitive_states[i].size);
|
||||
}
|
||||
|
||||
static void etnaviv_warn_if_non_sensitive(struct etna_validation_state *state,
|
||||
unsigned int buf_offset, unsigned int state_addr)
|
||||
{
|
||||
if (state->num_relocs && state->relocs->submit_offset < buf_offset) {
|
||||
dev_warn_once(state->gpu->dev,
|
||||
"%s: relocation for non-sensitive state 0x%x at offset %u\n",
|
||||
__func__, state_addr,
|
||||
state->relocs->submit_offset);
|
||||
while (state->num_relocs &&
|
||||
state->relocs->submit_offset < buf_offset) {
|
||||
state->relocs++;
|
||||
state->num_relocs--;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
static bool etnaviv_validate_load_state(struct etna_validation_state *state,
|
||||
u32 *ptr, unsigned int state_offset, unsigned int num)
|
||||
{
|
||||
unsigned int size = min(ETNAVIV_STATES_SIZE, state_offset + num);
|
||||
unsigned int st_offset = state_offset, buf_offset;
|
||||
|
||||
for_each_set_bit_from(st_offset, etnaviv_states, size) {
|
||||
buf_offset = (ptr - state->start +
|
||||
st_offset - state_offset) * 4;
|
||||
|
||||
etnaviv_warn_if_non_sensitive(state, buf_offset, st_offset * 4);
|
||||
if (state->num_relocs &&
|
||||
state->relocs->submit_offset == buf_offset) {
|
||||
state->relocs++;
|
||||
state->num_relocs--;
|
||||
continue;
|
||||
}
|
||||
|
||||
dev_warn_ratelimited(state->gpu->dev,
|
||||
"%s: load state touches restricted state 0x%x at offset %u\n",
|
||||
__func__, st_offset * 4, buf_offset);
|
||||
return false;
|
||||
}
|
||||
|
||||
if (state->num_relocs) {
|
||||
buf_offset = (ptr - state->start + num) * 4;
|
||||
etnaviv_warn_if_non_sensitive(state, buf_offset, st_offset * 4 +
|
||||
state->relocs->submit_offset -
|
||||
buf_offset);
|
||||
}
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
static uint8_t cmd_length[32] = {
|
||||
[FE_OPCODE_DRAW_PRIMITIVES] = 4,
|
||||
[FE_OPCODE_DRAW_INDEXED_PRIMITIVES] = 6,
|
||||
[FE_OPCODE_NOP] = 2,
|
||||
[FE_OPCODE_STALL] = 2,
|
||||
};
|
||||
|
||||
bool etnaviv_cmd_validate_one(struct etnaviv_gpu *gpu, u32 *stream,
|
||||
unsigned int size,
|
||||
struct drm_etnaviv_gem_submit_reloc *relocs,
|
||||
unsigned int reloc_size)
|
||||
{
|
||||
struct etna_validation_state state;
|
||||
u32 *buf = stream;
|
||||
u32 *end = buf + size;
|
||||
|
||||
state.gpu = gpu;
|
||||
state.relocs = relocs;
|
||||
state.num_relocs = reloc_size;
|
||||
state.start = stream;
|
||||
|
||||
while (buf < end) {
|
||||
u32 cmd = *buf;
|
||||
unsigned int len, n, off;
|
||||
unsigned int op = cmd >> 27;
|
||||
|
||||
switch (op) {
|
||||
case FE_OPCODE_LOAD_STATE:
|
||||
n = EXTRACT(cmd, VIV_FE_LOAD_STATE_HEADER_COUNT);
|
||||
len = ALIGN(1 + n, 2);
|
||||
if (buf + len > end)
|
||||
break;
|
||||
|
||||
off = EXTRACT(cmd, VIV_FE_LOAD_STATE_HEADER_OFFSET);
|
||||
if (!etnaviv_validate_load_state(&state, buf + 1,
|
||||
off, n))
|
||||
return false;
|
||||
break;
|
||||
|
||||
case FE_OPCODE_DRAW_2D:
|
||||
n = EXTRACT(cmd, VIV_FE_DRAW_2D_HEADER_COUNT);
|
||||
if (n == 0)
|
||||
n = 256;
|
||||
len = 2 + n * 2;
|
||||
break;
|
||||
|
||||
default:
|
||||
len = cmd_length[op];
|
||||
if (len == 0) {
|
||||
dev_err(gpu->dev, "%s: op %u not permitted at offset %tu\n",
|
||||
__func__, op, buf - state.start);
|
||||
return false;
|
||||
}
|
||||
break;
|
||||
}
|
||||
|
||||
buf += len;
|
||||
}
|
||||
|
||||
if (buf > end) {
|
||||
dev_err(gpu->dev, "%s: commands overflow end of buffer: %tu > %u\n",
|
||||
__func__, buf - state.start, size);
|
||||
return false;
|
||||
}
|
||||
|
||||
return true;
|
||||
}
|
707
drivers/gpu/drm/etnaviv/etnaviv_drv.c
Normal file
707
drivers/gpu/drm/etnaviv/etnaviv_drv.c
Normal file
|
@ -0,0 +1,707 @@
|
|||
/*
|
||||
* Copyright (C) 2015 Etnaviv Project
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#include <linux/component.h>
|
||||
#include <linux/of_platform.h>
|
||||
|
||||
#include "etnaviv_drv.h"
|
||||
#include "etnaviv_gpu.h"
|
||||
#include "etnaviv_gem.h"
|
||||
#include "etnaviv_mmu.h"
|
||||
#include "etnaviv_gem.h"
|
||||
|
||||
#ifdef CONFIG_DRM_ETNAVIV_REGISTER_LOGGING
|
||||
static bool reglog;
|
||||
MODULE_PARM_DESC(reglog, "Enable register read/write logging");
|
||||
module_param(reglog, bool, 0600);
|
||||
#else
|
||||
#define reglog 0
|
||||
#endif
|
||||
|
||||
void __iomem *etnaviv_ioremap(struct platform_device *pdev, const char *name,
|
||||
const char *dbgname)
|
||||
{
|
||||
struct resource *res;
|
||||
void __iomem *ptr;
|
||||
|
||||
if (name)
|
||||
res = platform_get_resource_byname(pdev, IORESOURCE_MEM, name);
|
||||
else
|
||||
res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
|
||||
|
||||
ptr = devm_ioremap_resource(&pdev->dev, res);
|
||||
if (IS_ERR(ptr)) {
|
||||
dev_err(&pdev->dev, "failed to ioremap %s: %ld\n", name,
|
||||
PTR_ERR(ptr));
|
||||
return ptr;
|
||||
}
|
||||
|
||||
if (reglog)
|
||||
dev_printk(KERN_DEBUG, &pdev->dev, "IO:region %s 0x%p %08zx\n",
|
||||
dbgname, ptr, (size_t)resource_size(res));
|
||||
|
||||
return ptr;
|
||||
}
|
||||
|
||||
void etnaviv_writel(u32 data, void __iomem *addr)
|
||||
{
|
||||
if (reglog)
|
||||
printk(KERN_DEBUG "IO:W %p %08x\n", addr, data);
|
||||
|
||||
writel(data, addr);
|
||||
}
|
||||
|
||||
u32 etnaviv_readl(const void __iomem *addr)
|
||||
{
|
||||
u32 val = readl(addr);
|
||||
|
||||
if (reglog)
|
||||
printk(KERN_DEBUG "IO:R %p %08x\n", addr, val);
|
||||
|
||||
return val;
|
||||
}
|
||||
|
||||
/*
|
||||
* DRM operations:
|
||||
*/
|
||||
|
||||
|
||||
static void load_gpu(struct drm_device *dev)
|
||||
{
|
||||
struct etnaviv_drm_private *priv = dev->dev_private;
|
||||
unsigned int i;
|
||||
|
||||
for (i = 0; i < ETNA_MAX_PIPES; i++) {
|
||||
struct etnaviv_gpu *g = priv->gpu[i];
|
||||
|
||||
if (g) {
|
||||
int ret;
|
||||
|
||||
ret = etnaviv_gpu_init(g);
|
||||
if (ret) {
|
||||
dev_err(g->dev, "hw init failed: %d\n", ret);
|
||||
priv->gpu[i] = NULL;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
static int etnaviv_open(struct drm_device *dev, struct drm_file *file)
|
||||
{
|
||||
struct etnaviv_file_private *ctx;
|
||||
|
||||
ctx = kzalloc(sizeof(*ctx), GFP_KERNEL);
|
||||
if (!ctx)
|
||||
return -ENOMEM;
|
||||
|
||||
file->driver_priv = ctx;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void etnaviv_preclose(struct drm_device *dev, struct drm_file *file)
|
||||
{
|
||||
struct etnaviv_drm_private *priv = dev->dev_private;
|
||||
struct etnaviv_file_private *ctx = file->driver_priv;
|
||||
unsigned int i;
|
||||
|
||||
for (i = 0; i < ETNA_MAX_PIPES; i++) {
|
||||
struct etnaviv_gpu *gpu = priv->gpu[i];
|
||||
|
||||
if (gpu) {
|
||||
mutex_lock(&gpu->lock);
|
||||
if (gpu->lastctx == ctx)
|
||||
gpu->lastctx = NULL;
|
||||
mutex_unlock(&gpu->lock);
|
||||
}
|
||||
}
|
||||
|
||||
kfree(ctx);
|
||||
}
|
||||
|
||||
/*
|
||||
* DRM debugfs:
|
||||
*/
|
||||
|
||||
#ifdef CONFIG_DEBUG_FS
|
||||
static int etnaviv_gem_show(struct drm_device *dev, struct seq_file *m)
|
||||
{
|
||||
struct etnaviv_drm_private *priv = dev->dev_private;
|
||||
|
||||
etnaviv_gem_describe_objects(priv, m);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int etnaviv_mm_show(struct drm_device *dev, struct seq_file *m)
|
||||
{
|
||||
int ret;
|
||||
|
||||
read_lock(&dev->vma_offset_manager->vm_lock);
|
||||
ret = drm_mm_dump_table(m, &dev->vma_offset_manager->vm_addr_space_mm);
|
||||
read_unlock(&dev->vma_offset_manager->vm_lock);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int etnaviv_mmu_show(struct etnaviv_gpu *gpu, struct seq_file *m)
|
||||
{
|
||||
seq_printf(m, "Active Objects (%s):\n", dev_name(gpu->dev));
|
||||
|
||||
mutex_lock(&gpu->mmu->lock);
|
||||
drm_mm_dump_table(m, &gpu->mmu->mm);
|
||||
mutex_unlock(&gpu->mmu->lock);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void etnaviv_buffer_dump(struct etnaviv_gpu *gpu, struct seq_file *m)
|
||||
{
|
||||
struct etnaviv_cmdbuf *buf = gpu->buffer;
|
||||
u32 size = buf->size;
|
||||
u32 *ptr = buf->vaddr;
|
||||
u32 i;
|
||||
|
||||
seq_printf(m, "virt %p - phys 0x%llx - free 0x%08x\n",
|
||||
buf->vaddr, (u64)buf->paddr, size - buf->user_size);
|
||||
|
||||
for (i = 0; i < size / 4; i++) {
|
||||
if (i && !(i % 4))
|
||||
seq_puts(m, "\n");
|
||||
if (i % 4 == 0)
|
||||
seq_printf(m, "\t0x%p: ", ptr + i);
|
||||
seq_printf(m, "%08x ", *(ptr + i));
|
||||
}
|
||||
seq_puts(m, "\n");
|
||||
}
|
||||
|
||||
static int etnaviv_ring_show(struct etnaviv_gpu *gpu, struct seq_file *m)
|
||||
{
|
||||
seq_printf(m, "Ring Buffer (%s): ", dev_name(gpu->dev));
|
||||
|
||||
mutex_lock(&gpu->lock);
|
||||
etnaviv_buffer_dump(gpu, m);
|
||||
mutex_unlock(&gpu->lock);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int show_unlocked(struct seq_file *m, void *arg)
|
||||
{
|
||||
struct drm_info_node *node = (struct drm_info_node *) m->private;
|
||||
struct drm_device *dev = node->minor->dev;
|
||||
int (*show)(struct drm_device *dev, struct seq_file *m) =
|
||||
node->info_ent->data;
|
||||
|
||||
return show(dev, m);
|
||||
}
|
||||
|
||||
static int show_each_gpu(struct seq_file *m, void *arg)
|
||||
{
|
||||
struct drm_info_node *node = (struct drm_info_node *) m->private;
|
||||
struct drm_device *dev = node->minor->dev;
|
||||
struct etnaviv_drm_private *priv = dev->dev_private;
|
||||
struct etnaviv_gpu *gpu;
|
||||
int (*show)(struct etnaviv_gpu *gpu, struct seq_file *m) =
|
||||
node->info_ent->data;
|
||||
unsigned int i;
|
||||
int ret = 0;
|
||||
|
||||
for (i = 0; i < ETNA_MAX_PIPES; i++) {
|
||||
gpu = priv->gpu[i];
|
||||
if (!gpu)
|
||||
continue;
|
||||
|
||||
ret = show(gpu, m);
|
||||
if (ret < 0)
|
||||
break;
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static struct drm_info_list etnaviv_debugfs_list[] = {
|
||||
{"gpu", show_each_gpu, 0, etnaviv_gpu_debugfs},
|
||||
{"gem", show_unlocked, 0, etnaviv_gem_show},
|
||||
{ "mm", show_unlocked, 0, etnaviv_mm_show },
|
||||
{"mmu", show_each_gpu, 0, etnaviv_mmu_show},
|
||||
{"ring", show_each_gpu, 0, etnaviv_ring_show},
|
||||
};
|
||||
|
||||
static int etnaviv_debugfs_init(struct drm_minor *minor)
|
||||
{
|
||||
struct drm_device *dev = minor->dev;
|
||||
int ret;
|
||||
|
||||
ret = drm_debugfs_create_files(etnaviv_debugfs_list,
|
||||
ARRAY_SIZE(etnaviv_debugfs_list),
|
||||
minor->debugfs_root, minor);
|
||||
|
||||
if (ret) {
|
||||
dev_err(dev->dev, "could not install etnaviv_debugfs_list\n");
|
||||
return ret;
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void etnaviv_debugfs_cleanup(struct drm_minor *minor)
|
||||
{
|
||||
drm_debugfs_remove_files(etnaviv_debugfs_list,
|
||||
ARRAY_SIZE(etnaviv_debugfs_list), minor);
|
||||
}
|
||||
#endif
|
||||
|
||||
/*
|
||||
* DRM ioctls:
|
||||
*/
|
||||
|
||||
static int etnaviv_ioctl_get_param(struct drm_device *dev, void *data,
|
||||
struct drm_file *file)
|
||||
{
|
||||
struct etnaviv_drm_private *priv = dev->dev_private;
|
||||
struct drm_etnaviv_param *args = data;
|
||||
struct etnaviv_gpu *gpu;
|
||||
|
||||
if (args->pipe >= ETNA_MAX_PIPES)
|
||||
return -EINVAL;
|
||||
|
||||
gpu = priv->gpu[args->pipe];
|
||||
if (!gpu)
|
||||
return -ENXIO;
|
||||
|
||||
return etnaviv_gpu_get_param(gpu, args->param, &args->value);
|
||||
}
|
||||
|
||||
static int etnaviv_ioctl_gem_new(struct drm_device *dev, void *data,
|
||||
struct drm_file *file)
|
||||
{
|
||||
struct drm_etnaviv_gem_new *args = data;
|
||||
|
||||
if (args->flags & ~(ETNA_BO_CACHED | ETNA_BO_WC | ETNA_BO_UNCACHED |
|
||||
ETNA_BO_FORCE_MMU))
|
||||
return -EINVAL;
|
||||
|
||||
return etnaviv_gem_new_handle(dev, file, args->size,
|
||||
args->flags, &args->handle);
|
||||
}
|
||||
|
||||
#define TS(t) ((struct timespec){ \
|
||||
.tv_sec = (t).tv_sec, \
|
||||
.tv_nsec = (t).tv_nsec \
|
||||
})
|
||||
|
||||
static int etnaviv_ioctl_gem_cpu_prep(struct drm_device *dev, void *data,
|
||||
struct drm_file *file)
|
||||
{
|
||||
struct drm_etnaviv_gem_cpu_prep *args = data;
|
||||
struct drm_gem_object *obj;
|
||||
int ret;
|
||||
|
||||
if (args->op & ~(ETNA_PREP_READ | ETNA_PREP_WRITE | ETNA_PREP_NOSYNC))
|
||||
return -EINVAL;
|
||||
|
||||
obj = drm_gem_object_lookup(dev, file, args->handle);
|
||||
if (!obj)
|
||||
return -ENOENT;
|
||||
|
||||
ret = etnaviv_gem_cpu_prep(obj, args->op, &TS(args->timeout));
|
||||
|
||||
drm_gem_object_unreference_unlocked(obj);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int etnaviv_ioctl_gem_cpu_fini(struct drm_device *dev, void *data,
|
||||
struct drm_file *file)
|
||||
{
|
||||
struct drm_etnaviv_gem_cpu_fini *args = data;
|
||||
struct drm_gem_object *obj;
|
||||
int ret;
|
||||
|
||||
if (args->flags)
|
||||
return -EINVAL;
|
||||
|
||||
obj = drm_gem_object_lookup(dev, file, args->handle);
|
||||
if (!obj)
|
||||
return -ENOENT;
|
||||
|
||||
ret = etnaviv_gem_cpu_fini(obj);
|
||||
|
||||
drm_gem_object_unreference_unlocked(obj);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int etnaviv_ioctl_gem_info(struct drm_device *dev, void *data,
|
||||
struct drm_file *file)
|
||||
{
|
||||
struct drm_etnaviv_gem_info *args = data;
|
||||
struct drm_gem_object *obj;
|
||||
int ret;
|
||||
|
||||
if (args->pad)
|
||||
return -EINVAL;
|
||||
|
||||
obj = drm_gem_object_lookup(dev, file, args->handle);
|
||||
if (!obj)
|
||||
return -ENOENT;
|
||||
|
||||
ret = etnaviv_gem_mmap_offset(obj, &args->offset);
|
||||
drm_gem_object_unreference_unlocked(obj);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int etnaviv_ioctl_wait_fence(struct drm_device *dev, void *data,
|
||||
struct drm_file *file)
|
||||
{
|
||||
struct drm_etnaviv_wait_fence *args = data;
|
||||
struct etnaviv_drm_private *priv = dev->dev_private;
|
||||
struct timespec *timeout = &TS(args->timeout);
|
||||
struct etnaviv_gpu *gpu;
|
||||
|
||||
if (args->flags & ~(ETNA_WAIT_NONBLOCK))
|
||||
return -EINVAL;
|
||||
|
||||
if (args->pipe >= ETNA_MAX_PIPES)
|
||||
return -EINVAL;
|
||||
|
||||
gpu = priv->gpu[args->pipe];
|
||||
if (!gpu)
|
||||
return -ENXIO;
|
||||
|
||||
if (args->flags & ETNA_WAIT_NONBLOCK)
|
||||
timeout = NULL;
|
||||
|
||||
return etnaviv_gpu_wait_fence_interruptible(gpu, args->fence,
|
||||
timeout);
|
||||
}
|
||||
|
||||
static int etnaviv_ioctl_gem_userptr(struct drm_device *dev, void *data,
|
||||
struct drm_file *file)
|
||||
{
|
||||
struct drm_etnaviv_gem_userptr *args = data;
|
||||
int access;
|
||||
|
||||
if (args->flags & ~(ETNA_USERPTR_READ|ETNA_USERPTR_WRITE) ||
|
||||
args->flags == 0)
|
||||
return -EINVAL;
|
||||
|
||||
if (offset_in_page(args->user_ptr | args->user_size) ||
|
||||
(uintptr_t)args->user_ptr != args->user_ptr ||
|
||||
(u32)args->user_size != args->user_size ||
|
||||
args->user_ptr & ~PAGE_MASK)
|
||||
return -EINVAL;
|
||||
|
||||
if (args->flags & ETNA_USERPTR_WRITE)
|
||||
access = VERIFY_WRITE;
|
||||
else
|
||||
access = VERIFY_READ;
|
||||
|
||||
if (!access_ok(access, (void __user *)(unsigned long)args->user_ptr,
|
||||
args->user_size))
|
||||
return -EFAULT;
|
||||
|
||||
return etnaviv_gem_new_userptr(dev, file, args->user_ptr,
|
||||
args->user_size, args->flags,
|
||||
&args->handle);
|
||||
}
|
||||
|
||||
static int etnaviv_ioctl_gem_wait(struct drm_device *dev, void *data,
|
||||
struct drm_file *file)
|
||||
{
|
||||
struct etnaviv_drm_private *priv = dev->dev_private;
|
||||
struct drm_etnaviv_gem_wait *args = data;
|
||||
struct timespec *timeout = &TS(args->timeout);
|
||||
struct drm_gem_object *obj;
|
||||
struct etnaviv_gpu *gpu;
|
||||
int ret;
|
||||
|
||||
if (args->flags & ~(ETNA_WAIT_NONBLOCK))
|
||||
return -EINVAL;
|
||||
|
||||
if (args->pipe >= ETNA_MAX_PIPES)
|
||||
return -EINVAL;
|
||||
|
||||
gpu = priv->gpu[args->pipe];
|
||||
if (!gpu)
|
||||
return -ENXIO;
|
||||
|
||||
obj = drm_gem_object_lookup(dev, file, args->handle);
|
||||
if (!obj)
|
||||
return -ENOENT;
|
||||
|
||||
if (args->flags & ETNA_WAIT_NONBLOCK)
|
||||
timeout = NULL;
|
||||
|
||||
ret = etnaviv_gem_wait_bo(gpu, obj, timeout);
|
||||
|
||||
drm_gem_object_unreference_unlocked(obj);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static const struct drm_ioctl_desc etnaviv_ioctls[] = {
|
||||
#define ETNA_IOCTL(n, func, flags) \
|
||||
DRM_IOCTL_DEF_DRV(ETNAVIV_##n, etnaviv_ioctl_##func, flags)
|
||||
ETNA_IOCTL(GET_PARAM, get_param, DRM_AUTH|DRM_RENDER_ALLOW),
|
||||
ETNA_IOCTL(GEM_NEW, gem_new, DRM_AUTH|DRM_RENDER_ALLOW),
|
||||
ETNA_IOCTL(GEM_INFO, gem_info, DRM_AUTH|DRM_RENDER_ALLOW),
|
||||
ETNA_IOCTL(GEM_CPU_PREP, gem_cpu_prep, DRM_AUTH|DRM_RENDER_ALLOW),
|
||||
ETNA_IOCTL(GEM_CPU_FINI, gem_cpu_fini, DRM_AUTH|DRM_RENDER_ALLOW),
|
||||
ETNA_IOCTL(GEM_SUBMIT, gem_submit, DRM_AUTH|DRM_RENDER_ALLOW),
|
||||
ETNA_IOCTL(WAIT_FENCE, wait_fence, DRM_AUTH|DRM_RENDER_ALLOW),
|
||||
ETNA_IOCTL(GEM_USERPTR, gem_userptr, DRM_AUTH|DRM_RENDER_ALLOW),
|
||||
ETNA_IOCTL(GEM_WAIT, gem_wait, DRM_AUTH|DRM_RENDER_ALLOW),
|
||||
};
|
||||
|
||||
static const struct vm_operations_struct vm_ops = {
|
||||
.fault = etnaviv_gem_fault,
|
||||
.open = drm_gem_vm_open,
|
||||
.close = drm_gem_vm_close,
|
||||
};
|
||||
|
||||
static const struct file_operations fops = {
|
||||
.owner = THIS_MODULE,
|
||||
.open = drm_open,
|
||||
.release = drm_release,
|
||||
.unlocked_ioctl = drm_ioctl,
|
||||
#ifdef CONFIG_COMPAT
|
||||
.compat_ioctl = drm_compat_ioctl,
|
||||
#endif
|
||||
.poll = drm_poll,
|
||||
.read = drm_read,
|
||||
.llseek = no_llseek,
|
||||
.mmap = etnaviv_gem_mmap,
|
||||
};
|
||||
|
||||
static struct drm_driver etnaviv_drm_driver = {
|
||||
.driver_features = DRIVER_HAVE_IRQ |
|
||||
DRIVER_GEM |
|
||||
DRIVER_PRIME |
|
||||
DRIVER_RENDER,
|
||||
.open = etnaviv_open,
|
||||
.preclose = etnaviv_preclose,
|
||||
.set_busid = drm_platform_set_busid,
|
||||
.gem_free_object = etnaviv_gem_free_object,
|
||||
.gem_vm_ops = &vm_ops,
|
||||
.prime_handle_to_fd = drm_gem_prime_handle_to_fd,
|
||||
.prime_fd_to_handle = drm_gem_prime_fd_to_handle,
|
||||
.gem_prime_export = drm_gem_prime_export,
|
||||
.gem_prime_import = drm_gem_prime_import,
|
||||
.gem_prime_pin = etnaviv_gem_prime_pin,
|
||||
.gem_prime_unpin = etnaviv_gem_prime_unpin,
|
||||
.gem_prime_get_sg_table = etnaviv_gem_prime_get_sg_table,
|
||||
.gem_prime_import_sg_table = etnaviv_gem_prime_import_sg_table,
|
||||
.gem_prime_vmap = etnaviv_gem_prime_vmap,
|
||||
.gem_prime_vunmap = etnaviv_gem_prime_vunmap,
|
||||
#ifdef CONFIG_DEBUG_FS
|
||||
.debugfs_init = etnaviv_debugfs_init,
|
||||
.debugfs_cleanup = etnaviv_debugfs_cleanup,
|
||||
#endif
|
||||
.ioctls = etnaviv_ioctls,
|
||||
.num_ioctls = DRM_ETNAVIV_NUM_IOCTLS,
|
||||
.fops = &fops,
|
||||
.name = "etnaviv",
|
||||
.desc = "etnaviv DRM",
|
||||
.date = "20151214",
|
||||
.major = 1,
|
||||
.minor = 0,
|
||||
};
|
||||
|
||||
/*
|
||||
* Platform driver:
|
||||
*/
|
||||
static int etnaviv_bind(struct device *dev)
|
||||
{
|
||||
struct etnaviv_drm_private *priv;
|
||||
struct drm_device *drm;
|
||||
int ret;
|
||||
|
||||
drm = drm_dev_alloc(&etnaviv_drm_driver, dev);
|
||||
if (!drm)
|
||||
return -ENOMEM;
|
||||
|
||||
drm->platformdev = to_platform_device(dev);
|
||||
|
||||
priv = kzalloc(sizeof(*priv), GFP_KERNEL);
|
||||
if (!priv) {
|
||||
dev_err(dev, "failed to allocate private data\n");
|
||||
ret = -ENOMEM;
|
||||
goto out_unref;
|
||||
}
|
||||
drm->dev_private = priv;
|
||||
|
||||
priv->wq = alloc_ordered_workqueue("etnaviv", 0);
|
||||
if (!priv->wq) {
|
||||
ret = -ENOMEM;
|
||||
goto out_wq;
|
||||
}
|
||||
|
||||
mutex_init(&priv->gem_lock);
|
||||
INIT_LIST_HEAD(&priv->gem_list);
|
||||
priv->num_gpus = 0;
|
||||
|
||||
dev_set_drvdata(dev, drm);
|
||||
|
||||
ret = component_bind_all(dev, drm);
|
||||
if (ret < 0)
|
||||
goto out_bind;
|
||||
|
||||
load_gpu(drm);
|
||||
|
||||
ret = drm_dev_register(drm, 0);
|
||||
if (ret)
|
||||
goto out_register;
|
||||
|
||||
return 0;
|
||||
|
||||
out_register:
|
||||
component_unbind_all(dev, drm);
|
||||
out_bind:
|
||||
flush_workqueue(priv->wq);
|
||||
destroy_workqueue(priv->wq);
|
||||
out_wq:
|
||||
kfree(priv);
|
||||
out_unref:
|
||||
drm_dev_unref(drm);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void etnaviv_unbind(struct device *dev)
|
||||
{
|
||||
struct drm_device *drm = dev_get_drvdata(dev);
|
||||
struct etnaviv_drm_private *priv = drm->dev_private;
|
||||
|
||||
drm_dev_unregister(drm);
|
||||
|
||||
flush_workqueue(priv->wq);
|
||||
destroy_workqueue(priv->wq);
|
||||
|
||||
component_unbind_all(dev, drm);
|
||||
|
||||
drm->dev_private = NULL;
|
||||
kfree(priv);
|
||||
|
||||
drm_put_dev(drm);
|
||||
}
|
||||
|
||||
static const struct component_master_ops etnaviv_master_ops = {
|
||||
.bind = etnaviv_bind,
|
||||
.unbind = etnaviv_unbind,
|
||||
};
|
||||
|
||||
static int compare_of(struct device *dev, void *data)
|
||||
{
|
||||
struct device_node *np = data;
|
||||
|
||||
return dev->of_node == np;
|
||||
}
|
||||
|
||||
static int compare_str(struct device *dev, void *data)
|
||||
{
|
||||
return !strcmp(dev_name(dev), data);
|
||||
}
|
||||
|
||||
static int etnaviv_pdev_probe(struct platform_device *pdev)
|
||||
{
|
||||
struct device *dev = &pdev->dev;
|
||||
struct device_node *node = dev->of_node;
|
||||
struct component_match *match = NULL;
|
||||
|
||||
dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
|
||||
|
||||
if (node) {
|
||||
struct device_node *core_node;
|
||||
int i;
|
||||
|
||||
for (i = 0; ; i++) {
|
||||
core_node = of_parse_phandle(node, "cores", i);
|
||||
if (!core_node)
|
||||
break;
|
||||
|
||||
component_match_add(&pdev->dev, &match, compare_of,
|
||||
core_node);
|
||||
of_node_put(core_node);
|
||||
}
|
||||
} else if (dev->platform_data) {
|
||||
char **names = dev->platform_data;
|
||||
unsigned i;
|
||||
|
||||
for (i = 0; names[i]; i++)
|
||||
component_match_add(dev, &match, compare_str, names[i]);
|
||||
}
|
||||
|
||||
return component_master_add_with_match(dev, &etnaviv_master_ops, match);
|
||||
}
|
||||
|
||||
static int etnaviv_pdev_remove(struct platform_device *pdev)
|
||||
{
|
||||
component_master_del(&pdev->dev, &etnaviv_master_ops);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static const struct of_device_id dt_match[] = {
|
||||
{ .compatible = "fsl,imx-gpu-subsystem" },
|
||||
{ .compatible = "marvell,dove-gpu-subsystem" },
|
||||
{}
|
||||
};
|
||||
MODULE_DEVICE_TABLE(of, dt_match);
|
||||
|
||||
static struct platform_driver etnaviv_platform_driver = {
|
||||
.probe = etnaviv_pdev_probe,
|
||||
.remove = etnaviv_pdev_remove,
|
||||
.driver = {
|
||||
.owner = THIS_MODULE,
|
||||
.name = "etnaviv",
|
||||
.of_match_table = dt_match,
|
||||
},
|
||||
};
|
||||
|
||||
static int __init etnaviv_init(void)
|
||||
{
|
||||
int ret;
|
||||
|
||||
etnaviv_validate_init();
|
||||
|
||||
ret = platform_driver_register(&etnaviv_gpu_driver);
|
||||
if (ret != 0)
|
||||
return ret;
|
||||
|
||||
ret = platform_driver_register(&etnaviv_platform_driver);
|
||||
if (ret != 0)
|
||||
platform_driver_unregister(&etnaviv_gpu_driver);
|
||||
|
||||
return ret;
|
||||
}
|
||||
module_init(etnaviv_init);
|
||||
|
||||
static void __exit etnaviv_exit(void)
|
||||
{
|
||||
platform_driver_unregister(&etnaviv_gpu_driver);
|
||||
platform_driver_unregister(&etnaviv_platform_driver);
|
||||
}
|
||||
module_exit(etnaviv_exit);
|
||||
|
||||
MODULE_AUTHOR("Christian Gmeiner <christian.gmeiner@gmail.com>");
|
||||
MODULE_AUTHOR("Russell King <rmk+kernel@arm.linux.org.uk>");
|
||||
MODULE_AUTHOR("Lucas Stach <l.stach@pengutronix.de>");
|
||||
MODULE_DESCRIPTION("etnaviv DRM Driver");
|
||||
MODULE_LICENSE("GPL v2");
|
||||
MODULE_ALIAS("platform:etnaviv");
|
161
drivers/gpu/drm/etnaviv/etnaviv_drv.h
Normal file
161
drivers/gpu/drm/etnaviv/etnaviv_drv.h
Normal file
|
@ -0,0 +1,161 @@
|
|||
/*
|
||||
* Copyright (C) 2015 Etnaviv Project
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#ifndef __ETNAVIV_DRV_H__
|
||||
#define __ETNAVIV_DRV_H__
|
||||
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/clk.h>
|
||||
#include <linux/cpufreq.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/platform_device.h>
|
||||
#include <linux/pm.h>
|
||||
#include <linux/pm_runtime.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/list.h>
|
||||
#include <linux/iommu.h>
|
||||
#include <linux/types.h>
|
||||
#include <linux/sizes.h>
|
||||
|
||||
#include <drm/drmP.h>
|
||||
#include <drm/drm_crtc_helper.h>
|
||||
#include <drm/drm_fb_helper.h>
|
||||
#include <drm/drm_gem.h>
|
||||
#include <drm/etnaviv_drm.h>
|
||||
|
||||
struct etnaviv_cmdbuf;
|
||||
struct etnaviv_gpu;
|
||||
struct etnaviv_mmu;
|
||||
struct etnaviv_gem_object;
|
||||
struct etnaviv_gem_submit;
|
||||
|
||||
struct etnaviv_file_private {
|
||||
/* currently we don't do anything useful with this.. but when
|
||||
* per-context address spaces are supported we'd keep track of
|
||||
* the context's page-tables here.
|
||||
*/
|
||||
int dummy;
|
||||
};
|
||||
|
||||
struct etnaviv_drm_private {
|
||||
int num_gpus;
|
||||
struct etnaviv_gpu *gpu[ETNA_MAX_PIPES];
|
||||
|
||||
/* list of GEM objects: */
|
||||
struct mutex gem_lock;
|
||||
struct list_head gem_list;
|
||||
|
||||
struct workqueue_struct *wq;
|
||||
};
|
||||
|
||||
static inline void etnaviv_queue_work(struct drm_device *dev,
|
||||
struct work_struct *w)
|
||||
{
|
||||
struct etnaviv_drm_private *priv = dev->dev_private;
|
||||
|
||||
queue_work(priv->wq, w);
|
||||
}
|
||||
|
||||
int etnaviv_ioctl_gem_submit(struct drm_device *dev, void *data,
|
||||
struct drm_file *file);
|
||||
|
||||
int etnaviv_gem_mmap(struct file *filp, struct vm_area_struct *vma);
|
||||
int etnaviv_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf);
|
||||
int etnaviv_gem_mmap_offset(struct drm_gem_object *obj, u64 *offset);
|
||||
int etnaviv_gem_get_iova(struct etnaviv_gpu *gpu,
|
||||
struct drm_gem_object *obj, u32 *iova);
|
||||
void etnaviv_gem_put_iova(struct etnaviv_gpu *gpu, struct drm_gem_object *obj);
|
||||
struct sg_table *etnaviv_gem_prime_get_sg_table(struct drm_gem_object *obj);
|
||||
void *etnaviv_gem_prime_vmap(struct drm_gem_object *obj);
|
||||
void etnaviv_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr);
|
||||
struct drm_gem_object *etnaviv_gem_prime_import_sg_table(struct drm_device *dev,
|
||||
struct dma_buf_attachment *attach, struct sg_table *sg);
|
||||
int etnaviv_gem_prime_pin(struct drm_gem_object *obj);
|
||||
void etnaviv_gem_prime_unpin(struct drm_gem_object *obj);
|
||||
void *etnaviv_gem_vaddr(struct drm_gem_object *obj);
|
||||
int etnaviv_gem_cpu_prep(struct drm_gem_object *obj, u32 op,
|
||||
struct timespec *timeout);
|
||||
int etnaviv_gem_cpu_fini(struct drm_gem_object *obj);
|
||||
void etnaviv_gem_free_object(struct drm_gem_object *obj);
|
||||
int etnaviv_gem_new_handle(struct drm_device *dev, struct drm_file *file,
|
||||
u32 size, u32 flags, u32 *handle);
|
||||
struct drm_gem_object *etnaviv_gem_new_locked(struct drm_device *dev,
|
||||
u32 size, u32 flags);
|
||||
struct drm_gem_object *etnaviv_gem_new(struct drm_device *dev,
|
||||
u32 size, u32 flags);
|
||||
int etnaviv_gem_new_userptr(struct drm_device *dev, struct drm_file *file,
|
||||
uintptr_t ptr, u32 size, u32 flags, u32 *handle);
|
||||
u16 etnaviv_buffer_init(struct etnaviv_gpu *gpu);
|
||||
void etnaviv_buffer_end(struct etnaviv_gpu *gpu);
|
||||
void etnaviv_buffer_queue(struct etnaviv_gpu *gpu, unsigned int event,
|
||||
struct etnaviv_cmdbuf *cmdbuf);
|
||||
void etnaviv_validate_init(void);
|
||||
bool etnaviv_cmd_validate_one(struct etnaviv_gpu *gpu,
|
||||
u32 *stream, unsigned int size,
|
||||
struct drm_etnaviv_gem_submit_reloc *relocs, unsigned int reloc_size);
|
||||
|
||||
#ifdef CONFIG_DEBUG_FS
|
||||
void etnaviv_gem_describe_objects(struct etnaviv_drm_private *priv,
|
||||
struct seq_file *m);
|
||||
#endif
|
||||
|
||||
void __iomem *etnaviv_ioremap(struct platform_device *pdev, const char *name,
|
||||
const char *dbgname);
|
||||
void etnaviv_writel(u32 data, void __iomem *addr);
|
||||
u32 etnaviv_readl(const void __iomem *addr);
|
||||
|
||||
#define DBG(fmt, ...) DRM_DEBUG(fmt"\n", ##__VA_ARGS__)
|
||||
#define VERB(fmt, ...) if (0) DRM_DEBUG(fmt"\n", ##__VA_ARGS__)
|
||||
|
||||
/*
|
||||
* Return the storage size of a structure with a variable length array.
|
||||
* The array is nelem elements of elem_size, where the base structure
|
||||
* is defined by base. If the size overflows size_t, return zero.
|
||||
*/
|
||||
static inline size_t size_vstruct(size_t nelem, size_t elem_size, size_t base)
|
||||
{
|
||||
if (elem_size && nelem > (SIZE_MAX - base) / elem_size)
|
||||
return 0;
|
||||
return base + nelem * elem_size;
|
||||
}
|
||||
|
||||
/* returns true if fence a comes after fence b */
|
||||
static inline bool fence_after(u32 a, u32 b)
|
||||
{
|
||||
return (s32)(a - b) > 0;
|
||||
}
|
||||
|
||||
static inline bool fence_after_eq(u32 a, u32 b)
|
||||
{
|
||||
return (s32)(a - b) >= 0;
|
||||
}
|
||||
|
||||
static inline unsigned long etnaviv_timeout_to_jiffies(
|
||||
const struct timespec *timeout)
|
||||
{
|
||||
unsigned long timeout_jiffies = timespec_to_jiffies(timeout);
|
||||
unsigned long start_jiffies = jiffies;
|
||||
unsigned long remaining_jiffies;
|
||||
|
||||
if (time_after(start_jiffies, timeout_jiffies))
|
||||
remaining_jiffies = 0;
|
||||
else
|
||||
remaining_jiffies = timeout_jiffies - start_jiffies;
|
||||
|
||||
return remaining_jiffies;
|
||||
}
|
||||
|
||||
#endif /* __ETNAVIV_DRV_H__ */
|
227
drivers/gpu/drm/etnaviv/etnaviv_dump.c
Normal file
227
drivers/gpu/drm/etnaviv/etnaviv_dump.c
Normal file
|
@ -0,0 +1,227 @@
|
|||
/*
|
||||
* Copyright (C) 2015 Etnaviv Project
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#include <linux/devcoredump.h>
|
||||
#include "etnaviv_dump.h"
|
||||
#include "etnaviv_gem.h"
|
||||
#include "etnaviv_gpu.h"
|
||||
#include "etnaviv_mmu.h"
|
||||
#include "state.xml.h"
|
||||
#include "state_hi.xml.h"
|
||||
|
||||
struct core_dump_iterator {
|
||||
void *start;
|
||||
struct etnaviv_dump_object_header *hdr;
|
||||
void *data;
|
||||
};
|
||||
|
||||
static const unsigned short etnaviv_dump_registers[] = {
|
||||
VIVS_HI_AXI_STATUS,
|
||||
VIVS_HI_CLOCK_CONTROL,
|
||||
VIVS_HI_IDLE_STATE,
|
||||
VIVS_HI_AXI_CONFIG,
|
||||
VIVS_HI_INTR_ENBL,
|
||||
VIVS_HI_CHIP_IDENTITY,
|
||||
VIVS_HI_CHIP_FEATURE,
|
||||
VIVS_HI_CHIP_MODEL,
|
||||
VIVS_HI_CHIP_REV,
|
||||
VIVS_HI_CHIP_DATE,
|
||||
VIVS_HI_CHIP_TIME,
|
||||
VIVS_HI_CHIP_MINOR_FEATURE_0,
|
||||
VIVS_HI_CACHE_CONTROL,
|
||||
VIVS_HI_AXI_CONTROL,
|
||||
VIVS_PM_POWER_CONTROLS,
|
||||
VIVS_PM_MODULE_CONTROLS,
|
||||
VIVS_PM_MODULE_STATUS,
|
||||
VIVS_PM_PULSE_EATER,
|
||||
VIVS_MC_MMU_FE_PAGE_TABLE,
|
||||
VIVS_MC_MMU_TX_PAGE_TABLE,
|
||||
VIVS_MC_MMU_PE_PAGE_TABLE,
|
||||
VIVS_MC_MMU_PEZ_PAGE_TABLE,
|
||||
VIVS_MC_MMU_RA_PAGE_TABLE,
|
||||
VIVS_MC_DEBUG_MEMORY,
|
||||
VIVS_MC_MEMORY_BASE_ADDR_RA,
|
||||
VIVS_MC_MEMORY_BASE_ADDR_FE,
|
||||
VIVS_MC_MEMORY_BASE_ADDR_TX,
|
||||
VIVS_MC_MEMORY_BASE_ADDR_PEZ,
|
||||
VIVS_MC_MEMORY_BASE_ADDR_PE,
|
||||
VIVS_MC_MEMORY_TIMING_CONTROL,
|
||||
VIVS_MC_BUS_CONFIG,
|
||||
VIVS_FE_DMA_STATUS,
|
||||
VIVS_FE_DMA_DEBUG_STATE,
|
||||
VIVS_FE_DMA_ADDRESS,
|
||||
VIVS_FE_DMA_LOW,
|
||||
VIVS_FE_DMA_HIGH,
|
||||
VIVS_FE_AUTO_FLUSH,
|
||||
};
|
||||
|
||||
static void etnaviv_core_dump_header(struct core_dump_iterator *iter,
|
||||
u32 type, void *data_end)
|
||||
{
|
||||
struct etnaviv_dump_object_header *hdr = iter->hdr;
|
||||
|
||||
hdr->magic = cpu_to_le32(ETDUMP_MAGIC);
|
||||
hdr->type = cpu_to_le32(type);
|
||||
hdr->file_offset = cpu_to_le32(iter->data - iter->start);
|
||||
hdr->file_size = cpu_to_le32(data_end - iter->data);
|
||||
|
||||
iter->hdr++;
|
||||
iter->data += hdr->file_size;
|
||||
}
|
||||
|
||||
static void etnaviv_core_dump_registers(struct core_dump_iterator *iter,
|
||||
struct etnaviv_gpu *gpu)
|
||||
{
|
||||
struct etnaviv_dump_registers *reg = iter->data;
|
||||
unsigned int i;
|
||||
|
||||
for (i = 0; i < ARRAY_SIZE(etnaviv_dump_registers); i++, reg++) {
|
||||
reg->reg = etnaviv_dump_registers[i];
|
||||
reg->value = gpu_read(gpu, etnaviv_dump_registers[i]);
|
||||
}
|
||||
|
||||
etnaviv_core_dump_header(iter, ETDUMP_BUF_REG, reg);
|
||||
}
|
||||
|
||||
static void etnaviv_core_dump_mmu(struct core_dump_iterator *iter,
|
||||
struct etnaviv_gpu *gpu, size_t mmu_size)
|
||||
{
|
||||
etnaviv_iommu_dump(gpu->mmu, iter->data);
|
||||
|
||||
etnaviv_core_dump_header(iter, ETDUMP_BUF_MMU, iter->data + mmu_size);
|
||||
}
|
||||
|
||||
static void etnaviv_core_dump_mem(struct core_dump_iterator *iter, u32 type,
|
||||
void *ptr, size_t size, u64 iova)
|
||||
{
|
||||
memcpy(iter->data, ptr, size);
|
||||
|
||||
iter->hdr->iova = cpu_to_le64(iova);
|
||||
|
||||
etnaviv_core_dump_header(iter, type, iter->data + size);
|
||||
}
|
||||
|
||||
void etnaviv_core_dump(struct etnaviv_gpu *gpu)
|
||||
{
|
||||
struct core_dump_iterator iter;
|
||||
struct etnaviv_vram_mapping *vram;
|
||||
struct etnaviv_gem_object *obj;
|
||||
struct etnaviv_cmdbuf *cmd;
|
||||
unsigned int n_obj, n_bomap_pages;
|
||||
size_t file_size, mmu_size;
|
||||
__le64 *bomap, *bomap_start;
|
||||
|
||||
mmu_size = etnaviv_iommu_dump_size(gpu->mmu);
|
||||
|
||||
/* We always dump registers, mmu, ring and end marker */
|
||||
n_obj = 4;
|
||||
n_bomap_pages = 0;
|
||||
file_size = ARRAY_SIZE(etnaviv_dump_registers) *
|
||||
sizeof(struct etnaviv_dump_registers) +
|
||||
mmu_size + gpu->buffer->size;
|
||||
|
||||
/* Add in the active command buffers */
|
||||
list_for_each_entry(cmd, &gpu->active_cmd_list, node) {
|
||||
file_size += cmd->size;
|
||||
n_obj++;
|
||||
}
|
||||
|
||||
/* Add in the active buffer objects */
|
||||
list_for_each_entry(vram, &gpu->mmu->mappings, mmu_node) {
|
||||
if (!vram->use)
|
||||
continue;
|
||||
|
||||
obj = vram->object;
|
||||
file_size += obj->base.size;
|
||||
n_bomap_pages += obj->base.size >> PAGE_SHIFT;
|
||||
n_obj++;
|
||||
}
|
||||
|
||||
/* If we have any buffer objects, add a bomap object */
|
||||
if (n_bomap_pages) {
|
||||
file_size += n_bomap_pages * sizeof(__le64);
|
||||
n_obj++;
|
||||
}
|
||||
|
||||
/* Add the size of the headers */
|
||||
file_size += sizeof(*iter.hdr) * n_obj;
|
||||
|
||||
/* Allocate the file in vmalloc memory, it's likely to be big */
|
||||
iter.start = vmalloc(file_size);
|
||||
if (!iter.start) {
|
||||
dev_warn(gpu->dev, "failed to allocate devcoredump file\n");
|
||||
return;
|
||||
}
|
||||
|
||||
/* Point the data member after the headers */
|
||||
iter.hdr = iter.start;
|
||||
iter.data = &iter.hdr[n_obj];
|
||||
|
||||
memset(iter.hdr, 0, iter.data - iter.start);
|
||||
|
||||
etnaviv_core_dump_registers(&iter, gpu);
|
||||
etnaviv_core_dump_mmu(&iter, gpu, mmu_size);
|
||||
etnaviv_core_dump_mem(&iter, ETDUMP_BUF_RING, gpu->buffer->vaddr,
|
||||
gpu->buffer->size, gpu->buffer->paddr);
|
||||
|
||||
list_for_each_entry(cmd, &gpu->active_cmd_list, node)
|
||||
etnaviv_core_dump_mem(&iter, ETDUMP_BUF_CMD, cmd->vaddr,
|
||||
cmd->size, cmd->paddr);
|
||||
|
||||
/* Reserve space for the bomap */
|
||||
if (n_bomap_pages) {
|
||||
bomap_start = bomap = iter.data;
|
||||
memset(bomap, 0, sizeof(*bomap) * n_bomap_pages);
|
||||
etnaviv_core_dump_header(&iter, ETDUMP_BUF_BOMAP,
|
||||
bomap + n_bomap_pages);
|
||||
} else {
|
||||
/* Silence warning */
|
||||
bomap_start = bomap = NULL;
|
||||
}
|
||||
|
||||
list_for_each_entry(vram, &gpu->mmu->mappings, mmu_node) {
|
||||
struct page **pages;
|
||||
void *vaddr;
|
||||
|
||||
if (vram->use == 0)
|
||||
continue;
|
||||
|
||||
obj = vram->object;
|
||||
|
||||
pages = etnaviv_gem_get_pages(obj);
|
||||
if (pages) {
|
||||
int j;
|
||||
|
||||
iter.hdr->data[0] = bomap - bomap_start;
|
||||
|
||||
for (j = 0; j < obj->base.size >> PAGE_SHIFT; j++)
|
||||
*bomap++ = cpu_to_le64(page_to_phys(*pages++));
|
||||
}
|
||||
|
||||
iter.hdr->iova = cpu_to_le64(vram->iova);
|
||||
|
||||
vaddr = etnaviv_gem_vaddr(&obj->base);
|
||||
if (vaddr && !IS_ERR(vaddr))
|
||||
memcpy(iter.data, vaddr, obj->base.size);
|
||||
|
||||
etnaviv_core_dump_header(&iter, ETDUMP_BUF_BO, iter.data +
|
||||
obj->base.size);
|
||||
}
|
||||
|
||||
etnaviv_core_dump_header(&iter, ETDUMP_BUF_END, iter.data);
|
||||
|
||||
dev_coredumpv(gpu->dev, iter.start, iter.data - iter.start, GFP_KERNEL);
|
||||
}
|
54
drivers/gpu/drm/etnaviv/etnaviv_dump.h
Normal file
54
drivers/gpu/drm/etnaviv/etnaviv_dump.h
Normal file
|
@ -0,0 +1,54 @@
|
|||
/*
|
||||
* Copyright (C) 2015 Etnaviv Project
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*
|
||||
* Etnaviv devcoredump file definitions
|
||||
*/
|
||||
#ifndef ETNAVIV_DUMP_H
|
||||
#define ETNAVIV_DUMP_H
|
||||
|
||||
#include <linux/types.h>
|
||||
|
||||
enum {
|
||||
ETDUMP_MAGIC = 0x414e5445,
|
||||
ETDUMP_BUF_REG = 0,
|
||||
ETDUMP_BUF_MMU,
|
||||
ETDUMP_BUF_RING,
|
||||
ETDUMP_BUF_CMD,
|
||||
ETDUMP_BUF_BOMAP,
|
||||
ETDUMP_BUF_BO,
|
||||
ETDUMP_BUF_END,
|
||||
};
|
||||
|
||||
struct etnaviv_dump_object_header {
|
||||
__le32 magic;
|
||||
__le32 type;
|
||||
__le32 file_offset;
|
||||
__le32 file_size;
|
||||
__le64 iova;
|
||||
__le32 data[2];
|
||||
};
|
||||
|
||||
/* Registers object, an array of these */
|
||||
struct etnaviv_dump_registers {
|
||||
__le32 reg;
|
||||
__le32 value;
|
||||
};
|
||||
|
||||
#ifdef __KERNEL__
|
||||
struct etnaviv_gpu;
|
||||
void etnaviv_core_dump(struct etnaviv_gpu *gpu);
|
||||
#endif
|
||||
|
||||
#endif
|
897
drivers/gpu/drm/etnaviv/etnaviv_gem.c
Normal file
897
drivers/gpu/drm/etnaviv/etnaviv_gem.c
Normal file
|
@ -0,0 +1,897 @@
|
|||
/*
|
||||
* Copyright (C) 2015 Etnaviv Project
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#include <linux/spinlock.h>
|
||||
#include <linux/shmem_fs.h>
|
||||
|
||||
#include "etnaviv_drv.h"
|
||||
#include "etnaviv_gem.h"
|
||||
#include "etnaviv_gpu.h"
|
||||
#include "etnaviv_mmu.h"
|
||||
|
||||
static void etnaviv_gem_scatter_map(struct etnaviv_gem_object *etnaviv_obj)
|
||||
{
|
||||
struct drm_device *dev = etnaviv_obj->base.dev;
|
||||
struct sg_table *sgt = etnaviv_obj->sgt;
|
||||
|
||||
/*
|
||||
* For non-cached buffers, ensure the new pages are clean
|
||||
* because display controller, GPU, etc. are not coherent.
|
||||
*/
|
||||
if (etnaviv_obj->flags & ETNA_BO_CACHE_MASK)
|
||||
dma_map_sg(dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL);
|
||||
}
|
||||
|
||||
static void etnaviv_gem_scatterlist_unmap(struct etnaviv_gem_object *etnaviv_obj)
|
||||
{
|
||||
struct drm_device *dev = etnaviv_obj->base.dev;
|
||||
struct sg_table *sgt = etnaviv_obj->sgt;
|
||||
|
||||
/*
|
||||
* For non-cached buffers, ensure the new pages are clean
|
||||
* because display controller, GPU, etc. are not coherent:
|
||||
*
|
||||
* WARNING: The DMA API does not support concurrent CPU
|
||||
* and device access to the memory area. With BIDIRECTIONAL,
|
||||
* we will clean the cache lines which overlap the region,
|
||||
* and invalidate all cache lines (partially) contained in
|
||||
* the region.
|
||||
*
|
||||
* If you have dirty data in the overlapping cache lines,
|
||||
* that will corrupt the GPU-written data. If you have
|
||||
* written into the remainder of the region, this can
|
||||
* discard those writes.
|
||||
*/
|
||||
if (etnaviv_obj->flags & ETNA_BO_CACHE_MASK)
|
||||
dma_unmap_sg(dev->dev, sgt->sgl, sgt->nents, DMA_BIDIRECTIONAL);
|
||||
}
|
||||
|
||||
/* called with etnaviv_obj->lock held */
|
||||
static int etnaviv_gem_shmem_get_pages(struct etnaviv_gem_object *etnaviv_obj)
|
||||
{
|
||||
struct drm_device *dev = etnaviv_obj->base.dev;
|
||||
struct page **p = drm_gem_get_pages(&etnaviv_obj->base);
|
||||
|
||||
if (IS_ERR(p)) {
|
||||
dev_err(dev->dev, "could not get pages: %ld\n", PTR_ERR(p));
|
||||
return PTR_ERR(p);
|
||||
}
|
||||
|
||||
etnaviv_obj->pages = p;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void put_pages(struct etnaviv_gem_object *etnaviv_obj)
|
||||
{
|
||||
if (etnaviv_obj->sgt) {
|
||||
etnaviv_gem_scatterlist_unmap(etnaviv_obj);
|
||||
sg_free_table(etnaviv_obj->sgt);
|
||||
kfree(etnaviv_obj->sgt);
|
||||
etnaviv_obj->sgt = NULL;
|
||||
}
|
||||
if (etnaviv_obj->pages) {
|
||||
drm_gem_put_pages(&etnaviv_obj->base, etnaviv_obj->pages,
|
||||
true, false);
|
||||
|
||||
etnaviv_obj->pages = NULL;
|
||||
}
|
||||
}
|
||||
|
||||
struct page **etnaviv_gem_get_pages(struct etnaviv_gem_object *etnaviv_obj)
|
||||
{
|
||||
int ret;
|
||||
|
||||
lockdep_assert_held(&etnaviv_obj->lock);
|
||||
|
||||
if (!etnaviv_obj->pages) {
|
||||
ret = etnaviv_obj->ops->get_pages(etnaviv_obj);
|
||||
if (ret < 0)
|
||||
return ERR_PTR(ret);
|
||||
}
|
||||
|
||||
if (!etnaviv_obj->sgt) {
|
||||
struct drm_device *dev = etnaviv_obj->base.dev;
|
||||
int npages = etnaviv_obj->base.size >> PAGE_SHIFT;
|
||||
struct sg_table *sgt;
|
||||
|
||||
sgt = drm_prime_pages_to_sg(etnaviv_obj->pages, npages);
|
||||
if (IS_ERR(sgt)) {
|
||||
dev_err(dev->dev, "failed to allocate sgt: %ld\n",
|
||||
PTR_ERR(sgt));
|
||||
return ERR_CAST(sgt);
|
||||
}
|
||||
|
||||
etnaviv_obj->sgt = sgt;
|
||||
|
||||
etnaviv_gem_scatter_map(etnaviv_obj);
|
||||
}
|
||||
|
||||
return etnaviv_obj->pages;
|
||||
}
|
||||
|
||||
void etnaviv_gem_put_pages(struct etnaviv_gem_object *etnaviv_obj)
|
||||
{
|
||||
lockdep_assert_held(&etnaviv_obj->lock);
|
||||
/* when we start tracking the pin count, then do something here */
|
||||
}
|
||||
|
||||
static int etnaviv_gem_mmap_obj(struct drm_gem_object *obj,
|
||||
struct vm_area_struct *vma)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
pgprot_t vm_page_prot;
|
||||
|
||||
vma->vm_flags &= ~VM_PFNMAP;
|
||||
vma->vm_flags |= VM_MIXEDMAP;
|
||||
|
||||
vm_page_prot = vm_get_page_prot(vma->vm_flags);
|
||||
|
||||
if (etnaviv_obj->flags & ETNA_BO_WC) {
|
||||
vma->vm_page_prot = pgprot_writecombine(vm_page_prot);
|
||||
} else if (etnaviv_obj->flags & ETNA_BO_UNCACHED) {
|
||||
vma->vm_page_prot = pgprot_noncached(vm_page_prot);
|
||||
} else {
|
||||
/*
|
||||
* Shunt off cached objs to shmem file so they have their own
|
||||
* address_space (so unmap_mapping_range does what we want,
|
||||
* in particular in the case of mmap'd dmabufs)
|
||||
*/
|
||||
fput(vma->vm_file);
|
||||
get_file(obj->filp);
|
||||
vma->vm_pgoff = 0;
|
||||
vma->vm_file = obj->filp;
|
||||
|
||||
vma->vm_page_prot = vm_page_prot;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
int etnaviv_gem_mmap(struct file *filp, struct vm_area_struct *vma)
|
||||
{
|
||||
struct etnaviv_gem_object *obj;
|
||||
int ret;
|
||||
|
||||
ret = drm_gem_mmap(filp, vma);
|
||||
if (ret) {
|
||||
DBG("mmap failed: %d", ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
obj = to_etnaviv_bo(vma->vm_private_data);
|
||||
return etnaviv_gem_mmap_obj(vma->vm_private_data, vma);
|
||||
}
|
||||
|
||||
int etnaviv_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
|
||||
{
|
||||
struct drm_gem_object *obj = vma->vm_private_data;
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
struct page **pages, *page;
|
||||
pgoff_t pgoff;
|
||||
int ret;
|
||||
|
||||
/*
|
||||
* Make sure we don't parallel update on a fault, nor move or remove
|
||||
* something from beneath our feet. Note that vm_insert_page() is
|
||||
* specifically coded to take care of this, so we don't have to.
|
||||
*/
|
||||
ret = mutex_lock_interruptible(&etnaviv_obj->lock);
|
||||
if (ret)
|
||||
goto out;
|
||||
|
||||
/* make sure we have pages attached now */
|
||||
pages = etnaviv_gem_get_pages(etnaviv_obj);
|
||||
mutex_unlock(&etnaviv_obj->lock);
|
||||
|
||||
if (IS_ERR(pages)) {
|
||||
ret = PTR_ERR(pages);
|
||||
goto out;
|
||||
}
|
||||
|
||||
/* We don't use vmf->pgoff since that has the fake offset: */
|
||||
pgoff = ((unsigned long)vmf->virtual_address -
|
||||
vma->vm_start) >> PAGE_SHIFT;
|
||||
|
||||
page = pages[pgoff];
|
||||
|
||||
VERB("Inserting %p pfn %lx, pa %lx", vmf->virtual_address,
|
||||
page_to_pfn(page), page_to_pfn(page) << PAGE_SHIFT);
|
||||
|
||||
ret = vm_insert_page(vma, (unsigned long)vmf->virtual_address, page);
|
||||
|
||||
out:
|
||||
switch (ret) {
|
||||
case -EAGAIN:
|
||||
case 0:
|
||||
case -ERESTARTSYS:
|
||||
case -EINTR:
|
||||
case -EBUSY:
|
||||
/*
|
||||
* EBUSY is ok: this just means that another thread
|
||||
* already did the job.
|
||||
*/
|
||||
return VM_FAULT_NOPAGE;
|
||||
case -ENOMEM:
|
||||
return VM_FAULT_OOM;
|
||||
default:
|
||||
return VM_FAULT_SIGBUS;
|
||||
}
|
||||
}
|
||||
|
||||
int etnaviv_gem_mmap_offset(struct drm_gem_object *obj, u64 *offset)
|
||||
{
|
||||
int ret;
|
||||
|
||||
/* Make it mmapable */
|
||||
ret = drm_gem_create_mmap_offset(obj);
|
||||
if (ret)
|
||||
dev_err(obj->dev->dev, "could not allocate mmap offset\n");
|
||||
else
|
||||
*offset = drm_vma_node_offset_addr(&obj->vma_node);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static struct etnaviv_vram_mapping *
|
||||
etnaviv_gem_get_vram_mapping(struct etnaviv_gem_object *obj,
|
||||
struct etnaviv_iommu *mmu)
|
||||
{
|
||||
struct etnaviv_vram_mapping *mapping;
|
||||
|
||||
list_for_each_entry(mapping, &obj->vram_list, obj_node) {
|
||||
if (mapping->mmu == mmu)
|
||||
return mapping;
|
||||
}
|
||||
|
||||
return NULL;
|
||||
}
|
||||
|
||||
int etnaviv_gem_get_iova(struct etnaviv_gpu *gpu,
|
||||
struct drm_gem_object *obj, u32 *iova)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
struct etnaviv_vram_mapping *mapping;
|
||||
struct page **pages;
|
||||
int ret = 0;
|
||||
|
||||
mutex_lock(&etnaviv_obj->lock);
|
||||
mapping = etnaviv_gem_get_vram_mapping(etnaviv_obj, gpu->mmu);
|
||||
if (mapping) {
|
||||
/*
|
||||
* Holding the object lock prevents the use count changing
|
||||
* beneath us. If the use count is zero, the MMU might be
|
||||
* reaping this object, so take the lock and re-check that
|
||||
* the MMU owns this mapping to close this race.
|
||||
*/
|
||||
if (mapping->use == 0) {
|
||||
mutex_lock(&gpu->mmu->lock);
|
||||
if (mapping->mmu == gpu->mmu)
|
||||
mapping->use += 1;
|
||||
else
|
||||
mapping = NULL;
|
||||
mutex_unlock(&gpu->mmu->lock);
|
||||
if (mapping)
|
||||
goto out;
|
||||
} else {
|
||||
mapping->use += 1;
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
|
||||
pages = etnaviv_gem_get_pages(etnaviv_obj);
|
||||
if (IS_ERR(pages)) {
|
||||
ret = PTR_ERR(pages);
|
||||
goto out;
|
||||
}
|
||||
|
||||
/*
|
||||
* See if we have a reaped vram mapping we can re-use before
|
||||
* allocating a fresh mapping.
|
||||
*/
|
||||
mapping = etnaviv_gem_get_vram_mapping(etnaviv_obj, NULL);
|
||||
if (!mapping) {
|
||||
mapping = kzalloc(sizeof(*mapping), GFP_KERNEL);
|
||||
if (!mapping)
|
||||
return -ENOMEM;
|
||||
|
||||
INIT_LIST_HEAD(&mapping->scan_node);
|
||||
mapping->object = etnaviv_obj;
|
||||
} else {
|
||||
list_del(&mapping->obj_node);
|
||||
}
|
||||
|
||||
mapping->mmu = gpu->mmu;
|
||||
mapping->use = 1;
|
||||
|
||||
ret = etnaviv_iommu_map_gem(gpu->mmu, etnaviv_obj, gpu->memory_base,
|
||||
mapping);
|
||||
if (ret < 0)
|
||||
kfree(mapping);
|
||||
else
|
||||
list_add_tail(&mapping->obj_node, &etnaviv_obj->vram_list);
|
||||
|
||||
out:
|
||||
mutex_unlock(&etnaviv_obj->lock);
|
||||
|
||||
if (!ret) {
|
||||
/* Take a reference on the object */
|
||||
drm_gem_object_reference(obj);
|
||||
*iova = mapping->iova;
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
void etnaviv_gem_put_iova(struct etnaviv_gpu *gpu, struct drm_gem_object *obj)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
struct etnaviv_vram_mapping *mapping;
|
||||
|
||||
mutex_lock(&etnaviv_obj->lock);
|
||||
mapping = etnaviv_gem_get_vram_mapping(etnaviv_obj, gpu->mmu);
|
||||
|
||||
WARN_ON(mapping->use == 0);
|
||||
mapping->use -= 1;
|
||||
mutex_unlock(&etnaviv_obj->lock);
|
||||
|
||||
drm_gem_object_unreference_unlocked(obj);
|
||||
}
|
||||
|
||||
void *etnaviv_gem_vaddr(struct drm_gem_object *obj)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
|
||||
mutex_lock(&etnaviv_obj->lock);
|
||||
if (!etnaviv_obj->vaddr) {
|
||||
struct page **pages = etnaviv_gem_get_pages(etnaviv_obj);
|
||||
|
||||
if (IS_ERR(pages))
|
||||
return ERR_CAST(pages);
|
||||
|
||||
etnaviv_obj->vaddr = vmap(pages, obj->size >> PAGE_SHIFT,
|
||||
VM_MAP, pgprot_writecombine(PAGE_KERNEL));
|
||||
}
|
||||
mutex_unlock(&etnaviv_obj->lock);
|
||||
|
||||
return etnaviv_obj->vaddr;
|
||||
}
|
||||
|
||||
static inline enum dma_data_direction etnaviv_op_to_dma_dir(u32 op)
|
||||
{
|
||||
if (op & ETNA_PREP_READ)
|
||||
return DMA_FROM_DEVICE;
|
||||
else if (op & ETNA_PREP_WRITE)
|
||||
return DMA_TO_DEVICE;
|
||||
else
|
||||
return DMA_BIDIRECTIONAL;
|
||||
}
|
||||
|
||||
int etnaviv_gem_cpu_prep(struct drm_gem_object *obj, u32 op,
|
||||
struct timespec *timeout)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
struct drm_device *dev = obj->dev;
|
||||
bool write = !!(op & ETNA_PREP_WRITE);
|
||||
int ret;
|
||||
|
||||
if (op & ETNA_PREP_NOSYNC) {
|
||||
if (!reservation_object_test_signaled_rcu(etnaviv_obj->resv,
|
||||
write))
|
||||
return -EBUSY;
|
||||
} else {
|
||||
unsigned long remain = etnaviv_timeout_to_jiffies(timeout);
|
||||
|
||||
ret = reservation_object_wait_timeout_rcu(etnaviv_obj->resv,
|
||||
write, true, remain);
|
||||
if (ret <= 0)
|
||||
return ret == 0 ? -ETIMEDOUT : ret;
|
||||
}
|
||||
|
||||
if (etnaviv_obj->flags & ETNA_BO_CACHED) {
|
||||
if (!etnaviv_obj->sgt) {
|
||||
void *ret;
|
||||
|
||||
mutex_lock(&etnaviv_obj->lock);
|
||||
ret = etnaviv_gem_get_pages(etnaviv_obj);
|
||||
mutex_unlock(&etnaviv_obj->lock);
|
||||
if (IS_ERR(ret))
|
||||
return PTR_ERR(ret);
|
||||
}
|
||||
|
||||
dma_sync_sg_for_cpu(dev->dev, etnaviv_obj->sgt->sgl,
|
||||
etnaviv_obj->sgt->nents,
|
||||
etnaviv_op_to_dma_dir(op));
|
||||
etnaviv_obj->last_cpu_prep_op = op;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
int etnaviv_gem_cpu_fini(struct drm_gem_object *obj)
|
||||
{
|
||||
struct drm_device *dev = obj->dev;
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
|
||||
if (etnaviv_obj->flags & ETNA_BO_CACHED) {
|
||||
/* fini without a prep is almost certainly a userspace error */
|
||||
WARN_ON(etnaviv_obj->last_cpu_prep_op == 0);
|
||||
dma_sync_sg_for_device(dev->dev, etnaviv_obj->sgt->sgl,
|
||||
etnaviv_obj->sgt->nents,
|
||||
etnaviv_op_to_dma_dir(etnaviv_obj->last_cpu_prep_op));
|
||||
etnaviv_obj->last_cpu_prep_op = 0;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
int etnaviv_gem_wait_bo(struct etnaviv_gpu *gpu, struct drm_gem_object *obj,
|
||||
struct timespec *timeout)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
|
||||
return etnaviv_gpu_wait_obj_inactive(gpu, etnaviv_obj, timeout);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_DEBUG_FS
|
||||
static void etnaviv_gem_describe_fence(struct fence *fence,
|
||||
const char *type, struct seq_file *m)
|
||||
{
|
||||
if (!test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags))
|
||||
seq_printf(m, "\t%9s: %s %s seq %u\n",
|
||||
type,
|
||||
fence->ops->get_driver_name(fence),
|
||||
fence->ops->get_timeline_name(fence),
|
||||
fence->seqno);
|
||||
}
|
||||
|
||||
static void etnaviv_gem_describe(struct drm_gem_object *obj, struct seq_file *m)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
struct reservation_object *robj = etnaviv_obj->resv;
|
||||
struct reservation_object_list *fobj;
|
||||
struct fence *fence;
|
||||
unsigned long off = drm_vma_node_start(&obj->vma_node);
|
||||
|
||||
seq_printf(m, "%08x: %c %2d (%2d) %08lx %p %zd\n",
|
||||
etnaviv_obj->flags, is_active(etnaviv_obj) ? 'A' : 'I',
|
||||
obj->name, obj->refcount.refcount.counter,
|
||||
off, etnaviv_obj->vaddr, obj->size);
|
||||
|
||||
rcu_read_lock();
|
||||
fobj = rcu_dereference(robj->fence);
|
||||
if (fobj) {
|
||||
unsigned int i, shared_count = fobj->shared_count;
|
||||
|
||||
for (i = 0; i < shared_count; i++) {
|
||||
fence = rcu_dereference(fobj->shared[i]);
|
||||
etnaviv_gem_describe_fence(fence, "Shared", m);
|
||||
}
|
||||
}
|
||||
|
||||
fence = rcu_dereference(robj->fence_excl);
|
||||
if (fence)
|
||||
etnaviv_gem_describe_fence(fence, "Exclusive", m);
|
||||
rcu_read_unlock();
|
||||
}
|
||||
|
||||
void etnaviv_gem_describe_objects(struct etnaviv_drm_private *priv,
|
||||
struct seq_file *m)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj;
|
||||
int count = 0;
|
||||
size_t size = 0;
|
||||
|
||||
mutex_lock(&priv->gem_lock);
|
||||
list_for_each_entry(etnaviv_obj, &priv->gem_list, gem_node) {
|
||||
struct drm_gem_object *obj = &etnaviv_obj->base;
|
||||
|
||||
seq_puts(m, " ");
|
||||
etnaviv_gem_describe(obj, m);
|
||||
count++;
|
||||
size += obj->size;
|
||||
}
|
||||
mutex_unlock(&priv->gem_lock);
|
||||
|
||||
seq_printf(m, "Total %d objects, %zu bytes\n", count, size);
|
||||
}
|
||||
#endif
|
||||
|
||||
static void etnaviv_gem_shmem_release(struct etnaviv_gem_object *etnaviv_obj)
|
||||
{
|
||||
if (etnaviv_obj->vaddr)
|
||||
vunmap(etnaviv_obj->vaddr);
|
||||
put_pages(etnaviv_obj);
|
||||
}
|
||||
|
||||
static const struct etnaviv_gem_ops etnaviv_gem_shmem_ops = {
|
||||
.get_pages = etnaviv_gem_shmem_get_pages,
|
||||
.release = etnaviv_gem_shmem_release,
|
||||
};
|
||||
|
||||
void etnaviv_gem_free_object(struct drm_gem_object *obj)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
struct etnaviv_vram_mapping *mapping, *tmp;
|
||||
|
||||
/* object should not be active */
|
||||
WARN_ON(is_active(etnaviv_obj));
|
||||
|
||||
list_del(&etnaviv_obj->gem_node);
|
||||
|
||||
list_for_each_entry_safe(mapping, tmp, &etnaviv_obj->vram_list,
|
||||
obj_node) {
|
||||
struct etnaviv_iommu *mmu = mapping->mmu;
|
||||
|
||||
WARN_ON(mapping->use);
|
||||
|
||||
if (mmu)
|
||||
etnaviv_iommu_unmap_gem(mmu, mapping);
|
||||
|
||||
list_del(&mapping->obj_node);
|
||||
kfree(mapping);
|
||||
}
|
||||
|
||||
drm_gem_free_mmap_offset(obj);
|
||||
etnaviv_obj->ops->release(etnaviv_obj);
|
||||
if (etnaviv_obj->resv == &etnaviv_obj->_resv)
|
||||
reservation_object_fini(&etnaviv_obj->_resv);
|
||||
drm_gem_object_release(obj);
|
||||
|
||||
kfree(etnaviv_obj);
|
||||
}
|
||||
|
||||
int etnaviv_gem_obj_add(struct drm_device *dev, struct drm_gem_object *obj)
|
||||
{
|
||||
struct etnaviv_drm_private *priv = dev->dev_private;
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
|
||||
mutex_lock(&priv->gem_lock);
|
||||
list_add_tail(&etnaviv_obj->gem_node, &priv->gem_list);
|
||||
mutex_unlock(&priv->gem_lock);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int etnaviv_gem_new_impl(struct drm_device *dev, u32 size, u32 flags,
|
||||
struct reservation_object *robj, const struct etnaviv_gem_ops *ops,
|
||||
struct drm_gem_object **obj)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj;
|
||||
unsigned sz = sizeof(*etnaviv_obj);
|
||||
bool valid = true;
|
||||
|
||||
/* validate flags */
|
||||
switch (flags & ETNA_BO_CACHE_MASK) {
|
||||
case ETNA_BO_UNCACHED:
|
||||
case ETNA_BO_CACHED:
|
||||
case ETNA_BO_WC:
|
||||
break;
|
||||
default:
|
||||
valid = false;
|
||||
}
|
||||
|
||||
if (!valid) {
|
||||
dev_err(dev->dev, "invalid cache flag: %x\n",
|
||||
(flags & ETNA_BO_CACHE_MASK));
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
etnaviv_obj = kzalloc(sz, GFP_KERNEL);
|
||||
if (!etnaviv_obj)
|
||||
return -ENOMEM;
|
||||
|
||||
etnaviv_obj->flags = flags;
|
||||
etnaviv_obj->ops = ops;
|
||||
if (robj) {
|
||||
etnaviv_obj->resv = robj;
|
||||
} else {
|
||||
etnaviv_obj->resv = &etnaviv_obj->_resv;
|
||||
reservation_object_init(&etnaviv_obj->_resv);
|
||||
}
|
||||
|
||||
mutex_init(&etnaviv_obj->lock);
|
||||
INIT_LIST_HEAD(&etnaviv_obj->vram_list);
|
||||
|
||||
*obj = &etnaviv_obj->base;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static struct drm_gem_object *__etnaviv_gem_new(struct drm_device *dev,
|
||||
u32 size, u32 flags)
|
||||
{
|
||||
struct drm_gem_object *obj = NULL;
|
||||
int ret;
|
||||
|
||||
size = PAGE_ALIGN(size);
|
||||
|
||||
ret = etnaviv_gem_new_impl(dev, size, flags, NULL,
|
||||
&etnaviv_gem_shmem_ops, &obj);
|
||||
if (ret)
|
||||
goto fail;
|
||||
|
||||
ret = drm_gem_object_init(dev, obj, size);
|
||||
if (ret == 0) {
|
||||
struct address_space *mapping;
|
||||
|
||||
/*
|
||||
* Our buffers are kept pinned, so allocating them
|
||||
* from the MOVABLE zone is a really bad idea, and
|
||||
* conflicts with CMA. See coments above new_inode()
|
||||
* why this is required _and_ expected if you're
|
||||
* going to pin these pages.
|
||||
*/
|
||||
mapping = file_inode(obj->filp)->i_mapping;
|
||||
mapping_set_gfp_mask(mapping, GFP_HIGHUSER);
|
||||
}
|
||||
|
||||
if (ret)
|
||||
goto fail;
|
||||
|
||||
return obj;
|
||||
|
||||
fail:
|
||||
if (obj)
|
||||
drm_gem_object_unreference_unlocked(obj);
|
||||
|
||||
return ERR_PTR(ret);
|
||||
}
|
||||
|
||||
/* convenience method to construct a GEM buffer object, and userspace handle */
|
||||
int etnaviv_gem_new_handle(struct drm_device *dev, struct drm_file *file,
|
||||
u32 size, u32 flags, u32 *handle)
|
||||
{
|
||||
struct drm_gem_object *obj;
|
||||
int ret;
|
||||
|
||||
obj = __etnaviv_gem_new(dev, size, flags);
|
||||
if (IS_ERR(obj))
|
||||
return PTR_ERR(obj);
|
||||
|
||||
ret = etnaviv_gem_obj_add(dev, obj);
|
||||
if (ret < 0) {
|
||||
drm_gem_object_unreference_unlocked(obj);
|
||||
return ret;
|
||||
}
|
||||
|
||||
ret = drm_gem_handle_create(file, obj, handle);
|
||||
|
||||
/* drop reference from allocate - handle holds it now */
|
||||
drm_gem_object_unreference_unlocked(obj);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
struct drm_gem_object *etnaviv_gem_new(struct drm_device *dev,
|
||||
u32 size, u32 flags)
|
||||
{
|
||||
struct drm_gem_object *obj;
|
||||
int ret;
|
||||
|
||||
obj = __etnaviv_gem_new(dev, size, flags);
|
||||
if (IS_ERR(obj))
|
||||
return obj;
|
||||
|
||||
ret = etnaviv_gem_obj_add(dev, obj);
|
||||
if (ret < 0) {
|
||||
drm_gem_object_unreference_unlocked(obj);
|
||||
return ERR_PTR(ret);
|
||||
}
|
||||
|
||||
return obj;
|
||||
}
|
||||
|
||||
int etnaviv_gem_new_private(struct drm_device *dev, size_t size, u32 flags,
|
||||
struct reservation_object *robj, const struct etnaviv_gem_ops *ops,
|
||||
struct etnaviv_gem_object **res)
|
||||
{
|
||||
struct drm_gem_object *obj;
|
||||
int ret;
|
||||
|
||||
ret = etnaviv_gem_new_impl(dev, size, flags, robj, ops, &obj);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
drm_gem_private_object_init(dev, obj, size);
|
||||
|
||||
*res = to_etnaviv_bo(obj);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
struct get_pages_work {
|
||||
struct work_struct work;
|
||||
struct mm_struct *mm;
|
||||
struct task_struct *task;
|
||||
struct etnaviv_gem_object *etnaviv_obj;
|
||||
};
|
||||
|
||||
static struct page **etnaviv_gem_userptr_do_get_pages(
|
||||
struct etnaviv_gem_object *etnaviv_obj, struct mm_struct *mm, struct task_struct *task)
|
||||
{
|
||||
int ret = 0, pinned, npages = etnaviv_obj->base.size >> PAGE_SHIFT;
|
||||
struct page **pvec;
|
||||
uintptr_t ptr;
|
||||
|
||||
pvec = drm_malloc_ab(npages, sizeof(struct page *));
|
||||
if (!pvec)
|
||||
return ERR_PTR(-ENOMEM);
|
||||
|
||||
pinned = 0;
|
||||
ptr = etnaviv_obj->userptr.ptr;
|
||||
|
||||
down_read(&mm->mmap_sem);
|
||||
while (pinned < npages) {
|
||||
ret = get_user_pages(task, mm, ptr, npages - pinned,
|
||||
!etnaviv_obj->userptr.ro, 0,
|
||||
pvec + pinned, NULL);
|
||||
if (ret < 0)
|
||||
break;
|
||||
|
||||
ptr += ret * PAGE_SIZE;
|
||||
pinned += ret;
|
||||
}
|
||||
up_read(&mm->mmap_sem);
|
||||
|
||||
if (ret < 0) {
|
||||
release_pages(pvec, pinned, 0);
|
||||
drm_free_large(pvec);
|
||||
return ERR_PTR(ret);
|
||||
}
|
||||
|
||||
return pvec;
|
||||
}
|
||||
|
||||
static void __etnaviv_gem_userptr_get_pages(struct work_struct *_work)
|
||||
{
|
||||
struct get_pages_work *work = container_of(_work, typeof(*work), work);
|
||||
struct etnaviv_gem_object *etnaviv_obj = work->etnaviv_obj;
|
||||
struct page **pvec;
|
||||
|
||||
pvec = etnaviv_gem_userptr_do_get_pages(etnaviv_obj, work->mm, work->task);
|
||||
|
||||
mutex_lock(&etnaviv_obj->lock);
|
||||
if (IS_ERR(pvec)) {
|
||||
etnaviv_obj->userptr.work = ERR_CAST(pvec);
|
||||
} else {
|
||||
etnaviv_obj->userptr.work = NULL;
|
||||
etnaviv_obj->pages = pvec;
|
||||
}
|
||||
|
||||
mutex_unlock(&etnaviv_obj->lock);
|
||||
drm_gem_object_unreference_unlocked(&etnaviv_obj->base);
|
||||
|
||||
mmput(work->mm);
|
||||
put_task_struct(work->task);
|
||||
kfree(work);
|
||||
}
|
||||
|
||||
static int etnaviv_gem_userptr_get_pages(struct etnaviv_gem_object *etnaviv_obj)
|
||||
{
|
||||
struct page **pvec = NULL;
|
||||
struct get_pages_work *work;
|
||||
struct mm_struct *mm;
|
||||
int ret, pinned, npages = etnaviv_obj->base.size >> PAGE_SHIFT;
|
||||
|
||||
if (etnaviv_obj->userptr.work) {
|
||||
if (IS_ERR(etnaviv_obj->userptr.work)) {
|
||||
ret = PTR_ERR(etnaviv_obj->userptr.work);
|
||||
etnaviv_obj->userptr.work = NULL;
|
||||
} else {
|
||||
ret = -EAGAIN;
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
mm = get_task_mm(etnaviv_obj->userptr.task);
|
||||
pinned = 0;
|
||||
if (mm == current->mm) {
|
||||
pvec = drm_malloc_ab(npages, sizeof(struct page *));
|
||||
if (!pvec) {
|
||||
mmput(mm);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
pinned = __get_user_pages_fast(etnaviv_obj->userptr.ptr, npages,
|
||||
!etnaviv_obj->userptr.ro, pvec);
|
||||
if (pinned < 0) {
|
||||
drm_free_large(pvec);
|
||||
mmput(mm);
|
||||
return pinned;
|
||||
}
|
||||
|
||||
if (pinned == npages) {
|
||||
etnaviv_obj->pages = pvec;
|
||||
mmput(mm);
|
||||
return 0;
|
||||
}
|
||||
}
|
||||
|
||||
release_pages(pvec, pinned, 0);
|
||||
drm_free_large(pvec);
|
||||
|
||||
work = kmalloc(sizeof(*work), GFP_KERNEL);
|
||||
if (!work) {
|
||||
mmput(mm);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
get_task_struct(current);
|
||||
drm_gem_object_reference(&etnaviv_obj->base);
|
||||
|
||||
work->mm = mm;
|
||||
work->task = current;
|
||||
work->etnaviv_obj = etnaviv_obj;
|
||||
|
||||
etnaviv_obj->userptr.work = &work->work;
|
||||
INIT_WORK(&work->work, __etnaviv_gem_userptr_get_pages);
|
||||
|
||||
etnaviv_queue_work(etnaviv_obj->base.dev, &work->work);
|
||||
|
||||
return -EAGAIN;
|
||||
}
|
||||
|
||||
static void etnaviv_gem_userptr_release(struct etnaviv_gem_object *etnaviv_obj)
|
||||
{
|
||||
if (etnaviv_obj->sgt) {
|
||||
etnaviv_gem_scatterlist_unmap(etnaviv_obj);
|
||||
sg_free_table(etnaviv_obj->sgt);
|
||||
kfree(etnaviv_obj->sgt);
|
||||
}
|
||||
if (etnaviv_obj->pages) {
|
||||
int npages = etnaviv_obj->base.size >> PAGE_SHIFT;
|
||||
|
||||
release_pages(etnaviv_obj->pages, npages, 0);
|
||||
drm_free_large(etnaviv_obj->pages);
|
||||
}
|
||||
put_task_struct(etnaviv_obj->userptr.task);
|
||||
}
|
||||
|
||||
static const struct etnaviv_gem_ops etnaviv_gem_userptr_ops = {
|
||||
.get_pages = etnaviv_gem_userptr_get_pages,
|
||||
.release = etnaviv_gem_userptr_release,
|
||||
};
|
||||
|
||||
int etnaviv_gem_new_userptr(struct drm_device *dev, struct drm_file *file,
|
||||
uintptr_t ptr, u32 size, u32 flags, u32 *handle)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj;
|
||||
int ret;
|
||||
|
||||
ret = etnaviv_gem_new_private(dev, size, ETNA_BO_CACHED, NULL,
|
||||
&etnaviv_gem_userptr_ops, &etnaviv_obj);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
etnaviv_obj->userptr.ptr = ptr;
|
||||
etnaviv_obj->userptr.task = current;
|
||||
etnaviv_obj->userptr.ro = !(flags & ETNA_USERPTR_WRITE);
|
||||
get_task_struct(current);
|
||||
|
||||
ret = etnaviv_gem_obj_add(dev, &etnaviv_obj->base);
|
||||
if (ret) {
|
||||
drm_gem_object_unreference_unlocked(&etnaviv_obj->base);
|
||||
return ret;
|
||||
}
|
||||
|
||||
ret = drm_gem_handle_create(file, &etnaviv_obj->base, handle);
|
||||
|
||||
/* drop reference from allocate - handle holds it now */
|
||||
drm_gem_object_unreference_unlocked(&etnaviv_obj->base);
|
||||
|
||||
return ret;
|
||||
}
|
117
drivers/gpu/drm/etnaviv/etnaviv_gem.h
Normal file
117
drivers/gpu/drm/etnaviv/etnaviv_gem.h
Normal file
|
@ -0,0 +1,117 @@
|
|||
/*
|
||||
* Copyright (C) 2015 Etnaviv Project
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#ifndef __ETNAVIV_GEM_H__
|
||||
#define __ETNAVIV_GEM_H__
|
||||
|
||||
#include <linux/reservation.h>
|
||||
#include "etnaviv_drv.h"
|
||||
|
||||
struct etnaviv_gem_ops;
|
||||
struct etnaviv_gem_object;
|
||||
|
||||
struct etnaviv_gem_userptr {
|
||||
uintptr_t ptr;
|
||||
struct task_struct *task;
|
||||
struct work_struct *work;
|
||||
bool ro;
|
||||
};
|
||||
|
||||
struct etnaviv_vram_mapping {
|
||||
struct list_head obj_node;
|
||||
struct list_head scan_node;
|
||||
struct list_head mmu_node;
|
||||
struct etnaviv_gem_object *object;
|
||||
struct etnaviv_iommu *mmu;
|
||||
struct drm_mm_node vram_node;
|
||||
unsigned int use;
|
||||
u32 iova;
|
||||
};
|
||||
|
||||
struct etnaviv_gem_object {
|
||||
struct drm_gem_object base;
|
||||
const struct etnaviv_gem_ops *ops;
|
||||
struct mutex lock;
|
||||
|
||||
u32 flags;
|
||||
|
||||
struct list_head gem_node;
|
||||
struct etnaviv_gpu *gpu; /* non-null if active */
|
||||
atomic_t gpu_active;
|
||||
u32 access;
|
||||
|
||||
struct page **pages;
|
||||
struct sg_table *sgt;
|
||||
void *vaddr;
|
||||
|
||||
/* normally (resv == &_resv) except for imported bo's */
|
||||
struct reservation_object *resv;
|
||||
struct reservation_object _resv;
|
||||
|
||||
struct list_head vram_list;
|
||||
|
||||
/* cache maintenance */
|
||||
u32 last_cpu_prep_op;
|
||||
|
||||
struct etnaviv_gem_userptr userptr;
|
||||
};
|
||||
|
||||
static inline
|
||||
struct etnaviv_gem_object *to_etnaviv_bo(struct drm_gem_object *obj)
|
||||
{
|
||||
return container_of(obj, struct etnaviv_gem_object, base);
|
||||
}
|
||||
|
||||
struct etnaviv_gem_ops {
|
||||
int (*get_pages)(struct etnaviv_gem_object *);
|
||||
void (*release)(struct etnaviv_gem_object *);
|
||||
};
|
||||
|
||||
static inline bool is_active(struct etnaviv_gem_object *etnaviv_obj)
|
||||
{
|
||||
return atomic_read(&etnaviv_obj->gpu_active) != 0;
|
||||
}
|
||||
|
||||
#define MAX_CMDS 4
|
||||
|
||||
/* Created per submit-ioctl, to track bo's and cmdstream bufs, etc,
|
||||
* associated with the cmdstream submission for synchronization (and
|
||||
* make it easier to unwind when things go wrong, etc). This only
|
||||
* lasts for the duration of the submit-ioctl.
|
||||
*/
|
||||
struct etnaviv_gem_submit {
|
||||
struct drm_device *dev;
|
||||
struct etnaviv_gpu *gpu;
|
||||
struct ww_acquire_ctx ticket;
|
||||
u32 fence;
|
||||
unsigned int nr_bos;
|
||||
struct {
|
||||
u32 flags;
|
||||
struct etnaviv_gem_object *obj;
|
||||
u32 iova;
|
||||
} bos[0];
|
||||
};
|
||||
|
||||
int etnaviv_gem_wait_bo(struct etnaviv_gpu *gpu, struct drm_gem_object *obj,
|
||||
struct timespec *timeout);
|
||||
int etnaviv_gem_new_private(struct drm_device *dev, size_t size, u32 flags,
|
||||
struct reservation_object *robj, const struct etnaviv_gem_ops *ops,
|
||||
struct etnaviv_gem_object **res);
|
||||
int etnaviv_gem_obj_add(struct drm_device *dev, struct drm_gem_object *obj);
|
||||
struct page **etnaviv_gem_get_pages(struct etnaviv_gem_object *obj);
|
||||
void etnaviv_gem_put_pages(struct etnaviv_gem_object *obj);
|
||||
|
||||
#endif /* __ETNAVIV_GEM_H__ */
|
122
drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c
Normal file
122
drivers/gpu/drm/etnaviv/etnaviv_gem_prime.c
Normal file
|
@ -0,0 +1,122 @@
|
|||
/*
|
||||
* Copyright (C) 2013 Red Hat
|
||||
* Author: Rob Clark <robdclark@gmail.com>
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#include <linux/dma-buf.h>
|
||||
#include "etnaviv_drv.h"
|
||||
#include "etnaviv_gem.h"
|
||||
|
||||
|
||||
struct sg_table *etnaviv_gem_prime_get_sg_table(struct drm_gem_object *obj)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
|
||||
BUG_ON(!etnaviv_obj->sgt); /* should have already pinned! */
|
||||
|
||||
return etnaviv_obj->sgt;
|
||||
}
|
||||
|
||||
void *etnaviv_gem_prime_vmap(struct drm_gem_object *obj)
|
||||
{
|
||||
return etnaviv_gem_vaddr(obj);
|
||||
}
|
||||
|
||||
void etnaviv_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr)
|
||||
{
|
||||
/* TODO msm_gem_vunmap() */
|
||||
}
|
||||
|
||||
int etnaviv_gem_prime_pin(struct drm_gem_object *obj)
|
||||
{
|
||||
if (!obj->import_attach) {
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
|
||||
mutex_lock(&etnaviv_obj->lock);
|
||||
etnaviv_gem_get_pages(etnaviv_obj);
|
||||
mutex_unlock(&etnaviv_obj->lock);
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
void etnaviv_gem_prime_unpin(struct drm_gem_object *obj)
|
||||
{
|
||||
if (!obj->import_attach) {
|
||||
struct etnaviv_gem_object *etnaviv_obj = to_etnaviv_bo(obj);
|
||||
|
||||
mutex_lock(&etnaviv_obj->lock);
|
||||
etnaviv_gem_put_pages(to_etnaviv_bo(obj));
|
||||
mutex_unlock(&etnaviv_obj->lock);
|
||||
}
|
||||
}
|
||||
|
||||
static void etnaviv_gem_prime_release(struct etnaviv_gem_object *etnaviv_obj)
|
||||
{
|
||||
if (etnaviv_obj->vaddr)
|
||||
dma_buf_vunmap(etnaviv_obj->base.import_attach->dmabuf,
|
||||
etnaviv_obj->vaddr);
|
||||
|
||||
/* Don't drop the pages for imported dmabuf, as they are not
|
||||
* ours, just free the array we allocated:
|
||||
*/
|
||||
if (etnaviv_obj->pages)
|
||||
drm_free_large(etnaviv_obj->pages);
|
||||
|
||||
drm_prime_gem_destroy(&etnaviv_obj->base, etnaviv_obj->sgt);
|
||||
}
|
||||
|
||||
static const struct etnaviv_gem_ops etnaviv_gem_prime_ops = {
|
||||
/* .get_pages should never be called */
|
||||
.release = etnaviv_gem_prime_release,
|
||||
};
|
||||
|
||||
struct drm_gem_object *etnaviv_gem_prime_import_sg_table(struct drm_device *dev,
|
||||
struct dma_buf_attachment *attach, struct sg_table *sgt)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj;
|
||||
size_t size = PAGE_ALIGN(attach->dmabuf->size);
|
||||
int ret, npages;
|
||||
|
||||
ret = etnaviv_gem_new_private(dev, size, ETNA_BO_WC,
|
||||
attach->dmabuf->resv,
|
||||
&etnaviv_gem_prime_ops, &etnaviv_obj);
|
||||
if (ret < 0)
|
||||
return ERR_PTR(ret);
|
||||
|
||||
npages = size / PAGE_SIZE;
|
||||
|
||||
etnaviv_obj->sgt = sgt;
|
||||
etnaviv_obj->pages = drm_malloc_ab(npages, sizeof(struct page *));
|
||||
if (!etnaviv_obj->pages) {
|
||||
ret = -ENOMEM;
|
||||
goto fail;
|
||||
}
|
||||
|
||||
ret = drm_prime_sg_to_page_addr_arrays(sgt, etnaviv_obj->pages,
|
||||
NULL, npages);
|
||||
if (ret)
|
||||
goto fail;
|
||||
|
||||
ret = etnaviv_gem_obj_add(dev, &etnaviv_obj->base);
|
||||
if (ret)
|
||||
goto fail;
|
||||
|
||||
return &etnaviv_obj->base;
|
||||
|
||||
fail:
|
||||
drm_gem_object_unreference_unlocked(&etnaviv_obj->base);
|
||||
|
||||
return ERR_PTR(ret);
|
||||
}
|
443
drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
Normal file
443
drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
Normal file
|
@ -0,0 +1,443 @@
|
|||
/*
|
||||
* Copyright (C) 2015 Etnaviv Project
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#include <linux/reservation.h>
|
||||
#include "etnaviv_drv.h"
|
||||
#include "etnaviv_gpu.h"
|
||||
#include "etnaviv_gem.h"
|
||||
|
||||
/*
|
||||
* Cmdstream submission:
|
||||
*/
|
||||
|
||||
#define BO_INVALID_FLAGS ~(ETNA_SUBMIT_BO_READ | ETNA_SUBMIT_BO_WRITE)
|
||||
/* make sure these don't conflict w/ ETNAVIV_SUBMIT_BO_x */
|
||||
#define BO_LOCKED 0x4000
|
||||
#define BO_PINNED 0x2000
|
||||
|
||||
static inline void __user *to_user_ptr(u64 address)
|
||||
{
|
||||
return (void __user *)(uintptr_t)address;
|
||||
}
|
||||
|
||||
static struct etnaviv_gem_submit *submit_create(struct drm_device *dev,
|
||||
struct etnaviv_gpu *gpu, size_t nr)
|
||||
{
|
||||
struct etnaviv_gem_submit *submit;
|
||||
size_t sz = size_vstruct(nr, sizeof(submit->bos[0]), sizeof(*submit));
|
||||
|
||||
submit = kmalloc(sz, GFP_TEMPORARY | __GFP_NOWARN | __GFP_NORETRY);
|
||||
if (submit) {
|
||||
submit->dev = dev;
|
||||
submit->gpu = gpu;
|
||||
|
||||
/* initially, until copy_from_user() and bo lookup succeeds: */
|
||||
submit->nr_bos = 0;
|
||||
|
||||
ww_acquire_init(&submit->ticket, &reservation_ww_class);
|
||||
}
|
||||
|
||||
return submit;
|
||||
}
|
||||
|
||||
static int submit_lookup_objects(struct etnaviv_gem_submit *submit,
|
||||
struct drm_file *file, struct drm_etnaviv_gem_submit_bo *submit_bos,
|
||||
unsigned nr_bos)
|
||||
{
|
||||
struct drm_etnaviv_gem_submit_bo *bo;
|
||||
unsigned i;
|
||||
int ret = 0;
|
||||
|
||||
spin_lock(&file->table_lock);
|
||||
|
||||
for (i = 0, bo = submit_bos; i < nr_bos; i++, bo++) {
|
||||
struct drm_gem_object *obj;
|
||||
|
||||
if (bo->flags & BO_INVALID_FLAGS) {
|
||||
DRM_ERROR("invalid flags: %x\n", bo->flags);
|
||||
ret = -EINVAL;
|
||||
goto out_unlock;
|
||||
}
|
||||
|
||||
submit->bos[i].flags = bo->flags;
|
||||
|
||||
/* normally use drm_gem_object_lookup(), but for bulk lookup
|
||||
* all under single table_lock just hit object_idr directly:
|
||||
*/
|
||||
obj = idr_find(&file->object_idr, bo->handle);
|
||||
if (!obj) {
|
||||
DRM_ERROR("invalid handle %u at index %u\n",
|
||||
bo->handle, i);
|
||||
ret = -EINVAL;
|
||||
goto out_unlock;
|
||||
}
|
||||
|
||||
/*
|
||||
* Take a refcount on the object. The file table lock
|
||||
* prevents the object_idr's refcount on this being dropped.
|
||||
*/
|
||||
drm_gem_object_reference(obj);
|
||||
|
||||
submit->bos[i].obj = to_etnaviv_bo(obj);
|
||||
}
|
||||
|
||||
out_unlock:
|
||||
submit->nr_bos = i;
|
||||
spin_unlock(&file->table_lock);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void submit_unlock_object(struct etnaviv_gem_submit *submit, int i)
|
||||
{
|
||||
if (submit->bos[i].flags & BO_LOCKED) {
|
||||
struct etnaviv_gem_object *etnaviv_obj = submit->bos[i].obj;
|
||||
|
||||
ww_mutex_unlock(&etnaviv_obj->resv->lock);
|
||||
submit->bos[i].flags &= ~BO_LOCKED;
|
||||
}
|
||||
}
|
||||
|
||||
static int submit_lock_objects(struct etnaviv_gem_submit *submit)
|
||||
{
|
||||
int contended, slow_locked = -1, i, ret = 0;
|
||||
|
||||
retry:
|
||||
for (i = 0; i < submit->nr_bos; i++) {
|
||||
struct etnaviv_gem_object *etnaviv_obj = submit->bos[i].obj;
|
||||
|
||||
if (slow_locked == i)
|
||||
slow_locked = -1;
|
||||
|
||||
contended = i;
|
||||
|
||||
if (!(submit->bos[i].flags & BO_LOCKED)) {
|
||||
ret = ww_mutex_lock_interruptible(&etnaviv_obj->resv->lock,
|
||||
&submit->ticket);
|
||||
if (ret == -EALREADY)
|
||||
DRM_ERROR("BO at index %u already on submit list\n",
|
||||
i);
|
||||
if (ret)
|
||||
goto fail;
|
||||
submit->bos[i].flags |= BO_LOCKED;
|
||||
}
|
||||
}
|
||||
|
||||
ww_acquire_done(&submit->ticket);
|
||||
|
||||
return 0;
|
||||
|
||||
fail:
|
||||
for (; i >= 0; i--)
|
||||
submit_unlock_object(submit, i);
|
||||
|
||||
if (slow_locked > 0)
|
||||
submit_unlock_object(submit, slow_locked);
|
||||
|
||||
if (ret == -EDEADLK) {
|
||||
struct etnaviv_gem_object *etnaviv_obj;
|
||||
|
||||
etnaviv_obj = submit->bos[contended].obj;
|
||||
|
||||
/* we lost out in a seqno race, lock and retry.. */
|
||||
ret = ww_mutex_lock_slow_interruptible(&etnaviv_obj->resv->lock,
|
||||
&submit->ticket);
|
||||
if (!ret) {
|
||||
submit->bos[contended].flags |= BO_LOCKED;
|
||||
slow_locked = contended;
|
||||
goto retry;
|
||||
}
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int submit_fence_sync(const struct etnaviv_gem_submit *submit)
|
||||
{
|
||||
unsigned int context = submit->gpu->fence_context;
|
||||
int i, ret = 0;
|
||||
|
||||
for (i = 0; i < submit->nr_bos; i++) {
|
||||
struct etnaviv_gem_object *etnaviv_obj = submit->bos[i].obj;
|
||||
bool write = submit->bos[i].flags & ETNA_SUBMIT_BO_WRITE;
|
||||
|
||||
ret = etnaviv_gpu_fence_sync_obj(etnaviv_obj, context, write);
|
||||
if (ret)
|
||||
break;
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void submit_unpin_objects(struct etnaviv_gem_submit *submit)
|
||||
{
|
||||
int i;
|
||||
|
||||
for (i = 0; i < submit->nr_bos; i++) {
|
||||
struct etnaviv_gem_object *etnaviv_obj = submit->bos[i].obj;
|
||||
|
||||
if (submit->bos[i].flags & BO_PINNED)
|
||||
etnaviv_gem_put_iova(submit->gpu, &etnaviv_obj->base);
|
||||
|
||||
submit->bos[i].iova = 0;
|
||||
submit->bos[i].flags &= ~BO_PINNED;
|
||||
}
|
||||
}
|
||||
|
||||
static int submit_pin_objects(struct etnaviv_gem_submit *submit)
|
||||
{
|
||||
int i, ret = 0;
|
||||
|
||||
for (i = 0; i < submit->nr_bos; i++) {
|
||||
struct etnaviv_gem_object *etnaviv_obj = submit->bos[i].obj;
|
||||
u32 iova;
|
||||
|
||||
ret = etnaviv_gem_get_iova(submit->gpu, &etnaviv_obj->base,
|
||||
&iova);
|
||||
if (ret)
|
||||
break;
|
||||
|
||||
submit->bos[i].flags |= BO_PINNED;
|
||||
submit->bos[i].iova = iova;
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int submit_bo(struct etnaviv_gem_submit *submit, u32 idx,
|
||||
struct etnaviv_gem_object **obj, u32 *iova)
|
||||
{
|
||||
if (idx >= submit->nr_bos) {
|
||||
DRM_ERROR("invalid buffer index: %u (out of %u)\n",
|
||||
idx, submit->nr_bos);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (obj)
|
||||
*obj = submit->bos[idx].obj;
|
||||
if (iova)
|
||||
*iova = submit->bos[idx].iova;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/* process the reloc's and patch up the cmdstream as needed: */
|
||||
static int submit_reloc(struct etnaviv_gem_submit *submit, void *stream,
|
||||
u32 size, const struct drm_etnaviv_gem_submit_reloc *relocs,
|
||||
u32 nr_relocs)
|
||||
{
|
||||
u32 i, last_offset = 0;
|
||||
u32 *ptr = stream;
|
||||
int ret;
|
||||
|
||||
for (i = 0; i < nr_relocs; i++) {
|
||||
const struct drm_etnaviv_gem_submit_reloc *r = relocs + i;
|
||||
struct etnaviv_gem_object *bobj;
|
||||
u32 iova, off;
|
||||
|
||||
if (unlikely(r->flags)) {
|
||||
DRM_ERROR("invalid reloc flags\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (r->submit_offset % 4) {
|
||||
DRM_ERROR("non-aligned reloc offset: %u\n",
|
||||
r->submit_offset);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
/* offset in dwords: */
|
||||
off = r->submit_offset / 4;
|
||||
|
||||
if ((off >= size ) ||
|
||||
(off < last_offset)) {
|
||||
DRM_ERROR("invalid offset %u at reloc %u\n", off, i);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
ret = submit_bo(submit, r->reloc_idx, &bobj, &iova);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
if (r->reloc_offset >=
|
||||
bobj->base.size - sizeof(*ptr)) {
|
||||
DRM_ERROR("relocation %u outside object", i);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
ptr[off] = iova + r->reloc_offset;
|
||||
|
||||
last_offset = off;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void submit_cleanup(struct etnaviv_gem_submit *submit)
|
||||
{
|
||||
unsigned i;
|
||||
|
||||
for (i = 0; i < submit->nr_bos; i++) {
|
||||
struct etnaviv_gem_object *etnaviv_obj = submit->bos[i].obj;
|
||||
|
||||
submit_unlock_object(submit, i);
|
||||
drm_gem_object_unreference_unlocked(&etnaviv_obj->base);
|
||||
}
|
||||
|
||||
ww_acquire_fini(&submit->ticket);
|
||||
kfree(submit);
|
||||
}
|
||||
|
||||
int etnaviv_ioctl_gem_submit(struct drm_device *dev, void *data,
|
||||
struct drm_file *file)
|
||||
{
|
||||
struct etnaviv_drm_private *priv = dev->dev_private;
|
||||
struct drm_etnaviv_gem_submit *args = data;
|
||||
struct drm_etnaviv_gem_submit_reloc *relocs;
|
||||
struct drm_etnaviv_gem_submit_bo *bos;
|
||||
struct etnaviv_gem_submit *submit;
|
||||
struct etnaviv_cmdbuf *cmdbuf;
|
||||
struct etnaviv_gpu *gpu;
|
||||
void *stream;
|
||||
int ret;
|
||||
|
||||
if (args->pipe >= ETNA_MAX_PIPES)
|
||||
return -EINVAL;
|
||||
|
||||
gpu = priv->gpu[args->pipe];
|
||||
if (!gpu)
|
||||
return -ENXIO;
|
||||
|
||||
if (args->stream_size % 4) {
|
||||
DRM_ERROR("non-aligned cmdstream buffer size: %u\n",
|
||||
args->stream_size);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (args->exec_state != ETNA_PIPE_3D &&
|
||||
args->exec_state != ETNA_PIPE_2D &&
|
||||
args->exec_state != ETNA_PIPE_VG) {
|
||||
DRM_ERROR("invalid exec_state: 0x%x\n", args->exec_state);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
/*
|
||||
* Copy the command submission and bo array to kernel space in
|
||||
* one go, and do this outside of any locks.
|
||||
*/
|
||||
bos = drm_malloc_ab(args->nr_bos, sizeof(*bos));
|
||||
relocs = drm_malloc_ab(args->nr_relocs, sizeof(*relocs));
|
||||
stream = drm_malloc_ab(1, args->stream_size);
|
||||
cmdbuf = etnaviv_gpu_cmdbuf_new(gpu, ALIGN(args->stream_size, 8) + 8,
|
||||
args->nr_bos);
|
||||
if (!bos || !relocs || !stream || !cmdbuf) {
|
||||
ret = -ENOMEM;
|
||||
goto err_submit_cmds;
|
||||
}
|
||||
|
||||
cmdbuf->exec_state = args->exec_state;
|
||||
cmdbuf->ctx = file->driver_priv;
|
||||
|
||||
ret = copy_from_user(bos, to_user_ptr(args->bos),
|
||||
args->nr_bos * sizeof(*bos));
|
||||
if (ret) {
|
||||
ret = -EFAULT;
|
||||
goto err_submit_cmds;
|
||||
}
|
||||
|
||||
ret = copy_from_user(relocs, to_user_ptr(args->relocs),
|
||||
args->nr_relocs * sizeof(*relocs));
|
||||
if (ret) {
|
||||
ret = -EFAULT;
|
||||
goto err_submit_cmds;
|
||||
}
|
||||
|
||||
ret = copy_from_user(stream, to_user_ptr(args->stream),
|
||||
args->stream_size);
|
||||
if (ret) {
|
||||
ret = -EFAULT;
|
||||
goto err_submit_cmds;
|
||||
}
|
||||
|
||||
submit = submit_create(dev, gpu, args->nr_bos);
|
||||
if (!submit) {
|
||||
ret = -ENOMEM;
|
||||
goto err_submit_cmds;
|
||||
}
|
||||
|
||||
ret = submit_lookup_objects(submit, file, bos, args->nr_bos);
|
||||
if (ret)
|
||||
goto err_submit_objects;
|
||||
|
||||
ret = submit_lock_objects(submit);
|
||||
if (ret)
|
||||
goto err_submit_objects;
|
||||
|
||||
if (!etnaviv_cmd_validate_one(gpu, stream, args->stream_size / 4,
|
||||
relocs, args->nr_relocs)) {
|
||||
ret = -EINVAL;
|
||||
goto err_submit_objects;
|
||||
}
|
||||
|
||||
ret = submit_fence_sync(submit);
|
||||
if (ret)
|
||||
goto err_submit_objects;
|
||||
|
||||
ret = submit_pin_objects(submit);
|
||||
if (ret)
|
||||
goto out;
|
||||
|
||||
ret = submit_reloc(submit, stream, args->stream_size / 4,
|
||||
relocs, args->nr_relocs);
|
||||
if (ret)
|
||||
goto out;
|
||||
|
||||
memcpy(cmdbuf->vaddr, stream, args->stream_size);
|
||||
cmdbuf->user_size = ALIGN(args->stream_size, 8);
|
||||
|
||||
ret = etnaviv_gpu_submit(gpu, submit, cmdbuf);
|
||||
if (ret == 0)
|
||||
cmdbuf = NULL;
|
||||
|
||||
args->fence = submit->fence;
|
||||
|
||||
out:
|
||||
submit_unpin_objects(submit);
|
||||
|
||||
/*
|
||||
* If we're returning -EAGAIN, it may be due to the userptr code
|
||||
* wanting to run its workqueue outside of any locks. Flush our
|
||||
* workqueue to ensure that it is run in a timely manner.
|
||||
*/
|
||||
if (ret == -EAGAIN)
|
||||
flush_workqueue(priv->wq);
|
||||
|
||||
err_submit_objects:
|
||||
submit_cleanup(submit);
|
||||
|
||||
err_submit_cmds:
|
||||
/* if we still own the cmdbuf */
|
||||
if (cmdbuf)
|
||||
etnaviv_gpu_cmdbuf_free(cmdbuf);
|
||||
if (stream)
|
||||
drm_free_large(stream);
|
||||
if (bos)
|
||||
drm_free_large(bos);
|
||||
if (relocs)
|
||||
drm_free_large(relocs);
|
||||
|
||||
return ret;
|
||||
}
|
1644
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
Normal file
1644
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
Normal file
File diff suppressed because it is too large
Load diff
209
drivers/gpu/drm/etnaviv/etnaviv_gpu.h
Normal file
209
drivers/gpu/drm/etnaviv/etnaviv_gpu.h
Normal file
|
@ -0,0 +1,209 @@
|
|||
/*
|
||||
* Copyright (C) 2015 Etnaviv Project
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#ifndef __ETNAVIV_GPU_H__
|
||||
#define __ETNAVIV_GPU_H__
|
||||
|
||||
#include <linux/clk.h>
|
||||
#include <linux/regulator/consumer.h>
|
||||
|
||||
#include "etnaviv_drv.h"
|
||||
|
||||
struct etnaviv_gem_submit;
|
||||
|
||||
struct etnaviv_chip_identity {
|
||||
/* Chip model. */
|
||||
u32 model;
|
||||
|
||||
/* Revision value.*/
|
||||
u32 revision;
|
||||
|
||||
/* Supported feature fields. */
|
||||
u32 features;
|
||||
|
||||
/* Supported minor feature fields. */
|
||||
u32 minor_features0;
|
||||
|
||||
/* Supported minor feature 1 fields. */
|
||||
u32 minor_features1;
|
||||
|
||||
/* Supported minor feature 2 fields. */
|
||||
u32 minor_features2;
|
||||
|
||||
/* Supported minor feature 3 fields. */
|
||||
u32 minor_features3;
|
||||
|
||||
/* Number of streams supported. */
|
||||
u32 stream_count;
|
||||
|
||||
/* Total number of temporary registers per thread. */
|
||||
u32 register_max;
|
||||
|
||||
/* Maximum number of threads. */
|
||||
u32 thread_count;
|
||||
|
||||
/* Number of shader cores. */
|
||||
u32 shader_core_count;
|
||||
|
||||
/* Size of the vertex cache. */
|
||||
u32 vertex_cache_size;
|
||||
|
||||
/* Number of entries in the vertex output buffer. */
|
||||
u32 vertex_output_buffer_size;
|
||||
|
||||
/* Number of pixel pipes. */
|
||||
u32 pixel_pipes;
|
||||
|
||||
/* Number of instructions. */
|
||||
u32 instruction_count;
|
||||
|
||||
/* Number of constants. */
|
||||
u32 num_constants;
|
||||
|
||||
/* Buffer size */
|
||||
u32 buffer_size;
|
||||
};
|
||||
|
||||
struct etnaviv_event {
|
||||
bool used;
|
||||
struct fence *fence;
|
||||
};
|
||||
|
||||
struct etnaviv_cmdbuf;
|
||||
|
||||
struct etnaviv_gpu {
|
||||
struct drm_device *drm;
|
||||
struct device *dev;
|
||||
struct mutex lock;
|
||||
struct etnaviv_chip_identity identity;
|
||||
struct etnaviv_file_private *lastctx;
|
||||
bool switch_context;
|
||||
|
||||
/* 'ring'-buffer: */
|
||||
struct etnaviv_cmdbuf *buffer;
|
||||
|
||||
/* bus base address of memory */
|
||||
u32 memory_base;
|
||||
|
||||
/* event management: */
|
||||
struct etnaviv_event event[30];
|
||||
struct completion event_free;
|
||||
spinlock_t event_spinlock;
|
||||
|
||||
/* list of currently in-flight command buffers */
|
||||
struct list_head active_cmd_list;
|
||||
|
||||
u32 idle_mask;
|
||||
|
||||
/* Fencing support */
|
||||
u32 next_fence;
|
||||
u32 active_fence;
|
||||
u32 completed_fence;
|
||||
u32 retired_fence;
|
||||
wait_queue_head_t fence_event;
|
||||
unsigned int fence_context;
|
||||
spinlock_t fence_spinlock;
|
||||
|
||||
/* worker for handling active-list retiring: */
|
||||
struct work_struct retire_work;
|
||||
|
||||
void __iomem *mmio;
|
||||
int irq;
|
||||
|
||||
struct etnaviv_iommu *mmu;
|
||||
|
||||
/* Power Control: */
|
||||
struct clk *clk_bus;
|
||||
struct clk *clk_core;
|
||||
struct clk *clk_shader;
|
||||
|
||||
/* Hang Detction: */
|
||||
#define DRM_ETNAVIV_HANGCHECK_PERIOD 500 /* in ms */
|
||||
#define DRM_ETNAVIV_HANGCHECK_JIFFIES msecs_to_jiffies(DRM_ETNAVIV_HANGCHECK_PERIOD)
|
||||
struct timer_list hangcheck_timer;
|
||||
u32 hangcheck_fence;
|
||||
u32 hangcheck_dma_addr;
|
||||
struct work_struct recover_work;
|
||||
};
|
||||
|
||||
struct etnaviv_cmdbuf {
|
||||
/* device this cmdbuf is allocated for */
|
||||
struct etnaviv_gpu *gpu;
|
||||
/* user context key, must be unique between all active users */
|
||||
struct etnaviv_file_private *ctx;
|
||||
/* cmdbuf properties */
|
||||
void *vaddr;
|
||||
dma_addr_t paddr;
|
||||
u32 size;
|
||||
u32 user_size;
|
||||
/* fence after which this buffer is to be disposed */
|
||||
struct fence *fence;
|
||||
/* target exec state */
|
||||
u32 exec_state;
|
||||
/* per GPU in-flight list */
|
||||
struct list_head node;
|
||||
/* BOs attached to this command buffer */
|
||||
unsigned int nr_bos;
|
||||
struct etnaviv_gem_object *bo[0];
|
||||
};
|
||||
|
||||
static inline void gpu_write(struct etnaviv_gpu *gpu, u32 reg, u32 data)
|
||||
{
|
||||
etnaviv_writel(data, gpu->mmio + reg);
|
||||
}
|
||||
|
||||
static inline u32 gpu_read(struct etnaviv_gpu *gpu, u32 reg)
|
||||
{
|
||||
return etnaviv_readl(gpu->mmio + reg);
|
||||
}
|
||||
|
||||
static inline bool fence_completed(struct etnaviv_gpu *gpu, u32 fence)
|
||||
{
|
||||
return fence_after_eq(gpu->completed_fence, fence);
|
||||
}
|
||||
|
||||
static inline bool fence_retired(struct etnaviv_gpu *gpu, u32 fence)
|
||||
{
|
||||
return fence_after_eq(gpu->retired_fence, fence);
|
||||
}
|
||||
|
||||
int etnaviv_gpu_get_param(struct etnaviv_gpu *gpu, u32 param, u64 *value);
|
||||
|
||||
int etnaviv_gpu_init(struct etnaviv_gpu *gpu);
|
||||
|
||||
#ifdef CONFIG_DEBUG_FS
|
||||
int etnaviv_gpu_debugfs(struct etnaviv_gpu *gpu, struct seq_file *m);
|
||||
#endif
|
||||
|
||||
int etnaviv_gpu_fence_sync_obj(struct etnaviv_gem_object *etnaviv_obj,
|
||||
unsigned int context, bool exclusive);
|
||||
|
||||
void etnaviv_gpu_retire(struct etnaviv_gpu *gpu);
|
||||
int etnaviv_gpu_wait_fence_interruptible(struct etnaviv_gpu *gpu,
|
||||
u32 fence, struct timespec *timeout);
|
||||
int etnaviv_gpu_wait_obj_inactive(struct etnaviv_gpu *gpu,
|
||||
struct etnaviv_gem_object *etnaviv_obj, struct timespec *timeout);
|
||||
int etnaviv_gpu_submit(struct etnaviv_gpu *gpu,
|
||||
struct etnaviv_gem_submit *submit, struct etnaviv_cmdbuf *cmdbuf);
|
||||
struct etnaviv_cmdbuf *etnaviv_gpu_cmdbuf_new(struct etnaviv_gpu *gpu,
|
||||
u32 size, size_t nr_bos);
|
||||
void etnaviv_gpu_cmdbuf_free(struct etnaviv_cmdbuf *cmdbuf);
|
||||
int etnaviv_gpu_pm_get_sync(struct etnaviv_gpu *gpu);
|
||||
void etnaviv_gpu_pm_put(struct etnaviv_gpu *gpu);
|
||||
|
||||
extern struct platform_driver etnaviv_gpu_driver;
|
||||
|
||||
#endif /* __ETNAVIV_GPU_H__ */
|
240
drivers/gpu/drm/etnaviv/etnaviv_iommu.c
Normal file
240
drivers/gpu/drm/etnaviv/etnaviv_iommu.c
Normal file
|
@ -0,0 +1,240 @@
|
|||
/*
|
||||
* Copyright (C) 2014 Christian Gmeiner <christian.gmeiner@gmail.com>
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#include <linux/iommu.h>
|
||||
#include <linux/platform_device.h>
|
||||
#include <linux/sizes.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/dma-mapping.h>
|
||||
#include <linux/bitops.h>
|
||||
|
||||
#include "etnaviv_gpu.h"
|
||||
#include "etnaviv_mmu.h"
|
||||
#include "etnaviv_iommu.h"
|
||||
#include "state_hi.xml.h"
|
||||
|
||||
#define PT_SIZE SZ_2M
|
||||
#define PT_ENTRIES (PT_SIZE / sizeof(u32))
|
||||
|
||||
#define GPU_MEM_START 0x80000000
|
||||
|
||||
struct etnaviv_iommu_domain_pgtable {
|
||||
u32 *pgtable;
|
||||
dma_addr_t paddr;
|
||||
};
|
||||
|
||||
struct etnaviv_iommu_domain {
|
||||
struct iommu_domain domain;
|
||||
struct device *dev;
|
||||
void *bad_page_cpu;
|
||||
dma_addr_t bad_page_dma;
|
||||
struct etnaviv_iommu_domain_pgtable pgtable;
|
||||
spinlock_t map_lock;
|
||||
};
|
||||
|
||||
static struct etnaviv_iommu_domain *to_etnaviv_domain(struct iommu_domain *domain)
|
||||
{
|
||||
return container_of(domain, struct etnaviv_iommu_domain, domain);
|
||||
}
|
||||
|
||||
static int pgtable_alloc(struct etnaviv_iommu_domain_pgtable *pgtable,
|
||||
size_t size)
|
||||
{
|
||||
pgtable->pgtable = dma_alloc_coherent(NULL, size, &pgtable->paddr, GFP_KERNEL);
|
||||
if (!pgtable->pgtable)
|
||||
return -ENOMEM;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void pgtable_free(struct etnaviv_iommu_domain_pgtable *pgtable,
|
||||
size_t size)
|
||||
{
|
||||
dma_free_coherent(NULL, size, pgtable->pgtable, pgtable->paddr);
|
||||
}
|
||||
|
||||
static u32 pgtable_read(struct etnaviv_iommu_domain_pgtable *pgtable,
|
||||
unsigned long iova)
|
||||
{
|
||||
/* calcuate index into page table */
|
||||
unsigned int index = (iova - GPU_MEM_START) / SZ_4K;
|
||||
phys_addr_t paddr;
|
||||
|
||||
paddr = pgtable->pgtable[index];
|
||||
|
||||
return paddr;
|
||||
}
|
||||
|
||||
static void pgtable_write(struct etnaviv_iommu_domain_pgtable *pgtable,
|
||||
unsigned long iova, phys_addr_t paddr)
|
||||
{
|
||||
/* calcuate index into page table */
|
||||
unsigned int index = (iova - GPU_MEM_START) / SZ_4K;
|
||||
|
||||
pgtable->pgtable[index] = paddr;
|
||||
}
|
||||
|
||||
static int __etnaviv_iommu_init(struct etnaviv_iommu_domain *etnaviv_domain)
|
||||
{
|
||||
u32 *p;
|
||||
int ret, i;
|
||||
|
||||
etnaviv_domain->bad_page_cpu = dma_alloc_coherent(etnaviv_domain->dev,
|
||||
SZ_4K,
|
||||
&etnaviv_domain->bad_page_dma,
|
||||
GFP_KERNEL);
|
||||
if (!etnaviv_domain->bad_page_cpu)
|
||||
return -ENOMEM;
|
||||
|
||||
p = etnaviv_domain->bad_page_cpu;
|
||||
for (i = 0; i < SZ_4K / 4; i++)
|
||||
*p++ = 0xdead55aa;
|
||||
|
||||
ret = pgtable_alloc(&etnaviv_domain->pgtable, PT_SIZE);
|
||||
if (ret < 0) {
|
||||
dma_free_coherent(etnaviv_domain->dev, SZ_4K,
|
||||
etnaviv_domain->bad_page_cpu,
|
||||
etnaviv_domain->bad_page_dma);
|
||||
return ret;
|
||||
}
|
||||
|
||||
for (i = 0; i < PT_ENTRIES; i++)
|
||||
etnaviv_domain->pgtable.pgtable[i] =
|
||||
etnaviv_domain->bad_page_dma;
|
||||
|
||||
spin_lock_init(&etnaviv_domain->map_lock);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void etnaviv_domain_free(struct iommu_domain *domain)
|
||||
{
|
||||
struct etnaviv_iommu_domain *etnaviv_domain = to_etnaviv_domain(domain);
|
||||
|
||||
pgtable_free(&etnaviv_domain->pgtable, PT_SIZE);
|
||||
|
||||
dma_free_coherent(etnaviv_domain->dev, SZ_4K,
|
||||
etnaviv_domain->bad_page_cpu,
|
||||
etnaviv_domain->bad_page_dma);
|
||||
|
||||
kfree(etnaviv_domain);
|
||||
}
|
||||
|
||||
static int etnaviv_iommuv1_map(struct iommu_domain *domain, unsigned long iova,
|
||||
phys_addr_t paddr, size_t size, int prot)
|
||||
{
|
||||
struct etnaviv_iommu_domain *etnaviv_domain = to_etnaviv_domain(domain);
|
||||
|
||||
if (size != SZ_4K)
|
||||
return -EINVAL;
|
||||
|
||||
spin_lock(&etnaviv_domain->map_lock);
|
||||
pgtable_write(&etnaviv_domain->pgtable, iova, paddr);
|
||||
spin_unlock(&etnaviv_domain->map_lock);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static size_t etnaviv_iommuv1_unmap(struct iommu_domain *domain,
|
||||
unsigned long iova, size_t size)
|
||||
{
|
||||
struct etnaviv_iommu_domain *etnaviv_domain = to_etnaviv_domain(domain);
|
||||
|
||||
if (size != SZ_4K)
|
||||
return -EINVAL;
|
||||
|
||||
spin_lock(&etnaviv_domain->map_lock);
|
||||
pgtable_write(&etnaviv_domain->pgtable, iova,
|
||||
etnaviv_domain->bad_page_dma);
|
||||
spin_unlock(&etnaviv_domain->map_lock);
|
||||
|
||||
return SZ_4K;
|
||||
}
|
||||
|
||||
static phys_addr_t etnaviv_iommu_iova_to_phys(struct iommu_domain *domain,
|
||||
dma_addr_t iova)
|
||||
{
|
||||
struct etnaviv_iommu_domain *etnaviv_domain = to_etnaviv_domain(domain);
|
||||
|
||||
return pgtable_read(&etnaviv_domain->pgtable, iova);
|
||||
}
|
||||
|
||||
static size_t etnaviv_iommuv1_dump_size(struct iommu_domain *domain)
|
||||
{
|
||||
return PT_SIZE;
|
||||
}
|
||||
|
||||
static void etnaviv_iommuv1_dump(struct iommu_domain *domain, void *buf)
|
||||
{
|
||||
struct etnaviv_iommu_domain *etnaviv_domain = to_etnaviv_domain(domain);
|
||||
|
||||
memcpy(buf, etnaviv_domain->pgtable.pgtable, PT_SIZE);
|
||||
}
|
||||
|
||||
static struct etnaviv_iommu_ops etnaviv_iommu_ops = {
|
||||
.ops = {
|
||||
.domain_free = etnaviv_domain_free,
|
||||
.map = etnaviv_iommuv1_map,
|
||||
.unmap = etnaviv_iommuv1_unmap,
|
||||
.iova_to_phys = etnaviv_iommu_iova_to_phys,
|
||||
.pgsize_bitmap = SZ_4K,
|
||||
},
|
||||
.dump_size = etnaviv_iommuv1_dump_size,
|
||||
.dump = etnaviv_iommuv1_dump,
|
||||
};
|
||||
|
||||
void etnaviv_iommu_domain_restore(struct etnaviv_gpu *gpu,
|
||||
struct iommu_domain *domain)
|
||||
{
|
||||
struct etnaviv_iommu_domain *etnaviv_domain = to_etnaviv_domain(domain);
|
||||
u32 pgtable;
|
||||
|
||||
/* set page table address in MC */
|
||||
pgtable = (u32)etnaviv_domain->pgtable.paddr;
|
||||
|
||||
gpu_write(gpu, VIVS_MC_MMU_FE_PAGE_TABLE, pgtable);
|
||||
gpu_write(gpu, VIVS_MC_MMU_TX_PAGE_TABLE, pgtable);
|
||||
gpu_write(gpu, VIVS_MC_MMU_PE_PAGE_TABLE, pgtable);
|
||||
gpu_write(gpu, VIVS_MC_MMU_PEZ_PAGE_TABLE, pgtable);
|
||||
gpu_write(gpu, VIVS_MC_MMU_RA_PAGE_TABLE, pgtable);
|
||||
}
|
||||
|
||||
struct iommu_domain *etnaviv_iommu_domain_alloc(struct etnaviv_gpu *gpu)
|
||||
{
|
||||
struct etnaviv_iommu_domain *etnaviv_domain;
|
||||
int ret;
|
||||
|
||||
etnaviv_domain = kzalloc(sizeof(*etnaviv_domain), GFP_KERNEL);
|
||||
if (!etnaviv_domain)
|
||||
return NULL;
|
||||
|
||||
etnaviv_domain->dev = gpu->dev;
|
||||
|
||||
etnaviv_domain->domain.type = __IOMMU_DOMAIN_PAGING;
|
||||
etnaviv_domain->domain.ops = &etnaviv_iommu_ops.ops;
|
||||
etnaviv_domain->domain.geometry.aperture_start = GPU_MEM_START;
|
||||
etnaviv_domain->domain.geometry.aperture_end = GPU_MEM_START + PT_ENTRIES * SZ_4K - 1;
|
||||
|
||||
ret = __etnaviv_iommu_init(etnaviv_domain);
|
||||
if (ret)
|
||||
goto out_free;
|
||||
|
||||
return &etnaviv_domain->domain;
|
||||
|
||||
out_free:
|
||||
kfree(etnaviv_domain);
|
||||
return NULL;
|
||||
}
|
28
drivers/gpu/drm/etnaviv/etnaviv_iommu.h
Normal file
28
drivers/gpu/drm/etnaviv/etnaviv_iommu.h
Normal file
|
@ -0,0 +1,28 @@
|
|||
/*
|
||||
* Copyright (C) 2014 Christian Gmeiner <christian.gmeiner@gmail.com>
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#ifndef __ETNAVIV_IOMMU_H__
|
||||
#define __ETNAVIV_IOMMU_H__
|
||||
|
||||
#include <linux/iommu.h>
|
||||
struct etnaviv_gpu;
|
||||
|
||||
struct iommu_domain *etnaviv_iommu_domain_alloc(struct etnaviv_gpu *gpu);
|
||||
void etnaviv_iommu_domain_restore(struct etnaviv_gpu *gpu,
|
||||
struct iommu_domain *domain);
|
||||
struct iommu_domain *etnaviv_iommu_v2_domain_alloc(struct etnaviv_gpu *gpu);
|
||||
|
||||
#endif /* __ETNAVIV_IOMMU_H__ */
|
33
drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c
Normal file
33
drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c
Normal file
|
@ -0,0 +1,33 @@
|
|||
/*
|
||||
* Copyright (C) 2014 Christian Gmeiner <christian.gmeiner@gmail.com>
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#include <linux/iommu.h>
|
||||
#include <linux/platform_device.h>
|
||||
#include <linux/sizes.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/dma-mapping.h>
|
||||
#include <linux/bitops.h>
|
||||
|
||||
#include "etnaviv_gpu.h"
|
||||
#include "etnaviv_iommu.h"
|
||||
#include "state_hi.xml.h"
|
||||
|
||||
|
||||
struct iommu_domain *etnaviv_iommu_v2_domain_alloc(struct etnaviv_gpu *gpu)
|
||||
{
|
||||
/* TODO */
|
||||
return NULL;
|
||||
}
|
25
drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.h
Normal file
25
drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.h
Normal file
|
@ -0,0 +1,25 @@
|
|||
/*
|
||||
* Copyright (C) 2014 Christian Gmeiner <christian.gmeiner@gmail.com>
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#ifndef __ETNAVIV_IOMMU_V2_H__
|
||||
#define __ETNAVIV_IOMMU_V2_H__
|
||||
|
||||
#include <linux/iommu.h>
|
||||
struct etnaviv_gpu;
|
||||
|
||||
struct iommu_domain *etnaviv_iommu_v2_domain_alloc(struct etnaviv_gpu *gpu);
|
||||
|
||||
#endif /* __ETNAVIV_IOMMU_V2_H__ */
|
299
drivers/gpu/drm/etnaviv/etnaviv_mmu.c
Normal file
299
drivers/gpu/drm/etnaviv/etnaviv_mmu.c
Normal file
|
@ -0,0 +1,299 @@
|
|||
/*
|
||||
* Copyright (C) 2015 Etnaviv Project
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#include "etnaviv_drv.h"
|
||||
#include "etnaviv_gem.h"
|
||||
#include "etnaviv_gpu.h"
|
||||
#include "etnaviv_mmu.h"
|
||||
|
||||
static int etnaviv_fault_handler(struct iommu_domain *iommu, struct device *dev,
|
||||
unsigned long iova, int flags, void *arg)
|
||||
{
|
||||
DBG("*** fault: iova=%08lx, flags=%d", iova, flags);
|
||||
return 0;
|
||||
}
|
||||
|
||||
int etnaviv_iommu_map(struct etnaviv_iommu *iommu, u32 iova,
|
||||
struct sg_table *sgt, unsigned len, int prot)
|
||||
{
|
||||
struct iommu_domain *domain = iommu->domain;
|
||||
struct scatterlist *sg;
|
||||
unsigned int da = iova;
|
||||
unsigned int i, j;
|
||||
int ret;
|
||||
|
||||
if (!domain || !sgt)
|
||||
return -EINVAL;
|
||||
|
||||
for_each_sg(sgt->sgl, sg, sgt->nents, i) {
|
||||
u32 pa = sg_dma_address(sg) - sg->offset;
|
||||
size_t bytes = sg_dma_len(sg) + sg->offset;
|
||||
|
||||
VERB("map[%d]: %08x %08x(%zx)", i, iova, pa, bytes);
|
||||
|
||||
ret = iommu_map(domain, da, pa, bytes, prot);
|
||||
if (ret)
|
||||
goto fail;
|
||||
|
||||
da += bytes;
|
||||
}
|
||||
|
||||
return 0;
|
||||
|
||||
fail:
|
||||
da = iova;
|
||||
|
||||
for_each_sg(sgt->sgl, sg, i, j) {
|
||||
size_t bytes = sg_dma_len(sg) + sg->offset;
|
||||
|
||||
iommu_unmap(domain, da, bytes);
|
||||
da += bytes;
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
int etnaviv_iommu_unmap(struct etnaviv_iommu *iommu, u32 iova,
|
||||
struct sg_table *sgt, unsigned len)
|
||||
{
|
||||
struct iommu_domain *domain = iommu->domain;
|
||||
struct scatterlist *sg;
|
||||
unsigned int da = iova;
|
||||
int i;
|
||||
|
||||
for_each_sg(sgt->sgl, sg, sgt->nents, i) {
|
||||
size_t bytes = sg_dma_len(sg) + sg->offset;
|
||||
size_t unmapped;
|
||||
|
||||
unmapped = iommu_unmap(domain, da, bytes);
|
||||
if (unmapped < bytes)
|
||||
return unmapped;
|
||||
|
||||
VERB("unmap[%d]: %08x(%zx)", i, iova, bytes);
|
||||
|
||||
BUG_ON(!PAGE_ALIGNED(bytes));
|
||||
|
||||
da += bytes;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void etnaviv_iommu_remove_mapping(struct etnaviv_iommu *mmu,
|
||||
struct etnaviv_vram_mapping *mapping)
|
||||
{
|
||||
struct etnaviv_gem_object *etnaviv_obj = mapping->object;
|
||||
|
||||
etnaviv_iommu_unmap(mmu, mapping->vram_node.start,
|
||||
etnaviv_obj->sgt, etnaviv_obj->base.size);
|
||||
drm_mm_remove_node(&mapping->vram_node);
|
||||
}
|
||||
|
||||
int etnaviv_iommu_map_gem(struct etnaviv_iommu *mmu,
|
||||
struct etnaviv_gem_object *etnaviv_obj, u32 memory_base,
|
||||
struct etnaviv_vram_mapping *mapping)
|
||||
{
|
||||
struct etnaviv_vram_mapping *free = NULL;
|
||||
struct sg_table *sgt = etnaviv_obj->sgt;
|
||||
struct drm_mm_node *node;
|
||||
int ret;
|
||||
|
||||
lockdep_assert_held(&etnaviv_obj->lock);
|
||||
|
||||
mutex_lock(&mmu->lock);
|
||||
|
||||
/* v1 MMU can optimize single entry (contiguous) scatterlists */
|
||||
if (sgt->nents == 1 && !(etnaviv_obj->flags & ETNA_BO_FORCE_MMU)) {
|
||||
u32 iova;
|
||||
|
||||
iova = sg_dma_address(sgt->sgl) - memory_base;
|
||||
if (iova < 0x80000000 - sg_dma_len(sgt->sgl)) {
|
||||
mapping->iova = iova;
|
||||
list_add_tail(&mapping->mmu_node, &mmu->mappings);
|
||||
mutex_unlock(&mmu->lock);
|
||||
return 0;
|
||||
}
|
||||
}
|
||||
|
||||
node = &mapping->vram_node;
|
||||
while (1) {
|
||||
struct etnaviv_vram_mapping *m, *n;
|
||||
struct list_head list;
|
||||
bool found;
|
||||
|
||||
ret = drm_mm_insert_node_in_range(&mmu->mm, node,
|
||||
etnaviv_obj->base.size, 0, mmu->last_iova, ~0UL,
|
||||
DRM_MM_SEARCH_DEFAULT);
|
||||
|
||||
if (ret != -ENOSPC)
|
||||
break;
|
||||
|
||||
/*
|
||||
* If we did not search from the start of the MMU region,
|
||||
* try again in case there are free slots.
|
||||
*/
|
||||
if (mmu->last_iova) {
|
||||
mmu->last_iova = 0;
|
||||
mmu->need_flush = true;
|
||||
continue;
|
||||
}
|
||||
|
||||
/* Try to retire some entries */
|
||||
drm_mm_init_scan(&mmu->mm, etnaviv_obj->base.size, 0, 0);
|
||||
|
||||
found = 0;
|
||||
INIT_LIST_HEAD(&list);
|
||||
list_for_each_entry(free, &mmu->mappings, mmu_node) {
|
||||
/* If this vram node has not been used, skip this. */
|
||||
if (!free->vram_node.mm)
|
||||
continue;
|
||||
|
||||
/*
|
||||
* If the iova is pinned, then it's in-use,
|
||||
* so we must keep its mapping.
|
||||
*/
|
||||
if (free->use)
|
||||
continue;
|
||||
|
||||
list_add(&free->scan_node, &list);
|
||||
if (drm_mm_scan_add_block(&free->vram_node)) {
|
||||
found = true;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (!found) {
|
||||
/* Nothing found, clean up and fail */
|
||||
list_for_each_entry_safe(m, n, &list, scan_node)
|
||||
BUG_ON(drm_mm_scan_remove_block(&m->vram_node));
|
||||
break;
|
||||
}
|
||||
|
||||
/*
|
||||
* drm_mm does not allow any other operations while
|
||||
* scanning, so we have to remove all blocks first.
|
||||
* If drm_mm_scan_remove_block() returns false, we
|
||||
* can leave the block pinned.
|
||||
*/
|
||||
list_for_each_entry_safe(m, n, &list, scan_node)
|
||||
if (!drm_mm_scan_remove_block(&m->vram_node))
|
||||
list_del_init(&m->scan_node);
|
||||
|
||||
/*
|
||||
* Unmap the blocks which need to be reaped from the MMU.
|
||||
* Clear the mmu pointer to prevent the get_iova finding
|
||||
* this mapping.
|
||||
*/
|
||||
list_for_each_entry_safe(m, n, &list, scan_node) {
|
||||
etnaviv_iommu_remove_mapping(mmu, m);
|
||||
m->mmu = NULL;
|
||||
list_del_init(&m->mmu_node);
|
||||
list_del_init(&m->scan_node);
|
||||
}
|
||||
|
||||
/*
|
||||
* We removed enough mappings so that the new allocation will
|
||||
* succeed. Ensure that the MMU will be flushed before the
|
||||
* associated commit requesting this mapping, and retry the
|
||||
* allocation one more time.
|
||||
*/
|
||||
mmu->need_flush = true;
|
||||
}
|
||||
|
||||
if (ret < 0) {
|
||||
mutex_unlock(&mmu->lock);
|
||||
return ret;
|
||||
}
|
||||
|
||||
mmu->last_iova = node->start + etnaviv_obj->base.size;
|
||||
mapping->iova = node->start;
|
||||
ret = etnaviv_iommu_map(mmu, node->start, sgt, etnaviv_obj->base.size,
|
||||
IOMMU_READ | IOMMU_WRITE);
|
||||
|
||||
if (ret < 0) {
|
||||
drm_mm_remove_node(node);
|
||||
mutex_unlock(&mmu->lock);
|
||||
return ret;
|
||||
}
|
||||
|
||||
list_add_tail(&mapping->mmu_node, &mmu->mappings);
|
||||
mutex_unlock(&mmu->lock);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
void etnaviv_iommu_unmap_gem(struct etnaviv_iommu *mmu,
|
||||
struct etnaviv_vram_mapping *mapping)
|
||||
{
|
||||
WARN_ON(mapping->use);
|
||||
|
||||
mutex_lock(&mmu->lock);
|
||||
|
||||
/* If the vram node is on the mm, unmap and remove the node */
|
||||
if (mapping->vram_node.mm == &mmu->mm)
|
||||
etnaviv_iommu_remove_mapping(mmu, mapping);
|
||||
|
||||
list_del(&mapping->mmu_node);
|
||||
mutex_unlock(&mmu->lock);
|
||||
}
|
||||
|
||||
void etnaviv_iommu_destroy(struct etnaviv_iommu *mmu)
|
||||
{
|
||||
drm_mm_takedown(&mmu->mm);
|
||||
iommu_domain_free(mmu->domain);
|
||||
kfree(mmu);
|
||||
}
|
||||
|
||||
struct etnaviv_iommu *etnaviv_iommu_new(struct etnaviv_gpu *gpu,
|
||||
struct iommu_domain *domain, enum etnaviv_iommu_version version)
|
||||
{
|
||||
struct etnaviv_iommu *mmu;
|
||||
|
||||
mmu = kzalloc(sizeof(*mmu), GFP_KERNEL);
|
||||
if (!mmu)
|
||||
return ERR_PTR(-ENOMEM);
|
||||
|
||||
mmu->domain = domain;
|
||||
mmu->gpu = gpu;
|
||||
mmu->version = version;
|
||||
mutex_init(&mmu->lock);
|
||||
INIT_LIST_HEAD(&mmu->mappings);
|
||||
|
||||
drm_mm_init(&mmu->mm, domain->geometry.aperture_start,
|
||||
domain->geometry.aperture_end -
|
||||
domain->geometry.aperture_start + 1);
|
||||
|
||||
iommu_set_fault_handler(domain, etnaviv_fault_handler, gpu->dev);
|
||||
|
||||
return mmu;
|
||||
}
|
||||
|
||||
size_t etnaviv_iommu_dump_size(struct etnaviv_iommu *iommu)
|
||||
{
|
||||
struct etnaviv_iommu_ops *ops;
|
||||
|
||||
ops = container_of(iommu->domain->ops, struct etnaviv_iommu_ops, ops);
|
||||
|
||||
return ops->dump_size(iommu->domain);
|
||||
}
|
||||
|
||||
void etnaviv_iommu_dump(struct etnaviv_iommu *iommu, void *buf)
|
||||
{
|
||||
struct etnaviv_iommu_ops *ops;
|
||||
|
||||
ops = container_of(iommu->domain->ops, struct etnaviv_iommu_ops, ops);
|
||||
|
||||
ops->dump(iommu->domain, buf);
|
||||
}
|
71
drivers/gpu/drm/etnaviv/etnaviv_mmu.h
Normal file
71
drivers/gpu/drm/etnaviv/etnaviv_mmu.h
Normal file
|
@ -0,0 +1,71 @@
|
|||
/*
|
||||
* Copyright (C) 2015 Etnaviv Project
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#ifndef __ETNAVIV_MMU_H__
|
||||
#define __ETNAVIV_MMU_H__
|
||||
|
||||
#include <linux/iommu.h>
|
||||
|
||||
enum etnaviv_iommu_version {
|
||||
ETNAVIV_IOMMU_V1 = 0,
|
||||
ETNAVIV_IOMMU_V2,
|
||||
};
|
||||
|
||||
struct etnaviv_gpu;
|
||||
struct etnaviv_vram_mapping;
|
||||
|
||||
struct etnaviv_iommu_ops {
|
||||
struct iommu_ops ops;
|
||||
size_t (*dump_size)(struct iommu_domain *);
|
||||
void (*dump)(struct iommu_domain *, void *);
|
||||
};
|
||||
|
||||
struct etnaviv_iommu {
|
||||
struct etnaviv_gpu *gpu;
|
||||
struct iommu_domain *domain;
|
||||
|
||||
enum etnaviv_iommu_version version;
|
||||
|
||||
/* memory manager for GPU address area */
|
||||
struct mutex lock;
|
||||
struct list_head mappings;
|
||||
struct drm_mm mm;
|
||||
u32 last_iova;
|
||||
bool need_flush;
|
||||
};
|
||||
|
||||
struct etnaviv_gem_object;
|
||||
|
||||
int etnaviv_iommu_attach(struct etnaviv_iommu *iommu, const char **names,
|
||||
int cnt);
|
||||
int etnaviv_iommu_map(struct etnaviv_iommu *iommu, u32 iova,
|
||||
struct sg_table *sgt, unsigned len, int prot);
|
||||
int etnaviv_iommu_unmap(struct etnaviv_iommu *iommu, u32 iova,
|
||||
struct sg_table *sgt, unsigned len);
|
||||
int etnaviv_iommu_map_gem(struct etnaviv_iommu *mmu,
|
||||
struct etnaviv_gem_object *etnaviv_obj, u32 memory_base,
|
||||
struct etnaviv_vram_mapping *mapping);
|
||||
void etnaviv_iommu_unmap_gem(struct etnaviv_iommu *mmu,
|
||||
struct etnaviv_vram_mapping *mapping);
|
||||
void etnaviv_iommu_destroy(struct etnaviv_iommu *iommu);
|
||||
|
||||
size_t etnaviv_iommu_dump_size(struct etnaviv_iommu *iommu);
|
||||
void etnaviv_iommu_dump(struct etnaviv_iommu *iommu, void *buf);
|
||||
|
||||
struct etnaviv_iommu *etnaviv_iommu_new(struct etnaviv_gpu *gpu,
|
||||
struct iommu_domain *domain, enum etnaviv_iommu_version version);
|
||||
|
||||
#endif /* __ETNAVIV_MMU_H__ */
|
351
drivers/gpu/drm/etnaviv/state.xml.h
Normal file
351
drivers/gpu/drm/etnaviv/state.xml.h
Normal file
|
@ -0,0 +1,351 @@
|
|||
#ifndef STATE_XML
|
||||
#define STATE_XML
|
||||
|
||||
/* Autogenerated file, DO NOT EDIT manually!
|
||||
|
||||
This file was generated by the rules-ng-ng headergen tool in this git repository:
|
||||
http://0x04.net/cgit/index.cgi/rules-ng-ng
|
||||
git clone git://0x04.net/rules-ng-ng
|
||||
|
||||
The rules-ng-ng source files this header was generated from are:
|
||||
- state.xml ( 18882 bytes, from 2015-03-25 11:42:32)
|
||||
- common.xml ( 18437 bytes, from 2015-03-25 11:27:41)
|
||||
- state_hi.xml ( 23420 bytes, from 2015-03-25 11:47:21)
|
||||
- state_2d.xml ( 51549 bytes, from 2015-03-25 11:25:06)
|
||||
- state_3d.xml ( 54600 bytes, from 2015-03-25 11:25:19)
|
||||
- state_vg.xml ( 5973 bytes, from 2015-03-25 11:26:01)
|
||||
|
||||
Copyright (C) 2015
|
||||
*/
|
||||
|
||||
|
||||
#define VARYING_COMPONENT_USE_UNUSED 0x00000000
|
||||
#define VARYING_COMPONENT_USE_USED 0x00000001
|
||||
#define VARYING_COMPONENT_USE_POINTCOORD_X 0x00000002
|
||||
#define VARYING_COMPONENT_USE_POINTCOORD_Y 0x00000003
|
||||
#define FE_VERTEX_STREAM_CONTROL_VERTEX_STRIDE__MASK 0x000000ff
|
||||
#define FE_VERTEX_STREAM_CONTROL_VERTEX_STRIDE__SHIFT 0
|
||||
#define FE_VERTEX_STREAM_CONTROL_VERTEX_STRIDE(x) (((x) << FE_VERTEX_STREAM_CONTROL_VERTEX_STRIDE__SHIFT) & FE_VERTEX_STREAM_CONTROL_VERTEX_STRIDE__MASK)
|
||||
#define VIVS_FE 0x00000000
|
||||
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG(i0) (0x00000600 + 0x4*(i0))
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG__ESIZE 0x00000004
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG__LEN 0x00000010
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE__MASK 0x0000000f
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE__SHIFT 0
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_BYTE 0x00000000
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_UNSIGNED_BYTE 0x00000001
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_SHORT 0x00000002
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_UNSIGNED_SHORT 0x00000003
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_INT 0x00000004
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_UNSIGNED_INT 0x00000005
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_FLOAT 0x00000008
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_HALF_FLOAT 0x00000009
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_FIXED 0x0000000b
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_INT_10_10_10_2 0x0000000c
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_TYPE_UNSIGNED_INT_10_10_10_2 0x0000000d
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_ENDIAN__MASK 0x00000030
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_ENDIAN__SHIFT 4
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_ENDIAN(x) (((x) << VIVS_FE_VERTEX_ELEMENT_CONFIG_ENDIAN__SHIFT) & VIVS_FE_VERTEX_ELEMENT_CONFIG_ENDIAN__MASK)
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_NONCONSECUTIVE 0x00000080
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_STREAM__MASK 0x00000700
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_STREAM__SHIFT 8
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_STREAM(x) (((x) << VIVS_FE_VERTEX_ELEMENT_CONFIG_STREAM__SHIFT) & VIVS_FE_VERTEX_ELEMENT_CONFIG_STREAM__MASK)
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_NUM__MASK 0x00003000
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_NUM__SHIFT 12
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_NUM(x) (((x) << VIVS_FE_VERTEX_ELEMENT_CONFIG_NUM__SHIFT) & VIVS_FE_VERTEX_ELEMENT_CONFIG_NUM__MASK)
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_NORMALIZE__MASK 0x0000c000
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_NORMALIZE__SHIFT 14
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_NORMALIZE_OFF 0x00000000
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_NORMALIZE_ON 0x00008000
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_START__MASK 0x00ff0000
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_START__SHIFT 16
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_START(x) (((x) << VIVS_FE_VERTEX_ELEMENT_CONFIG_START__SHIFT) & VIVS_FE_VERTEX_ELEMENT_CONFIG_START__MASK)
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_END__MASK 0xff000000
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_END__SHIFT 24
|
||||
#define VIVS_FE_VERTEX_ELEMENT_CONFIG_END(x) (((x) << VIVS_FE_VERTEX_ELEMENT_CONFIG_END__SHIFT) & VIVS_FE_VERTEX_ELEMENT_CONFIG_END__MASK)
|
||||
|
||||
#define VIVS_FE_CMD_STREAM_BASE_ADDR 0x00000640
|
||||
|
||||
#define VIVS_FE_INDEX_STREAM_BASE_ADDR 0x00000644
|
||||
|
||||
#define VIVS_FE_INDEX_STREAM_CONTROL 0x00000648
|
||||
#define VIVS_FE_INDEX_STREAM_CONTROL_TYPE__MASK 0x00000003
|
||||
#define VIVS_FE_INDEX_STREAM_CONTROL_TYPE__SHIFT 0
|
||||
#define VIVS_FE_INDEX_STREAM_CONTROL_TYPE_UNSIGNED_CHAR 0x00000000
|
||||
#define VIVS_FE_INDEX_STREAM_CONTROL_TYPE_UNSIGNED_SHORT 0x00000001
|
||||
#define VIVS_FE_INDEX_STREAM_CONTROL_TYPE_UNSIGNED_INT 0x00000002
|
||||
|
||||
#define VIVS_FE_VERTEX_STREAM_BASE_ADDR 0x0000064c
|
||||
|
||||
#define VIVS_FE_VERTEX_STREAM_CONTROL 0x00000650
|
||||
|
||||
#define VIVS_FE_COMMAND_ADDRESS 0x00000654
|
||||
|
||||
#define VIVS_FE_COMMAND_CONTROL 0x00000658
|
||||
#define VIVS_FE_COMMAND_CONTROL_PREFETCH__MASK 0x0000ffff
|
||||
#define VIVS_FE_COMMAND_CONTROL_PREFETCH__SHIFT 0
|
||||
#define VIVS_FE_COMMAND_CONTROL_PREFETCH(x) (((x) << VIVS_FE_COMMAND_CONTROL_PREFETCH__SHIFT) & VIVS_FE_COMMAND_CONTROL_PREFETCH__MASK)
|
||||
#define VIVS_FE_COMMAND_CONTROL_ENABLE 0x00010000
|
||||
|
||||
#define VIVS_FE_DMA_STATUS 0x0000065c
|
||||
|
||||
#define VIVS_FE_DMA_DEBUG_STATE 0x00000660
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE__MASK 0x0000001f
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE__SHIFT 0
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_IDLE 0x00000000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_DEC 0x00000001
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_ADR0 0x00000002
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_LOAD0 0x00000003
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_ADR1 0x00000004
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_LOAD1 0x00000005
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_3DADR 0x00000006
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_3DCMD 0x00000007
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_3DCNTL 0x00000008
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_3DIDXCNTL 0x00000009
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_INITREQDMA 0x0000000a
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_DRAWIDX 0x0000000b
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_DRAW 0x0000000c
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_2DRECT0 0x0000000d
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_2DRECT1 0x0000000e
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_2DDATA0 0x0000000f
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_2DDATA1 0x00000010
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_WAITFIFO 0x00000011
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_WAIT 0x00000012
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_LINK 0x00000013
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_END 0x00000014
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_STATE_STALL 0x00000015
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_DMA_STATE__MASK 0x00000300
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_DMA_STATE__SHIFT 8
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_DMA_STATE_IDLE 0x00000000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_DMA_STATE_START 0x00000100
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_DMA_STATE_REQ 0x00000200
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_DMA_STATE_END 0x00000300
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_FETCH_STATE__MASK 0x00000c00
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_FETCH_STATE__SHIFT 10
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_FETCH_STATE_IDLE 0x00000000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_FETCH_STATE_RAMVALID 0x00000400
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CMD_FETCH_STATE_VALID 0x00000800
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_REQ_DMA_STATE__MASK 0x00003000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_REQ_DMA_STATE__SHIFT 12
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_REQ_DMA_STATE_IDLE 0x00000000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_REQ_DMA_STATE_WAITIDX 0x00001000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_REQ_DMA_STATE_CAL 0x00002000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CAL_STATE__MASK 0x0000c000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CAL_STATE__SHIFT 14
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CAL_STATE_IDLE 0x00000000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CAL_STATE_LDADR 0x00004000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_CAL_STATE_IDXCALC 0x00008000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_VE_REQ_STATE__MASK 0x00030000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_VE_REQ_STATE__SHIFT 16
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_VE_REQ_STATE_IDLE 0x00000000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_VE_REQ_STATE_CKCACHE 0x00010000
|
||||
#define VIVS_FE_DMA_DEBUG_STATE_VE_REQ_STATE_MISS 0x00020000
|
||||
|
||||
#define VIVS_FE_DMA_ADDRESS 0x00000664
|
||||
|
||||
#define VIVS_FE_DMA_LOW 0x00000668
|
||||
|
||||
#define VIVS_FE_DMA_HIGH 0x0000066c
|
||||
|
||||
#define VIVS_FE_AUTO_FLUSH 0x00000670
|
||||
|
||||
#define VIVS_FE_UNK00678 0x00000678
|
||||
|
||||
#define VIVS_FE_UNK0067C 0x0000067c
|
||||
|
||||
#define VIVS_FE_VERTEX_STREAMS(i0) (0x00000000 + 0x4*(i0))
|
||||
#define VIVS_FE_VERTEX_STREAMS__ESIZE 0x00000004
|
||||
#define VIVS_FE_VERTEX_STREAMS__LEN 0x00000008
|
||||
|
||||
#define VIVS_FE_VERTEX_STREAMS_BASE_ADDR(i0) (0x00000680 + 0x4*(i0))
|
||||
|
||||
#define VIVS_FE_VERTEX_STREAMS_CONTROL(i0) (0x000006a0 + 0x4*(i0))
|
||||
|
||||
#define VIVS_FE_UNK00700(i0) (0x00000700 + 0x4*(i0))
|
||||
#define VIVS_FE_UNK00700__ESIZE 0x00000004
|
||||
#define VIVS_FE_UNK00700__LEN 0x00000010
|
||||
|
||||
#define VIVS_FE_UNK00740(i0) (0x00000740 + 0x4*(i0))
|
||||
#define VIVS_FE_UNK00740__ESIZE 0x00000004
|
||||
#define VIVS_FE_UNK00740__LEN 0x00000010
|
||||
|
||||
#define VIVS_FE_UNK00780(i0) (0x00000780 + 0x4*(i0))
|
||||
#define VIVS_FE_UNK00780__ESIZE 0x00000004
|
||||
#define VIVS_FE_UNK00780__LEN 0x00000010
|
||||
|
||||
#define VIVS_GL 0x00000000
|
||||
|
||||
#define VIVS_GL_PIPE_SELECT 0x00003800
|
||||
#define VIVS_GL_PIPE_SELECT_PIPE__MASK 0x00000001
|
||||
#define VIVS_GL_PIPE_SELECT_PIPE__SHIFT 0
|
||||
#define VIVS_GL_PIPE_SELECT_PIPE(x) (((x) << VIVS_GL_PIPE_SELECT_PIPE__SHIFT) & VIVS_GL_PIPE_SELECT_PIPE__MASK)
|
||||
|
||||
#define VIVS_GL_EVENT 0x00003804
|
||||
#define VIVS_GL_EVENT_EVENT_ID__MASK 0x0000001f
|
||||
#define VIVS_GL_EVENT_EVENT_ID__SHIFT 0
|
||||
#define VIVS_GL_EVENT_EVENT_ID(x) (((x) << VIVS_GL_EVENT_EVENT_ID__SHIFT) & VIVS_GL_EVENT_EVENT_ID__MASK)
|
||||
#define VIVS_GL_EVENT_FROM_FE 0x00000020
|
||||
#define VIVS_GL_EVENT_FROM_PE 0x00000040
|
||||
#define VIVS_GL_EVENT_SOURCE__MASK 0x00001f00
|
||||
#define VIVS_GL_EVENT_SOURCE__SHIFT 8
|
||||
#define VIVS_GL_EVENT_SOURCE(x) (((x) << VIVS_GL_EVENT_SOURCE__SHIFT) & VIVS_GL_EVENT_SOURCE__MASK)
|
||||
|
||||
#define VIVS_GL_SEMAPHORE_TOKEN 0x00003808
|
||||
#define VIVS_GL_SEMAPHORE_TOKEN_FROM__MASK 0x0000001f
|
||||
#define VIVS_GL_SEMAPHORE_TOKEN_FROM__SHIFT 0
|
||||
#define VIVS_GL_SEMAPHORE_TOKEN_FROM(x) (((x) << VIVS_GL_SEMAPHORE_TOKEN_FROM__SHIFT) & VIVS_GL_SEMAPHORE_TOKEN_FROM__MASK)
|
||||
#define VIVS_GL_SEMAPHORE_TOKEN_TO__MASK 0x00001f00
|
||||
#define VIVS_GL_SEMAPHORE_TOKEN_TO__SHIFT 8
|
||||
#define VIVS_GL_SEMAPHORE_TOKEN_TO(x) (((x) << VIVS_GL_SEMAPHORE_TOKEN_TO__SHIFT) & VIVS_GL_SEMAPHORE_TOKEN_TO__MASK)
|
||||
|
||||
#define VIVS_GL_FLUSH_CACHE 0x0000380c
|
||||
#define VIVS_GL_FLUSH_CACHE_DEPTH 0x00000001
|
||||
#define VIVS_GL_FLUSH_CACHE_COLOR 0x00000002
|
||||
#define VIVS_GL_FLUSH_CACHE_TEXTURE 0x00000004
|
||||
#define VIVS_GL_FLUSH_CACHE_PE2D 0x00000008
|
||||
#define VIVS_GL_FLUSH_CACHE_TEXTUREVS 0x00000010
|
||||
#define VIVS_GL_FLUSH_CACHE_SHADER_L1 0x00000020
|
||||
#define VIVS_GL_FLUSH_CACHE_SHADER_L2 0x00000040
|
||||
|
||||
#define VIVS_GL_FLUSH_MMU 0x00003810
|
||||
#define VIVS_GL_FLUSH_MMU_FLUSH_FEMMU 0x00000001
|
||||
#define VIVS_GL_FLUSH_MMU_FLUSH_UNK1 0x00000002
|
||||
#define VIVS_GL_FLUSH_MMU_FLUSH_UNK2 0x00000004
|
||||
#define VIVS_GL_FLUSH_MMU_FLUSH_PEMMU 0x00000008
|
||||
#define VIVS_GL_FLUSH_MMU_FLUSH_UNK4 0x00000010
|
||||
|
||||
#define VIVS_GL_VERTEX_ELEMENT_CONFIG 0x00003814
|
||||
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG 0x00003818
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_MSAA_SAMPLES__MASK 0x00000003
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_MSAA_SAMPLES__SHIFT 0
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_MSAA_SAMPLES_NONE 0x00000000
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_MSAA_SAMPLES_2X 0x00000001
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_MSAA_SAMPLES_4X 0x00000002
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_MSAA_SAMPLES_MASK 0x00000008
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_MSAA_ENABLES__MASK 0x000000f0
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_MSAA_ENABLES__SHIFT 4
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_MSAA_ENABLES(x) (((x) << VIVS_GL_MULTI_SAMPLE_CONFIG_MSAA_ENABLES__SHIFT) & VIVS_GL_MULTI_SAMPLE_CONFIG_MSAA_ENABLES__MASK)
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_MSAA_ENABLES_MASK 0x00000100
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_UNK12__MASK 0x00007000
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_UNK12__SHIFT 12
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_UNK12(x) (((x) << VIVS_GL_MULTI_SAMPLE_CONFIG_UNK12__SHIFT) & VIVS_GL_MULTI_SAMPLE_CONFIG_UNK12__MASK)
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_UNK12_MASK 0x00008000
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_UNK16__MASK 0x00030000
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_UNK16__SHIFT 16
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_UNK16(x) (((x) << VIVS_GL_MULTI_SAMPLE_CONFIG_UNK16__SHIFT) & VIVS_GL_MULTI_SAMPLE_CONFIG_UNK16__MASK)
|
||||
#define VIVS_GL_MULTI_SAMPLE_CONFIG_UNK16_MASK 0x00080000
|
||||
|
||||
#define VIVS_GL_VARYING_TOTAL_COMPONENTS 0x0000381c
|
||||
#define VIVS_GL_VARYING_TOTAL_COMPONENTS_NUM__MASK 0x000000ff
|
||||
#define VIVS_GL_VARYING_TOTAL_COMPONENTS_NUM__SHIFT 0
|
||||
#define VIVS_GL_VARYING_TOTAL_COMPONENTS_NUM(x) (((x) << VIVS_GL_VARYING_TOTAL_COMPONENTS_NUM__SHIFT) & VIVS_GL_VARYING_TOTAL_COMPONENTS_NUM__MASK)
|
||||
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS 0x00003820
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR0__MASK 0x00000007
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR0__SHIFT 0
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR0(x) (((x) << VIVS_GL_VARYING_NUM_COMPONENTS_VAR0__SHIFT) & VIVS_GL_VARYING_NUM_COMPONENTS_VAR0__MASK)
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR1__MASK 0x00000070
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR1__SHIFT 4
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR1(x) (((x) << VIVS_GL_VARYING_NUM_COMPONENTS_VAR1__SHIFT) & VIVS_GL_VARYING_NUM_COMPONENTS_VAR1__MASK)
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR2__MASK 0x00000700
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR2__SHIFT 8
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR2(x) (((x) << VIVS_GL_VARYING_NUM_COMPONENTS_VAR2__SHIFT) & VIVS_GL_VARYING_NUM_COMPONENTS_VAR2__MASK)
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR3__MASK 0x00007000
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR3__SHIFT 12
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR3(x) (((x) << VIVS_GL_VARYING_NUM_COMPONENTS_VAR3__SHIFT) & VIVS_GL_VARYING_NUM_COMPONENTS_VAR3__MASK)
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR4__MASK 0x00070000
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR4__SHIFT 16
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR4(x) (((x) << VIVS_GL_VARYING_NUM_COMPONENTS_VAR4__SHIFT) & VIVS_GL_VARYING_NUM_COMPONENTS_VAR4__MASK)
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR5__MASK 0x00700000
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR5__SHIFT 20
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR5(x) (((x) << VIVS_GL_VARYING_NUM_COMPONENTS_VAR5__SHIFT) & VIVS_GL_VARYING_NUM_COMPONENTS_VAR5__MASK)
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR6__MASK 0x07000000
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR6__SHIFT 24
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR6(x) (((x) << VIVS_GL_VARYING_NUM_COMPONENTS_VAR6__SHIFT) & VIVS_GL_VARYING_NUM_COMPONENTS_VAR6__MASK)
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR7__MASK 0x70000000
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR7__SHIFT 28
|
||||
#define VIVS_GL_VARYING_NUM_COMPONENTS_VAR7(x) (((x) << VIVS_GL_VARYING_NUM_COMPONENTS_VAR7__SHIFT) & VIVS_GL_VARYING_NUM_COMPONENTS_VAR7__MASK)
|
||||
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE(i0) (0x00003828 + 0x4*(i0))
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE__ESIZE 0x00000004
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE__LEN 0x00000002
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP0__MASK 0x00000003
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP0__SHIFT 0
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP0(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP0__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP0__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP1__MASK 0x0000000c
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP1__SHIFT 2
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP1(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP1__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP1__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP2__MASK 0x00000030
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP2__SHIFT 4
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP2(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP2__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP2__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP3__MASK 0x000000c0
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP3__SHIFT 6
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP3(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP3__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP3__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP4__MASK 0x00000300
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP4__SHIFT 8
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP4(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP4__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP4__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP5__MASK 0x00000c00
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP5__SHIFT 10
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP5(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP5__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP5__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP6__MASK 0x00003000
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP6__SHIFT 12
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP6(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP6__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP6__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP7__MASK 0x0000c000
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP7__SHIFT 14
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP7(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP7__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP7__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP8__MASK 0x00030000
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP8__SHIFT 16
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP8(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP8__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP8__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP9__MASK 0x000c0000
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP9__SHIFT 18
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP9(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP9__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP9__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP10__MASK 0x00300000
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP10__SHIFT 20
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP10(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP10__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP10__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP11__MASK 0x00c00000
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP11__SHIFT 22
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP11(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP11__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP11__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP12__MASK 0x03000000
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP12__SHIFT 24
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP12(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP12__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP12__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP13__MASK 0x0c000000
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP13__SHIFT 26
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP13(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP13__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP13__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP14__MASK 0x30000000
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP14__SHIFT 28
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP14(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP14__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP14__MASK)
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP15__MASK 0xc0000000
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP15__SHIFT 30
|
||||
#define VIVS_GL_VARYING_COMPONENT_USE_COMP15(x) (((x) << VIVS_GL_VARYING_COMPONENT_USE_COMP15__SHIFT) & VIVS_GL_VARYING_COMPONENT_USE_COMP15__MASK)
|
||||
|
||||
#define VIVS_GL_UNK03834 0x00003834
|
||||
|
||||
#define VIVS_GL_UNK03838 0x00003838
|
||||
|
||||
#define VIVS_GL_API_MODE 0x0000384c
|
||||
#define VIVS_GL_API_MODE_OPENGL 0x00000000
|
||||
#define VIVS_GL_API_MODE_OPENVG 0x00000001
|
||||
#define VIVS_GL_API_MODE_OPENCL 0x00000002
|
||||
|
||||
#define VIVS_GL_CONTEXT_POINTER 0x00003850
|
||||
|
||||
#define VIVS_GL_UNK03A00 0x00003a00
|
||||
|
||||
#define VIVS_GL_STALL_TOKEN 0x00003c00
|
||||
#define VIVS_GL_STALL_TOKEN_FROM__MASK 0x0000001f
|
||||
#define VIVS_GL_STALL_TOKEN_FROM__SHIFT 0
|
||||
#define VIVS_GL_STALL_TOKEN_FROM(x) (((x) << VIVS_GL_STALL_TOKEN_FROM__SHIFT) & VIVS_GL_STALL_TOKEN_FROM__MASK)
|
||||
#define VIVS_GL_STALL_TOKEN_TO__MASK 0x00001f00
|
||||
#define VIVS_GL_STALL_TOKEN_TO__SHIFT 8
|
||||
#define VIVS_GL_STALL_TOKEN_TO(x) (((x) << VIVS_GL_STALL_TOKEN_TO__SHIFT) & VIVS_GL_STALL_TOKEN_TO__MASK)
|
||||
#define VIVS_GL_STALL_TOKEN_FLIP0 0x40000000
|
||||
#define VIVS_GL_STALL_TOKEN_FLIP1 0x80000000
|
||||
|
||||
#define VIVS_DUMMY 0x00000000
|
||||
|
||||
#define VIVS_DUMMY_DUMMY 0x0003fffc
|
||||
|
||||
|
||||
#endif /* STATE_XML */
|
407
drivers/gpu/drm/etnaviv/state_hi.xml.h
Normal file
407
drivers/gpu/drm/etnaviv/state_hi.xml.h
Normal file
|
@ -0,0 +1,407 @@
|
|||
#ifndef STATE_HI_XML
|
||||
#define STATE_HI_XML
|
||||
|
||||
/* Autogenerated file, DO NOT EDIT manually!
|
||||
|
||||
This file was generated by the rules-ng-ng headergen tool in this git repository:
|
||||
http://0x04.net/cgit/index.cgi/rules-ng-ng
|
||||
git clone git://0x04.net/rules-ng-ng
|
||||
|
||||
The rules-ng-ng source files this header was generated from are:
|
||||
- state_hi.xml ( 23420 bytes, from 2015-03-25 11:47:21)
|
||||
- common.xml ( 18437 bytes, from 2015-03-25 11:27:41)
|
||||
|
||||
Copyright (C) 2015
|
||||
*/
|
||||
|
||||
|
||||
#define MMU_EXCEPTION_SLAVE_NOT_PRESENT 0x00000001
|
||||
#define MMU_EXCEPTION_PAGE_NOT_PRESENT 0x00000002
|
||||
#define MMU_EXCEPTION_WRITE_VIOLATION 0x00000003
|
||||
#define VIVS_HI 0x00000000
|
||||
|
||||
#define VIVS_HI_CLOCK_CONTROL 0x00000000
|
||||
#define VIVS_HI_CLOCK_CONTROL_CLK3D_DIS 0x00000001
|
||||
#define VIVS_HI_CLOCK_CONTROL_CLK2D_DIS 0x00000002
|
||||
#define VIVS_HI_CLOCK_CONTROL_FSCALE_VAL__MASK 0x000001fc
|
||||
#define VIVS_HI_CLOCK_CONTROL_FSCALE_VAL__SHIFT 2
|
||||
#define VIVS_HI_CLOCK_CONTROL_FSCALE_VAL(x) (((x) << VIVS_HI_CLOCK_CONTROL_FSCALE_VAL__SHIFT) & VIVS_HI_CLOCK_CONTROL_FSCALE_VAL__MASK)
|
||||
#define VIVS_HI_CLOCK_CONTROL_FSCALE_CMD_LOAD 0x00000200
|
||||
#define VIVS_HI_CLOCK_CONTROL_DISABLE_RAM_CLK_GATING 0x00000400
|
||||
#define VIVS_HI_CLOCK_CONTROL_DISABLE_DEBUG_REGISTERS 0x00000800
|
||||
#define VIVS_HI_CLOCK_CONTROL_SOFT_RESET 0x00001000
|
||||
#define VIVS_HI_CLOCK_CONTROL_IDLE_3D 0x00010000
|
||||
#define VIVS_HI_CLOCK_CONTROL_IDLE_2D 0x00020000
|
||||
#define VIVS_HI_CLOCK_CONTROL_IDLE_VG 0x00040000
|
||||
#define VIVS_HI_CLOCK_CONTROL_ISOLATE_GPU 0x00080000
|
||||
#define VIVS_HI_CLOCK_CONTROL_DEBUG_PIXEL_PIPE__MASK 0x00f00000
|
||||
#define VIVS_HI_CLOCK_CONTROL_DEBUG_PIXEL_PIPE__SHIFT 20
|
||||
#define VIVS_HI_CLOCK_CONTROL_DEBUG_PIXEL_PIPE(x) (((x) << VIVS_HI_CLOCK_CONTROL_DEBUG_PIXEL_PIPE__SHIFT) & VIVS_HI_CLOCK_CONTROL_DEBUG_PIXEL_PIPE__MASK)
|
||||
|
||||
#define VIVS_HI_IDLE_STATE 0x00000004
|
||||
#define VIVS_HI_IDLE_STATE_FE 0x00000001
|
||||
#define VIVS_HI_IDLE_STATE_DE 0x00000002
|
||||
#define VIVS_HI_IDLE_STATE_PE 0x00000004
|
||||
#define VIVS_HI_IDLE_STATE_SH 0x00000008
|
||||
#define VIVS_HI_IDLE_STATE_PA 0x00000010
|
||||
#define VIVS_HI_IDLE_STATE_SE 0x00000020
|
||||
#define VIVS_HI_IDLE_STATE_RA 0x00000040
|
||||
#define VIVS_HI_IDLE_STATE_TX 0x00000080
|
||||
#define VIVS_HI_IDLE_STATE_VG 0x00000100
|
||||
#define VIVS_HI_IDLE_STATE_IM 0x00000200
|
||||
#define VIVS_HI_IDLE_STATE_FP 0x00000400
|
||||
#define VIVS_HI_IDLE_STATE_TS 0x00000800
|
||||
#define VIVS_HI_IDLE_STATE_AXI_LP 0x80000000
|
||||
|
||||
#define VIVS_HI_AXI_CONFIG 0x00000008
|
||||
#define VIVS_HI_AXI_CONFIG_AWID__MASK 0x0000000f
|
||||
#define VIVS_HI_AXI_CONFIG_AWID__SHIFT 0
|
||||
#define VIVS_HI_AXI_CONFIG_AWID(x) (((x) << VIVS_HI_AXI_CONFIG_AWID__SHIFT) & VIVS_HI_AXI_CONFIG_AWID__MASK)
|
||||
#define VIVS_HI_AXI_CONFIG_ARID__MASK 0x000000f0
|
||||
#define VIVS_HI_AXI_CONFIG_ARID__SHIFT 4
|
||||
#define VIVS_HI_AXI_CONFIG_ARID(x) (((x) << VIVS_HI_AXI_CONFIG_ARID__SHIFT) & VIVS_HI_AXI_CONFIG_ARID__MASK)
|
||||
#define VIVS_HI_AXI_CONFIG_AWCACHE__MASK 0x00000f00
|
||||
#define VIVS_HI_AXI_CONFIG_AWCACHE__SHIFT 8
|
||||
#define VIVS_HI_AXI_CONFIG_AWCACHE(x) (((x) << VIVS_HI_AXI_CONFIG_AWCACHE__SHIFT) & VIVS_HI_AXI_CONFIG_AWCACHE__MASK)
|
||||
#define VIVS_HI_AXI_CONFIG_ARCACHE__MASK 0x0000f000
|
||||
#define VIVS_HI_AXI_CONFIG_ARCACHE__SHIFT 12
|
||||
#define VIVS_HI_AXI_CONFIG_ARCACHE(x) (((x) << VIVS_HI_AXI_CONFIG_ARCACHE__SHIFT) & VIVS_HI_AXI_CONFIG_ARCACHE__MASK)
|
||||
|
||||
#define VIVS_HI_AXI_STATUS 0x0000000c
|
||||
#define VIVS_HI_AXI_STATUS_WR_ERR_ID__MASK 0x0000000f
|
||||
#define VIVS_HI_AXI_STATUS_WR_ERR_ID__SHIFT 0
|
||||
#define VIVS_HI_AXI_STATUS_WR_ERR_ID(x) (((x) << VIVS_HI_AXI_STATUS_WR_ERR_ID__SHIFT) & VIVS_HI_AXI_STATUS_WR_ERR_ID__MASK)
|
||||
#define VIVS_HI_AXI_STATUS_RD_ERR_ID__MASK 0x000000f0
|
||||
#define VIVS_HI_AXI_STATUS_RD_ERR_ID__SHIFT 4
|
||||
#define VIVS_HI_AXI_STATUS_RD_ERR_ID(x) (((x) << VIVS_HI_AXI_STATUS_RD_ERR_ID__SHIFT) & VIVS_HI_AXI_STATUS_RD_ERR_ID__MASK)
|
||||
#define VIVS_HI_AXI_STATUS_DET_WR_ERR 0x00000100
|
||||
#define VIVS_HI_AXI_STATUS_DET_RD_ERR 0x00000200
|
||||
|
||||
#define VIVS_HI_INTR_ACKNOWLEDGE 0x00000010
|
||||
#define VIVS_HI_INTR_ACKNOWLEDGE_INTR_VEC__MASK 0x7fffffff
|
||||
#define VIVS_HI_INTR_ACKNOWLEDGE_INTR_VEC__SHIFT 0
|
||||
#define VIVS_HI_INTR_ACKNOWLEDGE_INTR_VEC(x) (((x) << VIVS_HI_INTR_ACKNOWLEDGE_INTR_VEC__SHIFT) & VIVS_HI_INTR_ACKNOWLEDGE_INTR_VEC__MASK)
|
||||
#define VIVS_HI_INTR_ACKNOWLEDGE_AXI_BUS_ERROR 0x80000000
|
||||
|
||||
#define VIVS_HI_INTR_ENBL 0x00000014
|
||||
#define VIVS_HI_INTR_ENBL_INTR_ENBL_VEC__MASK 0xffffffff
|
||||
#define VIVS_HI_INTR_ENBL_INTR_ENBL_VEC__SHIFT 0
|
||||
#define VIVS_HI_INTR_ENBL_INTR_ENBL_VEC(x) (((x) << VIVS_HI_INTR_ENBL_INTR_ENBL_VEC__SHIFT) & VIVS_HI_INTR_ENBL_INTR_ENBL_VEC__MASK)
|
||||
|
||||
#define VIVS_HI_CHIP_IDENTITY 0x00000018
|
||||
#define VIVS_HI_CHIP_IDENTITY_FAMILY__MASK 0xff000000
|
||||
#define VIVS_HI_CHIP_IDENTITY_FAMILY__SHIFT 24
|
||||
#define VIVS_HI_CHIP_IDENTITY_FAMILY(x) (((x) << VIVS_HI_CHIP_IDENTITY_FAMILY__SHIFT) & VIVS_HI_CHIP_IDENTITY_FAMILY__MASK)
|
||||
#define VIVS_HI_CHIP_IDENTITY_PRODUCT__MASK 0x00ff0000
|
||||
#define VIVS_HI_CHIP_IDENTITY_PRODUCT__SHIFT 16
|
||||
#define VIVS_HI_CHIP_IDENTITY_PRODUCT(x) (((x) << VIVS_HI_CHIP_IDENTITY_PRODUCT__SHIFT) & VIVS_HI_CHIP_IDENTITY_PRODUCT__MASK)
|
||||
#define VIVS_HI_CHIP_IDENTITY_REVISION__MASK 0x0000f000
|
||||
#define VIVS_HI_CHIP_IDENTITY_REVISION__SHIFT 12
|
||||
#define VIVS_HI_CHIP_IDENTITY_REVISION(x) (((x) << VIVS_HI_CHIP_IDENTITY_REVISION__SHIFT) & VIVS_HI_CHIP_IDENTITY_REVISION__MASK)
|
||||
|
||||
#define VIVS_HI_CHIP_FEATURE 0x0000001c
|
||||
|
||||
#define VIVS_HI_CHIP_MODEL 0x00000020
|
||||
|
||||
#define VIVS_HI_CHIP_REV 0x00000024
|
||||
|
||||
#define VIVS_HI_CHIP_DATE 0x00000028
|
||||
|
||||
#define VIVS_HI_CHIP_TIME 0x0000002c
|
||||
|
||||
#define VIVS_HI_CHIP_MINOR_FEATURE_0 0x00000034
|
||||
|
||||
#define VIVS_HI_CACHE_CONTROL 0x00000038
|
||||
|
||||
#define VIVS_HI_MEMORY_COUNTER_RESET 0x0000003c
|
||||
|
||||
#define VIVS_HI_PROFILE_READ_BYTES8 0x00000040
|
||||
|
||||
#define VIVS_HI_PROFILE_WRITE_BYTES8 0x00000044
|
||||
|
||||
#define VIVS_HI_CHIP_SPECS 0x00000048
|
||||
#define VIVS_HI_CHIP_SPECS_STREAM_COUNT__MASK 0x0000000f
|
||||
#define VIVS_HI_CHIP_SPECS_STREAM_COUNT__SHIFT 0
|
||||
#define VIVS_HI_CHIP_SPECS_STREAM_COUNT(x) (((x) << VIVS_HI_CHIP_SPECS_STREAM_COUNT__SHIFT) & VIVS_HI_CHIP_SPECS_STREAM_COUNT__MASK)
|
||||
#define VIVS_HI_CHIP_SPECS_REGISTER_MAX__MASK 0x000000f0
|
||||
#define VIVS_HI_CHIP_SPECS_REGISTER_MAX__SHIFT 4
|
||||
#define VIVS_HI_CHIP_SPECS_REGISTER_MAX(x) (((x) << VIVS_HI_CHIP_SPECS_REGISTER_MAX__SHIFT) & VIVS_HI_CHIP_SPECS_REGISTER_MAX__MASK)
|
||||
#define VIVS_HI_CHIP_SPECS_THREAD_COUNT__MASK 0x00000f00
|
||||
#define VIVS_HI_CHIP_SPECS_THREAD_COUNT__SHIFT 8
|
||||
#define VIVS_HI_CHIP_SPECS_THREAD_COUNT(x) (((x) << VIVS_HI_CHIP_SPECS_THREAD_COUNT__SHIFT) & VIVS_HI_CHIP_SPECS_THREAD_COUNT__MASK)
|
||||
#define VIVS_HI_CHIP_SPECS_VERTEX_CACHE_SIZE__MASK 0x0001f000
|
||||
#define VIVS_HI_CHIP_SPECS_VERTEX_CACHE_SIZE__SHIFT 12
|
||||
#define VIVS_HI_CHIP_SPECS_VERTEX_CACHE_SIZE(x) (((x) << VIVS_HI_CHIP_SPECS_VERTEX_CACHE_SIZE__SHIFT) & VIVS_HI_CHIP_SPECS_VERTEX_CACHE_SIZE__MASK)
|
||||
#define VIVS_HI_CHIP_SPECS_SHADER_CORE_COUNT__MASK 0x01f00000
|
||||
#define VIVS_HI_CHIP_SPECS_SHADER_CORE_COUNT__SHIFT 20
|
||||
#define VIVS_HI_CHIP_SPECS_SHADER_CORE_COUNT(x) (((x) << VIVS_HI_CHIP_SPECS_SHADER_CORE_COUNT__SHIFT) & VIVS_HI_CHIP_SPECS_SHADER_CORE_COUNT__MASK)
|
||||
#define VIVS_HI_CHIP_SPECS_PIXEL_PIPES__MASK 0x0e000000
|
||||
#define VIVS_HI_CHIP_SPECS_PIXEL_PIPES__SHIFT 25
|
||||
#define VIVS_HI_CHIP_SPECS_PIXEL_PIPES(x) (((x) << VIVS_HI_CHIP_SPECS_PIXEL_PIPES__SHIFT) & VIVS_HI_CHIP_SPECS_PIXEL_PIPES__MASK)
|
||||
#define VIVS_HI_CHIP_SPECS_VERTEX_OUTPUT_BUFFER_SIZE__MASK 0xf0000000
|
||||
#define VIVS_HI_CHIP_SPECS_VERTEX_OUTPUT_BUFFER_SIZE__SHIFT 28
|
||||
#define VIVS_HI_CHIP_SPECS_VERTEX_OUTPUT_BUFFER_SIZE(x) (((x) << VIVS_HI_CHIP_SPECS_VERTEX_OUTPUT_BUFFER_SIZE__SHIFT) & VIVS_HI_CHIP_SPECS_VERTEX_OUTPUT_BUFFER_SIZE__MASK)
|
||||
|
||||
#define VIVS_HI_PROFILE_WRITE_BURSTS 0x0000004c
|
||||
|
||||
#define VIVS_HI_PROFILE_WRITE_REQUESTS 0x00000050
|
||||
|
||||
#define VIVS_HI_PROFILE_READ_BURSTS 0x00000058
|
||||
|
||||
#define VIVS_HI_PROFILE_READ_REQUESTS 0x0000005c
|
||||
|
||||
#define VIVS_HI_PROFILE_READ_LASTS 0x00000060
|
||||
|
||||
#define VIVS_HI_GP_OUT0 0x00000064
|
||||
|
||||
#define VIVS_HI_GP_OUT1 0x00000068
|
||||
|
||||
#define VIVS_HI_GP_OUT2 0x0000006c
|
||||
|
||||
#define VIVS_HI_AXI_CONTROL 0x00000070
|
||||
#define VIVS_HI_AXI_CONTROL_WR_FULL_BURST_MODE 0x00000001
|
||||
|
||||
#define VIVS_HI_CHIP_MINOR_FEATURE_1 0x00000074
|
||||
|
||||
#define VIVS_HI_PROFILE_TOTAL_CYCLES 0x00000078
|
||||
|
||||
#define VIVS_HI_PROFILE_IDLE_CYCLES 0x0000007c
|
||||
|
||||
#define VIVS_HI_CHIP_SPECS_2 0x00000080
|
||||
#define VIVS_HI_CHIP_SPECS_2_BUFFER_SIZE__MASK 0x000000ff
|
||||
#define VIVS_HI_CHIP_SPECS_2_BUFFER_SIZE__SHIFT 0
|
||||
#define VIVS_HI_CHIP_SPECS_2_BUFFER_SIZE(x) (((x) << VIVS_HI_CHIP_SPECS_2_BUFFER_SIZE__SHIFT) & VIVS_HI_CHIP_SPECS_2_BUFFER_SIZE__MASK)
|
||||
#define VIVS_HI_CHIP_SPECS_2_INSTRUCTION_COUNT__MASK 0x0000ff00
|
||||
#define VIVS_HI_CHIP_SPECS_2_INSTRUCTION_COUNT__SHIFT 8
|
||||
#define VIVS_HI_CHIP_SPECS_2_INSTRUCTION_COUNT(x) (((x) << VIVS_HI_CHIP_SPECS_2_INSTRUCTION_COUNT__SHIFT) & VIVS_HI_CHIP_SPECS_2_INSTRUCTION_COUNT__MASK)
|
||||
#define VIVS_HI_CHIP_SPECS_2_NUM_CONSTANTS__MASK 0xffff0000
|
||||
#define VIVS_HI_CHIP_SPECS_2_NUM_CONSTANTS__SHIFT 16
|
||||
#define VIVS_HI_CHIP_SPECS_2_NUM_CONSTANTS(x) (((x) << VIVS_HI_CHIP_SPECS_2_NUM_CONSTANTS__SHIFT) & VIVS_HI_CHIP_SPECS_2_NUM_CONSTANTS__MASK)
|
||||
|
||||
#define VIVS_HI_CHIP_MINOR_FEATURE_2 0x00000084
|
||||
|
||||
#define VIVS_HI_CHIP_MINOR_FEATURE_3 0x00000088
|
||||
|
||||
#define VIVS_HI_CHIP_MINOR_FEATURE_4 0x00000094
|
||||
|
||||
#define VIVS_PM 0x00000000
|
||||
|
||||
#define VIVS_PM_POWER_CONTROLS 0x00000100
|
||||
#define VIVS_PM_POWER_CONTROLS_ENABLE_MODULE_CLOCK_GATING 0x00000001
|
||||
#define VIVS_PM_POWER_CONTROLS_DISABLE_STALL_MODULE_CLOCK_GATING 0x00000002
|
||||
#define VIVS_PM_POWER_CONTROLS_DISABLE_STARVE_MODULE_CLOCK_GATING 0x00000004
|
||||
#define VIVS_PM_POWER_CONTROLS_TURN_ON_COUNTER__MASK 0x000000f0
|
||||
#define VIVS_PM_POWER_CONTROLS_TURN_ON_COUNTER__SHIFT 4
|
||||
#define VIVS_PM_POWER_CONTROLS_TURN_ON_COUNTER(x) (((x) << VIVS_PM_POWER_CONTROLS_TURN_ON_COUNTER__SHIFT) & VIVS_PM_POWER_CONTROLS_TURN_ON_COUNTER__MASK)
|
||||
#define VIVS_PM_POWER_CONTROLS_TURN_OFF_COUNTER__MASK 0xffff0000
|
||||
#define VIVS_PM_POWER_CONTROLS_TURN_OFF_COUNTER__SHIFT 16
|
||||
#define VIVS_PM_POWER_CONTROLS_TURN_OFF_COUNTER(x) (((x) << VIVS_PM_POWER_CONTROLS_TURN_OFF_COUNTER__SHIFT) & VIVS_PM_POWER_CONTROLS_TURN_OFF_COUNTER__MASK)
|
||||
|
||||
#define VIVS_PM_MODULE_CONTROLS 0x00000104
|
||||
#define VIVS_PM_MODULE_CONTROLS_DISABLE_MODULE_CLOCK_GATING_FE 0x00000001
|
||||
#define VIVS_PM_MODULE_CONTROLS_DISABLE_MODULE_CLOCK_GATING_DE 0x00000002
|
||||
#define VIVS_PM_MODULE_CONTROLS_DISABLE_MODULE_CLOCK_GATING_PE 0x00000004
|
||||
|
||||
#define VIVS_PM_MODULE_STATUS 0x00000108
|
||||
#define VIVS_PM_MODULE_STATUS_MODULE_CLOCK_GATED_FE 0x00000001
|
||||
#define VIVS_PM_MODULE_STATUS_MODULE_CLOCK_GATED_DE 0x00000002
|
||||
#define VIVS_PM_MODULE_STATUS_MODULE_CLOCK_GATED_PE 0x00000004
|
||||
|
||||
#define VIVS_PM_PULSE_EATER 0x0000010c
|
||||
|
||||
#define VIVS_MMUv2 0x00000000
|
||||
|
||||
#define VIVS_MMUv2_SAFE_ADDRESS 0x00000180
|
||||
|
||||
#define VIVS_MMUv2_CONFIGURATION 0x00000184
|
||||
#define VIVS_MMUv2_CONFIGURATION_MODE__MASK 0x00000001
|
||||
#define VIVS_MMUv2_CONFIGURATION_MODE__SHIFT 0
|
||||
#define VIVS_MMUv2_CONFIGURATION_MODE_MODE4_K 0x00000000
|
||||
#define VIVS_MMUv2_CONFIGURATION_MODE_MODE1_K 0x00000001
|
||||
#define VIVS_MMUv2_CONFIGURATION_MODE_MASK 0x00000008
|
||||
#define VIVS_MMUv2_CONFIGURATION_FLUSH__MASK 0x00000010
|
||||
#define VIVS_MMUv2_CONFIGURATION_FLUSH__SHIFT 4
|
||||
#define VIVS_MMUv2_CONFIGURATION_FLUSH_FLUSH 0x00000010
|
||||
#define VIVS_MMUv2_CONFIGURATION_FLUSH_MASK 0x00000080
|
||||
#define VIVS_MMUv2_CONFIGURATION_ADDRESS_MASK 0x00000100
|
||||
#define VIVS_MMUv2_CONFIGURATION_ADDRESS__MASK 0xfffffc00
|
||||
#define VIVS_MMUv2_CONFIGURATION_ADDRESS__SHIFT 10
|
||||
#define VIVS_MMUv2_CONFIGURATION_ADDRESS(x) (((x) << VIVS_MMUv2_CONFIGURATION_ADDRESS__SHIFT) & VIVS_MMUv2_CONFIGURATION_ADDRESS__MASK)
|
||||
|
||||
#define VIVS_MMUv2_STATUS 0x00000188
|
||||
#define VIVS_MMUv2_STATUS_EXCEPTION0__MASK 0x00000003
|
||||
#define VIVS_MMUv2_STATUS_EXCEPTION0__SHIFT 0
|
||||
#define VIVS_MMUv2_STATUS_EXCEPTION0(x) (((x) << VIVS_MMUv2_STATUS_EXCEPTION0__SHIFT) & VIVS_MMUv2_STATUS_EXCEPTION0__MASK)
|
||||
#define VIVS_MMUv2_STATUS_EXCEPTION1__MASK 0x00000030
|
||||
#define VIVS_MMUv2_STATUS_EXCEPTION1__SHIFT 4
|
||||
#define VIVS_MMUv2_STATUS_EXCEPTION1(x) (((x) << VIVS_MMUv2_STATUS_EXCEPTION1__SHIFT) & VIVS_MMUv2_STATUS_EXCEPTION1__MASK)
|
||||
#define VIVS_MMUv2_STATUS_EXCEPTION2__MASK 0x00000300
|
||||
#define VIVS_MMUv2_STATUS_EXCEPTION2__SHIFT 8
|
||||
#define VIVS_MMUv2_STATUS_EXCEPTION2(x) (((x) << VIVS_MMUv2_STATUS_EXCEPTION2__SHIFT) & VIVS_MMUv2_STATUS_EXCEPTION2__MASK)
|
||||
#define VIVS_MMUv2_STATUS_EXCEPTION3__MASK 0x00003000
|
||||
#define VIVS_MMUv2_STATUS_EXCEPTION3__SHIFT 12
|
||||
#define VIVS_MMUv2_STATUS_EXCEPTION3(x) (((x) << VIVS_MMUv2_STATUS_EXCEPTION3__SHIFT) & VIVS_MMUv2_STATUS_EXCEPTION3__MASK)
|
||||
|
||||
#define VIVS_MMUv2_CONTROL 0x0000018c
|
||||
#define VIVS_MMUv2_CONTROL_ENABLE 0x00000001
|
||||
|
||||
#define VIVS_MMUv2_EXCEPTION_ADDR(i0) (0x00000190 + 0x4*(i0))
|
||||
#define VIVS_MMUv2_EXCEPTION_ADDR__ESIZE 0x00000004
|
||||
#define VIVS_MMUv2_EXCEPTION_ADDR__LEN 0x00000004
|
||||
|
||||
#define VIVS_MC 0x00000000
|
||||
|
||||
#define VIVS_MC_MMU_FE_PAGE_TABLE 0x00000400
|
||||
|
||||
#define VIVS_MC_MMU_TX_PAGE_TABLE 0x00000404
|
||||
|
||||
#define VIVS_MC_MMU_PE_PAGE_TABLE 0x00000408
|
||||
|
||||
#define VIVS_MC_MMU_PEZ_PAGE_TABLE 0x0000040c
|
||||
|
||||
#define VIVS_MC_MMU_RA_PAGE_TABLE 0x00000410
|
||||
|
||||
#define VIVS_MC_DEBUG_MEMORY 0x00000414
|
||||
#define VIVS_MC_DEBUG_MEMORY_SPECIAL_PATCH_GC320 0x00000008
|
||||
#define VIVS_MC_DEBUG_MEMORY_FAST_CLEAR_BYPASS 0x00100000
|
||||
#define VIVS_MC_DEBUG_MEMORY_COMPRESSION_BYPASS 0x00200000
|
||||
|
||||
#define VIVS_MC_MEMORY_BASE_ADDR_RA 0x00000418
|
||||
|
||||
#define VIVS_MC_MEMORY_BASE_ADDR_FE 0x0000041c
|
||||
|
||||
#define VIVS_MC_MEMORY_BASE_ADDR_TX 0x00000420
|
||||
|
||||
#define VIVS_MC_MEMORY_BASE_ADDR_PEZ 0x00000424
|
||||
|
||||
#define VIVS_MC_MEMORY_BASE_ADDR_PE 0x00000428
|
||||
|
||||
#define VIVS_MC_MEMORY_TIMING_CONTROL 0x0000042c
|
||||
|
||||
#define VIVS_MC_MEMORY_FLUSH 0x00000430
|
||||
|
||||
#define VIVS_MC_PROFILE_CYCLE_COUNTER 0x00000438
|
||||
|
||||
#define VIVS_MC_DEBUG_READ0 0x0000043c
|
||||
|
||||
#define VIVS_MC_DEBUG_READ1 0x00000440
|
||||
|
||||
#define VIVS_MC_DEBUG_WRITE 0x00000444
|
||||
|
||||
#define VIVS_MC_PROFILE_RA_READ 0x00000448
|
||||
|
||||
#define VIVS_MC_PROFILE_TX_READ 0x0000044c
|
||||
|
||||
#define VIVS_MC_PROFILE_FE_READ 0x00000450
|
||||
|
||||
#define VIVS_MC_PROFILE_PE_READ 0x00000454
|
||||
|
||||
#define VIVS_MC_PROFILE_DE_READ 0x00000458
|
||||
|
||||
#define VIVS_MC_PROFILE_SH_READ 0x0000045c
|
||||
|
||||
#define VIVS_MC_PROFILE_PA_READ 0x00000460
|
||||
|
||||
#define VIVS_MC_PROFILE_SE_READ 0x00000464
|
||||
|
||||
#define VIVS_MC_PROFILE_MC_READ 0x00000468
|
||||
|
||||
#define VIVS_MC_PROFILE_HI_READ 0x0000046c
|
||||
|
||||
#define VIVS_MC_PROFILE_CONFIG0 0x00000470
|
||||
#define VIVS_MC_PROFILE_CONFIG0_FE__MASK 0x0000000f
|
||||
#define VIVS_MC_PROFILE_CONFIG0_FE__SHIFT 0
|
||||
#define VIVS_MC_PROFILE_CONFIG0_FE_RESET 0x0000000f
|
||||
#define VIVS_MC_PROFILE_CONFIG0_DE__MASK 0x00000f00
|
||||
#define VIVS_MC_PROFILE_CONFIG0_DE__SHIFT 8
|
||||
#define VIVS_MC_PROFILE_CONFIG0_DE_RESET 0x00000f00
|
||||
#define VIVS_MC_PROFILE_CONFIG0_PE__MASK 0x000f0000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_PE__SHIFT 16
|
||||
#define VIVS_MC_PROFILE_CONFIG0_PE_PIXEL_COUNT_KILLED_BY_COLOR_PIPE 0x00000000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_PE_PIXEL_COUNT_KILLED_BY_DEPTH_PIPE 0x00010000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_PE_PIXEL_COUNT_DRAWN_BY_COLOR_PIPE 0x00020000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_PE_PIXEL_COUNT_DRAWN_BY_DEPTH_PIPE 0x00030000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_PE_PIXELS_RENDERED_2D 0x000b0000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_PE_RESET 0x000f0000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_SH__MASK 0x0f000000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_SH__SHIFT 24
|
||||
#define VIVS_MC_PROFILE_CONFIG0_SH_SHADER_CYCLES 0x04000000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_SH_PS_INST_COUNTER 0x07000000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_SH_RENDERED_PIXEL_COUNTER 0x08000000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_SH_VS_INST_COUNTER 0x09000000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_SH_RENDERED_VERTICE_COUNTER 0x0a000000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_SH_VTX_BRANCH_INST_COUNTER 0x0b000000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_SH_VTX_TEXLD_INST_COUNTER 0x0c000000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_SH_PXL_BRANCH_INST_COUNTER 0x0d000000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_SH_PXL_TEXLD_INST_COUNTER 0x0e000000
|
||||
#define VIVS_MC_PROFILE_CONFIG0_SH_RESET 0x0f000000
|
||||
|
||||
#define VIVS_MC_PROFILE_CONFIG1 0x00000474
|
||||
#define VIVS_MC_PROFILE_CONFIG1_PA__MASK 0x0000000f
|
||||
#define VIVS_MC_PROFILE_CONFIG1_PA__SHIFT 0
|
||||
#define VIVS_MC_PROFILE_CONFIG1_PA_INPUT_VTX_COUNTER 0x00000003
|
||||
#define VIVS_MC_PROFILE_CONFIG1_PA_INPUT_PRIM_COUNTER 0x00000004
|
||||
#define VIVS_MC_PROFILE_CONFIG1_PA_OUTPUT_PRIM_COUNTER 0x00000005
|
||||
#define VIVS_MC_PROFILE_CONFIG1_PA_DEPTH_CLIPPED_COUNTER 0x00000006
|
||||
#define VIVS_MC_PROFILE_CONFIG1_PA_TRIVIAL_REJECTED_COUNTER 0x00000007
|
||||
#define VIVS_MC_PROFILE_CONFIG1_PA_CULLED_COUNTER 0x00000008
|
||||
#define VIVS_MC_PROFILE_CONFIG1_PA_RESET 0x0000000f
|
||||
#define VIVS_MC_PROFILE_CONFIG1_SE__MASK 0x00000f00
|
||||
#define VIVS_MC_PROFILE_CONFIG1_SE__SHIFT 8
|
||||
#define VIVS_MC_PROFILE_CONFIG1_SE_CULLED_TRIANGLE_COUNT 0x00000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_SE_CULLED_LINES_COUNT 0x00000100
|
||||
#define VIVS_MC_PROFILE_CONFIG1_SE_RESET 0x00000f00
|
||||
#define VIVS_MC_PROFILE_CONFIG1_RA__MASK 0x000f0000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_RA__SHIFT 16
|
||||
#define VIVS_MC_PROFILE_CONFIG1_RA_VALID_PIXEL_COUNT 0x00000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_RA_TOTAL_QUAD_COUNT 0x00010000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_RA_VALID_QUAD_COUNT_AFTER_EARLY_Z 0x00020000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_RA_TOTAL_PRIMITIVE_COUNT 0x00030000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_RA_PIPE_CACHE_MISS_COUNTER 0x00090000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_RA_PREFETCH_CACHE_MISS_COUNTER 0x000a0000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_RA_CULLED_QUAD_COUNT 0x000b0000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_RA_RESET 0x000f0000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX__MASK 0x0f000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX__SHIFT 24
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX_TOTAL_BILINEAR_REQUESTS 0x00000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX_TOTAL_TRILINEAR_REQUESTS 0x01000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX_TOTAL_DISCARDED_TEXTURE_REQUESTS 0x02000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX_TOTAL_TEXTURE_REQUESTS 0x03000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX_UNKNOWN 0x04000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX_MEM_READ_COUNT 0x05000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX_MEM_READ_IN_8B_COUNT 0x06000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX_CACHE_MISS_COUNT 0x07000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX_CACHE_HIT_TEXEL_COUNT 0x08000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX_CACHE_MISS_TEXEL_COUNT 0x09000000
|
||||
#define VIVS_MC_PROFILE_CONFIG1_TX_RESET 0x0f000000
|
||||
|
||||
#define VIVS_MC_PROFILE_CONFIG2 0x00000478
|
||||
#define VIVS_MC_PROFILE_CONFIG2_MC__MASK 0x0000000f
|
||||
#define VIVS_MC_PROFILE_CONFIG2_MC__SHIFT 0
|
||||
#define VIVS_MC_PROFILE_CONFIG2_MC_TOTAL_READ_REQ_8B_FROM_PIPELINE 0x00000001
|
||||
#define VIVS_MC_PROFILE_CONFIG2_MC_TOTAL_READ_REQ_8B_FROM_IP 0x00000002
|
||||
#define VIVS_MC_PROFILE_CONFIG2_MC_TOTAL_WRITE_REQ_8B_FROM_PIPELINE 0x00000003
|
||||
#define VIVS_MC_PROFILE_CONFIG2_MC_RESET 0x0000000f
|
||||
#define VIVS_MC_PROFILE_CONFIG2_HI__MASK 0x00000f00
|
||||
#define VIVS_MC_PROFILE_CONFIG2_HI__SHIFT 8
|
||||
#define VIVS_MC_PROFILE_CONFIG2_HI_AXI_CYCLES_READ_REQUEST_STALLED 0x00000000
|
||||
#define VIVS_MC_PROFILE_CONFIG2_HI_AXI_CYCLES_WRITE_REQUEST_STALLED 0x00000100
|
||||
#define VIVS_MC_PROFILE_CONFIG2_HI_AXI_CYCLES_WRITE_DATA_STALLED 0x00000200
|
||||
#define VIVS_MC_PROFILE_CONFIG2_HI_RESET 0x00000f00
|
||||
|
||||
#define VIVS_MC_PROFILE_CONFIG3 0x0000047c
|
||||
|
||||
#define VIVS_MC_BUS_CONFIG 0x00000480
|
||||
#define VIVS_MC_BUS_CONFIG_FE_BUS_CONFIG__MASK 0x0000000f
|
||||
#define VIVS_MC_BUS_CONFIG_FE_BUS_CONFIG__SHIFT 0
|
||||
#define VIVS_MC_BUS_CONFIG_FE_BUS_CONFIG(x) (((x) << VIVS_MC_BUS_CONFIG_FE_BUS_CONFIG__SHIFT) & VIVS_MC_BUS_CONFIG_FE_BUS_CONFIG__MASK)
|
||||
#define VIVS_MC_BUS_CONFIG_TX_BUS_CONFIG__MASK 0x000000f0
|
||||
#define VIVS_MC_BUS_CONFIG_TX_BUS_CONFIG__SHIFT 4
|
||||
#define VIVS_MC_BUS_CONFIG_TX_BUS_CONFIG(x) (((x) << VIVS_MC_BUS_CONFIG_TX_BUS_CONFIG__SHIFT) & VIVS_MC_BUS_CONFIG_TX_BUS_CONFIG__MASK)
|
||||
|
||||
#define VIVS_MC_START_COMPOSITION 0x00000554
|
||||
|
||||
#define VIVS_MC_128B_MERGE 0x00000558
|
||||
|
||||
|
||||
#endif /* STATE_HI_XML */
|
222
include/uapi/drm/etnaviv_drm.h
Normal file
222
include/uapi/drm/etnaviv_drm.h
Normal file
|
@ -0,0 +1,222 @@
|
|||
/*
|
||||
* Copyright (C) 2015 Etnaviv Project
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms of the GNU General Public License version 2 as published by
|
||||
* the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope that it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
*/
|
||||
|
||||
#ifndef __ETNAVIV_DRM_H__
|
||||
#define __ETNAVIV_DRM_H__
|
||||
|
||||
#include "drm.h"
|
||||
|
||||
/* Please note that modifications to all structs defined here are
|
||||
* subject to backwards-compatibility constraints:
|
||||
* 1) Do not use pointers, use __u64 instead for 32 bit / 64 bit
|
||||
* user/kernel compatibility
|
||||
* 2) Keep fields aligned to their size
|
||||
* 3) Because of how drm_ioctl() works, we can add new fields at
|
||||
* the end of an ioctl if some care is taken: drm_ioctl() will
|
||||
* zero out the new fields at the tail of the ioctl, so a zero
|
||||
* value should have a backwards compatible meaning. And for
|
||||
* output params, userspace won't see the newly added output
|
||||
* fields.. so that has to be somehow ok.
|
||||
*/
|
||||
|
||||
/* timeouts are specified in clock-monotonic absolute times (to simplify
|
||||
* restarting interrupted ioctls). The following struct is logically the
|
||||
* same as 'struct timespec' but 32/64b ABI safe.
|
||||
*/
|
||||
struct drm_etnaviv_timespec {
|
||||
__s64 tv_sec; /* seconds */
|
||||
__s64 tv_nsec; /* nanoseconds */
|
||||
};
|
||||
|
||||
#define ETNAVIV_PARAM_GPU_MODEL 0x01
|
||||
#define ETNAVIV_PARAM_GPU_REVISION 0x02
|
||||
#define ETNAVIV_PARAM_GPU_FEATURES_0 0x03
|
||||
#define ETNAVIV_PARAM_GPU_FEATURES_1 0x04
|
||||
#define ETNAVIV_PARAM_GPU_FEATURES_2 0x05
|
||||
#define ETNAVIV_PARAM_GPU_FEATURES_3 0x06
|
||||
#define ETNAVIV_PARAM_GPU_FEATURES_4 0x07
|
||||
|
||||
#define ETNAVIV_PARAM_GPU_STREAM_COUNT 0x10
|
||||
#define ETNAVIV_PARAM_GPU_REGISTER_MAX 0x11
|
||||
#define ETNAVIV_PARAM_GPU_THREAD_COUNT 0x12
|
||||
#define ETNAVIV_PARAM_GPU_VERTEX_CACHE_SIZE 0x13
|
||||
#define ETNAVIV_PARAM_GPU_SHADER_CORE_COUNT 0x14
|
||||
#define ETNAVIV_PARAM_GPU_PIXEL_PIPES 0x15
|
||||
#define ETNAVIV_PARAM_GPU_VERTEX_OUTPUT_BUFFER_SIZE 0x16
|
||||
#define ETNAVIV_PARAM_GPU_BUFFER_SIZE 0x17
|
||||
#define ETNAVIV_PARAM_GPU_INSTRUCTION_COUNT 0x18
|
||||
#define ETNAVIV_PARAM_GPU_NUM_CONSTANTS 0x19
|
||||
|
||||
#define ETNA_MAX_PIPES 4
|
||||
|
||||
struct drm_etnaviv_param {
|
||||
__u32 pipe; /* in */
|
||||
__u32 param; /* in, ETNAVIV_PARAM_x */
|
||||
__u64 value; /* out (get_param) or in (set_param) */
|
||||
};
|
||||
|
||||
/*
|
||||
* GEM buffers:
|
||||
*/
|
||||
|
||||
#define ETNA_BO_CACHE_MASK 0x000f0000
|
||||
/* cache modes */
|
||||
#define ETNA_BO_CACHED 0x00010000
|
||||
#define ETNA_BO_WC 0x00020000
|
||||
#define ETNA_BO_UNCACHED 0x00040000
|
||||
/* map flags */
|
||||
#define ETNA_BO_FORCE_MMU 0x00100000
|
||||
|
||||
struct drm_etnaviv_gem_new {
|
||||
__u64 size; /* in */
|
||||
__u32 flags; /* in, mask of ETNA_BO_x */
|
||||
__u32 handle; /* out */
|
||||
};
|
||||
|
||||
struct drm_etnaviv_gem_info {
|
||||
__u32 handle; /* in */
|
||||
__u32 pad;
|
||||
__u64 offset; /* out, offset to pass to mmap() */
|
||||
};
|
||||
|
||||
#define ETNA_PREP_READ 0x01
|
||||
#define ETNA_PREP_WRITE 0x02
|
||||
#define ETNA_PREP_NOSYNC 0x04
|
||||
|
||||
struct drm_etnaviv_gem_cpu_prep {
|
||||
__u32 handle; /* in */
|
||||
__u32 op; /* in, mask of ETNA_PREP_x */
|
||||
struct drm_etnaviv_timespec timeout; /* in */
|
||||
};
|
||||
|
||||
struct drm_etnaviv_gem_cpu_fini {
|
||||
__u32 handle; /* in */
|
||||
__u32 flags; /* in, placeholder for now, no defined values */
|
||||
};
|
||||
|
||||
/*
|
||||
* Cmdstream Submission:
|
||||
*/
|
||||
|
||||
/* The value written into the cmdstream is logically:
|
||||
* relocbuf->gpuaddr + reloc_offset
|
||||
*
|
||||
* NOTE that reloc's must be sorted by order of increasing submit_offset,
|
||||
* otherwise EINVAL.
|
||||
*/
|
||||
struct drm_etnaviv_gem_submit_reloc {
|
||||
__u32 submit_offset; /* in, offset from submit_bo */
|
||||
__u32 reloc_idx; /* in, index of reloc_bo buffer */
|
||||
__u64 reloc_offset; /* in, offset from start of reloc_bo */
|
||||
__u32 flags; /* in, placeholder for now, no defined values */
|
||||
};
|
||||
|
||||
/* Each buffer referenced elsewhere in the cmdstream submit (ie. the
|
||||
* cmdstream buffer(s) themselves or reloc entries) has one (and only
|
||||
* one) entry in the submit->bos[] table.
|
||||
*
|
||||
* As a optimization, the current buffer (gpu virtual address) can be
|
||||
* passed back through the 'presumed' field. If on a subsequent reloc,
|
||||
* userspace passes back a 'presumed' address that is still valid,
|
||||
* then patching the cmdstream for this entry is skipped. This can
|
||||
* avoid kernel needing to map/access the cmdstream bo in the common
|
||||
* case.
|
||||
*/
|
||||
#define ETNA_SUBMIT_BO_READ 0x0001
|
||||
#define ETNA_SUBMIT_BO_WRITE 0x0002
|
||||
struct drm_etnaviv_gem_submit_bo {
|
||||
__u32 flags; /* in, mask of ETNA_SUBMIT_BO_x */
|
||||
__u32 handle; /* in, GEM handle */
|
||||
__u64 presumed; /* in/out, presumed buffer address */
|
||||
};
|
||||
|
||||
/* Each cmdstream submit consists of a table of buffers involved, and
|
||||
* one or more cmdstream buffers. This allows for conditional execution
|
||||
* (context-restore), and IB buffers needed for per tile/bin draw cmds.
|
||||
*/
|
||||
#define ETNA_PIPE_3D 0x00
|
||||
#define ETNA_PIPE_2D 0x01
|
||||
#define ETNA_PIPE_VG 0x02
|
||||
struct drm_etnaviv_gem_submit {
|
||||
__u32 fence; /* out */
|
||||
__u32 pipe; /* in */
|
||||
__u32 exec_state; /* in, initial execution state (ETNA_PIPE_x) */
|
||||
__u32 nr_bos; /* in, number of submit_bo's */
|
||||
__u32 nr_relocs; /* in, number of submit_reloc's */
|
||||
__u32 stream_size; /* in, cmdstream size */
|
||||
__u64 bos; /* in, ptr to array of submit_bo's */
|
||||
__u64 relocs; /* in, ptr to array of submit_reloc's */
|
||||
__u64 stream; /* in, ptr to cmdstream */
|
||||
};
|
||||
|
||||
/* The normal way to synchronize with the GPU is just to CPU_PREP on
|
||||
* a buffer if you need to access it from the CPU (other cmdstream
|
||||
* submission from same or other contexts, PAGE_FLIP ioctl, etc, all
|
||||
* handle the required synchronization under the hood). This ioctl
|
||||
* mainly just exists as a way to implement the gallium pipe_fence
|
||||
* APIs without requiring a dummy bo to synchronize on.
|
||||
*/
|
||||
#define ETNA_WAIT_NONBLOCK 0x01
|
||||
struct drm_etnaviv_wait_fence {
|
||||
__u32 pipe; /* in */
|
||||
__u32 fence; /* in */
|
||||
__u32 flags; /* in, mask of ETNA_WAIT_x */
|
||||
__u32 pad;
|
||||
struct drm_etnaviv_timespec timeout; /* in */
|
||||
};
|
||||
|
||||
#define ETNA_USERPTR_READ 0x01
|
||||
#define ETNA_USERPTR_WRITE 0x02
|
||||
struct drm_etnaviv_gem_userptr {
|
||||
__u64 user_ptr; /* in, page aligned user pointer */
|
||||
__u64 user_size; /* in, page aligned user size */
|
||||
__u32 flags; /* in, flags */
|
||||
__u32 handle; /* out, non-zero handle */
|
||||
};
|
||||
|
||||
struct drm_etnaviv_gem_wait {
|
||||
__u32 pipe; /* in */
|
||||
__u32 handle; /* in, bo to be waited for */
|
||||
__u32 flags; /* in, mask of ETNA_WAIT_x */
|
||||
__u32 pad;
|
||||
struct drm_etnaviv_timespec timeout; /* in */
|
||||
};
|
||||
|
||||
#define DRM_ETNAVIV_GET_PARAM 0x00
|
||||
/* placeholder:
|
||||
#define DRM_ETNAVIV_SET_PARAM 0x01
|
||||
*/
|
||||
#define DRM_ETNAVIV_GEM_NEW 0x02
|
||||
#define DRM_ETNAVIV_GEM_INFO 0x03
|
||||
#define DRM_ETNAVIV_GEM_CPU_PREP 0x04
|
||||
#define DRM_ETNAVIV_GEM_CPU_FINI 0x05
|
||||
#define DRM_ETNAVIV_GEM_SUBMIT 0x06
|
||||
#define DRM_ETNAVIV_WAIT_FENCE 0x07
|
||||
#define DRM_ETNAVIV_GEM_USERPTR 0x08
|
||||
#define DRM_ETNAVIV_GEM_WAIT 0x09
|
||||
#define DRM_ETNAVIV_NUM_IOCTLS 0x0a
|
||||
|
||||
#define DRM_IOCTL_ETNAVIV_GET_PARAM DRM_IOWR(DRM_COMMAND_BASE + DRM_ETNAVIV_GET_PARAM, struct drm_etnaviv_param)
|
||||
#define DRM_IOCTL_ETNAVIV_GEM_NEW DRM_IOWR(DRM_COMMAND_BASE + DRM_ETNAVIV_GEM_NEW, struct drm_etnaviv_gem_new)
|
||||
#define DRM_IOCTL_ETNAVIV_GEM_INFO DRM_IOWR(DRM_COMMAND_BASE + DRM_ETNAVIV_GEM_INFO, struct drm_etnaviv_gem_info)
|
||||
#define DRM_IOCTL_ETNAVIV_GEM_CPU_PREP DRM_IOW(DRM_COMMAND_BASE + DRM_ETNAVIV_GEM_CPU_PREP, struct drm_etnaviv_gem_cpu_prep)
|
||||
#define DRM_IOCTL_ETNAVIV_GEM_CPU_FINI DRM_IOW(DRM_COMMAND_BASE + DRM_ETNAVIV_GEM_CPU_FINI, struct drm_etnaviv_gem_cpu_fini)
|
||||
#define DRM_IOCTL_ETNAVIV_GEM_SUBMIT DRM_IOWR(DRM_COMMAND_BASE + DRM_ETNAVIV_GEM_SUBMIT, struct drm_etnaviv_gem_submit)
|
||||
#define DRM_IOCTL_ETNAVIV_WAIT_FENCE DRM_IOW(DRM_COMMAND_BASE + DRM_ETNAVIV_WAIT_FENCE, struct drm_etnaviv_wait_fence)
|
||||
#define DRM_IOCTL_ETNAVIV_GEM_USERPTR DRM_IOWR(DRM_COMMAND_BASE + DRM_ETNAVIV_GEM_USERPTR, struct drm_etnaviv_gem_userptr)
|
||||
#define DRM_IOCTL_ETNAVIV_GEM_WAIT DRM_IOW(DRM_COMMAND_BASE + DRM_ETNAVIV_GEM_WAIT, struct drm_etnaviv_gem_wait)
|
||||
|
||||
#endif /* __ETNAVIV_DRM_H__ */
|
Loading…
Reference in a new issue