linux-next: manual merge of the drm tree with Linus' tree

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Dave,

Today's linux-next merge of the drm tree got a conflict in
drivers/gpu/drm/radeon/evergreen_cs.c between commit de0babd60d8d
("drm/radeon: enforce use of radeon_get_ib_value when reading user cmd")
from Linus' tree and commit 0fcb6155cb5c ("radeon/kms: cleanup async dma
packet checking") from the drm tree.

I fixed it up (I think (I did it fairly mechanically) - see below) and
can carry the fix as necessary (no action is required - but it might be
worth doing this merge yourself before asking Linus to pull - you could
just *merge* the above commit from Linus' tree (or the head of the branch
that Linus merged)).

-- 
Cheers,
Stephen Rothwell                    sfr@xxxxxxxxxxxxxxxx

diff --cc drivers/gpu/drm/radeon/evergreen_cs.c
index ee4cff5,d8f5d5f..0000000
--- a/drivers/gpu/drm/radeon/evergreen_cs.c
+++ b/drivers/gpu/drm/radeon/evergreen_cs.c
@@@ -2908,15 -2708,19 +2708,19 @@@ int evergreen_dma_cs_parse(struct radeo
  				DRM_ERROR("bad DMA_PACKET_WRITE\n");
  				return -EINVAL;
  			}
- 			if (tiled) {
+ 			switch (sub_cmd) {
+ 			/* tiled */
+ 			case 8:
 -				dst_offset = ib[idx+1];
 +				dst_offset = radeon_get_ib_value(p, idx+1);
  				dst_offset <<= 8;
  
  				ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
  				p->idx += count + 7;
- 			} else {
+ 				break;
+ 			/* linear */
+ 			case 0:
 -				dst_offset = ib[idx+1];
 -				dst_offset |= ((u64)(ib[idx+2] & 0xff)) << 32;
 +				dst_offset = radeon_get_ib_value(p, idx+1);
 +				dst_offset |= ((u64)(radeon_get_ib_value(p, idx+2) & 0xff)) << 32;
  
  				ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset & 0xfffffffc);
  				ib[idx+2] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
@@@ -2939,338 -2747,330 +2747,330 @@@
  				DRM_ERROR("bad DMA_PACKET_COPY\n");
  				return -EINVAL;
  			}
- 			if (tiled) {
- 				idx_value = radeon_get_ib_value(p, idx + 2);
- 				if (new_cmd) {
- 					switch (misc) {
- 					case 0:
- 						/* L2T, frame to fields */
- 						if (idx_value & (1 << 31)) {
- 							DRM_ERROR("bad L2T, frame to fields DMA_PACKET_COPY\n");
- 							return -EINVAL;
- 						}
- 						r = r600_dma_cs_next_reloc(p, &dst2_reloc);
- 						if (r) {
- 							DRM_ERROR("bad L2T, frame to fields DMA_PACKET_COPY\n");
- 							return -EINVAL;
- 						}
- 						dst_offset = radeon_get_ib_value(p, idx+1);
- 						dst_offset <<= 8;
- 						dst2_offset = radeon_get_ib_value(p, idx+2);
- 						dst2_offset <<= 8;
- 						src_offset = radeon_get_ib_value(p, idx+8);
- 						src_offset |= ((u64)(radeon_get_ib_value(p, idx+9) & 0xff)) << 32;
- 						if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, frame to fields src buffer too small (%llu %lu)\n",
- 								 src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
- 							return -EINVAL;
- 						}
- 						if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, frame to fields buffer too small (%llu %lu)\n",
- 								 dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
- 							return -EINVAL;
- 						}
- 						if ((dst2_offset + (count * 4)) > radeon_bo_size(dst2_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, frame to fields buffer too small (%llu %lu)\n",
- 								 dst2_offset + (count * 4), radeon_bo_size(dst2_reloc->robj));
- 							return -EINVAL;
- 						}
- 						ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
- 						ib[idx+2] += (u32)(dst2_reloc->lobj.gpu_offset >> 8);
- 						ib[idx+8] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
- 						ib[idx+9] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
- 						p->idx += 10;
- 						break;
- 					case 1:
- 						/* L2T, T2L partial */
- 						if (p->family < CHIP_CAYMAN) {
- 							DRM_ERROR("L2T, T2L Partial is cayman only !\n");
- 							return -EINVAL;
- 						}
- 						/* detile bit */
- 						if (idx_value & (1 << 31)) {
- 							/* tiled src, linear dst */
- 							ib[idx+1] += (u32)(src_reloc->lobj.gpu_offset >> 8);
- 
- 							ib[idx+7] += (u32)(dst_reloc->lobj.gpu_offset & 0xfffffffc);
- 							ib[idx+8] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
- 						} else {
- 							/* linear src, tiled dst */
- 							ib[idx+7] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
- 							ib[idx+8] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
- 
- 							ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
- 						}
- 						p->idx += 12;
- 						break;
- 					case 3:
- 						/* L2T, broadcast */
- 						if (idx_value & (1 << 31)) {
- 							DRM_ERROR("bad L2T, broadcast DMA_PACKET_COPY\n");
- 							return -EINVAL;
- 						}
- 						r = r600_dma_cs_next_reloc(p, &dst2_reloc);
- 						if (r) {
- 							DRM_ERROR("bad L2T, broadcast DMA_PACKET_COPY\n");
- 							return -EINVAL;
- 						}
- 						dst_offset = radeon_get_ib_value(p, idx+1);
- 						dst_offset <<= 8;
- 						dst2_offset = radeon_get_ib_value(p, idx+2);
- 						dst2_offset <<= 8;
- 						src_offset = radeon_get_ib_value(p, idx+8);
- 						src_offset |= ((u64)(radeon_get_ib_value(p, idx+9) & 0xff)) << 32;
- 						if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, broadcast src buffer too small (%llu %lu)\n",
- 								 src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
- 							return -EINVAL;
- 						}
- 						if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, broadcast dst buffer too small (%llu %lu)\n",
- 								 dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
- 							return -EINVAL;
- 						}
- 						if ((dst2_offset + (count * 4)) > radeon_bo_size(dst2_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, broadcast dst2 buffer too small (%llu %lu)\n",
- 								 dst2_offset + (count * 4), radeon_bo_size(dst2_reloc->robj));
- 							return -EINVAL;
- 						}
- 						ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
- 						ib[idx+2] += (u32)(dst2_reloc->lobj.gpu_offset >> 8);
- 						ib[idx+8] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
- 						ib[idx+9] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
- 						p->idx += 10;
- 						break;
- 					case 4:
- 						/* L2T, T2L */
- 						/* detile bit */
- 						if (idx_value & (1 << 31)) {
- 							/* tiled src, linear dst */
- 							src_offset = radeon_get_ib_value(p, idx+1);
- 							src_offset <<= 8;
- 							ib[idx+1] += (u32)(src_reloc->lobj.gpu_offset >> 8);
- 
- 							dst_offset = radeon_get_ib_value(p, idx+7);
- 							dst_offset |= ((u64)(radeon_get_ib_value(p, idx+8) & 0xff)) << 32;
- 							ib[idx+7] += (u32)(dst_reloc->lobj.gpu_offset & 0xfffffffc);
- 							ib[idx+8] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
- 						} else {
- 							/* linear src, tiled dst */
- 							src_offset = radeon_get_ib_value(p, idx+7);
- 							src_offset |= ((u64)(radeon_get_ib_value(p, idx+8) & 0xff)) << 32;
- 							ib[idx+7] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
- 							ib[idx+8] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
- 
- 							dst_offset = radeon_get_ib_value(p, idx+1);
- 							dst_offset <<= 8;
- 							ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
- 						}
- 						if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, T2L src buffer too small (%llu %lu)\n",
- 								 src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
- 							return -EINVAL;
- 						}
- 						if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, T2L dst buffer too small (%llu %lu)\n",
- 								 dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
- 							return -EINVAL;
- 						}
- 						p->idx += 9;
- 						break;
- 					case 5:
- 						/* T2T partial */
- 						if (p->family < CHIP_CAYMAN) {
- 							DRM_ERROR("L2T, T2L Partial is cayman only !\n");
- 							return -EINVAL;
- 						}
- 						ib[idx+1] += (u32)(src_reloc->lobj.gpu_offset >> 8);
- 						ib[idx+4] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
- 						p->idx += 13;
- 						break;
- 					case 7:
- 						/* L2T, broadcast */
- 						if (idx_value & (1 << 31)) {
- 							DRM_ERROR("bad L2T, broadcast DMA_PACKET_COPY\n");
- 							return -EINVAL;
- 						}
- 						r = r600_dma_cs_next_reloc(p, &dst2_reloc);
- 						if (r) {
- 							DRM_ERROR("bad L2T, broadcast DMA_PACKET_COPY\n");
- 							return -EINVAL;
- 						}
- 						dst_offset = radeon_get_ib_value(p, idx+1);
- 						dst_offset <<= 8;
- 						dst2_offset = radeon_get_ib_value(p, idx+2);
- 						dst2_offset <<= 8;
- 						src_offset = radeon_get_ib_value(p, idx+8);
- 						src_offset |= ((u64)(radeon_get_ib_value(p, idx+9) & 0xff)) << 32;
- 						if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, broadcast src buffer too small (%llu %lu)\n",
- 								 src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
- 							return -EINVAL;
- 						}
- 						if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, broadcast dst buffer too small (%llu %lu)\n",
- 								 dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
- 							return -EINVAL;
- 						}
- 						if ((dst2_offset + (count * 4)) > radeon_bo_size(dst2_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, broadcast dst2 buffer too small (%llu %lu)\n",
- 								 dst2_offset + (count * 4), radeon_bo_size(dst2_reloc->robj));
- 							return -EINVAL;
- 						}
- 						ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
- 						ib[idx+2] += (u32)(dst2_reloc->lobj.gpu_offset >> 8);
- 						ib[idx+8] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
- 						ib[idx+9] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
- 						p->idx += 10;
- 						break;
- 					default:
- 						DRM_ERROR("bad DMA_PACKET_COPY misc %u\n", misc);
- 						return -EINVAL;
- 					}
+ 			switch (sub_cmd) {
+ 			/* Copy L2L, DW aligned */
+ 			case 0x00:
+ 				/* L2L, dw */
 -				src_offset = ib[idx+2];
 -				src_offset |= ((u64)(ib[idx+4] & 0xff)) << 32;
 -				dst_offset = ib[idx+1];
 -				dst_offset |= ((u64)(ib[idx+3] & 0xff)) << 32;
++				src_offset = radeon_get_ib_value(p, idx+2);
++				src_offset |= ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32;
++				dst_offset = radeon_get_ib_value(p, idx+1);
++				dst_offset |= ((u64)(radeon_get_ib_value(p, idx+3) & 0xff)) << 32;
+ 				if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2L, dw src buffer too small (%llu %lu)\n",
+ 							src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2L, dw dst buffer too small (%llu %lu)\n",
+ 							dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset & 0xfffffffc);
+ 				ib[idx+2] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
+ 				ib[idx+3] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
+ 				ib[idx+4] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
+ 				p->idx += 5;
+ 				break;
+ 			/* Copy L2T/T2L */
+ 			case 0x08:
+ 				/* detile bit */
 -				if (ib[idx + 2] & (1 << 31)) {
++				if (radeon_get_ib_value(p, idx + 2) & (1 << 31)) {
+ 					/* tiled src, linear dst */
 -					src_offset = ib[idx+1];
++					src_offset = radeon_get_ib_value(p, idx+1);
+ 					src_offset <<= 8;
+ 					ib[idx+1] += (u32)(src_reloc->lobj.gpu_offset >> 8);
+ 
+ 					dst_offset = radeon_get_ib_value(p, idx + 7);
 -					dst_offset |= ((u64)(ib[idx+8] & 0xff)) << 32;
++					dst_offset |= ((u64)(radeon_get_ib_value(p, idx+8) & 0xff)) << 32;
+ 					ib[idx+7] += (u32)(dst_reloc->lobj.gpu_offset & 0xfffffffc);
+ 					ib[idx+8] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
  				} else {
- 					switch (misc) {
- 					case 0:
- 						/* detile bit */
- 						if (idx_value & (1 << 31)) {
- 							/* tiled src, linear dst */
- 							src_offset = radeon_get_ib_value(p, idx+1);
- 							src_offset <<= 8;
- 							ib[idx+1] += (u32)(src_reloc->lobj.gpu_offset >> 8);
- 
- 							dst_offset = radeon_get_ib_value(p, idx+7);
- 							dst_offset |= ((u64)(radeon_get_ib_value(p, idx+8) & 0xff)) << 32;
- 							ib[idx+7] += (u32)(dst_reloc->lobj.gpu_offset & 0xfffffffc);
- 							ib[idx+8] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
- 						} else {
- 							/* linear src, tiled dst */
- 							src_offset = radeon_get_ib_value(p, idx+7);
- 							src_offset |= ((u64)(radeon_get_ib_value(p, idx+8) & 0xff)) << 32;
- 							ib[idx+7] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
- 							ib[idx+8] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
- 
- 							dst_offset = radeon_get_ib_value(p, idx+1);
- 							dst_offset <<= 8;
- 							ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
- 						}
- 						if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, broadcast src buffer too small (%llu %lu)\n",
- 								 src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
- 							return -EINVAL;
- 						}
- 						if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2T, broadcast dst buffer too small (%llu %lu)\n",
- 								 dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
- 							return -EINVAL;
- 						}
- 						p->idx += 9;
- 						break;
- 					default:
- 						DRM_ERROR("bad DMA_PACKET_COPY misc %u\n", misc);
- 						return -EINVAL;
- 					}
+ 					/* linear src, tiled dst */
 -					src_offset = ib[idx+7];
 -					src_offset |= ((u64)(ib[idx+8] & 0xff)) << 32;
++					src_offset = radeon_get_ib_value(p, idx+7);
++					src_offset |= ((u64)(radeon_get_ib_value(p, idx+8) & 0xff)) << 32;
+ 					ib[idx+7] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
+ 					ib[idx+8] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
+ 
 -					dst_offset = ib[idx+1];
++					dst_offset = radeon_get_ib_value(p, idx+1);
+ 					dst_offset <<= 8;
+ 					ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
  				}
- 			} else {
- 				if (new_cmd) {
- 					switch (misc) {
- 					case 0:
- 						/* L2L, byte */
- 						src_offset = radeon_get_ib_value(p, idx+2);
- 						src_offset |= ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32;
- 						dst_offset = radeon_get_ib_value(p, idx+1);
- 						dst_offset |= ((u64)(radeon_get_ib_value(p, idx+3) & 0xff)) << 32;
- 						if ((src_offset + count) > radeon_bo_size(src_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2L, byte src buffer too small (%llu %lu)\n",
- 								 src_offset + count, radeon_bo_size(src_reloc->robj));
- 							return -EINVAL;
- 						}
- 						if ((dst_offset + count) > radeon_bo_size(dst_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2L, byte dst buffer too small (%llu %lu)\n",
- 								 dst_offset + count, radeon_bo_size(dst_reloc->robj));
- 							return -EINVAL;
- 						}
- 						ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset & 0xffffffff);
- 						ib[idx+2] += (u32)(src_reloc->lobj.gpu_offset & 0xffffffff);
- 						ib[idx+3] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
- 						ib[idx+4] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
- 						p->idx += 5;
- 						break;
- 					case 1:
- 						/* L2L, partial */
- 						if (p->family < CHIP_CAYMAN) {
- 							DRM_ERROR("L2L Partial is cayman only !\n");
- 							return -EINVAL;
- 						}
- 						ib[idx+1] += (u32)(src_reloc->lobj.gpu_offset & 0xffffffff);
- 						ib[idx+2] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
- 						ib[idx+4] += (u32)(dst_reloc->lobj.gpu_offset & 0xffffffff);
- 						ib[idx+5] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
- 
- 						p->idx += 9;
- 						break;
- 					case 4:
- 						/* L2L, dw, broadcast */
- 						r = r600_dma_cs_next_reloc(p, &dst2_reloc);
- 						if (r) {
- 							DRM_ERROR("bad L2L, dw, broadcast DMA_PACKET_COPY\n");
- 							return -EINVAL;
- 						}
- 						dst_offset = radeon_get_ib_value(p, idx+1);
- 						dst_offset |= ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32;
- 						dst2_offset = radeon_get_ib_value(p, idx+2);
- 						dst2_offset |= ((u64)(radeon_get_ib_value(p, idx+5) & 0xff)) << 32;
- 						src_offset = radeon_get_ib_value(p, idx+3);
- 						src_offset |= ((u64)(radeon_get_ib_value(p, idx+6) & 0xff)) << 32;
- 						if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2L, dw, broadcast src buffer too small (%llu %lu)\n",
- 								 src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
- 							return -EINVAL;
- 						}
- 						if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2L, dw, broadcast dst buffer too small (%llu %lu)\n",
- 								 dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
- 							return -EINVAL;
- 						}
- 						if ((dst2_offset + (count * 4)) > radeon_bo_size(dst2_reloc->robj)) {
- 							dev_warn(p->dev, "DMA L2L, dw, broadcast dst2 buffer too small (%llu %lu)\n",
- 								 dst2_offset + (count * 4), radeon_bo_size(dst2_reloc->robj));
- 							return -EINVAL;
- 						}
- 						ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset & 0xfffffffc);
- 						ib[idx+2] += (u32)(dst2_reloc->lobj.gpu_offset & 0xfffffffc);
- 						ib[idx+3] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
- 						ib[idx+4] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
- 						ib[idx+5] += upper_32_bits(dst2_reloc->lobj.gpu_offset) & 0xff;
- 						ib[idx+6] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
- 						p->idx += 7;
- 						break;
- 					default:
- 						DRM_ERROR("bad DMA_PACKET_COPY misc %u\n", misc);
- 						return -EINVAL;
- 					}
+ 				if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, src buffer too small (%llu %lu)\n",
+ 							src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, dst buffer too small (%llu %lu)\n",
+ 							dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				p->idx += 9;
+ 				break;
+ 			/* Copy L2L, byte aligned */
+ 			case 0x40:
+ 				/* L2L, byte */
 -				src_offset = ib[idx+2];
 -				src_offset |= ((u64)(ib[idx+4] & 0xff)) << 32;
 -				dst_offset = ib[idx+1];
 -				dst_offset |= ((u64)(ib[idx+3] & 0xff)) << 32;
++				src_offset = radeon_get_ib_value(p, idx+2);
++				src_offset |= ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32;
++				dst_offset = radeon_get_ib_value(p, idx+1);
++				dst_offset |= ((u64)(radeon_get_ib_value(p, idx+3) & 0xff)) << 32;
+ 				if ((src_offset + count) > radeon_bo_size(src_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2L, byte src buffer too small (%llu %lu)\n",
+ 							src_offset + count, radeon_bo_size(src_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				if ((dst_offset + count) > radeon_bo_size(dst_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2L, byte dst buffer too small (%llu %lu)\n",
+ 							dst_offset + count, radeon_bo_size(dst_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset & 0xffffffff);
+ 				ib[idx+2] += (u32)(src_reloc->lobj.gpu_offset & 0xffffffff);
+ 				ib[idx+3] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
+ 				ib[idx+4] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
+ 				p->idx += 5;
+ 				break;
+ 			/* Copy L2L, partial */
+ 			case 0x41:
+ 				/* L2L, partial */
+ 				if (p->family < CHIP_CAYMAN) {
+ 					DRM_ERROR("L2L Partial is cayman only !\n");
+ 					return -EINVAL;
+ 				}
+ 				ib[idx+1] += (u32)(src_reloc->lobj.gpu_offset & 0xffffffff);
+ 				ib[idx+2] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
+ 				ib[idx+4] += (u32)(dst_reloc->lobj.gpu_offset & 0xffffffff);
+ 				ib[idx+5] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
+ 
+ 				p->idx += 9;
+ 				break;
+ 			/* Copy L2L, DW aligned, broadcast */
+ 			case 0x44:
+ 				/* L2L, dw, broadcast */
+ 				r = r600_dma_cs_next_reloc(p, &dst2_reloc);
+ 				if (r) {
+ 					DRM_ERROR("bad L2L, dw, broadcast DMA_PACKET_COPY\n");
+ 					return -EINVAL;
+ 				}
 -				dst_offset = ib[idx+1];
 -				dst_offset |= ((u64)(ib[idx+4] & 0xff)) << 32;
 -				dst2_offset = ib[idx+2];
 -				dst2_offset |= ((u64)(ib[idx+5] & 0xff)) << 32;
 -				src_offset = ib[idx+3];
 -				src_offset |= ((u64)(ib[idx+6] & 0xff)) << 32;
++				dst_offset = radeon_get_ib_value(p, idx+1);
++				dst_offset |= ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32;
++				dst2_offset = radeon_get_ib_value(p, idx+2);
++				dst2_offset |= ((u64)(radeon_get_ib_value(p, idx+5) & 0xff)) << 32;
++				src_offset = radeon_get_ib_value(p, idx+3);
++				src_offset |= ((u64)(radeon_get_ib_value(p, idx+6) & 0xff)) << 32;
+ 				if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2L, dw, broadcast src buffer too small (%llu %lu)\n",
+ 							src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2L, dw, broadcast dst buffer too small (%llu %lu)\n",
+ 							dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				if ((dst2_offset + (count * 4)) > radeon_bo_size(dst2_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2L, dw, broadcast dst2 buffer too small (%llu %lu)\n",
+ 							dst2_offset + (count * 4), radeon_bo_size(dst2_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset & 0xfffffffc);
+ 				ib[idx+2] += (u32)(dst2_reloc->lobj.gpu_offset & 0xfffffffc);
+ 				ib[idx+3] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
+ 				ib[idx+4] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
+ 				ib[idx+5] += upper_32_bits(dst2_reloc->lobj.gpu_offset) & 0xff;
+ 				ib[idx+6] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
+ 				p->idx += 7;
+ 				break;
+ 			/* Copy L2T Frame to Field */
+ 			case 0x48:
 -				if (ib[idx + 2] & (1 << 31)) {
++				if (radeon_get_ib_value(p, idx + 2) & (1 << 31)) {
+ 					DRM_ERROR("bad L2T, frame to fields DMA_PACKET_COPY\n");
+ 					return -EINVAL;
+ 				}
+ 				r = r600_dma_cs_next_reloc(p, &dst2_reloc);
+ 				if (r) {
+ 					DRM_ERROR("bad L2T, frame to fields DMA_PACKET_COPY\n");
+ 					return -EINVAL;
+ 				}
 -				dst_offset = ib[idx+1];
++				dst_offset = radeon_get_ib_value(p, idx+1);
+ 				dst_offset <<= 8;
 -				dst2_offset = ib[idx+2];
++				dst2_offset = radeon_get_ib_value(p, idx+2);
+ 				dst2_offset <<= 8;
 -				src_offset = ib[idx+8];
 -				src_offset |= ((u64)(ib[idx+9] & 0xff)) << 32;
++				src_offset = radeon_get_ib_value(p, idx+8);
++				src_offset |= ((u64)(radeon_get_ib_value(p, idx+9) & 0xff)) << 32;
+ 				if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, frame to fields src buffer too small (%llu %lu)\n",
+ 							src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, frame to fields buffer too small (%llu %lu)\n",
+ 							dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				if ((dst2_offset + (count * 4)) > radeon_bo_size(dst2_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, frame to fields buffer too small (%llu %lu)\n",
+ 							dst2_offset + (count * 4), radeon_bo_size(dst2_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
+ 				ib[idx+2] += (u32)(dst2_reloc->lobj.gpu_offset >> 8);
+ 				ib[idx+8] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
+ 				ib[idx+9] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
+ 				p->idx += 10;
+ 				break;
+ 			/* Copy L2T/T2L, partial */
+ 			case 0x49:
+ 				/* L2T, T2L partial */
+ 				if (p->family < CHIP_CAYMAN) {
+ 					DRM_ERROR("L2T, T2L Partial is cayman only !\n");
+ 					return -EINVAL;
+ 				}
+ 				/* detile bit */
 -				if (ib[idx + 2 ] & (1 << 31)) {
++				if (radeon_get_ib_value(p, idx + 2 ) & (1 << 31)) {
+ 					/* tiled src, linear dst */
+ 					ib[idx+1] += (u32)(src_reloc->lobj.gpu_offset >> 8);
+ 
+ 					ib[idx+7] += (u32)(dst_reloc->lobj.gpu_offset & 0xfffffffc);
+ 					ib[idx+8] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
+ 				} else {
+ 					/* linear src, tiled dst */
+ 					ib[idx+7] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
+ 					ib[idx+8] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
+ 
+ 					ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
+ 				}
+ 				p->idx += 12;
+ 				break;
+ 			/* Copy L2T broadcast */
+ 			case 0x4b:
+ 				/* L2T, broadcast */
 -				if (ib[idx + 2] & (1 << 31)) {
++				if (radeon_get_ib_value(p, idx + 2) & (1 << 31)) {
+ 					DRM_ERROR("bad L2T, broadcast DMA_PACKET_COPY\n");
+ 					return -EINVAL;
+ 				}
+ 				r = r600_dma_cs_next_reloc(p, &dst2_reloc);
+ 				if (r) {
+ 					DRM_ERROR("bad L2T, broadcast DMA_PACKET_COPY\n");
+ 					return -EINVAL;
+ 				}
 -				dst_offset = ib[idx+1];
++				dst_offset = radeon_get_ib_value(p, idx+1);
+ 				dst_offset <<= 8;
 -				dst2_offset = ib[idx+2];
++				dst2_offset = radeon_get_ib_value(p, idx+2);
+ 				dst2_offset <<= 8;
 -				src_offset = ib[idx+8];
 -				src_offset |= ((u64)(ib[idx+9] & 0xff)) << 32;
++				src_offset = radeon_get_ib_value(p, idx+8);
++				src_offset |= ((u64)(radeon_get_ib_value(p, idx+9) & 0xff)) << 32;
+ 				if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, broadcast src buffer too small (%llu %lu)\n",
+ 							src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, broadcast dst buffer too small (%llu %lu)\n",
+ 							dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				if ((dst2_offset + (count * 4)) > radeon_bo_size(dst2_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, broadcast dst2 buffer too small (%llu %lu)\n",
+ 							dst2_offset + (count * 4), radeon_bo_size(dst2_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
+ 				ib[idx+2] += (u32)(dst2_reloc->lobj.gpu_offset >> 8);
+ 				ib[idx+8] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
+ 				ib[idx+9] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
+ 				p->idx += 10;
+ 				break;
+ 			/* Copy L2T/T2L (tile units) */
+ 			case 0x4c:
+ 				/* L2T, T2L */
+ 				/* detile bit */
 -				if (ib[idx + 2] & (1 << 31)) {
++				if (radeon_get_ib_value(p, idx + 2) & (1 << 31)) {
+ 					/* tiled src, linear dst */
 -					src_offset = ib[idx+1];
++					src_offset = radeon_get_ib_value(p, idx+1);
+ 					src_offset <<= 8;
+ 					ib[idx+1] += (u32)(src_reloc->lobj.gpu_offset >> 8);
+ 
 -					dst_offset = ib[idx+7];
 -					dst_offset |= ((u64)(ib[idx+8] & 0xff)) << 32;
++					dst_offset = radeon_get_ib_value(p, idx+7);
++					dst_offset |= ((u64)(radeon_get_ib_value(p, idx+8) & 0xff)) << 32;
+ 					ib[idx+7] += (u32)(dst_reloc->lobj.gpu_offset & 0xfffffffc);
+ 					ib[idx+8] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
  				} else {
- 					/* L2L, dw */
- 					src_offset = radeon_get_ib_value(p, idx+2);
- 					src_offset |= ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32;
+ 					/* linear src, tiled dst */
 -					src_offset = ib[idx+7];
 -					src_offset |= ((u64)(ib[idx+8] & 0xff)) << 32;
++					src_offset = radeon_get_ib_value(p, idx+7);
++					src_offset |= ((u64)(radeon_get_ib_value(p, idx+8) & 0xff)) << 32;
+ 					ib[idx+7] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
+ 					ib[idx+8] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
+ 
 -					dst_offset = ib[idx+1];
 +					dst_offset = radeon_get_ib_value(p, idx+1);
- 					dst_offset |= ((u64)(radeon_get_ib_value(p, idx+3) & 0xff)) << 32;
- 					if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
- 						dev_warn(p->dev, "DMA L2L, dw src buffer too small (%llu %lu)\n",
- 							 src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
- 						return -EINVAL;
- 					}
- 					if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
- 						dev_warn(p->dev, "DMA L2L, dw dst buffer too small (%llu %lu)\n",
- 							 dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
- 						return -EINVAL;
- 					}
- 					ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset & 0xfffffffc);
- 					ib[idx+2] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
- 					ib[idx+3] += upper_32_bits(dst_reloc->lobj.gpu_offset) & 0xff;
- 					ib[idx+4] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
- 					p->idx += 5;
+ 					dst_offset <<= 8;
+ 					ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
  				}
+ 				if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, T2L src buffer too small (%llu %lu)\n",
+ 							src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, T2L dst buffer too small (%llu %lu)\n",
+ 							dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				p->idx += 9;
+ 				break;
+ 			/* Copy T2T, partial (tile units) */
+ 			case 0x4d:
+ 				/* T2T partial */
+ 				if (p->family < CHIP_CAYMAN) {
+ 					DRM_ERROR("L2T, T2L Partial is cayman only !\n");
+ 					return -EINVAL;
+ 				}
+ 				ib[idx+1] += (u32)(src_reloc->lobj.gpu_offset >> 8);
+ 				ib[idx+4] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
+ 				p->idx += 13;
+ 				break;
+ 			/* Copy L2T broadcast (tile units) */
+ 			case 0x4f:
+ 				/* L2T, broadcast */
 -				if (ib[idx + 2] & (1 << 31)) {
++				if (radeon_get_ib_value(p, idx + 2) & (1 << 31)) {
+ 					DRM_ERROR("bad L2T, broadcast DMA_PACKET_COPY\n");
+ 					return -EINVAL;
+ 				}
+ 				r = r600_dma_cs_next_reloc(p, &dst2_reloc);
+ 				if (r) {
+ 					DRM_ERROR("bad L2T, broadcast DMA_PACKET_COPY\n");
+ 					return -EINVAL;
+ 				}
 -				dst_offset = ib[idx+1];
++				dst_offset = radeon_get_ib_value(p, idx+1);
+ 				dst_offset <<= 8;
 -				dst2_offset = ib[idx+2];
++				dst2_offset = radeon_get_ib_value(p, idx+2);
+ 				dst2_offset <<= 8;
 -				src_offset = ib[idx+8];
 -				src_offset |= ((u64)(ib[idx+9] & 0xff)) << 32;
++				src_offset = radeon_get_ib_value(p, idx+8);
++				src_offset |= ((u64)(radeon_get_ib_value(p, idx+9) & 0xff)) << 32;
+ 				if ((src_offset + (count * 4)) > radeon_bo_size(src_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, broadcast src buffer too small (%llu %lu)\n",
+ 							src_offset + (count * 4), radeon_bo_size(src_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				if ((dst_offset + (count * 4)) > radeon_bo_size(dst_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, broadcast dst buffer too small (%llu %lu)\n",
+ 							dst_offset + (count * 4), radeon_bo_size(dst_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				if ((dst2_offset + (count * 4)) > radeon_bo_size(dst2_reloc->robj)) {
+ 					dev_warn(p->dev, "DMA L2T, broadcast dst2 buffer too small (%llu %lu)\n",
+ 							dst2_offset + (count * 4), radeon_bo_size(dst2_reloc->robj));
+ 					return -EINVAL;
+ 				}
+ 				ib[idx+1] += (u32)(dst_reloc->lobj.gpu_offset >> 8);
+ 				ib[idx+2] += (u32)(dst2_reloc->lobj.gpu_offset >> 8);
+ 				ib[idx+8] += (u32)(src_reloc->lobj.gpu_offset & 0xfffffffc);
+ 				ib[idx+9] += upper_32_bits(src_reloc->lobj.gpu_offset) & 0xff;
+ 				p->idx += 10;
+ 				break;
+ 			default:
 -				DRM_ERROR("bad DMA_PACKET_COPY [%6d] 0x%08x invalid sub cmd\n", idx, ib[idx+0]);
++				DRM_ERROR("bad DMA_PACKET_COPY [%6d] 0x%08x invalid sub cmd\n", idx, radeon_get_ib_value(p, idx+0));
+ 				return -EINVAL;
  			}
  			break;
  		case DMA_PACKET_CONSTANT_FILL:

Attachment: pgpoxaEOjRupV.pgp
Description: PGP signature


[Index of Archives]     [Linux Kernel]     [Linux USB Development]     [Yosemite News]     [Linux SCSI]

  Powered by Linux