2 * Copyright 2010 Red Hat Inc.
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
26 #include <core/gpuobj.h>
27 #include <subdev/fb.h>
28 #include <subdev/timer.h>
30 #include <nvif/class.h>
32 /* Map from compressed to corresponding uncompressed storage type.
33 * The value 0xff represents an invalid storage type.
35 const u8 gf100_pte_storage_type_map
[256] =
37 0x00, 0x01, 0x01, 0x01, 0x01, 0x01, 0xff, 0x01, /* 0x00 */
38 0x01, 0x01, 0x01, 0xff, 0xff, 0xff, 0xff, 0xff,
39 0xff, 0x11, 0xff, 0xff, 0xff, 0xff, 0xff, 0x11, /* 0x10 */
40 0x11, 0x11, 0x11, 0xff, 0xff, 0xff, 0xff, 0xff,
41 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0x26, 0x27, /* 0x20 */
42 0x28, 0x29, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
43 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, /* 0x30 */
44 0xff, 0xff, 0x26, 0x27, 0x28, 0x29, 0x26, 0x27,
45 0x28, 0x29, 0xff, 0xff, 0xff, 0xff, 0x46, 0xff, /* 0x40 */
46 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
47 0xff, 0x46, 0x46, 0x46, 0x46, 0xff, 0xff, 0xff, /* 0x50 */
48 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
49 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, /* 0x60 */
50 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
51 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, /* 0x70 */
52 0xff, 0xff, 0xff, 0x7b, 0xff, 0xff, 0xff, 0xff,
53 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0x7b, 0x7b, /* 0x80 */
54 0x7b, 0x7b, 0xff, 0x8b, 0x8c, 0x8d, 0x8e, 0xff,
55 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, /* 0x90 */
56 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff,
57 0xff, 0xff, 0xff, 0x8b, 0x8c, 0x8d, 0x8e, 0xa7, /* 0xa0 */
58 0xa8, 0xa9, 0xaa, 0xff, 0xff, 0xff, 0xff, 0xff,
59 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, /* 0xb0 */
60 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xa7,
61 0xa8, 0xa9, 0xaa, 0xc3, 0xff, 0xff, 0xff, 0xff, /* 0xc0 */
62 0xff, 0xff, 0xff, 0xff, 0xfe, 0xfe, 0xc3, 0xc3,
63 0xc3, 0xc3, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, /* 0xd0 */
64 0xfe, 0xff, 0xff, 0xfe, 0xff, 0xfe, 0xff, 0xfe,
65 0xfe, 0xff, 0xfe, 0xfe, 0xfe, 0xfe, 0xfe, 0xff, /* 0xe0 */
66 0xff, 0xfe, 0xff, 0xfe, 0xff, 0xfe, 0xfe, 0xff,
67 0xfe, 0xfe, 0xfe, 0xfe, 0xfe, 0xfe, 0xfe, 0xfe, /* 0xf0 */
68 0xfe, 0xfe, 0xfe, 0xfe, 0xff, 0xfd, 0xfe, 0xff
73 gf100_vm_map_pgt(struct nvkm_vmm
*vmm
, u32 index
, struct nvkm_memory
*pgt
[2])
75 struct nvkm_memory
*pgd
= vmm
->pd
->pt
[0]->memory
;
76 u32 pde
[2] = { 0, 0 };
79 pde
[1] = 0x00000001 | (nvkm_memory_addr(pgt
[0]) >> 8);
81 pde
[0] = 0x00000001 | (nvkm_memory_addr(pgt
[1]) >> 8);
84 nvkm_wo32(pgd
, (index
* 8) + 0, pde
[0]);
85 nvkm_wo32(pgd
, (index
* 8) + 4, pde
[1]);
90 gf100_vm_addr(struct nvkm_vma
*vma
, u64 phys
, u32 memtype
, u32 target
)
94 phys
|= 0x00000001; /* present */
95 if (vma
->access
& NV_MEM_ACCESS_SYS
)
98 phys
|= ((u64
)target
<< 32);
99 phys
|= ((u64
)memtype
<< 36);
104 gf100_vm_map(struct nvkm_vma
*vma
, struct nvkm_memory
*pgt
,
105 struct nvkm_mem
*mem
, u32 pte
, u32 cnt
, u64 phys
, u64 delta
)
107 u64 next
= 1 << (vma
->node
->type
- 8);
109 phys
= gf100_vm_addr(vma
, phys
, mem
->memtype
, 0);
113 u32 tag
= mem
->tag
->offset
+ (delta
>> 17);
114 phys
|= (u64
)tag
<< (32 + 12);
115 next
|= (u64
)1 << (32 + 12);
120 nvkm_wo32(pgt
, pte
+ 0, lower_32_bits(phys
));
121 nvkm_wo32(pgt
, pte
+ 4, upper_32_bits(phys
));
129 gf100_vm_map_sg(struct nvkm_vma
*vma
, struct nvkm_memory
*pgt
,
130 struct nvkm_mem
*mem
, u32 pte
, u32 cnt
, dma_addr_t
*list
)
132 u32 target
= (vma
->access
& NV_MEM_ACCESS_NOSNOOP
) ? 7 : 5;
133 /* compressed storage types are invalid for system memory */
134 u32 memtype
= gf100_pte_storage_type_map
[mem
->memtype
& 0xff];
139 u64 phys
= gf100_vm_addr(vma
, *list
++, memtype
, target
);
140 nvkm_wo32(pgt
, pte
+ 0, lower_32_bits(phys
));
141 nvkm_wo32(pgt
, pte
+ 4, upper_32_bits(phys
));
148 gf100_vm_unmap(struct nvkm_vma
*vma
, struct nvkm_memory
*pgt
, u32 pte
, u32 cnt
)
153 nvkm_wo32(pgt
, pte
+ 0, 0x00000000);
154 nvkm_wo32(pgt
, pte
+ 4, 0x00000000);
161 gf100_vm_flush(struct nvkm_vm
*vm
)
163 struct nvkm_mmu
*mmu
= vm
->mmu
;
164 struct nvkm_device
*device
= mmu
->subdev
.device
;
167 type
= 0x00000001; /* PAGE_ALL */
168 if (atomic_read(&vm
->engref
[NVKM_SUBDEV_BAR
]))
169 type
|= 0x00000004; /* HUB_ONLY */
171 mutex_lock(&mmu
->subdev
.mutex
);
172 /* looks like maybe a "free flush slots" counter, the
173 * faster you write to 0x100cbc to more it decreases
175 nvkm_msec(device
, 2000,
176 if (nvkm_rd32(device
, 0x100c80) & 0x00ff0000)
180 nvkm_wr32(device
, 0x100cb8, vm
->pd
->pt
[0]->addr
>> 8);
181 nvkm_wr32(device
, 0x100cbc, 0x80000000 | type
);
183 /* wait for flush to be queued? */
184 nvkm_msec(device
, 2000,
185 if (nvkm_rd32(device
, 0x100c80) & 0x00008000)
188 mutex_unlock(&mmu
->subdev
.mutex
);
191 static const struct nvkm_mmu_func
193 .limit
= (1ULL << 40),
198 .map_pgt
= gf100_vm_map_pgt
,
200 .map_sg
= gf100_vm_map_sg
,
201 .unmap
= gf100_vm_unmap
,
202 .flush
= gf100_vm_flush
,
203 .vmm
= {{ -1, -1, NVIF_CLASS_VMM_GF100
}, gf100_vmm_new
},
207 gf100_mmu_new(struct nvkm_device
*device
, int index
, struct nvkm_mmu
**pmmu
)
209 return nvkm_mmu_new_(&gf100_mmu
, device
, index
, pmmu
);