1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
|
/* $NetBSD: pmap.h,v 1.37.4.1 2023/12/29 20:21:39 martin Exp $ */
/*-
* Copyright (C) 1995, 1996 Wolfgang Solfrank.
* Copyright (C) 1995, 1996 TooLs GmbH.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. All advertising materials mentioning features or use of this software
* must display the following acknowledgement:
* This product includes software developed by TooLs GmbH.
* 4. The name of TooLs GmbH may not be used to endorse or promote products
* derived from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY TOOLS GMBH ``AS IS'' AND ANY EXPRESS OR
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
* IN NO EVENT SHALL TOOLS GMBH BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
* OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
* OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
* ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#ifndef _POWERPC_OEA_PMAP_H_
#define _POWERPC_OEA_PMAP_H_
#ifdef _LOCORE
#error use assym.h instead
#endif
#ifdef _MODULE
#error this file should not be included by loadable kernel modules
#endif
#ifdef _KERNEL_OPT
#include "opt_ppcarch.h"
#include "opt_modular.h"
#endif
#include <powerpc/oea/pte.h>
#define __HAVE_PMAP_PV_TRACK
#include <uvm/pmap/pmap_pvt.h>
/*
* Pmap stuff
*/
struct pmap {
#ifdef PPC_OEA64
struct steg *pm_steg_table; /* segment table pointer */
/* XXX need way to track exec pages */
#endif
#if defined(PPC_OEA) || defined (PPC_OEA64_BRIDGE)
register_t pm_sr[16]; /* segments used in this pmap */
int pm_exec[16]; /* counts of exec mappings */
#endif
register_t pm_vsid; /* VSID bits */
int pm_refs; /* ref count */
struct pmap_statistics pm_stats; /* pmap statistics */
unsigned int pm_evictions; /* pvo's not in page table */
#ifdef PPC_OEA64
unsigned int pm_ste_evictions;
#endif
};
struct pmap_ops {
int (*pmapop_pte_spill)(struct pmap *, vaddr_t, bool);
void (*pmapop_real_memory)(paddr_t *, psize_t *);
void (*pmapop_init)(void);
void (*pmapop_virtual_space)(vaddr_t *, vaddr_t *);
pmap_t (*pmapop_create)(void);
void (*pmapop_reference)(pmap_t);
void (*pmapop_destroy)(pmap_t);
void (*pmapop_copy)(pmap_t, pmap_t, vaddr_t, vsize_t, vaddr_t);
void (*pmapop_update)(pmap_t);
int (*pmapop_enter)(pmap_t, vaddr_t, paddr_t, vm_prot_t, u_int);
void (*pmapop_remove)(pmap_t, vaddr_t, vaddr_t);
void (*pmapop_kenter_pa)(vaddr_t, paddr_t, vm_prot_t, u_int);
void (*pmapop_kremove)(vaddr_t, vsize_t);
bool (*pmapop_extract)(pmap_t, vaddr_t, paddr_t *);
void (*pmapop_protect)(pmap_t, vaddr_t, vaddr_t, vm_prot_t);
void (*pmapop_unwire)(pmap_t, vaddr_t);
void (*pmapop_page_protect)(struct vm_page *, vm_prot_t);
void (*pmapop_pv_protect)(paddr_t, vm_prot_t);
bool (*pmapop_query_bit)(struct vm_page *, int);
bool (*pmapop_clear_bit)(struct vm_page *, int);
void (*pmapop_activate)(struct lwp *);
void (*pmapop_deactivate)(struct lwp *);
void (*pmapop_pinit)(pmap_t);
void (*pmapop_procwr)(struct proc *, vaddr_t, size_t);
void (*pmapop_pte_print)(volatile struct pte *);
void (*pmapop_pteg_check)(void);
void (*pmapop_print_mmuregs)(void);
void (*pmapop_print_pte)(pmap_t, vaddr_t);
void (*pmapop_pteg_dist)(void);
void (*pmapop_pvo_verify)(void);
vaddr_t (*pmapop_steal_memory)(vsize_t, vaddr_t *, vaddr_t *);
void (*pmapop_bootstrap)(paddr_t, paddr_t);
void (*pmapop_bootstrap1)(paddr_t, paddr_t);
void (*pmapop_bootstrap2)(void);
};
#ifdef _KERNEL
#include <sys/cdefs.h>
__BEGIN_DECLS
#include <sys/param.h>
#include <sys/systm.h>
/*
* For OEA and OEA64_BRIDGE, we guarantee that pa below USER_ADDR
* (== 3GB < VM_MIN_KERNEL_ADDRESS) is direct-mapped.
*/
#if defined(PPC_OEA) || defined(PPC_OEA64_BRIDGE)
#define PMAP_DIRECT_MAPPED_SR (USER_SR - 1)
#define PMAP_DIRECT_MAPPED_LEN \
((vaddr_t)SEGMENT_LENGTH * (PMAP_DIRECT_MAPPED_SR + 1))
#endif
#if defined (PPC_OEA) || defined (PPC_OEA64_BRIDGE)
extern register_t iosrtable[];
#endif
extern int pmap_use_altivec;
#define pmap_clear_modify(pg) (pmap_clear_bit((pg), PTE_CHG))
#define pmap_clear_reference(pg) (pmap_clear_bit((pg), PTE_REF))
#define pmap_is_modified(pg) (pmap_query_bit((pg), PTE_CHG))
#define pmap_is_referenced(pg) (pmap_query_bit((pg), PTE_REF))
#define pmap_resident_count(pmap) ((pmap)->pm_stats.resident_count)
#define pmap_wired_count(pmap) ((pmap)->pm_stats.wired_count)
/* ARGSUSED */
static __inline bool
pmap_remove_all(struct pmap *pmap)
{
/* Nothing. */
return false;
}
#if (defined(PPC_OEA) + defined(PPC_OEA64) + defined(PPC_OEA64_BRIDGE)) != 1
#define PMAP_NEEDS_FIXUP
#endif
extern volatile struct pteg *pmap_pteg_table;
extern unsigned int pmap_pteg_cnt;
extern unsigned int pmap_pteg_mask;
void pmap_bootstrap(vaddr_t, vaddr_t);
void pmap_bootstrap1(vaddr_t, vaddr_t);
void pmap_bootstrap2(void);
bool pmap_extract(pmap_t, vaddr_t, paddr_t *);
bool pmap_query_bit(struct vm_page *, int);
bool pmap_clear_bit(struct vm_page *, int);
void pmap_real_memory(paddr_t *, psize_t *);
void pmap_procwr(struct proc *, vaddr_t, size_t);
int pmap_pte_spill(pmap_t, vaddr_t, bool);
int pmap_ste_spill(pmap_t, vaddr_t, bool);
void pmap_pinit(pmap_t);
#ifdef PPC_OEA601
bool pmap_extract_ioseg601(vaddr_t, paddr_t *);
#endif /* PPC_OEA601 */
#ifdef PPC_OEA
bool pmap_extract_battable(vaddr_t, paddr_t *);
#endif /* PPC_OEA */
u_int powerpc_mmap_flags(paddr_t);
#define POWERPC_MMAP_FLAG_MASK 0xf
#define POWERPC_MMAP_FLAG_PREFETCHABLE 0x1
#define POWERPC_MMAP_FLAG_CACHEABLE 0x2
#define pmap_phys_address(ppn) (ppn & ~POWERPC_MMAP_FLAG_MASK)
#define pmap_mmap_flags(ppn) powerpc_mmap_flags(ppn)
static __inline paddr_t vtophys (vaddr_t);
/*
* Alternate mapping hooks for pool pages. Avoids thrashing the TLB.
*
* Note: This won't work if we have more memory than can be direct-mapped
* VA==PA all at once. But pmap_copy_page() and pmap_zero_page() will have
* this problem, too.
*/
#if !defined(PPC_OEA64)
#define PMAP_MAP_POOLPAGE(pa) (pa)
#define PMAP_UNMAP_POOLPAGE(pa) (pa)
#define POOL_VTOPHYS(va) vtophys((vaddr_t) va)
#define PMAP_ALLOC_POOLPAGE(flags) pmap_alloc_poolpage(flags)
struct vm_page *pmap_alloc_poolpage(int);
#endif
static __inline paddr_t
vtophys(vaddr_t va)
{
paddr_t pa;
if (pmap_extract(pmap_kernel(), va, &pa))
return pa;
KASSERTMSG(0, "vtophys: pmap_extract of %#"PRIxVADDR" failed", va);
return (paddr_t) -1;
}
#ifdef PMAP_NEEDS_FIXUP
extern const struct pmap_ops *pmapops;
extern const struct pmap_ops pmap32_ops;
extern const struct pmap_ops pmap64_ops;
extern const struct pmap_ops pmap64bridge_ops;
static __inline void
pmap_setup32(void)
{
pmapops = &pmap32_ops;
}
static __inline void
pmap_setup64(void)
{
pmapops = &pmap64_ops;
}
static __inline void
pmap_setup64bridge(void)
{
pmapops = &pmap64bridge_ops;
}
#endif
bool pmap_pageidlezero (paddr_t);
void pmap_syncicache (paddr_t, psize_t);
#ifdef PPC_OEA64
vaddr_t pmap_setusr (vaddr_t);
vaddr_t pmap_unsetusr (void);
#endif
#ifdef PPC_OEA64_BRIDGE
int pmap_setup_segment0_map(int use_large_pages, ...);
#endif
#define PMAP_MD_PREFETCHABLE 0x2000000
#define PMAP_STEAL_MEMORY
#define PMAP_NEED_PROCWR
void pmap_zero_page(paddr_t);
void pmap_copy_page(paddr_t, paddr_t);
LIST_HEAD(pvo_head, pvo_entry);
#define __HAVE_VM_PAGE_MD
struct pmap_page {
unsigned int pp_attrs;
struct pvo_head pp_pvoh;
#ifdef MODULAR
uintptr_t pp_dummy[3];
#endif
};
struct vm_page_md {
struct pmap_page mdpg_pp;
#define mdpg_attrs mdpg_pp.pp_attrs
#define mdpg_pvoh mdpg_pp.pp_pvoh
#ifdef MODULAR
#define mdpg_dummy mdpg_pp.pp_dummy
#endif
};
#define VM_MDPAGE_INIT(pg) do { \
(pg)->mdpage.mdpg_attrs = 0; \
LIST_INIT(&(pg)->mdpage.mdpg_pvoh); \
} while (/*CONSTCOND*/0)
__END_DECLS
#endif /* _KERNEL */
#endif /* _POWERPC_OEA_PMAP_H_ */
|