1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
|
/* $OpenBSD: cpu.h,v 1.123 2018/01/29 14:53:01 visa Exp $ */
/*-
* Copyright (c) 1992, 1993
* The Regents of the University of California. All rights reserved.
*
* This code is derived from software contributed to Berkeley by
* Ralph Campbell and Rick Macklem.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. Neither the name of the University nor the names of its contributors
* may be used to endorse or promote products derived from this software
* without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*
* Copyright (C) 1989 Digital Equipment Corporation.
* Permission to use, copy, modify, and distribute this software and
* its documentation for any purpose and without fee is hereby granted,
* provided that the above copyright notice appears in all copies.
* Digital Equipment Corporation makes no representations about the
* suitability of this software for any purpose. It is provided "as is"
* without express or implied warranty.
*
* from: @(#)cpu.h 8.4 (Berkeley) 1/4/94
*/
#ifndef _MIPS64_CPU_H_
#define _MIPS64_CPU_H_
#ifndef _LOCORE
/*
* MIPS32-style segment definitions.
* They only cover the first 512MB of physical addresses.
*/
#define CKSEG0_BASE 0xffffffff80000000UL
#define CKSEG1_BASE 0xffffffffa0000000UL
#define CKSSEG_BASE 0xffffffffc0000000UL
#define CKSEG3_BASE 0xffffffffe0000000UL
#define CKSEG_SIZE 0x0000000020000000UL
#define CKSEG0_TO_PHYS(x) ((u_long)(x) & (CKSEG_SIZE - 1))
#define CKSEG1_TO_PHYS(x) ((u_long)(x) & (CKSEG_SIZE - 1))
#define PHYS_TO_CKSEG0(x) ((u_long)(x) | CKSEG0_BASE)
#define PHYS_TO_CKSEG1(x) ((u_long)(x) | CKSEG1_BASE)
/*
* MIPS64-style segment definitions.
* These allow for 36 bits of addressable physical memory, thus 64GB.
*/
/*
* Cache Coherency Attributes.
*/
/* r8k only */
#define CCA_NC_COPROCESSOR 0UL /* uncached, coprocessor ordered */
/* common to r4, r5k, r8k and r1xk */
#define CCA_NC 2UL /* uncached, write-around */
#define CCA_NONCOHERENT 3UL /* cached, non-coherent, write-back */
/* r8k, r1xk only */
#define CCA_COHERENT_EXCL 4UL /* cached, coherent, exclusive */
#define CCA_COHERENT_EXCLWRITE 5UL /* cached, coherent, exclusive write */
/* r4k only */
#define CCA_COHERENT_UPDWRITE 6UL /* cached, coherent, update on write */
/* r1xk only */
#define CCA_NC_ACCELERATED 7UL /* uncached accelerated */
#ifdef TGT_COHERENT
#define CCA_CACHED CCA_COHERENT_EXCLWRITE
#else
#define CCA_CACHED CCA_NONCOHERENT
#endif
/*
* Uncached spaces.
* R1x000 processors use bits 58:57 of uncached virtual addresses (CCA_NC)
* to select different spaces. Unfortunately, other processors need these
* bits to be zero, so uncached address have to be decided at runtime.
*/
#define SP_HUB 0UL /* Hub space */
#define SP_IO 1UL /* I/O space */
#define SP_SPECIAL 2UL /* Memory Special space */
#define SP_NC 3UL /* Memory Uncached space */
#define XKSSSEG_BASE 0x4000000000000000UL
#define XKPHYS_BASE 0x8000000000000000UL
#define XKSSEG_BASE 0xc000000000000000UL
#define XKPHYS_TO_PHYS(x) ((paddr_t)(x) & 0x0000000fffffffffUL)
#define PHYS_TO_XKPHYS(x,c) ((paddr_t)(x) | XKPHYS_BASE | ((c) << 59))
#define PHYS_TO_XKPHYS_UNCACHED(x,s) \
(PHYS_TO_XKPHYS(x, CCA_NC) | ((s) << 57))
#define IS_XKPHYS(va) (((va) >> 62) == 2)
#define XKPHYS_TO_CCA(x) (((x) >> 59) & 0x07)
#define XKPHYS_TO_SP(x) (((x) >> 57) & 0x03)
#endif /* _LOCORE */
/*
* Exported definitions unique to mips cpu support.
*/
#if defined(_KERNEL) && !defined(_LOCORE)
#include <sys/device.h>
#include <machine/intr.h>
#include <sys/sched.h>
struct cpu_hwinfo {
uint32_t c0prid;
uint32_t c1prid;
uint32_t clock; /* Hz */
uint32_t tlbsize;
uint type;
uint32_t l2size;
};
/*
* Cache memory configuration. One struct per cache.
*/
struct cache_info {
uint size; /* total cache size */
uint linesize; /* line size */
uint setsize; /* set size */
uint sets; /* number of sets */
};
struct cpu_info {
struct device *ci_dev; /* our device */
struct cpu_info *ci_self; /* pointer to this structure */
struct cpu_info *ci_next; /* next cpu */
struct proc *ci_curproc;
struct user *ci_curprocpaddr;
struct proc *ci_fpuproc; /* pointer to last proc to use FP */
uint32_t ci_delayconst;
struct cpu_hwinfo
ci_hw;
#if defined(MULTIPROCESSOR)
struct srp_hazard ci_srp_hazards[SRP_HAZARD_NUM];
#endif
/* cache information and pending flush state */
uint ci_cacheconfiguration;
uint64_t ci_cachepending_l1i;
struct cache_info
ci_l1inst,
ci_l1data,
ci_l2,
ci_l3;
/* function pointers for the cache handling routines */
void (*ci_SyncCache)(struct cpu_info *);
void (*ci_InvalidateICache)(struct cpu_info *, vaddr_t,
size_t);
void (*ci_InvalidateICachePage)(struct cpu_info *, vaddr_t);
void (*ci_SyncICache)(struct cpu_info *);
void (*ci_SyncDCachePage)(struct cpu_info *, vaddr_t,
paddr_t);
void (*ci_HitSyncDCachePage)(struct cpu_info *, vaddr_t,
paddr_t);
void (*ci_HitSyncDCache)(struct cpu_info *, vaddr_t, size_t);
void (*ci_HitInvalidateDCache)(struct cpu_info *, vaddr_t,
size_t);
void (*ci_IOSyncDCache)(struct cpu_info *, vaddr_t, size_t,
int);
struct schedstate_percpu
ci_schedstate;
int ci_want_resched; /* need_resched() invoked */
cpuid_t ci_cpuid; /* our CPU ID */
uint32_t ci_randseed; /* per cpu random seed */
int ci_ipl; /* software IPL */
uint32_t ci_softpending; /* pending soft interrupts */
int ci_clock_started;
u_int32_t ci_cpu_counter_last; /* last compare value loaded */
u_int32_t ci_cpu_counter_interval; /* # of counter ticks/tick */
u_int32_t ci_pendingticks;
#ifdef TGT_ORIGIN
u_int16_t ci_nasid;
u_int16_t ci_slice;
#endif
struct pmap *ci_curpmap;
uint ci_intrdepth; /* interrupt depth */
#ifdef MULTIPROCESSOR
u_long ci_flags; /* flags; see below */
#endif
volatile int ci_ddb;
#define CI_DDB_RUNNING 0
#define CI_DDB_SHOULDSTOP 1
#define CI_DDB_STOPPED 2
#define CI_DDB_ENTERDDB 3
#define CI_DDB_INDDB 4
#ifdef DIAGNOSTIC
int ci_mutex_level;
#endif
#ifdef GPROF
struct gmonparam *ci_gmon;
#endif
};
#define CPUF_PRIMARY 0x01 /* CPU is primary CPU */
#define CPUF_PRESENT 0x02 /* CPU is present */
#define CPUF_RUNNING 0x04 /* CPU is running */
extern struct cpu_info cpu_info_primary;
extern struct cpu_info *cpu_info_list;
#define CPU_INFO_ITERATOR int
#define CPU_INFO_FOREACH(cii, ci) for (cii = 0, ci = cpu_info_list; \
ci != NULL; ci = ci->ci_next)
#define CPU_INFO_UNIT(ci) ((ci)->ci_dev ? (ci)->ci_dev->dv_unit : 0)
extern void (*cpu_idle_cycle_func)(void);
#define cpu_idle_cycle() (*cpu_idle_cycle_func)()
#ifdef MULTIPROCESSOR
#define getcurcpu() hw_getcurcpu()
#define setcurcpu(ci) hw_setcurcpu(ci)
extern struct cpu_info *get_cpu_info(int);
#define curcpu() getcurcpu()
#define CPU_IS_PRIMARY(ci) ((ci)->ci_flags & CPUF_PRIMARY)
#define cpu_number() (curcpu()->ci_cpuid)
extern struct cpuset cpus_running;
void cpu_unidle(struct cpu_info *);
void cpu_boot_secondary_processors(void);
#define cpu_boot_secondary(ci) hw_cpu_boot_secondary(ci)
#define cpu_hatch(ci) hw_cpu_hatch(ci)
vaddr_t alloc_contiguous_pages(size_t);
#define MIPS64_IPI_NOP 0x00000001
#define MIPS64_IPI_RENDEZVOUS 0x00000002
#define MIPS64_IPI_DDB 0x00000004
#define MIPS64_NIPIS 3 /* must not exceed 32 */
void mips64_ipi_init(void);
void mips64_send_ipi(unsigned int, unsigned int);
void smp_rendezvous_cpus(unsigned long, void (*)(void *), void *arg);
#include <sys/mplock.h>
#else
#define MAXCPUS 1
#define curcpu() (&cpu_info_primary)
#define CPU_IS_PRIMARY(ci) 1
#define cpu_number() 0
#define cpu_unidle(ci)
#define get_cpu_info(i) (&cpu_info_primary)
#endif
#define CPU_BUSY_CYCLE() do {} while (0)
extern void (*md_startclock)(struct cpu_info *);
void cp0_calibrate(struct cpu_info *);
#include <machine/frame.h>
/*
* Arguments to hardclock encapsulate the previous machine state in
* an opaque clockframe.
*/
#define clockframe trapframe /* Use normal trap frame */
#define SR_KSU_USER 0x00000010
#define CLKF_USERMODE(framep) ((framep)->sr & SR_KSU_USER)
#define CLKF_PC(framep) ((framep)->pc)
#define CLKF_INTR(framep) (curcpu()->ci_intrdepth > 1) /* XXX */
/*
* This is used during profiling to integrate system time.
*/
#define PROC_PC(p) ((p)->p_md.md_regs->pc)
#define PROC_STACK(p) ((p)->p_md.md_regs->sp)
/*
* Preempt the current process if in interrupt from user mode,
* or after the current trap/syscall if in system mode.
*/
#define need_resched(ci) \
do { \
(ci)->ci_want_resched = 1; \
if ((ci)->ci_curproc != NULL) \
aston((ci)->ci_curproc); \
} while(0)
#define clear_resched(ci) (ci)->ci_want_resched = 0
/*
* Give a profiling tick to the current process when the user profiling
* buffer pages are invalid. On MIPS designs, request an ast to send us
* through trap, marking the proc as needing a profiling tick.
*/
#define need_proftick(p) aston(p)
/*
* Notify the current process (p) that it has a signal pending,
* process as soon as possible.
*/
#ifdef MULTIPROCESSOR
#define signotify(p) (aston(p), cpu_unidle((p)->p_cpu))
#else
#define signotify(p) aston(p)
#endif
#define aston(p) ((p)->p_md.md_astpending = 1)
#ifdef CPU_R8000
#define mips_sync() __asm__ volatile ("lw $0, 0(%0)" :: \
"r" (PHYS_TO_XKPHYS(0, CCA_NC)) : "memory")
#else
#define mips_sync() __asm__ volatile ("sync" ::: "memory")
#endif
#endif /* _KERNEL && !_LOCORE */
#ifdef _KERNEL
/*
* Values for the code field in a break instruction.
*/
#define BREAK_INSTR 0x0000000d
#define BREAK_VAL_MASK 0x03ff0000
#define BREAK_VAL_SHIFT 16
#define BREAK_KDB_VAL 512
#define BREAK_SSTEP_VAL 513
#define BREAK_BRKPT_VAL 514
#define BREAK_SOVER_VAL 515
#define BREAK_DDB_VAL 516
#define BREAK_FPUEMUL_VAL 517
#define BREAK_KDB (BREAK_INSTR | (BREAK_KDB_VAL << BREAK_VAL_SHIFT))
#define BREAK_SSTEP (BREAK_INSTR | (BREAK_SSTEP_VAL << BREAK_VAL_SHIFT))
#define BREAK_BRKPT (BREAK_INSTR | (BREAK_BRKPT_VAL << BREAK_VAL_SHIFT))
#define BREAK_SOVER (BREAK_INSTR | (BREAK_SOVER_VAL << BREAK_VAL_SHIFT))
#define BREAK_DDB (BREAK_INSTR | (BREAK_DDB_VAL << BREAK_VAL_SHIFT))
#define BREAK_FPUEMUL (BREAK_INSTR | (BREAK_FPUEMUL_VAL << BREAK_VAL_SHIFT))
#endif /* _KERNEL */
/*
* CTL_MACHDEP definitions.
*/
#define CPU_ALLOWAPERTURE 1 /* allow mmap of /dev/xf86 */
/* 2 formerly: keyboard reset */
/* 3 formerly: CPU_LIDSUSPEND */
#define CPU_LIDACTION 4 /* action caused by lid close */
#define CPU_MAXID 5 /* number of valid machdep ids */
#define CTL_MACHDEP_NAMES { \
{ 0, 0 }, \
{ "allowaperture", CTLTYPE_INT }, \
{ 0, 0 }, \
{ 0, 0 }, \
{ "lidaction", CTLTYPE_INT }, \
}
/*
* MIPS CPU types (cp_imp).
*/
#define MIPS_R2000 0x01 /* MIPS R2000 CPU ISA I */
#define MIPS_R3000 0x02 /* MIPS R3000 CPU ISA I */
#define MIPS_R6000 0x03 /* MIPS R6000 CPU ISA II */
#define MIPS_R4000 0x04 /* MIPS R4000/4400 CPU ISA III */
#define MIPS_R3LSI 0x05 /* LSI Logic R3000 derivate ISA I */
#define MIPS_R6000A 0x06 /* MIPS R6000A CPU ISA II */
#define MIPS_CN50XX 0x06 /* Cavium OCTEON CN50xx MIPS64R2*/
#define MIPS_R3IDT 0x07 /* IDT R3000 derivate ISA I */
#define MIPS_R10000 0x09 /* MIPS R10000/T5 CPU ISA IV */
#define MIPS_R4200 0x0a /* MIPS R4200 CPU (ICE) ISA III */
#define MIPS_R4300 0x0b /* NEC VR4300 CPU ISA III */
#define MIPS_R4100 0x0c /* NEC VR41xx CPU MIPS-16 ISA III */
#define MIPS_R12000 0x0e /* MIPS R12000 ISA IV */
#define MIPS_R14000 0x0f /* MIPS R14000 ISA IV */
#define MIPS_R8000 0x10 /* MIPS R8000 Blackbird/TFP ISA IV */
#define MIPS_R4600 0x20 /* PMCS R4600 Orion ISA III */
#define MIPS_R4700 0x21 /* PMCS R4700 Orion ISA III */
#define MIPS_R3TOSH 0x22 /* Toshiba R3000 based CPU ISA I */
#define MIPS_R5000 0x23 /* MIPS R5000 CPU ISA IV */
#define MIPS_RM7000 0x27 /* PMCS RM7000 CPU ISA IV */
#define MIPS_RM52X0 0x28 /* PMCS RM52X0 CPU ISA IV */
#define MIPS_RM9000 0x34 /* PMCS RM9000 CPU ISA IV */
#define MIPS_LOONGSON 0x42 /* STC LoongSon CPU ISA III */
#define MIPS_VR5400 0x54 /* NEC Vr5400 CPU ISA IV+ */
#define MIPS_LOONGSON2 0x63 /* STC LoongSon2/3 CPU ISA III+ */
#define MIPS_CN61XX 0x93 /* Cavium OCTEON II CN6[01]xx MIPS64R2 */
#define MIPS_CN71XX 0x96 /* Cavium OCTEON III CN7[01]xx MIPS64R2 */
#define MIPS_CN73XX 0x97 /* Cavium OCTEON III CN7[23]xx MIPS64R2 */
/*
* MIPS FPU types. Only soft, rest is the same as cpu type.
*/
#define MIPS_SOFT 0x00 /* Software emulation ISA I */
#if defined(_KERNEL) && !defined(_LOCORE)
extern register_t protosr;
extern int cpu_has_userlocal;
#ifdef FPUEMUL
#define CPU_HAS_FPU(ci) ((ci)->ci_hw.c1prid != 0)
#else
#define CPU_HAS_FPU(ci) 1
#endif
struct exec_package;
struct user;
void tlb_asid_wrap(struct cpu_info *);
void tlb_flush(int);
void tlb_flush_addr(vaddr_t);
void tlb_init(unsigned int);
int64_t tlb_probe(vaddr_t);
void tlb_set_gbase(vaddr_t, vsize_t);
void tlb_set_page_mask(uint32_t);
void tlb_set_pid(u_int);
void tlb_set_wired(uint32_t);
int tlb_update(vaddr_t, register_t);
void tlb_update_indexed(vaddr_t, register_t, register_t, uint);
void build_trampoline(vaddr_t, vaddr_t);
void cpu_switchto_asm(struct proc *, struct proc *);
int exec_md_map(struct proc *, struct exec_package *);
void savectx(struct user *, int);
void enable_fpu(struct proc *);
void save_fpu(void);
int fpe_branch_emulate(struct proc *, struct trapframe *, uint32_t,
vaddr_t);
void MipsSaveCurFPState(struct proc *);
void MipsSaveCurFPState16(struct proc *);
void MipsSwitchFPState(struct proc *, struct trapframe *);
void MipsSwitchFPState16(struct proc *, struct trapframe *);
int guarded_read_1(paddr_t, uint8_t *);
int guarded_read_2(paddr_t, uint16_t *);
int guarded_read_4(paddr_t, uint32_t *);
int guarded_write_4(paddr_t, uint32_t);
void MipsFPTrap(struct trapframe *);
register_t MipsEmulateBranch(struct trapframe *, vaddr_t, uint32_t, uint32_t);
int classify_insn(uint32_t);
#define INSNCLASS_NEUTRAL 0
#define INSNCLASS_CALL 1
#define INSNCLASS_BRANCH 2
/*
* R4000 end-of-page errata workaround routines
*/
extern int r4000_errata;
u_int eop_page_check(paddr_t);
void eop_tlb_flush_addr(struct pmap *, vaddr_t, u_long);
int eop_tlb_miss_handler(struct trapframe *, struct cpu_info *,
struct proc *);
void eop_cleanup(struct trapframe *, struct proc *);
/*
* Low level access routines to CPU registers
*/
void setsoftintr0(void);
void clearsoftintr0(void);
void setsoftintr1(void);
void clearsoftintr1(void);
register_t enableintr(void);
register_t disableintr(void);
register_t getsr(void);
register_t setsr(register_t);
u_int cp0_get_count(void);
register_t cp0_get_config(void);
uint32_t cp0_get_config_1(void);
uint32_t cp0_get_config_2(void);
uint32_t cp0_get_config_3(void);
uint32_t cp0_get_config_4(void);
uint32_t cp0_get_pagegrain(void);
register_t cp0_get_prid(void);
void cp0_reset_cause(register_t);
void cp0_set_compare(u_int);
void cp0_set_config(register_t);
void cp0_set_pagegrain(uint32_t);
void cp0_set_trapbase(register_t);
u_int cp1_get_prid(void);
static inline uint32_t
cp0_get_hwrena(void)
{
uint32_t value;
__asm__ volatile ("mfc0 %0, $7" : "=r" (value));
return value;
}
static inline void
cp0_set_hwrena(uint32_t value)
{
__asm__ volatile ("mtc0 %0, $7" : : "r" (value));
}
static inline void
cp0_set_userlocal(void *value)
{
__asm__ volatile (
" .set push\n"
" .set mips64r2\n"
" dmtc0 %0, $4, 2\n"
" .set pop\n"
: : "r" (value));
}
static inline u_long
intr_disable(void)
{
return disableintr();
}
static inline void
intr_restore(u_long sr)
{
setsr(sr);
}
/*
* Cache routines (may be overridden)
*/
#ifndef Mips_SyncCache
#define Mips_SyncCache(ci) \
((ci)->ci_SyncCache)(ci)
#endif
#ifndef Mips_InvalidateICache
#define Mips_InvalidateICache(ci, va, l) \
((ci)->ci_InvalidateICache)(ci, va, l)
#endif
#ifndef Mips_InvalidateICachePage
#define Mips_InvalidateICachePage(ci, va) \
((ci)->ci_InvalidateICachePage)(ci, va)
#endif
#ifndef Mips_SyncICache
#define Mips_SyncICache(ci) \
((ci)->ci_SyncICache)(ci)
#endif
#ifndef Mips_SyncDCachePage
#define Mips_SyncDCachePage(ci, va, pa) \
((ci)->ci_SyncDCachePage)(ci, va, pa)
#endif
#ifndef Mips_HitSyncDCachePage
#define Mips_HitSyncDCachePage(ci, va, pa) \
((ci)->ci_HitSyncDCachePage)(ci, va, pa)
#endif
#ifndef Mips_HitSyncDCache
#define Mips_HitSyncDCache(ci, va, l) \
((ci)->ci_HitSyncDCache)(ci, va, l)
#endif
#ifndef Mips_HitInvalidateDCache
#define Mips_HitInvalidateDCache(ci, va, l) \
((ci)->ci_HitInvalidateDCache)(ci, va, l)
#endif
#ifndef Mips_IOSyncDCache
#define Mips_IOSyncDCache(ci, va, l, h) \
((ci)->ci_IOSyncDCache)(ci, va, l, h)
#endif
#endif /* _KERNEL && !_LOCORE */
#endif /* !_MIPS64_CPU_H_ */
|