1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
|
/*
* Copyright (c) 2023-2024 Ian Marco Moffett and the Osmora Team.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
*
* 1. Redistributions of source code must retain the above copyright notice,
* this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. Neither the name of Hyra nor the names of its
* contributors may be used to endorse or promote products derived from
* this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/
#ifndef _AMD64_CPU_H_
#define _AMD64_CPU_H_
#include <sys/types.h>
#include <sys/cdefs.h>
#include <sys/spinlock.h>
#include <sys/sched_state.h>
#include <sys/queue.h>
#include <machine/tss.h>
#include <machine/msr.h>
/*
* XXX: We are not using the PAUSE instruction for the sake of
* ensuring compatibility... PAUSE is F3 90, REP NOP is
* F3 90... REP NOP will be read as a PAUSE on processors
* that support it.
*/
#define hint_spinwait() __ASMV("rep; nop")
#define this_cpu() amd64_this_cpu()
#define get_bsp() amd64_get_bsp()
#define is_intr_mask() amd64_is_intr_mask()
#define CPU_INFO_LOCK(info) spinlock_acquire(&(info->lock))
#define CPU_INFO_UNLOCK(info) spinlock_release(&(info->lock))
/*
* Info about a specific processor.
*
* XXX: Spinlock must be acquired outside of this module!
* None of these module's internal functions should
* acquire the spinlock themselves!
*/
struct cpu_info {
/* Per-arch fields */
void *pmap; /* Current pmap */
uint32_t id;
uint32_t idx;
struct spinlock lock;
struct sched_state sched_state;
TAILQ_ENTRY(cpu_info) link;
/* AMD64 */
volatile size_t lapic_tmr_freq;
volatile void *lapic_base;
volatile bool has_x2apic;
volatile struct tss_entry *tss;
};
/*
* Contains information for the current
* core. Stored in %GS.
*
* MUST REMAIN IN ORDER!!!
*/
struct cpu_ctx {
struct cpu_info *ci;
};
/*
* Returns true for this core if maskable
* interrupts are masked (CLI) and false if
* they aren't (STI).
*/
static inline bool
amd64_is_intr_mask(void)
{
uint64_t flags;
__ASMV("pushfq; pop %0" : "=rm" (flags) :: "memory");
return __TEST(flags, 1 << 9);
}
static inline void
amd64_write_gs_base(uintptr_t val)
{
wrmsr(IA32_KERNEL_GS_BASE, val);
}
static inline uintptr_t
amd64_read_gs_base(void)
{
return rdmsr(IA32_KERNEL_GS_BASE);
}
struct cpu_info *amd64_this_cpu(void);
#endif /* !_AMD64_CPU_H_ */
|