summaryrefslogtreecommitdiff
path: root/sys/dev/pv/hypervvar.h
blob: 41c973bef46d3ccace7f71b31dad3f8b23993ed3 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
/*
 * Copyright (c) 2016 Mike Belopuhov <mike@esdenera.com>
 *
 * Permission to use, copy, modify, and distribute this software for any
 * purpose with or without fee is hereby granted, provided that the above
 * copyright notice and this permission notice appear in all copies.
 *
 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
 */

#ifndef _HYPERVVAR_H_
#define _HYPERVVAR_H_

#define HYPERV_DEBUG

#ifdef HYPERV_DEBUG
#define DPRINTF(x...)		printf(x)
#else
#define DPRINTF(x...)
#endif

struct hv_softc;

struct hv_msg {
	uint64_t			 msg_flags;
#define  MSGF_NOSLEEP			  0x0001
#define  MSGF_NOQUEUE			  0x0002
#define  MSGF_ORPHANED			  0x0004
	struct hv_input_post_message	 msg_req;
	void				*msg_rsp;
	size_t				 msg_rsplen;
	TAILQ_ENTRY(hv_msg)		 msg_entry;
};
TAILQ_HEAD(hv_queue, hv_msg);

struct hv_offer {
	struct hv_channel_offer_channel	 co_chan;
	SIMPLEQ_ENTRY(hv_offer)		 co_entry;
};
SIMPLEQ_HEAD(hv_offers, hv_offer);

struct hv_ring_data {
	struct hv_ring_buffer		*rd_ring;
	uint32_t			 rd_size;
	struct mutex			 rd_lock;
	uint32_t			 rd_prod;
	uint32_t			 rd_cons;
	uint32_t			 rd_data_size;
	uint32_t			 rd_data_offset;
};

struct hv_channel {
	struct hv_softc			*ch_sc;

	int				 ch_state;
#define  HV_CHANSTATE_OFFERED		  1
#define  HV_CHANSTATE_OPENED		  2
#define  HV_CHANSTATE_CLOSING		  3
#define  HV_CHANSTATE_CLOSED		  4
	uint32_t			 ch_relid;

	struct hv_guid			 ch_type;
	struct hv_guid			 ch_inst;
	char				 ch_ident[38];

	uint8_t				 ch_mgroup;
	uint8_t				 ch_mindex;

	void				*ch_ring;
	uint32_t			 ch_ring_hndl;
	uint32_t			 ch_ring_npg;
	ulong				 ch_ring_size;

	struct hv_ring_data		 ch_wrd;
	struct hv_ring_data		 ch_rrd;

	uint32_t			 ch_vcpu;

	void				(*ch_handler)(void *);
	void				 *ch_ctx;
	uint8_t				 *ch_buf;
	int				  ch_buflen;
	struct evcount			  ch_evcnt;

	uint32_t			 ch_flags;
#define  CHF_BATCHED			  0x0001
#define  CHF_DEDICATED			  0x0002
#define  CHF_MONITOR			  0x0004

	struct hv_input_signal_event	 ch_sigevt __attribute__((aligned(8)));

	TAILQ_ENTRY(hv_channel)		 ch_entry;
};
TAILQ_HEAD(hv_channels, hv_channel);

struct hv_attach_args {
	void				*aa_parent;
	bus_dma_tag_t			 aa_dmat;
	struct hv_guid			*aa_type;
	struct hv_guid			*aa_inst;
	char				*aa_ident;
	struct hv_channel		*aa_chan;
};

struct hv_dev {
	struct hv_attach_args		 dv_aa;
	SLIST_ENTRY(hv_dev)		 dv_entry;
};
SLIST_HEAD(hv_devices, hv_dev);

struct hv_softc {
	struct device			 sc_dev;
	struct pvbus_hv			*sc_pvbus;
	struct bus_dma_tag		*sc_dmat;

	void				*sc_hc;
	uint32_t			 sc_features;

	uint32_t			 sc_flags;
#define  HSF_CONNECTED			  0x0001
#define  HSF_OFFERS_DELIVERED		  0x0002

	int				 sc_idtvec;
	int				 sc_proto;

	/* CPU id to VCPU id mapping */
	uint32_t			 sc_vcpus[1];	/* XXX: per-cpu */
	/* Synthetic Interrupt Message Page (SIMP) */
	void				*sc_simp[1];	/* XXX: per-cpu */
	/* Synthetic Interrupt Event Flags Page (SIEFP) */
	void				*sc_siep[1];	/* XXX: per-cpu */

	/* Channel port events page */
	void				*sc_events;
	uint32_t			*sc_wevents;	/* Write events */
	uint32_t			*sc_revents;	/* Read events */

	/* Monitor pages for parent<->child notifications */
	struct hv_monitor_page		*sc_monitor[2];

	struct hv_queue	 		 sc_reqs;	/* Request queue */
	struct mutex			 sc_reqlck;
	struct hv_queue	 		 sc_rsps;	/* Response queue */
	struct mutex			 sc_rsplck;

	struct hv_offers		 sc_offers;
	struct mutex			 sc_offerlck;

	struct hv_channels		 sc_channels;
	struct mutex			 sc_channelck;

	volatile uint32_t		 sc_handle;

	struct hv_devices		 sc_devs;
	struct mutex			 sc_devlck;

	struct task			 sc_sdtask;	/* shutdown */

	struct ksensordev		 sc_sensordev;
	struct ksensor			 sc_sensor;
};

static inline int
atomic_setbit_ptr(volatile void *ptr, int bit)
{
	int obit;

	__asm__ __volatile__ ("lock btsl %2,%1; sbbl %0,%0" :
	    "=r" (obit), "=m" (*(volatile long *)ptr) : "Ir" (bit) :
	    "memory");

	return (obit);
}

static inline int
atomic_clearbit_ptr(volatile void *ptr, int bit)
{
	int obit;

	__asm__ __volatile__ ("lock btrl %2,%1; sbbl %0,%0" :
	    "=r" (obit), "=m" (*(volatile long *)ptr) : "Ir" (bit) :
	    "memory");

	return (obit);
}

int	hv_handle_alloc(struct hv_channel *, void *, uint32_t, uint32_t *);
void	hv_handle_free(struct hv_channel *, uint32_t);
int	hv_channel_open(struct hv_channel *, void *, size_t, void (*)(void *),
	    void *);
int	hv_channel_close(struct hv_channel *);
int	hv_channel_send(struct hv_channel *, void *, uint32_t, uint64_t,
	    int, uint32_t);
int	hv_channel_sendbuf(struct hv_channel *, struct hv_page_buffer *,
	    uint32_t, void *, uint32_t, uint64_t);
int	hv_channel_recv(struct hv_channel *, void *, uint32_t, uint32_t *,
	    uint64_t *, int);

#endif	/* _HYPERVVAR_H_ */