forked from t2linux/apple-bce-drv
-
Notifications
You must be signed in to change notification settings - Fork 2
/
queue_dma.c
222 lines (199 loc) · 7.79 KB
/
queue_dma.c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
#include "queue_dma.h"
#include <linux/vmalloc.h>
#include <linux/mm.h>
#include "queue.h"
static int bce_alloc_scatterlist_from_vm(struct sg_table *tbl, void *data, size_t len);
static struct bce_segment_list_element_hostinfo *bce_map_segment_list(
struct device *dev, struct scatterlist *pages, int pagen);
static void bce_unmap_segement_list(struct device *dev, struct bce_segment_list_element_hostinfo *list);
int bce_map_dma_buffer(struct device *dev, struct bce_dma_buffer *buf, struct sg_table scatterlist,
enum dma_data_direction dir)
{
int cnt;
buf->direction = dir;
buf->scatterlist = scatterlist;
buf->seglist_hostinfo = NULL;
cnt = dma_map_sg(dev, buf->scatterlist.sgl, buf->scatterlist.nents, dir);
if (cnt != buf->scatterlist.nents) {
pr_err("apple-bce: DMA scatter list mapping returned an unexpected count: %i\n", cnt);
dma_unmap_sg(dev, buf->scatterlist.sgl, buf->scatterlist.nents, dir);
return -EIO;
}
if (cnt == 1)
return 0;
buf->seglist_hostinfo = bce_map_segment_list(dev, buf->scatterlist.sgl, buf->scatterlist.nents);
if (!buf->seglist_hostinfo) {
pr_err("apple-bce: Creating segment list failed\n");
dma_unmap_sg(dev, buf->scatterlist.sgl, buf->scatterlist.nents, dir);
return -EIO;
}
return 0;
}
int bce_map_dma_buffer_vm(struct device *dev, struct bce_dma_buffer *buf, void *data, size_t len,
enum dma_data_direction dir)
{
int status;
struct sg_table scatterlist;
if ((status = bce_alloc_scatterlist_from_vm(&scatterlist, data, len)))
return status;
if ((status = bce_map_dma_buffer(dev, buf, scatterlist, dir))) {
sg_free_table(&scatterlist);
return status;
}
return 0;
}
int bce_map_dma_buffer_km(struct device *dev, struct bce_dma_buffer *buf, void *data, size_t len,
enum dma_data_direction dir)
{
/* Kernel memory is continuous which is great for us. */
int status;
struct sg_table scatterlist;
if ((status = sg_alloc_table(&scatterlist, 1, GFP_KERNEL))) {
sg_free_table(&scatterlist);
return status;
}
sg_set_buf(scatterlist.sgl, data, (uint) len);
if ((status = bce_map_dma_buffer(dev, buf, scatterlist, dir))) {
sg_free_table(&scatterlist);
return status;
}
return 0;
}
void bce_unmap_dma_buffer(struct device *dev, struct bce_dma_buffer *buf)
{
dma_unmap_sg(dev, buf->scatterlist.sgl, buf->scatterlist.nents, buf->direction);
bce_unmap_segement_list(dev, buf->seglist_hostinfo);
}
static int bce_alloc_scatterlist_from_vm(struct sg_table *tbl, void *data, size_t len)
{
int status, i;
struct page **pages;
size_t off, start_page, end_page, page_count;
off = (size_t) data % PAGE_SIZE;
start_page = (size_t) data / PAGE_SIZE;
end_page = ((size_t) data + len - 1) / PAGE_SIZE;
page_count = end_page - start_page + 1;
// Ensure all our pointers to page structs fit in a continuous bit of memory
if (page_count > PAGE_SIZE / sizeof(struct page *))
pages = vmalloc(page_count * sizeof(struct page *)); //TODO: are there flags to make this continuois?
else
pages = kmalloc(page_count * sizeof(struct page *), GFP_KERNEL);
for (i = 0; i < page_count; i++)
pages[i] = vmalloc_to_page((void *) ((start_page + i) * PAGE_SIZE));
if ((status = sg_alloc_table_from_pages(tbl, pages, page_count, (unsigned int) off, len, GFP_KERNEL))) {
sg_free_table(tbl);
}
if (page_count > PAGE_SIZE / sizeof(struct page *))
vfree(pages);
else
kfree(pages);
return status;
}
#define BCE_ELEMENTS_PER_PAGE ((PAGE_SIZE - sizeof(struct bce_segment_list_header)) \
/ sizeof(struct bce_segment_list_element))
#define BCE_ELEMENTS_PER_ADDITIONAL_PAGE (PAGE_SIZE / sizeof(struct bce_segment_list_element))
static struct bce_segment_list_element_hostinfo *bce_map_segment_list(
struct device *dev, struct scatterlist *pages, int pagen)
{
size_t ptr, pptr = 0;
struct bce_segment_list_header theader; /* a temp header, to store the initial seg */
struct bce_segment_list_header *header;
struct bce_segment_list_element *el, *el_end;
struct bce_segment_list_element_hostinfo *out, *pout, *out_root;
struct scatterlist *sg;
int i;
header = &theader;
out = out_root = NULL;
el = el_end = NULL;
for_each_sg(pages, sg, pagen, i) {
if (el >= el_end) {
/* allocate a new page, this will be also done for the first element */
ptr = __get_free_page(GFP_KERNEL); //TODO when is this free'd
if (pptr && ptr == pptr + PAGE_SIZE) { // Got adjacent page
out->page_count++;
header->element_count += BCE_ELEMENTS_PER_ADDITIONAL_PAGE;
el_end += BCE_ELEMENTS_PER_ADDITIONAL_PAGE;
} else {
header = (void *) ptr;
header->element_count = BCE_ELEMENTS_PER_PAGE;
header->data_size = 0;
header->next_segl_addr = 0;
header->next_segl_length = 0;
el = (void *) (header + 1);
el_end = el + BCE_ELEMENTS_PER_PAGE;
if (out) {
out->next = kmalloc(sizeof(struct bce_segment_list_element_hostinfo), GFP_KERNEL);
out = out->next;
} else {
out_root = out = kmalloc(sizeof(struct bce_segment_list_element_hostinfo), GFP_KERNEL);
}
out->page_start = (void *) ptr;
out->page_count = 1;
out->dma_start = DMA_MAPPING_ERROR;
out->next = NULL;
}
pptr = ptr;
}
el->addr = sg->dma_address;
el->length = sg->length;
header->data_size += el->length;
//header->element_count++? increasing it above means we can have half filled pages leading to incorrect count
}
/* DMA map */
out = out_root;
pout = NULL;
while (out) {
out->dma_start = dma_map_single(dev, out->page_start, out->page_count * PAGE_SIZE, DMA_TO_DEVICE);
if (dma_mapping_error(dev, out->dma_start))
goto error;
if (pout) {
header = pout->page_start;
header->next_segl_addr = out->dma_start;
header->next_segl_length = out->page_count * PAGE_SIZE;
}
pout = out;
out = out->next;
}
return out_root;
error:
bce_unmap_segement_list(dev, out_root);
return NULL;
}
static void bce_unmap_segement_list(struct device *dev, struct bce_segment_list_element_hostinfo *list)
{
struct bce_segment_list_element_hostinfo *next;
while (list) {
if (list->dma_start != DMA_MAPPING_ERROR)
dma_unmap_single(dev, list->dma_start, list->page_count * PAGE_SIZE, DMA_TO_DEVICE);
next = list->next;
kfree(list);
list = next;
}
}
int bce_set_submission_buf(struct bce_qe_submission *element, struct bce_dma_buffer *buf, size_t offset, size_t length)
{
struct bce_segment_list_element_hostinfo *seg;
struct bce_segment_list_header *seg_header;
seg = buf->seglist_hostinfo;
if (!seg) {
element->addr = buf->scatterlist.sgl->dma_address + offset;
element->length = length;
element->segl_addr = 0;
element->segl_length = 0;
return 0;
}
while (seg) {
seg_header = seg->page_start;
if (offset <= seg_header->data_size)
break;
offset -= seg_header->data_size;
seg = seg->next;
}
if (!seg)
return -EINVAL;
element->addr = offset;
element->length = buf->scatterlist.sgl->dma_length;
element->segl_addr = seg->dma_start;
element->segl_length = seg->page_count * PAGE_SIZE;
return 0;
}