vchiq_2835_arm.c revision 278615
1/**
2 * Copyright (c) 2010-2012 Broadcom. All rights reserved.
3 *
4 * Redistribution and use in source and binary forms, with or without
5 * modification, are permitted provided that the following conditions
6 * are met:
7 * 1. Redistributions of source code must retain the above copyright
8 *    notice, this list of conditions, and the following disclaimer,
9 *    without modification.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 *    notice, this list of conditions and the following disclaimer in the
12 *    documentation and/or other materials provided with the distribution.
13 * 3. The names of the above-listed copyright holders may not be used
14 *    to endorse or promote products derived from this software without
15 *    specific prior written permission.
16 *
17 * ALTERNATIVELY, this software may be distributed under the terms of the
18 * GNU General Public License ("GPL") version 2, as published by the Free
19 * Software Foundation.
20 *
21 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS
22 * IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
23 * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
24 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
25 * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
26 * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
27 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
28 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
29 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
30 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
31 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
32 */
33
34#include <interface/compat/vchi_bsd.h>
35
36#include <sys/malloc.h>
37#include <sys/rwlock.h>
38
39#include <vm/vm.h>
40#include <vm/pmap.h>
41#include <vm/vm_extern.h>
42#include <vm/vm_kern.h>
43#include <vm/vm_map.h>
44#include <vm/vm_object.h>
45#include <vm/vm_page.h>
46#include <vm/vm_pager.h>
47#include <vm/vm_param.h>
48#include <vm/vm_phys.h>
49
50#include <machine/bus.h>
51#include <arm/broadcom/bcm2835/bcm2835_mbox.h>
52#include <arm/broadcom/bcm2835/bcm2835_vcbus.h>
53
54MALLOC_DEFINE(M_VCPAGELIST, "vcpagelist", "VideoCore pagelist memory");
55
56#define TOTAL_SLOTS (VCHIQ_SLOT_ZERO_SLOTS + 2 * 32)
57
58#define VCHIQ_DOORBELL_IRQ IRQ_ARM_DOORBELL_0
59#define VCHIQ_ARM_ADDRESS(x) ((void *)PHYS_TO_VCBUS(pmap_kextract((vm_offset_t)(x))))
60
61#include "vchiq_arm.h"
62#include "vchiq_2835.h"
63#include "vchiq_connected.h"
64
65#define MAX_FRAGMENTS (VCHIQ_NUM_CURRENT_BULKS * 2)
66
67typedef struct vchiq_2835_state_struct {
68   int inited;
69   VCHIQ_ARM_STATE_T arm_state;
70} VCHIQ_2835_ARM_STATE_T;
71
72static char *g_slot_mem;
73static int g_slot_mem_size;
74vm_paddr_t g_slot_phys;
75/* BSD DMA */
76bus_dma_tag_t bcm_slots_dma_tag;
77bus_dmamap_t bcm_slots_dma_map;
78
79static FRAGMENTS_T *g_fragments_base;
80static FRAGMENTS_T *g_free_fragments;
81struct semaphore g_free_fragments_sema;
82
83static DEFINE_SEMAPHORE(g_free_fragments_mutex);
84
85typedef struct bulkinfo_struct {
86	PAGELIST_T	*pagelist;
87	bus_dma_tag_t	pagelist_dma_tag;
88	bus_dmamap_t	pagelist_dma_map;
89	void		*buf;
90	size_t		size;
91} BULKINFO_T;
92
93static int
94create_pagelist(char __user *buf, size_t count, unsigned short type,
95                struct proc *p, BULKINFO_T *bi);
96
97static void
98free_pagelist(BULKINFO_T *bi, int actual);
99
100static void
101vchiq_dmamap_cb(void *arg, bus_dma_segment_t *segs, int nseg, int err)
102{
103	bus_addr_t *addr;
104
105	if (err)
106		return;
107
108	addr = (bus_addr_t*)arg;
109	*addr = PHYS_TO_VCBUS(segs[0].ds_addr);
110}
111
112int __init
113vchiq_platform_init(VCHIQ_STATE_T *state)
114{
115	VCHIQ_SLOT_ZERO_T *vchiq_slot_zero;
116	int frag_mem_size;
117	int err;
118	int i;
119
120	/* Allocate space for the channels in coherent memory */
121	g_slot_mem_size = PAGE_ALIGN(TOTAL_SLOTS * VCHIQ_SLOT_SIZE);
122	frag_mem_size = PAGE_ALIGN(sizeof(FRAGMENTS_T) * MAX_FRAGMENTS);
123
124	err = bus_dma_tag_create(
125	    NULL,
126	    PAGE_SIZE, 0,	       /* alignment, boundary */
127	    BUS_SPACE_MAXADDR_32BIT,    /* lowaddr */
128	    BUS_SPACE_MAXADDR,	  /* highaddr */
129	    NULL, NULL,		 /* filter, filterarg */
130	    g_slot_mem_size + frag_mem_size, 1,		/* maxsize, nsegments */
131	    g_slot_mem_size + frag_mem_size, 0,		/* maxsegsize, flags */
132	    NULL, NULL,		 /* lockfunc, lockarg */
133	    &bcm_slots_dma_tag);
134
135	err = bus_dmamem_alloc(bcm_slots_dma_tag, (void **)&g_slot_mem,
136	    BUS_DMA_COHERENT | BUS_DMA_WAITOK, &bcm_slots_dma_map);
137	if (err) {
138		vchiq_log_error(vchiq_core_log_level, "Unable to allocate channel memory");
139		err = -ENOMEM;
140		goto failed_alloc;
141	}
142
143	err = bus_dmamap_load(bcm_slots_dma_tag, bcm_slots_dma_map, g_slot_mem,
144	    g_slot_mem_size + frag_mem_size, vchiq_dmamap_cb,
145	    &g_slot_phys, 0);
146
147	if (err) {
148		vchiq_log_error(vchiq_core_log_level, "cannot load DMA map");
149		err = -ENOMEM;
150		goto failed_load;
151	}
152
153	WARN_ON(((int)g_slot_mem & (PAGE_SIZE - 1)) != 0);
154
155	vchiq_slot_zero = vchiq_init_slots(g_slot_mem, g_slot_mem_size);
156	if (!vchiq_slot_zero) {
157		err = -EINVAL;
158		goto failed_init_slots;
159	}
160
161	vchiq_slot_zero->platform_data[VCHIQ_PLATFORM_FRAGMENTS_OFFSET_IDX] =
162		(int)g_slot_phys + g_slot_mem_size;
163	vchiq_slot_zero->platform_data[VCHIQ_PLATFORM_FRAGMENTS_COUNT_IDX] =
164		MAX_FRAGMENTS;
165
166	g_fragments_base = (FRAGMENTS_T *)(g_slot_mem + g_slot_mem_size);
167	g_slot_mem_size += frag_mem_size;
168
169	g_free_fragments = g_fragments_base;
170	for (i = 0; i < (MAX_FRAGMENTS - 1); i++) {
171		*(FRAGMENTS_T **)&g_fragments_base[i] =
172			&g_fragments_base[i + 1];
173	}
174	*(FRAGMENTS_T **)&g_fragments_base[i] = NULL;
175	_sema_init(&g_free_fragments_sema, MAX_FRAGMENTS);
176
177	if (vchiq_init_state(state, vchiq_slot_zero, 0/*slave*/) !=
178		VCHIQ_SUCCESS) {
179		err = -EINVAL;
180		goto failed_vchiq_init;
181	}
182
183	bcm_mbox_write(BCM2835_MBOX_CHAN_VCHIQ, (unsigned int)g_slot_phys);
184
185	vchiq_log_info(vchiq_arm_log_level,
186		"vchiq_init - done (slots %x, phys %x)",
187		(unsigned int)vchiq_slot_zero, g_slot_phys);
188
189   vchiq_call_connected_callbacks();
190
191   return 0;
192
193failed_vchiq_init:
194failed_init_slots:
195failed_load:
196	bus_dmamap_unload(bcm_slots_dma_tag, bcm_slots_dma_map);
197failed_alloc:
198	bus_dmamap_destroy(bcm_slots_dma_tag, bcm_slots_dma_map);
199	bus_dma_tag_destroy(bcm_slots_dma_tag);
200
201   return err;
202}
203
204void __exit
205vchiq_platform_exit(VCHIQ_STATE_T *state)
206{
207
208	bus_dmamap_unload(bcm_slots_dma_tag, bcm_slots_dma_map);
209	bus_dmamap_destroy(bcm_slots_dma_tag, bcm_slots_dma_map);
210	bus_dma_tag_destroy(bcm_slots_dma_tag);
211}
212
213VCHIQ_STATUS_T
214vchiq_platform_init_state(VCHIQ_STATE_T *state)
215{
216   VCHIQ_STATUS_T status = VCHIQ_SUCCESS;
217   state->platform_state = kzalloc(sizeof(VCHIQ_2835_ARM_STATE_T), GFP_KERNEL);
218   ((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->inited = 1;
219   status = vchiq_arm_init_state(state, &((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->arm_state);
220   if(status != VCHIQ_SUCCESS)
221   {
222      ((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->inited = 0;
223   }
224   return status;
225}
226
227VCHIQ_ARM_STATE_T*
228vchiq_platform_get_arm_state(VCHIQ_STATE_T *state)
229{
230   if(!((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->inited)
231   {
232      BUG();
233   }
234   return &((VCHIQ_2835_ARM_STATE_T*)state->platform_state)->arm_state;
235}
236
237int
238vchiq_copy_from_user(void *dst, const void *src, int size)
239{
240
241	if (((vm_offset_t)(src)) < VM_MIN_KERNEL_ADDRESS) {
242		int error = copyin(src, dst, size);
243		return error ? VCHIQ_ERROR : VCHIQ_SUCCESS;
244	}
245	else
246		bcopy(src, dst, size);
247
248	return 0;
249}
250
251VCHIQ_STATUS_T
252vchiq_prepare_bulk_data(VCHIQ_BULK_T *bulk, VCHI_MEM_HANDLE_T memhandle,
253	void *offset, int size, int dir)
254{
255	BULKINFO_T *bi;
256	int ret;
257
258	WARN_ON(memhandle != VCHI_MEM_HANDLE_INVALID);
259	bi = malloc(sizeof(*bi), M_VCPAGELIST, M_WAITOK | M_ZERO);
260	if (bi == NULL)
261		return VCHIQ_ERROR;
262
263	ret = create_pagelist((char __user *)offset, size,
264			(dir == VCHIQ_BULK_RECEIVE)
265			? PAGELIST_READ
266			: PAGELIST_WRITE,
267			current,
268			bi);
269	if (ret != 0)
270		return VCHIQ_ERROR;
271
272	bulk->handle = memhandle;
273	bulk->data = VCHIQ_ARM_ADDRESS(bi->pagelist);
274
275	/* Store the pagelist address in remote_data, which isn't used by the
276	   slave. */
277	bulk->remote_data = bi;
278
279	return VCHIQ_SUCCESS;
280}
281
282void
283vchiq_complete_bulk(VCHIQ_BULK_T *bulk)
284{
285	if (bulk && bulk->remote_data && bulk->actual)
286		free_pagelist((BULKINFO_T *)bulk->remote_data, bulk->actual);
287}
288
289void
290vchiq_transfer_bulk(VCHIQ_BULK_T *bulk)
291{
292	/*
293	 * This should only be called on the master (VideoCore) side, but
294	 * provide an implementation to avoid the need for ifdefery.
295	 */
296	BUG();
297}
298
299void
300vchiq_dump_platform_state(void *dump_context)
301{
302	char buf[80];
303	int len;
304	len = snprintf(buf, sizeof(buf),
305		"  Platform: 2835 (VC master)");
306	vchiq_dump(dump_context, buf, len + 1);
307}
308
309VCHIQ_STATUS_T
310vchiq_platform_suspend(VCHIQ_STATE_T *state)
311{
312   return VCHIQ_ERROR;
313}
314
315VCHIQ_STATUS_T
316vchiq_platform_resume(VCHIQ_STATE_T *state)
317{
318   return VCHIQ_SUCCESS;
319}
320
321void
322vchiq_platform_paused(VCHIQ_STATE_T *state)
323{
324}
325
326void
327vchiq_platform_resumed(VCHIQ_STATE_T *state)
328{
329}
330
331int
332vchiq_platform_videocore_wanted(VCHIQ_STATE_T* state)
333{
334   return 1; // autosuspend not supported - videocore always wanted
335}
336
337int
338vchiq_platform_use_suspend_timer(void)
339{
340   return 0;
341}
342void
343vchiq_dump_platform_use_state(VCHIQ_STATE_T *state)
344{
345	vchiq_log_info(vchiq_arm_log_level, "Suspend timer not in use");
346}
347void
348vchiq_platform_handle_timeout(VCHIQ_STATE_T *state)
349{
350	(void)state;
351}
352/*
353 * Local functions
354 */
355
356static void
357pagelist_page_free(vm_page_t pp)
358{
359	vm_page_lock(pp);
360	vm_page_unwire(pp, PQ_INACTIVE);
361	if (pp->wire_count == 0 && pp->object == NULL)
362		vm_page_free(pp);
363	vm_page_unlock(pp);
364}
365
366/* There is a potential problem with partial cache lines (pages?)
367** at the ends of the block when reading. If the CPU accessed anything in
368** the same line (page?) then it may have pulled old data into the cache,
369** obscuring the new data underneath. We can solve this by transferring the
370** partial cache lines separately, and allowing the ARM to copy into the
371** cached area.
372
373** N.B. This implementation plays slightly fast and loose with the Linux
374** driver programming rules, e.g. its use of __virt_to_bus instead of
375** dma_map_single, but it isn't a multi-platform driver and it benefits
376** from increased speed as a result.
377*/
378
379static int
380create_pagelist(char __user *buf, size_t count, unsigned short type,
381	struct proc *p, BULKINFO_T *bi)
382{
383	PAGELIST_T *pagelist;
384	vm_page_t* pages;
385	unsigned long *addrs;
386	unsigned int num_pages, i;
387	vm_offset_t offset;
388	int pagelist_size;
389	char *addr, *base_addr, *next_addr;
390	int run, addridx, actual_pages;
391	int err;
392	vm_paddr_t pagelist_phys;
393
394	offset = (vm_offset_t)buf & (PAGE_SIZE - 1);
395	num_pages = (count + offset + PAGE_SIZE - 1) / PAGE_SIZE;
396
397	bi->pagelist = NULL;
398	bi->buf = buf;
399	bi->size = count;
400
401	/* Allocate enough storage to hold the page pointers and the page
402	** list
403	*/
404	pagelist_size = sizeof(PAGELIST_T) +
405		(num_pages * sizeof(unsigned long)) +
406		(num_pages * sizeof(pages[0]));
407
408	err = bus_dma_tag_create(
409	    NULL,
410	    PAGE_SIZE, 0,	       /* alignment, boundary */
411	    BUS_SPACE_MAXADDR_32BIT,    /* lowaddr */
412	    BUS_SPACE_MAXADDR,	  /* highaddr */
413	    NULL, NULL,		 /* filter, filterarg */
414	    pagelist_size, 1,		/* maxsize, nsegments */
415	    pagelist_size, 0,		/* maxsegsize, flags */
416	    NULL, NULL,		 /* lockfunc, lockarg */
417	    &bi->pagelist_dma_tag);
418
419	err = bus_dmamem_alloc(bi->pagelist_dma_tag, (void **)&pagelist,
420	    BUS_DMA_COHERENT | BUS_DMA_WAITOK, &bi->pagelist_dma_map);
421	if (err) {
422		vchiq_log_error(vchiq_core_log_level, "Unable to allocate pagelist memory");
423		err = -ENOMEM;
424		goto failed_alloc;
425	}
426
427	err = bus_dmamap_load(bi->pagelist_dma_tag, bi->pagelist_dma_map, pagelist,
428	    pagelist_size, vchiq_dmamap_cb,
429	    &pagelist_phys, 0);
430
431	if (err) {
432		vchiq_log_error(vchiq_core_log_level, "cannot load DMA map for pagelist memory");
433		err = -ENOMEM;
434		goto failed_load;
435	}
436
437	vchiq_log_trace(vchiq_arm_log_level,
438		"create_pagelist - %x", (unsigned int)pagelist);
439	if (!pagelist)
440		return -ENOMEM;
441
442	addrs = pagelist->addrs;
443	pages = (vm_page_t*)(addrs + num_pages);
444
445	actual_pages = vm_fault_quick_hold_pages(&p->p_vmspace->vm_map,
446	    (vm_offset_t)buf, count,
447	    (type == PAGELIST_READ ? VM_PROT_WRITE : 0 ) | VM_PROT_READ, pages, num_pages);
448
449	if (actual_pages != num_pages) {
450		vm_page_unhold_pages(pages, actual_pages);
451		free(pagelist, M_VCPAGELIST);
452		return (-ENOMEM);
453	}
454
455	for (i = 0; i < actual_pages; i++) {
456		vm_page_lock(pages[i]);
457		vm_page_wire(pages[i]);
458		vm_page_unhold(pages[i]);
459		vm_page_unlock(pages[i]);
460	}
461
462	pagelist->length = count;
463	pagelist->type = type;
464	pagelist->offset = offset;
465
466	/* Group the pages into runs of contiguous pages */
467
468	base_addr = (void *)PHYS_TO_VCBUS(VM_PAGE_TO_PHYS(pages[0]));
469	next_addr = base_addr + PAGE_SIZE;
470	addridx = 0;
471	run = 0;
472
473	for (i = 1; i < num_pages; i++) {
474		addr = (void *)PHYS_TO_VCBUS(VM_PAGE_TO_PHYS(pages[i]));
475		if ((addr == next_addr) && (run < (PAGE_SIZE - 1))) {
476			next_addr += PAGE_SIZE;
477			run++;
478		} else {
479			addrs[addridx] = (unsigned long)base_addr + run;
480			addridx++;
481			base_addr = addr;
482			next_addr = addr + PAGE_SIZE;
483			run = 0;
484		}
485	}
486
487	addrs[addridx] = (unsigned long)base_addr + run;
488	addridx++;
489
490	/* Partial cache lines (fragments) require special measures */
491	if ((type == PAGELIST_READ) &&
492		((pagelist->offset & (CACHE_LINE_SIZE - 1)) ||
493		((pagelist->offset + pagelist->length) &
494		(CACHE_LINE_SIZE - 1)))) {
495		FRAGMENTS_T *fragments;
496
497		if (down_interruptible(&g_free_fragments_sema) != 0) {
498      			free(pagelist, M_VCPAGELIST);
499			return -EINTR;
500		}
501
502		WARN_ON(g_free_fragments == NULL);
503
504		down(&g_free_fragments_mutex);
505		fragments = (FRAGMENTS_T *) g_free_fragments;
506		WARN_ON(fragments == NULL);
507		g_free_fragments = *(FRAGMENTS_T **) g_free_fragments;
508		up(&g_free_fragments_mutex);
509		pagelist->type =
510			 PAGELIST_READ_WITH_FRAGMENTS + (fragments -
511							 g_fragments_base);
512	}
513
514	cpu_dcache_wbinv_range((vm_offset_t)buf, count);
515
516	bus_dmamap_sync(bi->pagelist_dma_tag, bi->pagelist_dma_map, BUS_DMASYNC_PREWRITE);
517
518	bi->pagelist = pagelist;
519
520	return 0;
521
522failed_load:
523	bus_dmamap_unload(bi->pagelist_dma_tag, bi->pagelist_dma_map);
524failed_alloc:
525	bus_dmamap_destroy(bi->pagelist_dma_tag, bi->pagelist_dma_map);
526	bus_dma_tag_destroy(bi->pagelist_dma_tag);
527
528	return err;
529}
530
531static void
532free_pagelist(BULKINFO_T *bi, int actual)
533{
534	vm_page_t*pages;
535	unsigned int num_pages, i;
536	PAGELIST_T *pagelist;
537
538	pagelist = bi->pagelist;
539
540	vchiq_log_trace(vchiq_arm_log_level,
541		"free_pagelist - %x, %d", (unsigned int)pagelist, actual);
542
543	num_pages =
544		(pagelist->length + pagelist->offset + PAGE_SIZE - 1) /
545		PAGE_SIZE;
546
547	pages = (vm_page_t*)(pagelist->addrs + num_pages);
548
549	/* Deal with any partial cache lines (fragments) */
550	if (pagelist->type >= PAGELIST_READ_WITH_FRAGMENTS) {
551		FRAGMENTS_T *fragments = g_fragments_base +
552			(pagelist->type - PAGELIST_READ_WITH_FRAGMENTS);
553		int head_bytes, tail_bytes;
554		head_bytes = (CACHE_LINE_SIZE - pagelist->offset) &
555			(CACHE_LINE_SIZE - 1);
556		tail_bytes = (pagelist->offset + actual) &
557			(CACHE_LINE_SIZE - 1);
558
559		if ((actual >= 0) && (head_bytes != 0)) {
560			if (head_bytes > actual)
561				head_bytes = actual;
562
563			memcpy((char *)bi->buf,
564				fragments->headbuf,
565				head_bytes);
566		}
567
568		if ((actual >= 0) && (head_bytes < actual) &&
569			(tail_bytes != 0)) {
570			memcpy((char *)bi->buf + actual - tail_bytes,
571					 fragments->tailbuf, tail_bytes);
572		}
573
574		down(&g_free_fragments_mutex);
575		*(FRAGMENTS_T **) fragments = g_free_fragments;
576		g_free_fragments = fragments;
577		up(&g_free_fragments_mutex);
578		up(&g_free_fragments_sema);
579	}
580
581	for (i = 0; i < num_pages; i++) {
582		if (pagelist->type != PAGELIST_WRITE) {
583			vm_page_dirty(pages[i]);
584			pagelist_page_free(pages[i]);
585		}
586	}
587
588	bus_dmamap_unload(bi->pagelist_dma_tag, bi->pagelist_dma_map);
589	bus_dmamem_free(bi->pagelist_dma_tag, bi->pagelist, bi->pagelist_dma_map);
590	bus_dmamap_destroy(bi->pagelist_dma_tag, bi->pagelist_dma_map);
591	bus_dma_tag_destroy(bi->pagelist_dma_tag);
592
593	free(bi, M_VCPAGELIST);
594}
595