mirror of https://github.com/ipxe/ipxe.git
[xhci] Update driver to use DMA API
Signed-off-by: Michael Brown <mcb30@ipxe.org>pull/181/head
parent
6e01b74a8a
commit
13a6d17296
|
@ -31,7 +31,6 @@ FILE_LICENCE ( GPL2_OR_LATER_OR_UBDL );
|
|||
#include <errno.h>
|
||||
#include <byteswap.h>
|
||||
#include <ipxe/malloc.h>
|
||||
#include <ipxe/umalloc.h>
|
||||
#include <ipxe/pci.h>
|
||||
#include <ipxe/usb.h>
|
||||
#include <ipxe/init.h>
|
||||
|
@ -294,9 +293,9 @@ static void xhci_init ( struct xhci_device *xhci, void *regs ) {
|
|||
|
||||
/* Read structural parameters 2 */
|
||||
hcsparams2 = readl ( xhci->cap + XHCI_CAP_HCSPARAMS2 );
|
||||
xhci->scratchpads = XHCI_HCSPARAMS2_SCRATCHPADS ( hcsparams2 );
|
||||
xhci->scratch.count = XHCI_HCSPARAMS2_SCRATCHPADS ( hcsparams2 );
|
||||
DBGC2 ( xhci, "XHCI %s needs %d scratchpads\n",
|
||||
xhci->name, xhci->scratchpads );
|
||||
xhci->name, xhci->scratch.count );
|
||||
|
||||
/* Read capability parameters 1 */
|
||||
hccparams1 = readl ( xhci->cap + XHCI_CAP_HCCPARAMS1 );
|
||||
|
@ -918,27 +917,29 @@ static int xhci_dcbaa_alloc ( struct xhci_device *xhci ) {
|
|||
* align on its own size (rounded up to a power of two and
|
||||
* with a minimum of 64 bytes).
|
||||
*/
|
||||
len = ( ( xhci->slots + 1 ) * sizeof ( xhci->dcbaa[0] ) );
|
||||
xhci->dcbaa = malloc_phys ( len, xhci_align ( len ) );
|
||||
if ( ! xhci->dcbaa ) {
|
||||
len = ( ( xhci->slots + 1 ) * sizeof ( xhci->dcbaa.context[0] ) );
|
||||
xhci->dcbaa.context = dma_alloc ( xhci->dma, &xhci->dcbaa.map, len,
|
||||
xhci_align ( len ) );
|
||||
if ( ! xhci->dcbaa.context ) {
|
||||
DBGC ( xhci, "XHCI %s could not allocate DCBAA\n", xhci->name );
|
||||
rc = -ENOMEM;
|
||||
goto err_alloc;
|
||||
}
|
||||
memset ( xhci->dcbaa, 0, len );
|
||||
memset ( xhci->dcbaa.context, 0, len );
|
||||
|
||||
/* Program DCBAA pointer */
|
||||
dcbaap = virt_to_phys ( xhci->dcbaa );
|
||||
dcbaap = dma ( &xhci->dcbaa.map, xhci->dcbaa.context );
|
||||
if ( ( rc = xhci_writeq ( xhci, dcbaap,
|
||||
xhci->op + XHCI_OP_DCBAAP ) ) != 0 )
|
||||
goto err_writeq;
|
||||
|
||||
DBGC2 ( xhci, "XHCI %s DCBAA at [%08lx,%08lx)\n",
|
||||
xhci->name, dcbaap, ( dcbaap + len ) );
|
||||
DBGC2 ( xhci, "XHCI %s DCBAA at [%08lx,%08lx)\n", xhci->name,
|
||||
virt_to_phys ( xhci->dcbaa.context ),
|
||||
( virt_to_phys ( xhci->dcbaa.context ) + len ) );
|
||||
return 0;
|
||||
|
||||
err_writeq:
|
||||
free_phys ( xhci->dcbaa, len );
|
||||
dma_free ( &xhci->dcbaa.map, xhci->dcbaa.context, len );
|
||||
err_alloc:
|
||||
return rc;
|
||||
}
|
||||
|
@ -954,14 +955,14 @@ static void xhci_dcbaa_free ( struct xhci_device *xhci ) {
|
|||
|
||||
/* Sanity check */
|
||||
for ( i = 0 ; i <= xhci->slots ; i++ )
|
||||
assert ( xhci->dcbaa[i] == 0 );
|
||||
assert ( xhci->dcbaa.context[i] == 0 );
|
||||
|
||||
/* Clear DCBAA pointer */
|
||||
xhci_writeq ( xhci, 0, xhci->op + XHCI_OP_DCBAAP );
|
||||
|
||||
/* Free DCBAA */
|
||||
len = ( ( xhci->slots + 1 ) * sizeof ( xhci->dcbaa[0] ) );
|
||||
free_phys ( xhci->dcbaa, len );
|
||||
len = ( ( xhci->slots + 1 ) * sizeof ( xhci->dcbaa.context[0] ) );
|
||||
dma_free ( &xhci->dcbaa.map, xhci->dcbaa.context, len );
|
||||
}
|
||||
|
||||
/******************************************************************************
|
||||
|
@ -978,32 +979,34 @@ static void xhci_dcbaa_free ( struct xhci_device *xhci ) {
|
|||
* @ret rc Return status code
|
||||
*/
|
||||
static int xhci_scratchpad_alloc ( struct xhci_device *xhci ) {
|
||||
struct xhci_scratchpad *scratch = &xhci->scratch;
|
||||
size_t buffer_len;
|
||||
size_t array_len;
|
||||
size_t len;
|
||||
physaddr_t phys;
|
||||
physaddr_t addr;
|
||||
unsigned int i;
|
||||
int rc;
|
||||
|
||||
/* Do nothing if no scratchpad buffers are used */
|
||||
if ( ! xhci->scratchpads )
|
||||
if ( ! scratch->count )
|
||||
return 0;
|
||||
|
||||
/* Allocate scratchpads */
|
||||
len = ( xhci->scratchpads * xhci->pagesize );
|
||||
xhci->scratchpad = umalloc ( len );
|
||||
if ( ! xhci->scratchpad ) {
|
||||
/* Allocate scratchpad buffers */
|
||||
buffer_len = ( scratch->count * xhci->pagesize );
|
||||
scratch->buffer = dma_umalloc ( xhci->dma, &scratch->buffer_map,
|
||||
buffer_len, xhci->pagesize );
|
||||
if ( ! scratch->buffer ) {
|
||||
DBGC ( xhci, "XHCI %s could not allocate scratchpad buffers\n",
|
||||
xhci->name );
|
||||
rc = -ENOMEM;
|
||||
goto err_alloc;
|
||||
}
|
||||
memset_user ( xhci->scratchpad, 0, 0, len );
|
||||
memset_user ( scratch->buffer, 0, 0, buffer_len );
|
||||
|
||||
/* Allocate scratchpad array */
|
||||
array_len = ( xhci->scratchpads * sizeof ( xhci->scratchpad_array[0] ));
|
||||
xhci->scratchpad_array =
|
||||
malloc_phys ( array_len, xhci_align ( array_len ) );
|
||||
if ( ! xhci->scratchpad_array ) {
|
||||
array_len = ( scratch->count * sizeof ( scratch->array[0] ) );
|
||||
scratch->array = dma_alloc ( xhci->dma, &scratch->array_map,
|
||||
array_len, xhci_align ( array_len ) );
|
||||
if ( ! scratch->array ) {
|
||||
DBGC ( xhci, "XHCI %s could not allocate scratchpad buffer "
|
||||
"array\n", xhci->name );
|
||||
rc = -ENOMEM;
|
||||
|
@ -1011,25 +1014,28 @@ static int xhci_scratchpad_alloc ( struct xhci_device *xhci ) {
|
|||
}
|
||||
|
||||
/* Populate scratchpad array */
|
||||
for ( i = 0 ; i < xhci->scratchpads ; i++ ) {
|
||||
phys = user_to_phys ( xhci->scratchpad, ( i * xhci->pagesize ));
|
||||
xhci->scratchpad_array[i] = phys;
|
||||
addr = dma_phys ( &scratch->buffer_map,
|
||||
user_to_phys ( scratch->buffer, 0 ) );
|
||||
for ( i = 0 ; i < scratch->count ; i++ ) {
|
||||
scratch->array[i] = cpu_to_le64 ( addr );
|
||||
addr += xhci->pagesize;
|
||||
}
|
||||
|
||||
/* Set scratchpad array pointer */
|
||||
assert ( xhci->dcbaa != NULL );
|
||||
xhci->dcbaa[0] = cpu_to_le64 ( virt_to_phys ( xhci->scratchpad_array ));
|
||||
assert ( xhci->dcbaa.context != NULL );
|
||||
xhci->dcbaa.context[0] = cpu_to_le64 ( dma ( &scratch->array_map,
|
||||
scratch->array ) );
|
||||
|
||||
DBGC2 ( xhci, "XHCI %s scratchpad [%08lx,%08lx) array [%08lx,%08lx)\n",
|
||||
xhci->name, user_to_phys ( xhci->scratchpad, 0 ),
|
||||
user_to_phys ( xhci->scratchpad, len ),
|
||||
virt_to_phys ( xhci->scratchpad_array ),
|
||||
( virt_to_phys ( xhci->scratchpad_array ) + array_len ) );
|
||||
xhci->name, user_to_phys ( scratch->buffer, 0 ),
|
||||
user_to_phys ( scratch->buffer, buffer_len ),
|
||||
virt_to_phys ( scratch->array ),
|
||||
( virt_to_phys ( scratch->array ) + array_len ) );
|
||||
return 0;
|
||||
|
||||
free_phys ( xhci->scratchpad_array, array_len );
|
||||
dma_free ( &scratch->array_map, scratch->array, array_len );
|
||||
err_alloc_array:
|
||||
ufree ( xhci->scratchpad );
|
||||
dma_ufree ( &scratch->buffer_map, scratch->buffer, buffer_len );
|
||||
err_alloc:
|
||||
return rc;
|
||||
}
|
||||
|
@ -1040,22 +1046,25 @@ static int xhci_scratchpad_alloc ( struct xhci_device *xhci ) {
|
|||
* @v xhci xHCI device
|
||||
*/
|
||||
static void xhci_scratchpad_free ( struct xhci_device *xhci ) {
|
||||
struct xhci_scratchpad *scratch = &xhci->scratch;
|
||||
size_t array_len;
|
||||
size_t buffer_len;
|
||||
|
||||
/* Do nothing if no scratchpad buffers are used */
|
||||
if ( ! xhci->scratchpads )
|
||||
if ( ! scratch->count )
|
||||
return;
|
||||
|
||||
/* Clear scratchpad array pointer */
|
||||
assert ( xhci->dcbaa != NULL );
|
||||
xhci->dcbaa[0] = 0;
|
||||
assert ( xhci->dcbaa.context != NULL );
|
||||
xhci->dcbaa.context[0] = 0;
|
||||
|
||||
/* Free scratchpad array */
|
||||
array_len = ( xhci->scratchpads * sizeof ( xhci->scratchpad_array[0] ));
|
||||
free_phys ( xhci->scratchpad_array, array_len );
|
||||
array_len = ( scratch->count * sizeof ( scratch->array[0] ) );
|
||||
dma_free ( &scratch->array_map, scratch->array, array_len );
|
||||
|
||||
/* Free scratchpads */
|
||||
ufree ( xhci->scratchpad );
|
||||
/* Free scratchpad buffers */
|
||||
buffer_len = ( scratch->count * xhci->pagesize );
|
||||
dma_ufree ( &scratch->buffer_map, scratch->buffer, buffer_len );
|
||||
}
|
||||
|
||||
/******************************************************************************
|
||||
|
@ -1202,7 +1211,8 @@ static int xhci_ring_alloc ( struct xhci_device *xhci,
|
|||
}
|
||||
|
||||
/* Allocate TRBs */
|
||||
ring->trb = malloc_phys ( ring->len, xhci_align ( ring->len ) );
|
||||
ring->trb = dma_alloc ( xhci->dma, &ring->map, ring->len,
|
||||
xhci_align ( ring->len ) );
|
||||
if ( ! ring->trb ) {
|
||||
rc = -ENOMEM;
|
||||
goto err_alloc_trb;
|
||||
|
@ -1211,14 +1221,14 @@ static int xhci_ring_alloc ( struct xhci_device *xhci,
|
|||
|
||||
/* Initialise Link TRB */
|
||||
link = &ring->trb[count].link;
|
||||
link->next = cpu_to_le64 ( virt_to_phys ( ring->trb ) );
|
||||
link->next = cpu_to_le64 ( dma ( &ring->map, ring->trb ) );
|
||||
link->flags = XHCI_TRB_TC;
|
||||
link->type = XHCI_TRB_LINK;
|
||||
ring->link = link;
|
||||
|
||||
return 0;
|
||||
|
||||
free_phys ( ring->trb, ring->len );
|
||||
dma_free ( &ring->map, ring->trb, ring->len );
|
||||
err_alloc_trb:
|
||||
free ( ring->iobuf );
|
||||
err_alloc_iobuf:
|
||||
|
@ -1256,7 +1266,7 @@ static void xhci_ring_free ( struct xhci_trb_ring *ring ) {
|
|||
assert ( ring->iobuf[i] == NULL );
|
||||
|
||||
/* Free TRBs */
|
||||
free_phys ( ring->trb, ring->len );
|
||||
dma_free ( &ring->map, ring->trb, ring->len );
|
||||
|
||||
/* Free I/O buffers */
|
||||
free ( ring->iobuf );
|
||||
|
@ -1422,13 +1432,14 @@ static int xhci_command_alloc ( struct xhci_device *xhci ) {
|
|||
goto err_ring_alloc;
|
||||
|
||||
/* Program command ring control register */
|
||||
crp = virt_to_phys ( xhci->command.trb );
|
||||
crp = dma ( &xhci->command.map, xhci->command.trb );
|
||||
if ( ( rc = xhci_writeq ( xhci, ( crp | XHCI_CRCR_RCS ),
|
||||
xhci->op + XHCI_OP_CRCR ) ) != 0 )
|
||||
goto err_writeq;
|
||||
|
||||
DBGC2 ( xhci, "XHCI %s CRCR at [%08lx,%08lx)\n",
|
||||
xhci->name, crp, ( crp + xhci->command.len ) );
|
||||
DBGC2 ( xhci, "XHCI %s CRCR at [%08lx,%08lx)\n", xhci->name,
|
||||
virt_to_phys ( xhci->command.trb ),
|
||||
( virt_to_phys ( xhci->command.trb ) + xhci->command.len ) );
|
||||
return 0;
|
||||
|
||||
err_writeq:
|
||||
|
@ -1469,7 +1480,8 @@ static int xhci_event_alloc ( struct xhci_device *xhci ) {
|
|||
/* Allocate event ring */
|
||||
count = ( 1 << XHCI_EVENT_TRBS_LOG2 );
|
||||
len = ( count * sizeof ( event->trb[0] ) );
|
||||
event->trb = malloc_phys ( len, xhci_align ( len ) );
|
||||
event->trb = dma_alloc ( xhci->dma, &event->trb_map, len,
|
||||
xhci_align ( len ) );
|
||||
if ( ! event->trb ) {
|
||||
rc = -ENOMEM;
|
||||
goto err_alloc_trb;
|
||||
|
@ -1477,22 +1489,25 @@ static int xhci_event_alloc ( struct xhci_device *xhci ) {
|
|||
memset ( event->trb, 0, len );
|
||||
|
||||
/* Allocate event ring segment table */
|
||||
event->segment = malloc_phys ( sizeof ( event->segment[0] ),
|
||||
xhci_align ( sizeof(event->segment[0])));
|
||||
event->segment = dma_alloc ( xhci->dma, &event->segment_map,
|
||||
sizeof ( event->segment[0] ),
|
||||
xhci_align ( sizeof (event->segment[0])));
|
||||
if ( ! event->segment ) {
|
||||
rc = -ENOMEM;
|
||||
goto err_alloc_segment;
|
||||
}
|
||||
memset ( event->segment, 0, sizeof ( event->segment[0] ) );
|
||||
event->segment[0].base = cpu_to_le64 ( virt_to_phys ( event->trb ) );
|
||||
event->segment[0].base = cpu_to_le64 ( dma ( &event->trb_map,
|
||||
event->trb ) );
|
||||
event->segment[0].count = cpu_to_le32 ( count );
|
||||
|
||||
/* Program event ring registers */
|
||||
writel ( 1, xhci->run + XHCI_RUN_ERSTSZ ( 0 ) );
|
||||
if ( ( rc = xhci_writeq ( xhci, virt_to_phys ( event->trb ),
|
||||
if ( ( rc = xhci_writeq ( xhci, dma ( &event->trb_map, event->trb ),
|
||||
xhci->run + XHCI_RUN_ERDP ( 0 ) ) ) != 0 )
|
||||
goto err_writeq_erdp;
|
||||
if ( ( rc = xhci_writeq ( xhci, virt_to_phys ( event->segment ),
|
||||
if ( ( rc = xhci_writeq ( xhci,
|
||||
dma ( &event->segment_map, event->segment ),
|
||||
xhci->run + XHCI_RUN_ERSTBA ( 0 ) ) ) != 0 )
|
||||
goto err_writeq_erstba;
|
||||
|
||||
|
@ -1501,16 +1516,17 @@ static int xhci_event_alloc ( struct xhci_device *xhci ) {
|
|||
( virt_to_phys ( event->trb ) + len ),
|
||||
virt_to_phys ( event->segment ),
|
||||
( virt_to_phys ( event->segment ) +
|
||||
sizeof (event->segment[0] ) ) );
|
||||
sizeof ( event->segment[0] ) ) );
|
||||
return 0;
|
||||
|
||||
xhci_writeq ( xhci, 0, xhci->run + XHCI_RUN_ERSTBA ( 0 ) );
|
||||
err_writeq_erstba:
|
||||
xhci_writeq ( xhci, 0, xhci->run + XHCI_RUN_ERDP ( 0 ) );
|
||||
err_writeq_erdp:
|
||||
free_phys ( event->trb, len );
|
||||
dma_free ( &event->segment_map, event->segment,
|
||||
sizeof ( event->segment[0] ) );
|
||||
err_alloc_segment:
|
||||
free_phys ( event->segment, sizeof ( event->segment[0] ) );
|
||||
dma_free ( &event->trb_map, event->trb, len );
|
||||
err_alloc_trb:
|
||||
return rc;
|
||||
}
|
||||
|
@ -1531,12 +1547,13 @@ static void xhci_event_free ( struct xhci_device *xhci ) {
|
|||
xhci_writeq ( xhci, 0, xhci->run + XHCI_RUN_ERDP ( 0 ) );
|
||||
|
||||
/* Free event ring segment table */
|
||||
free_phys ( event->segment, sizeof ( event->segment[0] ) );
|
||||
dma_free ( &event->segment_map, event->segment,
|
||||
sizeof ( event->segment[0] ) );
|
||||
|
||||
/* Free event ring */
|
||||
count = ( 1 << XHCI_EVENT_TRBS_LOG2 );
|
||||
len = ( count * sizeof ( event->trb[0] ) );
|
||||
free_phys ( event->trb, len );
|
||||
dma_free ( &event->trb_map, event->trb, len );
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -1577,6 +1594,9 @@ static void xhci_transfer ( struct xhci_device *xhci,
|
|||
iobuf = xhci_dequeue_multi ( &endpoint->ring );
|
||||
assert ( iobuf != NULL );
|
||||
|
||||
/* Unmap I/O buffer */
|
||||
iob_unmap ( iobuf );
|
||||
|
||||
/* Check for errors */
|
||||
if ( ! ( ( trb->code == XHCI_CMPLT_SUCCESS ) ||
|
||||
( trb->code == XHCI_CMPLT_SHORT ) ) ) {
|
||||
|
@ -1745,7 +1765,7 @@ static void xhci_event_poll ( struct xhci_device *xhci ) {
|
|||
|
||||
/* Update dequeue pointer if applicable */
|
||||
if ( consumed ) {
|
||||
xhci_writeq ( xhci, virt_to_phys ( trb ),
|
||||
xhci_writeq ( xhci, dma ( &event->trb_map, trb ),
|
||||
xhci->run + XHCI_RUN_ERDP ( 0 ) );
|
||||
profile_stop ( &xhci_event_profiler );
|
||||
}
|
||||
|
@ -1774,7 +1794,7 @@ static void xhci_abort ( struct xhci_device *xhci ) {
|
|||
|
||||
/* Reset the command ring control register */
|
||||
xhci_ring_reset ( &xhci->command );
|
||||
crp = virt_to_phys ( xhci->command.trb );
|
||||
crp = dma ( &xhci->command.map, xhci->command.trb );
|
||||
xhci_writeq ( xhci, ( crp | XHCI_CRCR_RCS ), xhci->op + XHCI_OP_CRCR );
|
||||
}
|
||||
|
||||
|
@ -1942,13 +1962,15 @@ static int xhci_context ( struct xhci_device *xhci, struct xhci_slot *slot,
|
|||
void *input ) ) {
|
||||
union xhci_trb trb;
|
||||
struct xhci_trb_context *context = &trb.context;
|
||||
struct dma_mapping map;
|
||||
size_t len;
|
||||
void *input;
|
||||
int rc;
|
||||
|
||||
/* Allocate an input context */
|
||||
memset ( &map, 0, sizeof ( map ) );
|
||||
len = xhci_input_context_offset ( xhci, XHCI_CTX_END );
|
||||
input = malloc_phys ( len, xhci_align ( len ) );
|
||||
input = dma_alloc ( xhci->dma, &map, len, xhci_align ( len ) );
|
||||
if ( ! input ) {
|
||||
rc = -ENOMEM;
|
||||
goto err_alloc;
|
||||
|
@ -1961,7 +1983,7 @@ static int xhci_context ( struct xhci_device *xhci, struct xhci_slot *slot,
|
|||
/* Construct command */
|
||||
memset ( context, 0, sizeof ( *context ) );
|
||||
context->type = type;
|
||||
context->input = cpu_to_le64 ( virt_to_phys ( input ) );
|
||||
context->input = cpu_to_le64 ( dma ( &map, input ) );
|
||||
context->slot = slot->id;
|
||||
|
||||
/* Issue command and wait for completion */
|
||||
|
@ -1969,7 +1991,7 @@ static int xhci_context ( struct xhci_device *xhci, struct xhci_slot *slot,
|
|||
goto err_command;
|
||||
|
||||
err_command:
|
||||
free_phys ( input, len );
|
||||
dma_free ( &map, input, len );
|
||||
err_alloc:
|
||||
return rc;
|
||||
}
|
||||
|
@ -1986,6 +2008,7 @@ static void xhci_address_device_input ( struct xhci_device *xhci,
|
|||
struct xhci_slot *slot,
|
||||
struct xhci_endpoint *endpoint,
|
||||
void *input ) {
|
||||
struct xhci_trb_ring *ring = &endpoint->ring;
|
||||
struct xhci_control_context *control_ctx;
|
||||
struct xhci_slot_context *slot_ctx;
|
||||
struct xhci_endpoint_context *ep_ctx;
|
||||
|
@ -2011,7 +2034,7 @@ static void xhci_address_device_input ( struct xhci_device *xhci,
|
|||
ep_ctx->type = XHCI_EP_TYPE_CONTROL;
|
||||
ep_ctx->burst = endpoint->ep->burst;
|
||||
ep_ctx->mtu = cpu_to_le16 ( endpoint->ep->mtu );
|
||||
ep_ctx->dequeue = cpu_to_le64 ( virt_to_phys ( endpoint->ring.trb ) |
|
||||
ep_ctx->dequeue = cpu_to_le64 ( dma ( &ring->map, ring->trb ) |
|
||||
XHCI_EP_DCS );
|
||||
ep_ctx->trb_len = cpu_to_le16 ( XHCI_EP0_TRB_LEN );
|
||||
}
|
||||
|
@ -2057,6 +2080,7 @@ static void xhci_configure_endpoint_input ( struct xhci_device *xhci,
|
|||
struct xhci_slot *slot,
|
||||
struct xhci_endpoint *endpoint,
|
||||
void *input ) {
|
||||
struct xhci_trb_ring *ring = &endpoint->ring;
|
||||
struct xhci_control_context *control_ctx;
|
||||
struct xhci_slot_context *slot_ctx;
|
||||
struct xhci_endpoint_context *ep_ctx;
|
||||
|
@ -2079,7 +2103,7 @@ static void xhci_configure_endpoint_input ( struct xhci_device *xhci,
|
|||
ep_ctx->type = endpoint->type;
|
||||
ep_ctx->burst = endpoint->ep->burst;
|
||||
ep_ctx->mtu = cpu_to_le16 ( endpoint->ep->mtu );
|
||||
ep_ctx->dequeue = cpu_to_le64 ( virt_to_phys ( endpoint->ring.trb ) |
|
||||
ep_ctx->dequeue = cpu_to_le64 ( dma ( &ring->map, ring->trb ) |
|
||||
XHCI_EP_DCS );
|
||||
ep_ctx->trb_len = cpu_to_le16 ( endpoint->ep->mtu ); /* best guess */
|
||||
}
|
||||
|
@ -2297,6 +2321,7 @@ xhci_set_tr_dequeue_pointer ( struct xhci_device *xhci,
|
|||
unsigned int mask;
|
||||
unsigned int index;
|
||||
unsigned int dcs;
|
||||
physaddr_t addr;
|
||||
int rc;
|
||||
|
||||
/* Construct command */
|
||||
|
@ -2305,8 +2330,8 @@ xhci_set_tr_dequeue_pointer ( struct xhci_device *xhci,
|
|||
mask = ring->mask;
|
||||
dcs = ( ( ~( cons >> ring->shift ) ) & XHCI_EP_DCS );
|
||||
index = ( cons & mask );
|
||||
dequeue->dequeue =
|
||||
cpu_to_le64 ( virt_to_phys ( &ring->trb[index] ) | dcs );
|
||||
addr = dma ( &ring->map, &ring->trb[index] );
|
||||
dequeue->dequeue = cpu_to_le64 ( addr | dcs );
|
||||
dequeue->slot = slot->id;
|
||||
dequeue->endpoint = endpoint->ctx;
|
||||
dequeue->type = XHCI_TRB_SET_TR_DEQUEUE_POINTER;
|
||||
|
@ -2425,6 +2450,7 @@ static void xhci_endpoint_close ( struct usb_endpoint *ep ) {
|
|||
/* Cancel any incomplete transfers */
|
||||
while ( xhci_ring_fill ( &endpoint->ring ) ) {
|
||||
iobuf = xhci_dequeue_multi ( &endpoint->ring );
|
||||
iob_unmap ( iobuf );
|
||||
usb_complete_err ( ep, iobuf, -ECANCELED );
|
||||
}
|
||||
|
||||
|
@ -2491,6 +2517,7 @@ static int xhci_endpoint_mtu ( struct usb_endpoint *ep ) {
|
|||
static int xhci_endpoint_message ( struct usb_endpoint *ep,
|
||||
struct io_buffer *iobuf ) {
|
||||
struct xhci_endpoint *endpoint = usb_endpoint_get_hostdata ( ep );
|
||||
struct xhci_device *xhci = endpoint->xhci;
|
||||
struct usb_setup_packet *packet;
|
||||
unsigned int input;
|
||||
size_t len;
|
||||
|
@ -2520,10 +2547,15 @@ static int xhci_endpoint_message ( struct usb_endpoint *ep,
|
|||
if ( len )
|
||||
setup->direction = ( input ? XHCI_SETUP_IN : XHCI_SETUP_OUT );
|
||||
|
||||
/* Map I/O buffer */
|
||||
if ( ( rc = iob_map ( iobuf, xhci->dma, len,
|
||||
( input ? DMA_RX : DMA_TX ) ) ) != 0 )
|
||||
goto err_map;
|
||||
|
||||
/* Construct data stage TRB, if applicable */
|
||||
if ( len ) {
|
||||
data = &(trb++)->data;
|
||||
data->data = cpu_to_le64 ( virt_to_phys ( iobuf->data ) );
|
||||
data->data = cpu_to_le64 ( iob_dma ( iobuf ) );
|
||||
data->len = cpu_to_le32 ( len );
|
||||
data->type = XHCI_TRB_DATA;
|
||||
data->direction = ( input ? XHCI_DATA_IN : XHCI_DATA_OUT );
|
||||
|
@ -2539,13 +2571,18 @@ static int xhci_endpoint_message ( struct usb_endpoint *ep,
|
|||
/* Enqueue TRBs */
|
||||
if ( ( rc = xhci_enqueue_multi ( &endpoint->ring, iobuf, trbs,
|
||||
( trb - trbs ) ) ) != 0 )
|
||||
return rc;
|
||||
goto err_enqueue;
|
||||
|
||||
/* Ring the doorbell */
|
||||
xhci_doorbell ( &endpoint->ring );
|
||||
|
||||
profile_stop ( &xhci_message_profiler );
|
||||
return 0;
|
||||
|
||||
err_enqueue:
|
||||
iob_unmap ( iobuf );
|
||||
err_map:
|
||||
return rc;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -2579,12 +2616,13 @@ static unsigned int xhci_endpoint_count ( size_t len, int zlp ) {
|
|||
static int xhci_endpoint_stream ( struct usb_endpoint *ep,
|
||||
struct io_buffer *iobuf, int zlp ) {
|
||||
struct xhci_endpoint *endpoint = usb_endpoint_get_hostdata ( ep );
|
||||
void *data = iobuf->data;
|
||||
struct xhci_device *xhci = endpoint->xhci;
|
||||
size_t len = iob_len ( iobuf );
|
||||
unsigned int count = xhci_endpoint_count ( len, zlp );
|
||||
union xhci_trb trbs[count];
|
||||
union xhci_trb *trb = trbs;
|
||||
struct xhci_trb_normal *normal;
|
||||
physaddr_t data;
|
||||
unsigned int i;
|
||||
size_t trb_len;
|
||||
int rc;
|
||||
|
@ -2592,6 +2630,13 @@ static int xhci_endpoint_stream ( struct usb_endpoint *ep,
|
|||
/* Profile stream transfers */
|
||||
profile_start ( &xhci_stream_profiler );
|
||||
|
||||
/* Map I/O buffer */
|
||||
if ( ( rc = iob_map ( iobuf, xhci->dma, len,
|
||||
( ( ep->address & USB_DIR_IN ) ?
|
||||
DMA_RX : DMA_TX ) ) ) != 0 )
|
||||
goto err_map;
|
||||
data = iob_dma ( iobuf );
|
||||
|
||||
/* Construct normal TRBs */
|
||||
memset ( &trbs, 0, sizeof ( trbs ) );
|
||||
for ( i = 0 ; i < count ; i ++ ) {
|
||||
|
@ -2603,7 +2648,7 @@ static int xhci_endpoint_stream ( struct usb_endpoint *ep,
|
|||
|
||||
/* Construct normal TRB */
|
||||
normal = &trb->normal;
|
||||
normal->data = cpu_to_le64 ( virt_to_phys ( data ) );
|
||||
normal->data = cpu_to_le64 ( data );
|
||||
normal->len = cpu_to_le32 ( trb_len );
|
||||
normal->type = XHCI_TRB_NORMAL;
|
||||
normal->flags = XHCI_TRB_CH;
|
||||
|
@ -2624,13 +2669,18 @@ static int xhci_endpoint_stream ( struct usb_endpoint *ep,
|
|||
/* Enqueue TRBs */
|
||||
if ( ( rc = xhci_enqueue_multi ( &endpoint->ring, iobuf, trbs,
|
||||
count ) ) != 0 )
|
||||
return rc;
|
||||
goto err_enqueue;
|
||||
|
||||
/* Ring the doorbell */
|
||||
xhci_doorbell ( &endpoint->ring );
|
||||
|
||||
profile_stop ( &xhci_stream_profiler );
|
||||
return 0;
|
||||
|
||||
err_enqueue:
|
||||
iob_unmap ( iobuf );
|
||||
err_map:
|
||||
return rc;
|
||||
}
|
||||
|
||||
/******************************************************************************
|
||||
|
@ -2693,7 +2743,8 @@ static int xhci_device_open ( struct usb_device *usb ) {
|
|||
|
||||
/* Allocate a device context */
|
||||
len = xhci_device_context_offset ( xhci, XHCI_CTX_END );
|
||||
slot->context = malloc_phys ( len, xhci_align ( len ) );
|
||||
slot->context = dma_alloc ( xhci->dma, &slot->map, len,
|
||||
xhci_align ( len ) );
|
||||
if ( ! slot->context ) {
|
||||
rc = -ENOMEM;
|
||||
goto err_alloc_context;
|
||||
|
@ -2701,16 +2752,17 @@ static int xhci_device_open ( struct usb_device *usb ) {
|
|||
memset ( slot->context, 0, len );
|
||||
|
||||
/* Set device context base address */
|
||||
assert ( xhci->dcbaa[id] == 0 );
|
||||
xhci->dcbaa[id] = cpu_to_le64 ( virt_to_phys ( slot->context ) );
|
||||
assert ( xhci->dcbaa.context[id] == 0 );
|
||||
xhci->dcbaa.context[id] = cpu_to_le64 ( dma ( &slot->map,
|
||||
slot->context ) );
|
||||
|
||||
DBGC2 ( xhci, "XHCI %s slot %d device context [%08lx,%08lx) for %s\n",
|
||||
xhci->name, slot->id, virt_to_phys ( slot->context ),
|
||||
( virt_to_phys ( slot->context ) + len ), usb->name );
|
||||
return 0;
|
||||
|
||||
xhci->dcbaa[id] = 0;
|
||||
free_phys ( slot->context, len );
|
||||
xhci->dcbaa.context[id] = 0;
|
||||
dma_free ( &slot->map, slot->context, len );
|
||||
err_alloc_context:
|
||||
xhci->slot[id] = NULL;
|
||||
free ( slot );
|
||||
|
@ -2750,8 +2802,8 @@ static void xhci_device_close ( struct usb_device *usb ) {
|
|||
|
||||
/* Free slot */
|
||||
if ( slot->context ) {
|
||||
free_phys ( slot->context, len );
|
||||
xhci->dcbaa[id] = 0;
|
||||
dma_free ( &slot->map, slot->context, len );
|
||||
xhci->dcbaa.context[id] = 0;
|
||||
}
|
||||
xhci->slot[id] = NULL;
|
||||
free ( slot );
|
||||
|
@ -3270,6 +3322,11 @@ static int xhci_probe ( struct pci_device *pci ) {
|
|||
/* Initialise xHCI device */
|
||||
xhci_init ( xhci, xhci->regs );
|
||||
|
||||
/* Configure DMA device */
|
||||
xhci->dma = &pci->dma;
|
||||
if ( xhci->addr64 )
|
||||
dma_set_mask_64bit ( xhci->dma );
|
||||
|
||||
/* Initialise USB legacy support and claim ownership */
|
||||
xhci_legacy_init ( xhci );
|
||||
xhci_legacy_claim ( xhci );
|
||||
|
|
|
@ -854,6 +854,8 @@ struct xhci_trb_ring {
|
|||
union xhci_trb *trb;
|
||||
/** Length of transfer request blocks */
|
||||
size_t len;
|
||||
/** DMA mapping */
|
||||
struct dma_mapping map;
|
||||
/** Link TRB (if applicable) */
|
||||
struct xhci_trb_link *link;
|
||||
|
||||
|
@ -869,8 +871,12 @@ struct xhci_event_ring {
|
|||
unsigned int cons;
|
||||
/** Event ring segment table */
|
||||
struct xhci_event_ring_segment *segment;
|
||||
/** Event ring segment table DMA mapping */
|
||||
struct dma_mapping segment_map;
|
||||
/** Transfer request blocks */
|
||||
union xhci_trb *trb;
|
||||
/** Transfer request blocks DMA mapping */
|
||||
struct dma_mapping trb_map;
|
||||
};
|
||||
|
||||
/**
|
||||
|
@ -1035,10 +1041,34 @@ struct xhci_pch {
|
|||
/** Invalid protocol speed ID values quirk */
|
||||
#define XHCI_BAD_PSIV 0x0002
|
||||
|
||||
/** Device context base address array */
|
||||
struct xhci_dcbaa {
|
||||
/** Context base addresses */
|
||||
uint64_t *context;
|
||||
/** DMA mapping */
|
||||
struct dma_mapping map;
|
||||
};
|
||||
|
||||
/** Scratchpad buffer */
|
||||
struct xhci_scratchpad {
|
||||
/** Number of page-sized scratchpad buffers */
|
||||
unsigned int count;
|
||||
/** Scratchpad buffer area */
|
||||
userptr_t buffer;
|
||||
/** Buffer DMA mapping */
|
||||
struct dma_mapping buffer_map;
|
||||
/** Scratchpad array */
|
||||
uint64_t *array;
|
||||
/** Array DMA mapping */
|
||||
struct dma_mapping array_map;
|
||||
};
|
||||
|
||||
/** An xHCI device */
|
||||
struct xhci_device {
|
||||
/** Registers */
|
||||
void *regs;
|
||||
/** DMA device */
|
||||
struct dma_device *dma;
|
||||
/** Name */
|
||||
const char *name;
|
||||
/** Quirks */
|
||||
|
@ -1060,9 +1090,6 @@ struct xhci_device {
|
|||
/** Number of ports */
|
||||
unsigned int ports;
|
||||
|
||||
/** Number of page-sized scratchpad buffers */
|
||||
unsigned int scratchpads;
|
||||
|
||||
/** 64-bit addressing capability */
|
||||
int addr64;
|
||||
/** Context size shift */
|
||||
|
@ -1077,12 +1104,10 @@ struct xhci_device {
|
|||
unsigned int legacy;
|
||||
|
||||
/** Device context base address array */
|
||||
uint64_t *dcbaa;
|
||||
struct xhci_dcbaa dcbaa;
|
||||
|
||||
/** Scratchpad buffer area */
|
||||
userptr_t scratchpad;
|
||||
/** Scratchpad buffer array */
|
||||
uint64_t *scratchpad_array;
|
||||
/** Scratchpad buffer */
|
||||
struct xhci_scratchpad scratch;
|
||||
|
||||
/** Command ring */
|
||||
struct xhci_trb_ring command;
|
||||
|
@ -1111,6 +1136,8 @@ struct xhci_slot {
|
|||
unsigned int id;
|
||||
/** Slot context */
|
||||
struct xhci_slot_context *context;
|
||||
/** DMA mapping */
|
||||
struct dma_mapping map;
|
||||
/** Route string */
|
||||
unsigned int route;
|
||||
/** Root hub port number */
|
||||
|
|
|
@ -214,6 +214,22 @@ static inline void iob_populate ( struct io_buffer *iobuf,
|
|||
(iobuf) = NULL; \
|
||||
__iobuf; } )
|
||||
|
||||
/**
|
||||
* Map I/O buffer for DMA
|
||||
*
|
||||
* @v iobuf I/O buffer
|
||||
* @v dma DMA device
|
||||
* @v len Length to map
|
||||
* @v flags Mapping flags
|
||||
* @ret rc Return status code
|
||||
*/
|
||||
static inline __always_inline int iob_map ( struct io_buffer *iobuf,
|
||||
struct dma_device *dma,
|
||||
size_t len, int flags ) {
|
||||
return dma_map ( dma, &iobuf->map, virt_to_phys ( iobuf->data ),
|
||||
len, flags );
|
||||
}
|
||||
|
||||
/**
|
||||
* Map I/O buffer for transmit DMA
|
||||
*
|
||||
|
@ -223,8 +239,7 @@ static inline void iob_populate ( struct io_buffer *iobuf,
|
|||
*/
|
||||
static inline __always_inline int iob_map_tx ( struct io_buffer *iobuf,
|
||||
struct dma_device *dma ) {
|
||||
return dma_map ( dma, &iobuf->map, virt_to_phys ( iobuf->data ),
|
||||
iob_len ( iobuf ), DMA_TX );
|
||||
return iob_map ( iobuf, dma, iob_len ( iobuf ), DMA_TX );
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -237,8 +252,7 @@ static inline __always_inline int iob_map_tx ( struct io_buffer *iobuf,
|
|||
static inline __always_inline int iob_map_rx ( struct io_buffer *iobuf,
|
||||
struct dma_device *dma ) {
|
||||
assert ( iob_len ( iobuf ) == 0 );
|
||||
return dma_map ( dma, &iobuf->map, virt_to_phys ( iobuf->data ),
|
||||
iob_tailroom ( iobuf ), DMA_RX );
|
||||
return iob_map ( iobuf, dma, iob_tailroom ( iobuf ), DMA_RX );
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
Loading…
Reference in New Issue