#define _PCILIB_DMA_IPE_C #define _BSD_SOURCE #define _DEFAULT_SOURCE #define _POSIX_C_SOURCE 199309L #include #include #include #include #include #include #include #include #include "pci.h" #include "pcilib.h" #include "error.h" #include "tools.h" #include "debug.h" #include "bar.h" #include "ipe.h" #include "ipe_private.h" pcilib_dma_context_t *dma_ipe_init(pcilib_t *pcilib, const char *model, const void *arg) { int err = 0; pcilib_register_value_t version_value; // const pcilib_model_description_t *model_info = pcilib_get_model_description(pcilib); ipe_dma_t *ctx = malloc(sizeof(ipe_dma_t)); if (ctx) { memset(ctx, 0, sizeof(ipe_dma_t)); ctx->dmactx.pcilib = pcilib; pcilib_register_bank_t dma_bankc = pcilib_find_register_bank_by_addr(pcilib, PCILIB_REGISTER_BANK_DMACONF); pcilib_register_bank_t dma_bank0 = pcilib_find_register_bank_by_addr(pcilib, PCILIB_REGISTER_BANK_DMA0); pcilib_register_bank_t dma_bank1 = pcilib_find_register_bank_by_addr(pcilib, PCILIB_REGISTER_BANK_DMA1); if ((dma_bankc == PCILIB_REGISTER_BANK_INVALID)||(dma_bank0 == PCILIB_REGISTER_BANK_INVALID)||(dma_bank1 == PCILIB_REGISTER_BANK_INVALID)) { free(ctx); pcilib_error("DMA Register Bank could not be found"); return NULL; } ctx->base_addr[0] = (void*)pcilib_resolve_bank_address_by_id(pcilib, 0, dma_bank0); ctx->base_addr[1] = (void*)pcilib_resolve_bank_address_by_id(pcilib, 0, dma_bank1); ctx->base_addr[2] = (void*)pcilib_resolve_bank_address_by_id(pcilib, 0, dma_bankc); RD(IPEDMA_REG_VERSION, version_value); ctx->version = IPEDMA_VERSION(version_value); if ((model)&&(!strcasecmp(model, "ipecamera"))) { if (IPEDMA_GENERATION(version_value) > 2) { ctx->gen = 3; } else { ctx->gen = 2; } } else { if (IPEDMA_GENERATION(version_value) > 2) { ctx->gen = 3; } else { ctx->gen = 2; } err = pcilib_add_registers(pcilib, PCILIB_MODEL_MODIFICATON_FLAGS_DEFAULT, 0, ipe_dma_app_registers, NULL); } if (ctx->gen > 2) { ctx->mode64 = 1; ctx->addr64 = 1; #ifdef IPEDMA_STREAMING_MODE if (IPEDMA_STREAMING(version_value)) ctx->streaming = 1; #endif /* IPEDMA_STREAMING_MODE */ ctx->reg_last_read = IPEDMA_REG3_LAST_READ; if (!err) err = pcilib_add_registers(pcilib, PCILIB_MODEL_MODIFICATON_FLAGS_DEFAULT, 0, ipe_dma_v3_registers, NULL); } else { #ifdef IPEDMA_ENFORCE_64BIT_MODE // According to Lorenzo, some gen2 boards have problems with 64-bit addressing. Therefore, we only enable it for gen3 boards unless enforced ctx->mode64 = 1; #endif /* IPEDMA_ENFORCE_64BIT_MODE */ ctx->addr64 = 0; ctx->streaming = 0; ctx->reg_last_read = IPEDMA_REG2_LAST_READ; if (!err) err = pcilib_add_registers(pcilib, PCILIB_MODEL_MODIFICATON_FLAGS_DEFAULT, 0, ipe_dma_v2_registers, NULL); } pcilib_info("IPEDMA gen%lu version %lu (64-bit mode: %u, 64-bit addressing: %u, streaming: %u)", ctx->gen, ctx->version, ctx->mode64, ctx->addr64, ctx->streaming); if (err) { free(ctx); pcilib_error("Error (%i) registering firmware-dependent IPEDMA registers", err); return NULL; } } return (pcilib_dma_context_t*)ctx; } void dma_ipe_free(pcilib_dma_context_t *vctx) { ipe_dma_t *ctx = (ipe_dma_t*)vctx; if (ctx) { dma_ipe_stop(vctx, PCILIB_DMA_ENGINE_ALL, PCILIB_DMA_FLAGS_DEFAULT); free(ctx); } } static void dma_ipe_disable(ipe_dma_t *ctx) { // Disable DMA WR(IPEDMA_REG_CONTROL, 0x0); usleep(IPEDMA_RESET_DELAY); // Reset DMA engine WR(IPEDMA_REG_RESET, 0x1); usleep(IPEDMA_RESET_DELAY); WR(IPEDMA_REG_RESET, 0x0); usleep(IPEDMA_RESET_DELAY); // Reseting configured DMA pages if (ctx->gen < 3) { WR(IPEDMA_REG2_PAGE_COUNT, 0); } usleep(IPEDMA_RESET_DELAY); } int dma_ipe_start(pcilib_dma_context_t *vctx, pcilib_dma_engine_t dma, pcilib_dma_flags_t flags) { int err; int mask = 32; size_t i, num_pages; ipe_dma_t *ctx = (ipe_dma_t*)vctx; pcilib_kmem_handle_t *desc = NULL; pcilib_kmem_handle_t *pages = NULL; #ifndef IPEDMA_TLP_SIZE const pcilib_pcie_link_info_t *link_info; #endif /* ! IPEDMA_TLP_SIZE */ int preserve = 0; pcilib_kmem_flags_t kflags; pcilib_kmem_reuse_state_t reuse_desc, reuse_pages; volatile void *desc_va; volatile void *last_written_addr_ptr; pcilib_register_value_t value; uintptr_t dma_region = 0; int tlp_size; uint32_t address64; if (dma == PCILIB_DMA_ENGINE_INVALID) return 0; else if (dma > 1) return PCILIB_ERROR_INVALID_BANK; if (!ctx->started) ctx->started = 1; if (flags&PCILIB_DMA_FLAG_PERSISTENT) ctx->preserve = 1; if (ctx->pages) return 0; #ifdef IPEDMA_TLP_SIZE tlp_size = IPEDMA_TLP_SIZE; #else /* IPEDMA_TLP_SIZE */ link_info = pcilib_get_pcie_link_info(vctx->pcilib); if (link_info) { tlp_size = 1<payload; # ifdef IPEDMA_MAX_TLP_SIZE if (tlp_size > IPEDMA_MAX_TLP_SIZE) tlp_size = IPEDMA_MAX_TLP_SIZE; # endif /* IPEDMA_MAX_TLP_SIZE */ } else tlp_size = 128; #endif /* IPEDMA_TLP_SIZE */ if (!pcilib_read_register(ctx->dmactx.pcilib, "dmaconf", "dma_timeout", &value)) ctx->dma_timeout = value; else ctx->dma_timeout = IPEDMA_DMA_TIMEOUT; if (!pcilib_read_register(ctx->dmactx.pcilib, "dmaconf", "dma_page_size", &value)) { if (value % IPEDMA_PAGE_SIZE) { pcilib_error("Invalid DMA page size (%lu) is configured", value); return PCILIB_ERROR_INVALID_ARGUMENT; } //if ((value)&&((value / (tlp_size * IPEDMA_CORES)) > ...seems no limit...)) { ... fail ... } ctx->page_size = value; } else ctx->page_size = IPEDMA_PAGE_SIZE; if ((!pcilib_read_register(ctx->dmactx.pcilib, "dmaconf", "dma_pages", &value))&&(value > 0)) ctx->ring_size = value; else ctx->ring_size = IPEDMA_DMA_PAGES; if (!pcilib_read_register(ctx->dmactx.pcilib, "dmaconf", "dma_region_low", &value)) { dma_region = value; if (!pcilib_read_register(ctx->dmactx.pcilib, "dmaconf", "dma_region_low", &value)) dma_region |= ((uintptr_t)value)<<32; } if (!pcilib_read_register(ctx->dmactx.pcilib, "dmaconf", "ipedma_flags", &value)) ctx->dma_flags = value; else ctx->dma_flags = 0; #ifdef IPEDMA_CONFIGURE_DMA_MASK if (ctx->addr64) mask = 64; err = pcilib_set_dma_mask(ctx->dmactx.pcilib, mask); if (err) { pcilib_error("Error (%i) configuring dma mask (%i)", err, mask); return err; } #endif /* IPEDMA_CONFIGURE_DMA_MASK */ kflags = PCILIB_KMEM_FLAG_REUSE|PCILIB_KMEM_FLAG_EXCLUSIVE|PCILIB_KMEM_FLAG_HARDWARE|(ctx->preserve?PCILIB_KMEM_FLAG_PERSISTENT:0); desc = pcilib_alloc_kernel_memory(ctx->dmactx.pcilib, PCILIB_KMEM_TYPE_CONSISTENT, 1, IPEDMA_DESCRIPTOR_SIZE, IPEDMA_DESCRIPTOR_ALIGNMENT, PCILIB_KMEM_USE(PCILIB_KMEM_USE_DMA_RING, 0x00), kflags); if (dma_region) pages = pcilib_alloc_kernel_memory(ctx->dmactx.pcilib, PCILIB_KMEM_TYPE_REGION_C2S, ctx->ring_size, ctx->page_size, dma_region, PCILIB_KMEM_USE(PCILIB_KMEM_USE_DMA_PAGES, 0x00), kflags); else pages = pcilib_alloc_kernel_memory(ctx->dmactx.pcilib, PCILIB_KMEM_TYPE_DMA_C2S_PAGE, ctx->ring_size, ctx->page_size, 0, PCILIB_KMEM_USE(PCILIB_KMEM_USE_DMA_PAGES, 0x00), kflags); if (!desc||!pages) { if (pages) pcilib_free_kernel_memory(ctx->dmactx.pcilib, pages, KMEM_FLAG_REUSE); if (desc) pcilib_free_kernel_memory(ctx->dmactx.pcilib, desc, KMEM_FLAG_REUSE); pcilib_error("Can't allocate required kernel memory for IPEDMA engine (%lu pages of %lu bytes + %lu byte descriptor)", ctx->ring_size, ctx->page_size, (unsigned long)IPEDMA_DESCRIPTOR_SIZE); return PCILIB_ERROR_MEMORY; } reuse_desc = pcilib_kmem_is_reused(ctx->dmactx.pcilib, desc); reuse_pages = pcilib_kmem_is_reused(ctx->dmactx.pcilib, pages); if ((reuse_pages & PCILIB_KMEM_REUSE_PARTIAL)||(reuse_desc & PCILIB_KMEM_REUSE_PARTIAL)) { dma_ipe_disable(ctx); pcilib_free_kernel_memory(ctx->dmactx.pcilib, pages, KMEM_FLAG_REUSE); pcilib_free_kernel_memory(ctx->dmactx.pcilib, desc, KMEM_FLAG_REUSE); if (((flags&PCILIB_DMA_FLAG_STOP) == 0)||(dma_region)) { pcilib_error("Inconsistent DMA buffers are found (buffers are only partially re-used). This is very wrong, please stop DMA engine and correct configuration..."); return PCILIB_ERROR_INVALID_STATE; } pcilib_warning("Inconsistent DMA buffers are found (buffers are only partially re-used), reinitializing..."); desc = pcilib_alloc_kernel_memory(ctx->dmactx.pcilib, PCILIB_KMEM_TYPE_CONSISTENT, 1, IPEDMA_DESCRIPTOR_SIZE, IPEDMA_DESCRIPTOR_ALIGNMENT, PCILIB_KMEM_USE(PCILIB_KMEM_USE_DMA_RING, 0x00), kflags|PCILIB_KMEM_FLAG_MASS); pages = pcilib_alloc_kernel_memory(ctx->dmactx.pcilib, PCILIB_KMEM_TYPE_DMA_C2S_PAGE, ctx->ring_size, ctx->page_size, 0, PCILIB_KMEM_USE(PCILIB_KMEM_USE_DMA_PAGES, 0x00), kflags|PCILIB_KMEM_FLAG_MASS); if (!desc||!pages) { if (pages) pcilib_free_kernel_memory(ctx->dmactx.pcilib, pages, KMEM_FLAG_REUSE); if (desc) pcilib_free_kernel_memory(ctx->dmactx.pcilib, desc, KMEM_FLAG_REUSE); return PCILIB_ERROR_MEMORY; } } else if (reuse_desc != reuse_pages) { pcilib_warning("Inconsistent DMA buffers (modes of ring and page buffers does not match), reinitializing...."); } else if (reuse_desc & PCILIB_KMEM_REUSE_REUSED) { if ((reuse_desc & PCILIB_KMEM_REUSE_PERSISTENT) == 0) pcilib_warning("Lost DMA buffers are found (non-persistent mode), reinitializing..."); else if ((reuse_desc & PCILIB_KMEM_REUSE_HARDWARE) == 0) pcilib_warning("Lost DMA buffers are found (missing HW reference), reinitializing..."); else { if (ctx->streaming) preserve = 1; else { RD(IPEDMA_REG2_PAGE_COUNT, value); if (value != ctx->ring_size) pcilib_warning("Inconsistent DMA buffers are found (Number of allocated buffers (%lu) does not match current request (%lu)), reinitializing...", value + 1, IPEDMA_DMA_PAGES); else preserve = 1; } } } desc_va = pcilib_kmem_get_ua(ctx->dmactx.pcilib, desc); if (ctx->addr64) last_written_addr_ptr = desc_va + 2 * sizeof(uint32_t); else if (ctx->mode64) last_written_addr_ptr = desc_va + 3 * sizeof(uint32_t); else last_written_addr_ptr = desc_va + 4 * sizeof(uint32_t); // get page size if default size was used if (!ctx->page_size) { ctx->page_size = pcilib_kmem_get_block_size(ctx->dmactx.pcilib, pages, 0); } if (preserve) { ctx->reused = 1; ctx->preserve = 1; // Detect the current state of DMA engine RD(ctx->reg_last_read, value); // Numbered from 1 in FPGA # ifdef IPEDMA_BUG_LAST_READ if (value == ctx->ring_size) value = 0; # else /* IPEDMA_BUG_LAST_READ */ value--; # endif /* IPEDMA_BUG_LAST_READ */ ctx->last_read = value; } else { ctx->reused = 0; dma_ipe_disable(ctx); // Verify PCIe link status RD(IPEDMA_REG_RESET, value); if ((value != 0x14031700)&&(value != 0x14021700)) pcilib_warning("PCIe is not ready, code is %lx", value); // Enable 64 bit addressing and configure TLP and PACKET sizes (40 bit mode can be used with big pre-allocated buffers later) if (ctx->mode64) address64 = 0x8000 | (0<<24); else address64 = 0; WR(IPEDMA_REG_TLP_SIZE, address64 | (tlp_size>>2)); WR(IPEDMA_REG_TLP_COUNT, ctx->page_size / (tlp_size * IPEDMA_CORES)); // Setting progress register threshold WR(IPEDMA_REG_UPDATE_THRESHOLD, IPEDMA_DMA_PROGRESS_THRESHOLD); // Reseting configured DMA pages if (ctx->gen < 3) { WR(IPEDMA_REG2_PAGE_COUNT, 0); } // Setting current read position and configuring progress register #ifdef IPEDMA_BUG_LAST_READ WR(ctx->reg_last_read, ctx->ring_size - 1); #else /* IPEDMA_BUG_LAST_READ */ WR(ctx->reg_last_read, ctx->ring_size); #endif /* IPEDMA_BUG_LAST_READ */ // Instructing DMA engine that writting should start from the first DMA page if (ctx->addr64) { WR64(IPEDMA_REG3_UPDATE_ADDR, pcilib_kmem_get_block_ba(ctx->dmactx.pcilib, desc, 0)); *(uint64_t*)last_written_addr_ptr = 0; } else { WR(IPEDMA_REG2_UPDATE_ADDR, pcilib_kmem_get_block_ba(ctx->dmactx.pcilib, desc, 0)); *(uint32_t*)last_written_addr_ptr = 0; } // In ring buffer mode, the hardware taking care to preserve an empty buffer to help distinguish between // completely empty and completely full cases. In streaming mode, it is our responsibility to track this // information. Therefore, we always keep the last buffer free num_pages = ctx->ring_size; if (ctx->streaming) num_pages--; for (i = 0; i < num_pages; i++) { uintptr_t bus_addr_check, bus_addr = pcilib_kmem_get_block_ba(ctx->dmactx.pcilib, pages, i); if (ctx->addr64) { WR64(IPEDMA_REG3_PAGE_ADDR, bus_addr); } else { WR(IPEDMA_REG2_PAGE_ADDR, bus_addr); } if (bus_addr%4096) printf("Bad address %lu: %lx\n", i, bus_addr); if ((!ctx->addr64)&&(!ctx->streaming)) { RD(IPEDMA_REG2_PAGE_ADDR, bus_addr_check); if (bus_addr_check != bus_addr) { pcilib_error("Written (%x) and read (%x) bus addresses does not match\n", bus_addr, bus_addr_check); } } usleep(IPEDMA_ADD_PAGE_DELAY); } // Enable DMA WR(IPEDMA_REG_CONTROL, 0x1); ctx->last_read = ctx->ring_size - 1; } ctx->last_read_addr = pcilib_kmem_get_block_ba(ctx->dmactx.pcilib, pages, ctx->last_read); ctx->desc = desc; ctx->pages = pages; return 0; } int dma_ipe_stop(pcilib_dma_context_t *vctx, pcilib_dma_engine_t dma, pcilib_dma_flags_t flags) { pcilib_kmem_flags_t kflags; ipe_dma_t *ctx = (ipe_dma_t*)vctx; if (!ctx->started) return 0; if ((dma != PCILIB_DMA_ENGINE_INVALID)&&(dma > 1)) return PCILIB_ERROR_INVALID_BANK; // ignoring previous setting if flag specified if (flags&PCILIB_DMA_FLAG_PERSISTENT) { ctx->preserve = 0; } if (ctx->preserve) { kflags = PCILIB_KMEM_FLAG_REUSE; } else { kflags = PCILIB_KMEM_FLAG_HARDWARE|PCILIB_KMEM_FLAG_PERSISTENT; ctx->started = 0; dma_ipe_disable(ctx); } // Clean buffers if (ctx->desc) { pcilib_free_kernel_memory(ctx->dmactx.pcilib, ctx->desc, kflags); ctx->desc = NULL; } if (ctx->pages) { pcilib_free_kernel_memory(ctx->dmactx.pcilib, ctx->pages, kflags); ctx->pages = NULL; } return 0; } static size_t dma_ipe_find_buffer_by_bus_addr(ipe_dma_t *ctx, uintptr_t bus_addr) { size_t i; for (i = 0; i < ctx->ring_size; i++) { uintptr_t buf_addr = pcilib_kmem_get_block_ba(ctx->dmactx.pcilib, ctx->pages, i); if (bus_addr == buf_addr) return i; } return (size_t)-1; } int dma_ipe_get_status(pcilib_dma_context_t *vctx, pcilib_dma_engine_t dma, pcilib_dma_engine_status_t *status, size_t n_buffers, pcilib_dma_buffer_status_t *buffers) { size_t i; ipe_dma_t *ctx = (ipe_dma_t*)vctx; void *desc_va = (void*)pcilib_kmem_get_ua(ctx->dmactx.pcilib, ctx->desc); volatile void *last_written_addr_ptr; uint64_t last_written_addr; if (!status) return -1; if (ctx->addr64) { last_written_addr_ptr = desc_va + 2 * sizeof(uint32_t); last_written_addr = *(uint64_t*)last_written_addr_ptr; } else { if (ctx->mode64) last_written_addr_ptr = desc_va + 3 * sizeof(uint32_t); else last_written_addr_ptr = desc_va + 4 * sizeof(uint32_t); last_written_addr = *(uint32_t*)last_written_addr_ptr; } pcilib_debug(DMA, "Current DMA status - last read: %4u, last_read_addr: %4u (0x%x), last_written: %4lu (0x%lx)", ctx->last_read, dma_ipe_find_buffer_by_bus_addr(ctx, ctx->last_read_addr), ctx->last_read_addr, dma_ipe_find_buffer_by_bus_addr(ctx, last_written_addr), last_written_addr ); status->started = ctx->started; status->ring_size = ctx->ring_size; status->buffer_size = ctx->page_size; status->written_buffers = 0; status->written_bytes = 0; // For simplicity, we keep last_read here, and fix in the end status->ring_tail = ctx->last_read; status->ring_head = dma_ipe_find_buffer_by_bus_addr(ctx, last_written_addr); if (status->ring_head == (size_t)-1) { if (last_written_addr) { pcilib_warning("DMA is in unknown state, last_written_addr does not correspond any of available buffers"); return PCILIB_ERROR_FAILED; } status->ring_head = 0; status->ring_tail = 0; } if (n_buffers > ctx->ring_size) n_buffers = ctx->ring_size; if (buffers) memset(buffers, 0, n_buffers * sizeof(pcilib_dma_buffer_status_t)); if (status->ring_head >= status->ring_tail) { for (i = status->ring_tail + 1; i <= status->ring_head; i++) { status->written_buffers++; status->written_bytes += ctx->page_size; if ((buffers)&&(i < n_buffers)) { buffers[i].used = 1; buffers[i].size = ctx->page_size; buffers[i].first = 1; buffers[i].last = 1; } } } else { for (i = 0; i <= status->ring_head; i++) { status->written_buffers++; status->written_bytes += ctx->page_size; if ((buffers)&&(i < n_buffers)) { buffers[i].used = 1; buffers[i].size = ctx->page_size; buffers[i].first = 1; buffers[i].last = 1; } } for (i = status->ring_tail + 1; i < status->ring_size; i++) { status->written_buffers++; status->written_bytes += ctx->page_size; if ((buffers)&&(i < n_buffers)) { buffers[i].used = 1; buffers[i].size = ctx->page_size; buffers[i].first = 1; buffers[i].last = 1; } } } // We actually keep last_read in the ring_tail, so need to increase if (status->ring_tail != status->ring_head) { status->ring_tail++; if (status->ring_tail == status->ring_size) status->ring_tail = 0; } return 0; } int dma_ipe_stream_read(pcilib_dma_context_t *vctx, pcilib_dma_engine_t dma, uintptr_t addr, size_t size, pcilib_dma_flags_t flags, pcilib_timeout_t timeout, pcilib_dma_callback_t cb, void *cbattr) { int err, ret = PCILIB_STREAMING_REQ_PACKET; pcilib_timeout_t wait = 0; struct timeval start, cur; volatile void *desc_va; volatile void *last_written_addr_ptr; // uint32_t empty_detected_dummy = 0; volatile uint32_t *empty_detected_ptr; pcilib_dma_flags_t packet_flags = PCILIB_DMA_FLAG_EOP; size_t nodata_sleep; struct timespec sleep_ts = {0}; size_t cur_read; ipe_dma_t *ctx = (ipe_dma_t*)vctx; err = dma_ipe_start(vctx, dma, PCILIB_DMA_FLAGS_DEFAULT); if (err) return err; desc_va = (void*)pcilib_kmem_get_ua(ctx->dmactx.pcilib, ctx->desc); if (ctx->addr64) { last_written_addr_ptr = desc_va + 2 * sizeof(uint32_t); empty_detected_ptr = desc_va + sizeof(uint32_t); // empty_detected_ptr = &empty_detected_dummy; } else { if (ctx->mode64) last_written_addr_ptr = desc_va + 3 * sizeof(uint32_t); else last_written_addr_ptr = desc_va + 4 * sizeof(uint32_t); empty_detected_ptr = NULL; // Not properly supported // empty_detected_ptr = last_written_addr_ptr - 2; } switch (sched_getscheduler(0)) { case SCHED_FIFO: case SCHED_RR: if (ctx->dma_flags&IPEDMA_FLAG_NOSLEEP) nodata_sleep = 0; else nodata_sleep = IPEDMA_NODATA_SLEEP; break; default: pcilib_info_once("Streaming DMA data using non real-time thread (may cause extra CPU load)", errno); nodata_sleep = 0; } do { switch (ret&PCILIB_STREAMING_TIMEOUT_MASK) { case PCILIB_STREAMING_CONTINUE: // Hardware indicates that there is no more data pending and we can safely stop if there is no data in the kernel buffers already #ifdef IPEDMA_SUPPORT_EMPTY_DETECTED if ((empty_detected_ptr)&&(*empty_detected_ptr)) wait = 0; else #endif /* IPEDMA_SUPPORT_EMPTY_DETECTED */ wait = ctx->dma_timeout; break; case PCILIB_STREAMING_WAIT: wait = (timeout > ctx->dma_timeout)?timeout:ctx->dma_timeout; break; // case PCILIB_STREAMING_CHECK: wait = 0; break; } pcilib_debug(DMA, "Waiting for data in %4u - last_read: %4u, last_read_addr: %4u (0x%08x), last_written: %4u (0x%08x)", ctx->last_read + 1, ctx->last_read, dma_ipe_find_buffer_by_bus_addr(ctx, ctx->last_read_addr), ctx->last_read_addr, dma_ipe_find_buffer_by_bus_addr(ctx, DEREF(last_written_addr_ptr)), DEREF(last_written_addr_ptr) ); gettimeofday(&start, NULL); memcpy(&cur, &start, sizeof(struct timeval)); while (((DEREF(last_written_addr_ptr) == 0)||(ctx->last_read_addr == DEREF(last_written_addr_ptr)))&&((wait == PCILIB_TIMEOUT_INFINITE)||(((cur.tv_sec - start.tv_sec)*1000000 + (cur.tv_usec - start.tv_usec)) < wait))) { if (nodata_sleep) { sleep_ts.tv_nsec = nodata_sleep; nanosleep(&sleep_ts, NULL); } #ifdef IPEDMA_SUPPORT_EMPTY_DETECTED if ((ret != PCILIB_STREAMING_REQ_PACKET)&&(empty_detected_ptr)&&(*empty_detected_ptr)) break; #endif /* IPEDMA_SUPPORT_EMPTY_DETECTED */ gettimeofday(&cur, NULL); } // Failing out if we exited on timeout if ((ctx->last_read_addr == DEREF(last_written_addr_ptr))||(DEREF(last_written_addr_ptr) == 0)) { #ifdef IPEDMA_SUPPORT_EMPTY_DETECTED # ifdef PCILIB_DEBUG_DMA if ((wait)&&(empty_detected_ptr)&&(DEREF(last_written_addr_ptr))&&(!*empty_detected_ptr)) pcilib_debug(DMA, "The empty_detected flag is not set, but no data arrived within %lu us", wait); # endif /* PCILIB_DEBUG_DMA */ #endif /* IPEDMA_SUPPORT_EMPTY_DETECTED */ return (ret&PCILIB_STREAMING_FAIL)?PCILIB_ERROR_TIMEOUT:0; } // Getting next page to read cur_read = ctx->last_read + 1; if (cur_read == ctx->ring_size) cur_read = 0; pcilib_debug(DMA, "Got buffer %4u - last read: %4u, last_read_addr: %4u (0x%x), last_written: %4u (0x%x)", cur_read, ctx->last_read, dma_ipe_find_buffer_by_bus_addr(ctx, ctx->last_read_addr), ctx->last_read_addr, dma_ipe_find_buffer_by_bus_addr(ctx, DEREF(last_written_addr_ptr)), DEREF(last_written_addr_ptr) ); #ifdef IPEDMA_DETECT_PACKETS if ((empty_detected_ptr)&&(*empty_detected_ptr)&&(pcilib_kmem_get_block_ba(ctx->dmactx.pcilib, ctx->pages, cur_read) == DEREF(last_written_addr_ptr))) packet_flags = PCILIB_DMA_FLAG_EOP; else packet_flags = 0; #endif /* IPEDMA_DETECT_PACKETS */ if ((ctx->dma_flags&IPEDMA_FLAG_NOSYNC) == 0) pcilib_kmem_sync_block(ctx->dmactx.pcilib, ctx->pages, PCILIB_KMEM_SYNC_FROMDEVICE, cur_read); void *buf = (void*)pcilib_kmem_get_block_ua(ctx->dmactx.pcilib, ctx->pages, cur_read); ret = cb(cbattr, packet_flags, ctx->page_size, buf); if (ret < 0) return -ret; // We don't need this because hardware does not intend to read anything from the memory //pcilib_kmem_sync_block(ctx->dmactx.pcilib, ctx->pages, PCILIB_KMEM_SYNC_TODEVICE, cur_read); // Return buffer into the DMA pool when processed if (ctx->streaming) { size_t last_free; // We always keep 1 buffer free to distinguish between completely full and empty cases if (cur_read) last_free = cur_read - 1; else last_free = ctx->ring_size - 1; uintptr_t buf_ba = pcilib_kmem_get_block_ba(ctx->dmactx.pcilib, ctx->pages, last_free); if (ctx->addr64) { WR64(IPEDMA_REG3_PAGE_ADDR, buf_ba); } else { WR(IPEDMA_REG2_PAGE_ADDR, buf_ba); } # ifdef IPEDMA_STREAMING_CHECKS pcilib_register_value_t streaming_status; RD(IPEDMA_REG_STREAMING_STATUS, streaming_status); if (streaming_status) pcilib_error("Invalid status (0x%lx) adding a DMA buffer into the queue", streaming_status); # endif /* IPEDMA_STREAMING_MODE */ } // Numbered from 1 #ifdef IPEDMA_BUG_LAST_READ WR(ctx->reg_last_read, cur_read?cur_read:ctx->ring_size); #else /* IPEDMA_BUG_LAST_READ */ WR(ctx->reg_last_read, cur_read + 1); #endif /* IPEDMA_BUG_LAST_READ */ pcilib_debug(DMA, "Buffer returned %4u - last read: %4u, last_read_addr: %4u (0x%x), last_written: %4u (0x%x)", cur_read, ctx->last_read, dma_ipe_find_buffer_by_bus_addr(ctx, ctx->last_read_addr), ctx->last_read_addr, dma_ipe_find_buffer_by_bus_addr(ctx, DEREF(last_written_addr_ptr)), DEREF(last_written_addr_ptr) ); ctx->last_read = cur_read; ctx->last_read_addr = pcilib_kmem_get_block_ba(ctx->dmactx.pcilib, ctx->pages, cur_read); } while (ret); return 0; }