| #ifndef HTP_DMA_H |
| #define HTP_DMA_H |
|
|
| #include <HAP_farf.h> |
| #include <hexagon_types.h> |
| #include <stdbool.h> |
| #include <stdint.h> |
|
|
| #ifdef __cplusplus |
| extern "C" { |
| #endif |
|
|
| |
| typedef struct dma_descriptor_1d_s { |
| void * next; |
| uint32_t size:24; |
| uint32_t desc_size:2; |
| uint32_t dst_comp:1; |
| uint32_t src_comp:1; |
| uint32_t dst_bypass:1; |
| uint32_t src_bypass:1; |
| uint32_t order:1; |
| uint32_t done:1; |
| void * src; |
| void * dst; |
| } dma_descriptor_1d; |
|
|
| #if __HVX_ARCH__ < 75 |
|
|
| typedef struct dma_descriptor_2d_s { |
| void * next; |
| uint32_t reserved0:24; |
| uint32_t desc_size:2; |
| uint32_t dst_comp:1; |
| uint32_t src_comp:1; |
| uint32_t dst_bypass:1; |
| uint32_t src_bypass:1; |
| uint32_t order:1; |
| uint32_t done:1; |
| void * src; |
| void * dst; |
| uint32_t desc_type:8; |
| uint32_t reserved1:24; |
| uint32_t row_size:16; |
| uint32_t nrows:16; |
| uint32_t src_stride:16; |
| uint32_t dst_stride:16; |
| uint32_t src_offset:16; |
| uint32_t dst_offset:16; |
| } dma_descriptor_2d; |
|
|
| #else |
|
|
| typedef struct dma_descriptor_2d_s { |
| void * next; |
| uint32_t dst_stride:24; |
| uint32_t desc_size:2; |
| uint32_t dst_comp:1; |
| uint32_t src_comp:1; |
| uint32_t dst_bypass:1; |
| uint32_t src_bypass:1; |
| uint32_t order:1; |
| uint32_t done:1; |
| void * src; |
| void * dst; |
| uint32_t desc_type:8; |
| uint32_t reserved0:24; |
| uint32_t row_size:24; |
| uint32_t nrows_lo:8; |
| uint32_t nrows_hi:8; |
| uint32_t src_stride:24; |
| uint32_t offset:24; |
| uint32_t reserved1:8; |
| } dma_descriptor_2d; |
|
|
| #endif |
|
|
| typedef struct { |
| void *dst; |
| const void *src; |
| } dma_ptr; |
|
|
| typedef struct { |
| dma_descriptor_2d * desc; |
| dma_descriptor_2d * tail; |
| dma_ptr * dptr; |
| uint32_t push_idx; |
| uint32_t pop_idx; |
| uint32_t capacity; |
| uint32_t idx_mask; |
| } dma_queue; |
|
|
| dma_queue * dma_queue_create(size_t capacity); |
| void dma_queue_delete(dma_queue * q); |
| void dma_queue_flush(dma_queue * q); |
|
|
| |
| |
| static inline void dmstart(void * next) { |
| asm volatile(" release(%0):at" : : "r"(next)); |
| asm volatile(" dmstart(%0)" : : "r"(next)); |
| } |
|
|
| static inline void dmlink(void * cur, void * next) { |
| asm volatile(" release(%0):at" : : "r"(next)); |
| asm volatile(" dmlink(%0, %1)" : : "r"(cur), "r"(next)); |
| } |
|
|
| static inline unsigned int dmpoll(void) { |
| unsigned int ret = 0; |
| asm volatile(" %0 = dmpoll" : "=r"(ret) : : "memory"); |
| return ret; |
| } |
|
|
| static inline unsigned int dmwait(void) { |
| unsigned int ret = 0; |
| asm volatile(" %0 = dmwait" : "=r"(ret) : : "memory"); |
| return ret; |
| } |
|
|
| static inline dma_ptr dma_make_ptr(void *dst, const void *src) |
| { |
| dma_ptr p = { dst, src }; |
| return p; |
| } |
|
|
| #if __HVX_ARCH__ < 73 |
| static const uint32_t dma_src_l2_bypass_on = 1; |
| static const uint32_t dma_dst_l2_bypass_on = 0; |
| #else |
| static const uint32_t dma_src_l2_bypass_on = 1; |
| static const uint32_t dma_dst_l2_bypass_on = 1; |
| #endif |
|
|
| static inline bool dma_queue_push_single_1d(dma_queue * q, dma_ptr dptr, size_t size) { |
| if (((q->push_idx + 1) & q->idx_mask) == q->pop_idx) { |
| FARF(HIGH, "dma-push: queue full\n"); |
| return false; |
| } |
|
|
| dma_descriptor_1d * desc = (dma_descriptor_1d *) &q->desc[q->push_idx]; |
| desc->next = NULL; |
| desc->desc_size = 0; |
| desc->src_bypass = dma_src_l2_bypass_on; |
| desc->dst_bypass = dma_dst_l2_bypass_on; |
| desc->order = 0; |
| desc->done = 0; |
| desc->src = (void *) dptr.src; |
| desc->dst = (void *) dptr.dst; |
| desc->size = size; |
|
|
| q->dptr[q->push_idx] = dptr; |
|
|
| if (size) { |
| dmlink(q->tail, desc); |
| q->tail = (dma_descriptor_2d *) desc; |
| } else { |
| desc->done = 1; |
| } |
|
|
| |
| q->push_idx = (q->push_idx + 1) & q->idx_mask; |
| return true; |
| } |
|
|
| static inline bool dma_queue_push_single_2d(dma_queue * q, dma_ptr dptr, size_t dst_stride, size_t src_stride, size_t row_size, size_t nrows) { |
| if (((q->push_idx + 1) & q->idx_mask) == q->pop_idx) { |
| FARF(HIGH, "dma-push: queue full\n"); |
| return false; |
| } |
|
|
| dma_descriptor_2d * desc = &q->desc[q->push_idx]; |
|
|
| desc->next = NULL; |
| desc->reserved0 = 0; |
| desc->reserved1 = 0; |
| desc->desc_size = 1; |
| desc->src_bypass = dma_src_l2_bypass_on; |
| desc->dst_bypass = dma_dst_l2_bypass_on; |
| desc->src_comp = 0; |
| desc->dst_comp = 0; |
| desc->order = 0; |
| desc->done = 0; |
| desc->src_stride = src_stride; |
| desc->dst_stride = dst_stride; |
| desc->src = (void *) dptr.src; |
| desc->dst = (void *) dptr.dst; |
| desc->row_size = row_size; |
|
|
| #if __HVX_ARCH__ < 75 |
| desc->desc_type = 0; |
| desc->nrows = nrows; |
| desc->src_offset = 0; |
| desc->dst_offset = 0; |
| #else |
| desc->desc_type = 9; |
| desc->nrows_lo = (nrows & 0xff); |
| desc->nrows_hi = (nrows >> 8); |
| desc->offset = 0; |
| #endif |
|
|
| q->dptr[q->push_idx] = dptr; |
|
|
| if (nrows) { |
| dmlink(q->tail, desc); |
| q->tail = desc; |
| } else { |
| desc->done = 1; |
| } |
|
|
| |
| q->push_idx = (q->push_idx + 1) & q->idx_mask; |
| return true; |
| } |
|
|
| static inline dma_ptr dma_queue_pop(dma_queue * q) { |
| dma_ptr dptr = { NULL }; |
|
|
| if (q->push_idx == q->pop_idx) { |
| return dptr; |
| } |
|
|
| dma_descriptor_2d * desc = &q->desc[q->pop_idx]; |
|
|
| |
| while (!desc->done) { |
| |
| dmpoll(); |
| } |
|
|
| dptr = q->dptr[q->pop_idx]; |
|
|
| |
| q->pop_idx = (q->pop_idx + 1) & q->idx_mask; |
| return dptr; |
| } |
|
|
| static inline dma_ptr dma_queue_pop_nowait(dma_queue * q) { |
| dma_ptr dptr = { NULL }; |
|
|
| if (q->push_idx == q->pop_idx) { |
| return dptr; |
| } |
|
|
| dptr = q->dptr[q->pop_idx]; |
|
|
| |
| q->pop_idx = (q->pop_idx + 1) & q->idx_mask; |
| return dptr; |
| } |
|
|
| static inline bool dma_queue_empty(dma_queue * q) { |
| return q->push_idx == q->pop_idx; |
| } |
|
|
| static inline uint32_t dma_queue_depth(dma_queue * q) { |
| return (q->push_idx - q->pop_idx) & q->idx_mask; |
| } |
|
|
| static inline uint32_t dma_queue_capacity(dma_queue * q) { |
| return q->capacity; |
| } |
|
|
| #if __HVX_ARCH__ < 75 |
|
|
| |
| |
|
|
| #define DMA_MAX_FIELD_VAL 65535u |
|
|
| static inline bool dma_queue_push(dma_queue *q, dma_ptr dptr, size_t dst_stride, size_t src_stride, size_t row_size, size_t nrows) { |
| |
| if (nrows == 0 || __builtin_expect( |
| row_size <= DMA_MAX_FIELD_VAL && |
| nrows <= DMA_MAX_FIELD_VAL && |
| src_stride <= DMA_MAX_FIELD_VAL && |
| dst_stride <= DMA_MAX_FIELD_VAL, 1)) { |
| return dma_queue_push_single_2d(q, dptr, dst_stride, src_stride, row_size, nrows); |
| } |
|
|
| |
| |
| if (nrows == 1 || (row_size == src_stride && row_size == dst_stride)) { |
| size_t total = row_size * nrows; |
| return dma_queue_push_single_1d(q, dptr, total); |
| } |
|
|
| |
| { |
| const uint8_t *src = (const uint8_t *) dptr.src; |
| uint8_t *dst = (uint8_t *) dptr.dst; |
| for (size_t r = 0; r < nrows; ++r) { |
| dma_ptr p = dma_make_ptr(dst + r * dst_stride, src + r * src_stride); |
| if (!dma_queue_push_single_1d(q, p, row_size)) |
| return false; |
| if (r + 1 < nrows) |
| dma_queue_pop(q); |
| } |
| return true; |
| } |
| } |
|
|
| #else |
|
|
| static inline bool dma_queue_push(dma_queue *q, dma_ptr dptr, size_t dst_stride, size_t src_stride, size_t row_size, size_t nrows) { |
| |
| return dma_queue_push_single_2d(q, dptr, dst_stride, src_stride, row_size, nrows); |
| } |
|
|
| #endif |
|
|
| static inline bool dma_queue_push_ddr_to_vtcm(dma_queue * q, dma_ptr dptr, size_t dst_row_size, size_t src_row_size, size_t nrows) { |
| return dma_queue_push(q, dptr, dst_row_size, src_row_size, src_row_size, nrows); |
| } |
|
|
| static inline bool dma_queue_push_vtcm_to_ddr(dma_queue * q, dma_ptr dptr, size_t dst_row_size, size_t src_row_size, size_t nrows) { |
| return dma_queue_push(q, dptr, dst_row_size, src_row_size, dst_row_size, nrows); |
| } |
|
|
| #define DMA_CACHE_MAX_SIZE 64U |
|
|
| typedef struct { |
| uint8_t *base; |
| uint32_t line_size; |
| uint32_t capacity; |
| uint32_t src[DMA_CACHE_MAX_SIZE]; |
| uint16_t age[DMA_CACHE_MAX_SIZE]; |
| } dma_cache; |
|
|
| static inline void dma_cache_init(dma_cache *c, uint8_t *base, uint32_t line_size, uint32_t capacity) |
| { |
| c->capacity = (capacity > DMA_CACHE_MAX_SIZE) ? DMA_CACHE_MAX_SIZE : capacity; |
| c->base = base; |
| c->line_size = line_size; |
|
|
| for (unsigned i=0; i < c->capacity; i++) { |
| c->src[i] = 0; |
| c->age[i] = 0; |
| } |
| } |
|
|
| static inline bool dma_cache_push(dma_queue *q, dma_cache *c, const uint8_t * src, uint32_t dst_stride, uint32_t src_stride, uint32_t row_size, uint32_t nrows) |
| { |
| uint32_t o_idx = 0; |
| uint16_t o_age = 0; |
| uint8_t * dst = 0; |
|
|
| for (unsigned i=0; i < c->capacity; i++) { |
| if (c->src[i] == (uint32_t) src) { |
| c->age[i] = 0; |
| dst = c->base + (i * c->line_size); nrows = 0; |
| |
| } else { |
| c->age[i]++; |
| if (c->age[i] > o_age) { o_age = c->age[i]; o_idx = i; } |
| } |
| } |
| if (!dst) { |
| |
| c->age[o_idx] = 0; |
| c->src[o_idx] = (uint32_t) src; |
| dst = c->base + o_idx * c->line_size; |
| } |
|
|
| return dma_queue_push(q, dma_make_ptr(dst, src), dst_stride, src_stride, row_size, nrows); |
| } |
|
|
| #ifdef __cplusplus |
| } |
| #endif |
|
|
| #endif |
|
|