void init_mii_mem() { #ifdef ETHERNET_USE_HARDWARE_LOCKS ethernet_memory_lock = hwlock_alloc(); #endif for (int i=0; i<NUM_ETHERNET_PORTS; ++i) { #if ETHERNET_RX_HP_QUEUE rx_mem_hp[i] = (mii_mempool_t) &rx_hp_data[i][0]; mii_init_mempool(rx_mem_hp[i], ETHERNET_RX_HP_MEMSIZE*4); #endif rx_mem_lp[i] = (mii_mempool_t) &rx_lp_data[i][0]; mii_init_mempool(rx_mem_lp[i], ETHERNET_RX_LP_MEMSIZE*4); #if !ETHERNET_TX_NO_BUFFERING #if ETHERNET_TX_HP_QUEUE tx_mem_hp[i] = (mii_mempool_t) &tx_hp_data[i][0]; mii_init_mempool(tx_mem_hp[i], ETHERNET_TX_HP_MEMSIZE*4); #endif tx_mem_lp[i] = (mii_mempool_t) &tx_lp_data[i][0]; mii_init_mempool(tx_mem_lp[i], ETHERNET_TX_LP_MEMSIZE*4); init_ts_queue(&ts_queue[i]); #endif } return; }
void init_avb_1722_router_table_simple() { int i; table_lock = hwlock_alloc(); for(i=0;i<AVB_NUM_SINKS;i++) { router_table[i].id[0] = 0; router_table[i].id[1] = 0; } }
void init_media_input_fifos(media_input_fifo_t ififos[], media_input_fifo_data_t ififo_data[], int n) { enable_lock = hwlock_alloc(); for(int i=0;i<n;i++) { ififos[i] = (unsigned int) &ififo_data[i]; } }
void init_avb_1722_router_table_hash() { int i; table_lock = hwlock_alloc(); num_entries = 0; for(i=0;i<AVB_1722_ROUTER_TABLE_SIZE;i++) { router_table[i].id[0] = 0; router_table[i].id[1] = 0; backup_table[i].id[0] = 0; backup_table[i].id[1] = 0; } }
void init_locks(void) // Initialise Hardware locks (MutEx's) { hwlock = hwlock_alloc(); } // init_lock
void analyse_init() { // Allocate the hardware lock that will be used to guard access to shared state lock = hwlock_alloc(); }