29 static void bfa_ioc_cb_firmware_unlock(
struct bfa_ioc_s *
ioc);
33 static void bfa_ioc_cb_notify_fail(
struct bfa_ioc_s *
ioc);
34 static void bfa_ioc_cb_ownership_reset(
struct bfa_ioc_s *
ioc);
37 static void bfa_ioc_cb_sync_leave(
struct bfa_ioc_s *
ioc);
50 hwif_cb.ioc_firmware_lock = bfa_ioc_cb_firmware_lock;
51 hwif_cb.ioc_firmware_unlock = bfa_ioc_cb_firmware_unlock;
52 hwif_cb.ioc_reg_init = bfa_ioc_cb_reg_init;
53 hwif_cb.ioc_map_port = bfa_ioc_cb_map_port;
54 hwif_cb.ioc_isr_mode_set = bfa_ioc_cb_isr_mode_set;
55 hwif_cb.ioc_notify_fail = bfa_ioc_cb_notify_fail;
56 hwif_cb.ioc_ownership_reset = bfa_ioc_cb_ownership_reset;
57 hwif_cb.ioc_sync_start = bfa_ioc_cb_sync_start;
58 hwif_cb.ioc_sync_join = bfa_ioc_cb_sync_join;
59 hwif_cb.ioc_sync_leave = bfa_ioc_cb_sync_leave;
60 hwif_cb.ioc_sync_ack = bfa_ioc_cb_sync_ack;
61 hwif_cb.ioc_sync_complete = bfa_ioc_cb_sync_complete;
115 ioc->
ioc_regs.hfn_mbox = rb + iocreg_fnreg[pcifn].hfn_mbox;
116 ioc->
ioc_regs.lpu_mbox = rb + iocreg_fnreg[pcifn].lpu_mbox;
117 ioc->
ioc_regs.host_page_num_fn = rb + iocreg_fnreg[pcifn].hfn_pgn;
132 ioc->
ioc_regs.hfn_mbox_cmd = rb + iocreg_mbcmd[pcifn].hfn;
133 ioc->
ioc_regs.lpu_mbox_cmd = rb + iocreg_mbcmd[pcifn].lpu;
166 bfa_ioc_cb_map_port(
struct bfa_ioc_s *ioc)
188 bfa_ioc_cb_sync_start(
struct bfa_ioc_s *ioc)
190 return bfa_ioc_cb_sync_complete(ioc);
197 bfa_ioc_cb_ownership_reset(
struct bfa_ioc_s *ioc)
213 bfa_ioc_cb_sync_join(
struct bfa_ioc_s *ioc)
218 bfa_ioc_cb_sync_leave(
struct bfa_ioc_s *ioc)
223 bfa_ioc_cb_sync_ack(
struct bfa_ioc_s *ioc)
229 bfa_ioc_cb_sync_complete(
struct bfa_ioc_s *ioc)
275 u32 pll_sclk, pll_fclk;