|
2 | 2 |
|
3 | 3 | exi.c -- EXI subsystem |
4 | 4 |
|
5 | | -Copyright (C) 2004 - 2025 |
| 5 | +Copyright (C) 2004 - 2026 |
6 | 6 | Michael Wiedenbauer (shagkur) |
7 | 7 | Dave Murphy (WinterMute) |
8 | 8 | Extrems' Corner.org |
@@ -130,9 +130,9 @@ static __inline__ void __exi_setinterrupts(s32 nChn,exibus_priv *exi) |
130 | 130 | printf("__exi_setinterrupts(%d,%p)\n",nChn,exi); |
131 | 131 | #endif |
132 | 132 | if(nChn==EXI_CHANNEL_0) { |
133 | | - __MaskIrq((IRQMASK(IRQ_EXI0_EXI)|IRQMASK(IRQ_EXI2_EXI))); |
| 133 | + __MaskIrq(IRQMASK(IRQ_EXI0_EXI)|IRQMASK(IRQ_EXI2_EXI)); |
134 | 134 | if(!(exi->flags&EXI_FLAG_LOCKED) && (exi->CallbackEXI || pexi->CallbackEXI)) |
135 | | - __UnmaskIrq((IRQMASK(IRQ_EXI0_EXI)|IRQMASK(IRQ_EXI2_EXI))); |
| 135 | + __UnmaskIrq(IRQMASK(IRQ_EXI0_EXI)|IRQMASK(IRQ_EXI2_EXI)); |
136 | 136 | } else if(nChn==EXI_CHANNEL_1) { |
137 | 137 | __MaskIrq(IRQMASK(IRQ_EXI1_EXI)); |
138 | 138 | if(!(exi->flags&EXI_FLAG_LOCKED) && exi->CallbackEXI) __UnmaskIrq(IRQMASK(IRQ_EXI1_EXI)); |
@@ -227,7 +227,7 @@ static s32 __exi_attach(s32 nChn,EXICallback ext_cb) |
227 | 227 | if(__exi_probe(nChn)==1) { |
228 | 228 | __exi_clearirqs(nChn,1,0,0); |
229 | 229 | exi->CallbackEXT = ext_cb; |
230 | | - __UnmaskIrq(((IRQMASK(IRQ_EXI0_EXT))>>(nChn*3))); |
| 230 | + __UnmaskIrq(IRQMASK(IRQ_EXI0_EXT)>>(nChn*3)); |
231 | 231 | exi->flags |= EXI_FLAG_ATTACH; |
232 | 232 | ret = 1; |
233 | 233 | } |
@@ -575,7 +575,7 @@ s32 EXI_Imm(s32 nChn,void *pData,u32 nLen,u32 nMode,EXICallback tc_cb) |
575 | 575 | exi->CallbackTC = tc_cb; |
576 | 576 | if(tc_cb) { |
577 | 577 | __exi_clearirqs(nChn,0,1,0); |
578 | | - __UnmaskIrq(IRQMASK((IRQ_EXI0_TC+(nChn*3)))); |
| 578 | + __UnmaskIrq(IRQMASK(IRQ_EXI0_TC)>>(nChn*3)); |
579 | 579 | } |
580 | 580 | exi->flags |= EXI_FLAG_IMM; |
581 | 581 |
|
@@ -632,7 +632,7 @@ s32 EXI_Dma(s32 nChn,void *pData,u32 nLen,u32 nMode,EXICallback tc_cb) |
632 | 632 | exi->CallbackTC = tc_cb; |
633 | 633 | if(tc_cb) { |
634 | 634 | __exi_clearirqs(nChn,0,1,0); |
635 | | - __UnmaskIrq((IRQMASK((IRQ_EXI0_TC+(nChn*3))))); |
| 635 | + __UnmaskIrq(IRQMASK(IRQ_EXI0_TC)>>(nChn*3)); |
636 | 636 | } |
637 | 637 | exi->flags |= EXI_FLAG_DMA; |
638 | 638 |
|
@@ -923,7 +923,7 @@ s32 EXI_Detach(s32 nChn) |
923 | 923 | if(exi->flags&EXI_FLAG_LOCKED && exi->lockeddev==EXI_DEVICE_0) ret = 0; |
924 | 924 | else { |
925 | 925 | exi->flags &= ~EXI_FLAG_ATTACH; |
926 | | - __MaskIrq(((IRQMASK(IRQ_EXI0_EXI)|IRQMASK(IRQ_EXI0_EXT))>>(nChn*3))); |
| 926 | + __MaskIrq((IRQMASK(IRQ_EXI0_EXI)|IRQMASK(IRQ_EXI0_EXT))>>(nChn*3)); |
927 | 927 | } |
928 | 928 | } |
929 | 929 | _CPU_ISR_Restore(level); |
@@ -1081,7 +1081,7 @@ void __ext_irq_handler(u32 nIrq,frame_context *pCtx) |
1081 | 1081 |
|
1082 | 1082 | exi = &eximap[chan]; |
1083 | 1083 | exi->flags &= ~EXI_FLAG_ATTACH; |
1084 | | - __MaskIrq(((IRQMASK(IRQ_EXI0_EXI)|IRQMASK(IRQ_EXI0_EXT))>>(chan*3))); |
| 1084 | + __MaskIrq((IRQMASK(IRQ_EXI0_EXI)|IRQMASK(IRQ_EXI0_EXT))>>(chan*3)); |
1085 | 1085 |
|
1086 | 1086 | cb = exi->CallbackEXT; |
1087 | 1087 | #ifdef _EXI_DEBUG |
|
0 commit comments