madwifi: fix a few locking mistakes
[librecmc/librecmc.git] / package / madwifi / patches / 300-napi_polling.patch
1 diff -ur madwifi.old/ath/if_ath.c madwifi.dev/ath/if_ath.c
2 --- madwifi.old/ath/if_ath.c    2007-05-31 06:48:28.561525488 +0200
3 +++ madwifi.dev/ath/if_ath.c    2007-05-31 06:50:07.143538744 +0200
4 @@ -167,7 +167,7 @@
5         int, u_int32_t);
6  static void ath_setdefantenna(struct ath_softc *, u_int);
7  static struct ath_txq *ath_txq_setup(struct ath_softc *, int, int);
8 -static void ath_rx_tasklet(TQUEUE_ARG);
9 +static int ath_rx_poll(struct net_device *dev, int *budget);
10  static int ath_hardstart(struct sk_buff *, struct net_device *);
11  static int ath_mgtstart(struct ieee80211com *, struct sk_buff *);
12  #ifdef ATH_SUPERG_COMP
13 @@ -429,7 +429,6 @@
14         ATH_TXBUF_LOCK_INIT(sc);
15         ATH_RXBUF_LOCK_INIT(sc);
16  
17 -       ATH_INIT_TQUEUE(&sc->sc_rxtq,     ath_rx_tasklet,       dev);
18         ATH_INIT_TQUEUE(&sc->sc_txtq,     ath_tx_tasklet,       dev);
19         ATH_INIT_TQUEUE(&sc->sc_bmisstq,  ath_bmiss_tasklet,    dev);
20         ATH_INIT_TQUEUE(&sc->sc_bstucktq, ath_bstuck_tasklet,   dev);
21 @@ -685,6 +684,8 @@
22         dev->set_mac_address = ath_set_mac_address;
23         dev->change_mtu = ath_change_mtu;
24         dev->tx_queue_len = ATH_TXBUF - 1;              /* 1 for mgmt frame */
25 +       dev->poll = ath_rx_poll;
26 +       dev->weight = 64;
27  #ifdef USE_HEADERLEN_RESV
28         dev->hard_header_len += sizeof(struct ieee80211_qosframe) +
29                                 sizeof(struct llc) +
30 @@ -1643,6 +1644,7 @@
31          */
32         ath_hal_getisr(ah, &status);            /* NB: clears ISR too */
33         DPRINTF(sc, ATH_DEBUG_INTR, "%s: status 0x%x\n", __func__, status);
34 +       sc->sc_isr = status;
35         status &= sc->sc_imask;                 /* discard unasked for bits */
36         if (status & HAL_INT_FATAL) {
37                 sc->sc_stats.ast_hardware++;
38 @@ -1678,7 +1680,12 @@
39                 if (status & HAL_INT_RX) {
40                         sc->sc_tsf = ath_hal_gettsf64(ah);
41                         ath_uapsd_processtriggers(sc);
42 -                       ATH_SCHEDULE_TQUEUE(&sc->sc_rxtq, &needmark);
43 +                       sc->sc_isr &= ~HAL_INT_RX;
44 +                       if (netif_rx_schedule_prep(dev)) {
45 +                               sc->sc_imask &= ~HAL_INT_RX;
46 +                               ath_hal_intrset(ah, sc->sc_imask);
47 +                               __netif_rx_schedule(dev);
48 +                       }
49                 }
50                 if (status & HAL_INT_TX) {
51  #ifdef ATH_SUPERG_DYNTURBO
52 @@ -1704,6 +1711,11 @@
53                                 }
54                         } 
55  #endif
56 +                       /* disable transmit interrupt */
57 +                       sc->sc_isr &= ~HAL_INT_TX;
58 +                       ath_hal_intrset(ah, sc->sc_imask & ~HAL_INT_TX);
59 +                       sc->sc_imask &= ~HAL_INT_TX;
60 +
61                         ATH_SCHEDULE_TQUEUE(&sc->sc_txtq, &needmark);
62                 }
63                 if (status & HAL_INT_BMISS) {
64 @@ -2166,12 +2178,13 @@
65          * Insert the frame on the outbound list and
66          * pass it on to the hardware.
67          */
68 -       ATH_TXQ_LOCK(txq);
69 +       ATH_TXQ_LOCK_IRQ(txq);
70         if (ni && ni->ni_vap && txq == &ATH_VAP(ni->ni_vap)->av_mcastq) {
71                 /*
72                  * The CAB queue is started from the SWBA handler since
73                  * frames only go out on DTIM and to avoid possible races.
74                  */
75 +               sc->sc_imask &= ~HAL_INT_SWBA;
76                 ath_hal_intrset(ah, sc->sc_imask & ~HAL_INT_SWBA);
77                 ATH_TXQ_INSERT_TAIL(txq, bf, bf_list);
78                 DPRINTF(sc, ATH_DEBUG_TX_PROC, "%s: txq depth = %d\n", __func__, txq->axq_depth);
79 @@ -2187,6 +2200,7 @@
80                                 ito64(bf->bf_daddr), bf->bf_desc);
81                 }
82                 txq->axq_link = &lastds->ds_link;
83 +               sc->sc_imask |= HAL_INT_SWBA;
84                 ath_hal_intrset(ah, sc->sc_imask);
85         } else {
86                 ATH_TXQ_INSERT_TAIL(txq, bf, bf_list);
87 @@ -2222,7 +2236,7 @@
88                         }
89                 }
90         }
91 -       ATH_TXQ_UNLOCK(txq);
92 +       ATH_TXQ_UNLOCK_IRQ(txq);
93  
94         sc->sc_devstats.tx_packets++;
95         sc->sc_devstats.tx_bytes += framelen;
96 @@ -2373,12 +2387,18 @@
97         unsigned int pktlen;
98         int framecnt;
99  
100 +       /*
101 +        * NB: using _BH style locking even though this function may be called
102 +        *     at interrupt time (within tasklet or bh). This should be harmless
103 +        *     and this function calls others (i.e., ath_tx_start()) which do
104 +        *     the same.
105 +        */
106         for (;;) {
107 -               ATH_TXQ_LOCK(txq);
108 +               ATH_TXQ_LOCK_BH(txq);
109  
110                 bf_ff = TAILQ_LAST(&txq->axq_stageq, axq_headtype);
111                 if ((!bf_ff) || ath_ff_flushdonetest(txq, bf_ff)) {
112 -                       ATH_TXQ_UNLOCK(txq);
113 +                       ATH_TXQ_UNLOCK_BH(txq);
114                         break;
115                 }
116  
117 @@ -2388,7 +2408,7 @@
118                 ATH_NODE(ni)->an_tx_ffbuf[bf_ff->bf_skb->priority] = NULL;
119                 TAILQ_REMOVE(&txq->axq_stageq, bf_ff, bf_stagelist);
120  
121 -               ATH_TXQ_UNLOCK(txq);
122 +               ATH_TXQ_UNLOCK_BH(txq);
123  
124                 /* encap and xmit */
125                 bf_ff->bf_skb = ieee80211_encap(ni, bf_ff->bf_skb, &framecnt);
126 @@ -2409,15 +2429,16 @@
127                 }
128                 bf_ff->bf_node = NULL;
129  
130 -               ATH_TXBUF_LOCK_IRQ(sc);
131 +               ATH_TXBUF_LOCK_BH(sc);
132                 STAILQ_INSERT_TAIL(&sc->sc_txbuf, bf_ff, bf_list);
133 -               ATH_TXBUF_UNLOCK_IRQ(sc);
134 +               ATH_TXBUF_UNLOCK_BH(sc);
135         }
136 +       ATH_TXQ_UNLOCK_BH(txq);
137  }
138  #endif
139  
140  #define ATH_HARDSTART_GET_TX_BUF_WITH_LOCK                             \
141 -       ATH_TXBUF_LOCK_IRQ(sc);                                         \
142 +       ATH_TXBUF_LOCK_BH(sc);                                          \
143         bf = STAILQ_FIRST(&sc->sc_txbuf);                               \
144         if (bf != NULL) {                                               \
145                 STAILQ_REMOVE_HEAD(&sc->sc_txbuf, bf_list);             \
146 @@ -2432,11 +2453,23 @@
147                 sc->sc_devstopped = 1;                                  \
148                 ATH_SCHEDULE_TQUEUE(&sc->sc_txtq, NULL);                \
149         }                                                               \
150 -       ATH_TXBUF_UNLOCK_IRQ(sc);                                       \
151 +
152 +#define ATH_HARDSTART_REL_TX_BUF_WITH_TXQLOCK_OFF                      \
153 +       ATH_TXBUF_UNLOCK_BH(sc);                                        \
154 +       if (bf == NULL) {               /* NB: should not happen */     \
155 +           DPRINTF(sc,ATH_DEBUG_XMIT,"%s: discard, no xmit buf\n", __func__);                                                                          \
156 +           sc->sc_stats.ast_tx_nobuf++;                                \
157 +           goto hardstart_fail;                                        \
158 +       }
159 +
160 +#define ATH_HARDSTART_REL_TX_BUF_WITH_TXQLOCK_ON                       \
161 +       ATH_TXBUF_UNLOCK_BH(sc);                                        \
162         if (bf == NULL) {               /* NB: should not happen */     \
163                 DPRINTF(sc,ATH_DEBUG_XMIT,                              \
164                         "%s: discard, no xmit buf\n", __func__);        \
165 +           ATH_TXQ_UNLOCK_BH(txq);                                     \
166                 sc->sc_stats.ast_tx_nobuf++;                            \
167 +               goto hardstart_fail;                                    \
168         }
169  
170  /*
171 @@ -2498,6 +2531,7 @@
172         if (M_FLAG_GET(skb, M_UAPSD)) {
173                 /* bypass FF handling */
174                 ATH_HARDSTART_GET_TX_BUF_WITH_LOCK;
175 +               ATH_HARDSTART_REL_TX_BUF_WITH_TXQLOCK_OFF;
176                 if (bf == NULL)
177                         goto hardstart_fail;
178                 goto ff_bypass;
179 @@ -2519,7 +2553,7 @@
180         /* NB: use this lock to protect an->an_ff_txbuf in athff_can_aggregate()
181          *     call too.
182          */
183 -       ATH_TXQ_LOCK(txq);
184 +       ATH_TXQ_LOCK_BH(txq);
185         if (athff_can_aggregate(sc, eh, an, skb, vap->iv_fragthreshold, &ff_flush)) {
186  
187                 if (an->an_tx_ffbuf[skb->priority]) { /* i.e., frame on the staging queue */
188 @@ -2529,7 +2563,7 @@
189                         TAILQ_REMOVE(&txq->axq_stageq, bf, bf_stagelist);
190                         an->an_tx_ffbuf[skb->priority] = NULL;
191  
192 -                       ATH_TXQ_UNLOCK(txq);
193 +                       ATH_TXQ_UNLOCK_BH(txq);
194  
195                         /*
196                          * chain skbs and add FF magic
197 @@ -2556,8 +2590,9 @@
198                          *     to give the buffer back.
199                          */
200                         ATH_HARDSTART_GET_TX_BUF_WITH_LOCK;
201 +                       ATH_HARDSTART_REL_TX_BUF_WITH_TXQLOCK_ON;
202                         if (bf == NULL) {
203 -                               ATH_TXQ_UNLOCK(txq);
204 +                               ATH_TXQ_UNLOCK_BH(txq);
205                                 goto hardstart_fail;
206                         }
207                         DPRINTF(sc, ATH_DEBUG_XMIT | ATH_DEBUG_FF,
208 @@ -2570,7 +2605,7 @@
209  
210                         TAILQ_INSERT_HEAD(&txq->axq_stageq, bf, bf_stagelist);
211  
212 -                       ATH_TXQ_UNLOCK(txq);
213 +                       ATH_TXQ_UNLOCK_BH(txq);
214  
215                         return 0;
216                 }
217 @@ -2581,7 +2616,7 @@
218                         TAILQ_REMOVE(&txq->axq_stageq, bf_ff, bf_stagelist);
219                         an->an_tx_ffbuf[skb->priority] = NULL;
220  
221 -                       ATH_TXQ_UNLOCK(txq);
222 +                       ATH_TXQ_UNLOCK_BH(txq);
223  
224                         /* encap and xmit */
225                         bf_ff->bf_skb = ieee80211_encap(ni, bf_ff->bf_skb, &framecnt);
226 @@ -2611,9 +2646,9 @@
227                         }
228                         bf_ff->bf_node = NULL;
229  
230 -                       ATH_TXBUF_LOCK(sc);
231 +                       ATH_TXBUF_LOCK_BH(sc);
232                         STAILQ_INSERT_TAIL(&sc->sc_txbuf, bf_ff, bf_list);
233 -                       ATH_TXBUF_UNLOCK(sc);
234 +                       ATH_TXBUF_UNLOCK_BH(sc);
235                         goto ff_flushdone;
236                 }
237                 /*
238 @@ -2623,14 +2658,13 @@
239                 else if (an->an_tx_ffbuf[skb->priority]) {
240                         DPRINTF(sc, ATH_DEBUG_XMIT | ATH_DEBUG_FF,
241                                 "%s: Out-Of-Order fast-frame\n", __func__);
242 -                       ATH_TXQ_UNLOCK(txq);
243 +                       ATH_TXQ_UNLOCK_BH(txq);
244                 } else
245 -                       ATH_TXQ_UNLOCK(txq);
246 +                       ATH_TXQ_UNLOCK_BH(txq);
247  
248         ff_flushdone:
249                 ATH_HARDSTART_GET_TX_BUF_WITH_LOCK;
250 -               if (bf == NULL)
251 -                       goto hardstart_fail;
252 +               ATH_HARDSTART_REL_TX_BUF_WITH_TXQLOCK_OFF;
253         }
254  
255  ff_bypass:
256 @@ -2638,6 +2672,7 @@
257  #else /* ATH_SUPERG_FF */
258  
259         ATH_HARDSTART_GET_TX_BUF_WITH_LOCK;
260 +       ATH_HARDSTART_REL_TX_BUF_WITH_TXQLOCK_OFF;
261  
262  #endif /* ATH_SUPERG_FF */
263  
264 @@ -2659,7 +2694,7 @@
265                  *  Allocate 1 ath_buf for each frame given 1 was 
266                  *  already alloc'd
267                  */
268 -               ATH_TXBUF_LOCK(sc);
269 +               ATH_TXBUF_LOCK_BH(sc);
270                 for (bfcnt = 1; bfcnt < framecnt; ++bfcnt) {
271                         if ((tbf = STAILQ_FIRST(&sc->sc_txbuf)) != NULL) {
272                                 STAILQ_REMOVE_HEAD(&sc->sc_txbuf, bf_list);
273 @@ -2680,11 +2715,11 @@
274                                         STAILQ_INSERT_TAIL(&sc->sc_txbuf, tbf, bf_list);
275                                 }
276                         }
277 -                       ATH_TXBUF_UNLOCK(sc);
278 +                       ATH_TXBUF_UNLOCK_BH(sc);
279                         STAILQ_INIT(&bf_head);
280                         goto hardstart_fail;
281                 }
282 -               ATH_TXBUF_UNLOCK(sc);
283 +               ATH_TXBUF_UNLOCK_BH(sc);
284  
285                 while ((bf = STAILQ_FIRST(&bf_head)) != NULL && skb != NULL) {
286                         unsigned int nextfraglen = 0;
287 @@ -2720,7 +2755,7 @@
288  
289  hardstart_fail:
290         if (!STAILQ_EMPTY(&bf_head)) {
291 -               ATH_TXBUF_LOCK(sc);
292 +               ATH_TXBUF_LOCK_BH(sc);
293                 STAILQ_FOREACH_SAFE(tbf, &bf_head, bf_list, tempbf) {
294                         tbf->bf_skb = NULL;
295                         tbf->bf_node = NULL;
296 @@ -2730,7 +2765,7 @@
297  
298                         STAILQ_INSERT_TAIL(&sc->sc_txbuf, tbf, bf_list);
299                 }
300 -               ATH_TXBUF_UNLOCK(sc);
301 +               ATH_TXBUF_UNLOCK_BH(sc);
302         }
303  
304         /* free sk_buffs */
305 @@ -2773,7 +2808,7 @@
306         /*
307          * Grab a TX buffer and associated resources.
308          */
309 -       ATH_TXBUF_LOCK_IRQ(sc);
310 +       ATH_TXBUF_LOCK_BH(sc);
311         bf = STAILQ_FIRST(&sc->sc_txbuf);
312         if (bf != NULL)
313                 STAILQ_REMOVE_HEAD(&sc->sc_txbuf, bf_list);
314 @@ -2784,7 +2819,7 @@
315                 sc->sc_devstopped=1;
316                 ATH_SCHEDULE_TQUEUE(&sc->sc_txtq, NULL);
317         }
318 -       ATH_TXBUF_UNLOCK_IRQ(sc);
319 +       ATH_TXBUF_UNLOCK_BH(sc);
320         if (bf == NULL) {
321                 printk("ath_mgtstart: discard, no xmit buf\n");
322                 sc->sc_stats.ast_tx_nobufmgt++;
323 @@ -2813,9 +2848,9 @@
324                 bf->bf_skb = NULL;
325                 bf->bf_node = NULL;
326  
327 -               ATH_TXBUF_LOCK_IRQ(sc);
328 +               ATH_TXBUF_LOCK_BH(sc);
329                 STAILQ_INSERT_TAIL(&sc->sc_txbuf, bf, bf_list);
330 -               ATH_TXBUF_UNLOCK_IRQ(sc);
331 +               ATH_TXBUF_UNLOCK_BH(sc);
332         }
333         dev_kfree_skb_any(skb);
334         skb = NULL;
335 @@ -3283,10 +3318,10 @@
336          *
337          * XXX Using in_softirq is not right since we might
338          * be called from other soft irq contexts than
339 -        * ath_rx_tasklet.
340 +        * ath_rx_poll
341          */
342         if (!in_softirq())
343 -               tasklet_disable(&sc->sc_rxtq);
344 +               netif_poll_disable(dev);
345         netif_stop_queue(dev);
346  }
347  
348 @@ -3299,7 +3334,7 @@
349         DPRINTF(sc, ATH_DEBUG_KEYCACHE, "%s:\n", __func__);
350         netif_start_queue(dev);
351         if (!in_softirq())              /* NB: see above */
352 -               tasklet_enable(&sc->sc_rxtq);
353 +               netif_poll_enable(dev);
354  }
355  
356  /*
357 @@ -4866,9 +4901,9 @@
358                 bf->bf_node = NULL;
359                 bf->bf_desc->ds_link = 0;
360                 
361 -               ATH_TXBUF_LOCK_IRQ(sc);
362 +               ATH_TXBUF_LOCK_BH(sc);
363                 STAILQ_INSERT_TAIL(&sc->sc_txbuf, bf, bf_list);
364 -               ATH_TXBUF_UNLOCK_IRQ(sc);
365 +               ATH_TXBUF_UNLOCK_BH(sc);
366  
367                 an->an_uapsd_overflowqdepth--;
368         }
369 @@ -5544,13 +5579,12 @@
370         sc->sc_rxotherant = 0;
371  }
372  
373 -static void
374 -ath_rx_tasklet(TQUEUE_ARG data)
375 +static int
376 +ath_rx_poll(struct net_device *dev, int *budget)
377  {
378  #define        PA2DESC(_sc, _pa) \
379         ((struct ath_desc *)((caddr_t)(_sc)->sc_rxdma.dd_desc + \
380                 ((_pa) - (_sc)->sc_rxdma.dd_desc_paddr)))
381 -       struct net_device *dev = (struct net_device *)data;
382         struct ath_buf *bf;
383         struct ath_softc *sc = dev->priv;
384         struct ieee80211com *ic = &sc->sc_ic;
385 @@ -5562,12 +5596,15 @@
386         unsigned int len;
387         int type;
388         u_int phyerr;
389 +       u_int processed = 0, early_stop = 0;
390 +       u_int rx_limit = dev->quota;
391  
392         /* Let the 802.11 layer know about the new noise floor */
393         sc->sc_channoise = ath_hal_get_channel_noise(ah, &(sc->sc_curchan));
394         ic->ic_channoise = sc->sc_channoise;
395         
396         DPRINTF(sc, ATH_DEBUG_RX_PROC, "%s\n", __func__);
397 +process_rx_again:
398         do {
399                 bf = STAILQ_FIRST(&sc->sc_rxbuf);
400                 if (bf == NULL) {               /* XXX ??? can this happen */
401 @@ -5591,6 +5628,13 @@
402                         /* NB: never process the self-linked entry at the end */
403                         break;
404                 }
405 +
406 +               processed++;
407 +               if (rx_limit-- < 0) {
408 +                       early_stop = 1;
409 +                       break;
410 +               }
411 +
412                 skb = bf->bf_skb;
413                 if (skb == NULL) {              /* XXX ??? can this happen */
414                         printk("%s: no skbuff (%s)\n", dev->name, __func__);
415 @@ -5629,6 +5673,7 @@
416                                 sc->sc_stats.ast_rx_phyerr++;
417                                 phyerr = rs->rs_phyerr & 0x1f;
418                                 sc->sc_stats.ast_rx_phy[phyerr]++;
419 +                               goto rx_next;
420                         }
421                         if (rs->rs_status & HAL_RXERR_DECRYPT) {
422                                 /*
423 @@ -5840,9 +5885,29 @@
424                 STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list);
425                 ATH_RXBUF_UNLOCK_IRQ(sc);
426         } while (ath_rxbuf_init(sc, bf) == 0);
427 +       if (!early_stop) {
428 +               /* Check if more data is received while we were
429 +                * processing the descriptor chain.
430 +                */
431 +               ATH_DISABLE_INTR();
432 +               if (sc->sc_isr & HAL_INT_RX) {
433 +                       sc->sc_isr &= ~HAL_INT_RX;
434 +                       ATH_ENABLE_INTR();
435 +                       ath_uapsd_processtriggers(sc);
436 +                       goto process_rx_again;
437 +               }
438 +               netif_rx_complete(dev);
439 +
440 +               sc->sc_imask |= HAL_INT_RX;
441 +               ath_hal_intrset(ah, sc->sc_imask);
442 +               ATH_ENABLE_INTR();
443 +       }
444 +
445 +       *budget -= processed;
446         
447         /* rx signal state monitoring */
448         ath_hal_rxmonitor(ah, &sc->sc_halstats, &sc->sc_curchan);
449 +       return early_stop;
450  #undef PA2DESC
451  }
452  
453 @@ -6118,22 +6183,22 @@
454                                 }
455                         }
456  
457 -                       ATH_TXBUF_LOCK_IRQ(sc);                                 
458 +                       ATH_TXBUF_LOCK_BH(sc);                                  
459                         bf = STAILQ_FIRST(&sc->sc_grppollbuf);
460                         if (bf != NULL)
461                                 STAILQ_REMOVE_HEAD(&sc->sc_grppollbuf, bf_list);
462                         else {
463                                 DPRINTF(sc, ATH_DEBUG_XMIT, "%s: No more TxBufs\n", __func__);
464 -                               ATH_TXBUF_UNLOCK_IRQ_EARLY(sc);
465 +                               ATH_TXBUF_UNLOCK_BH(sc);
466                                 return;
467                         }
468                         /* XXX use a counter and leave at least one for mgmt frames */
469                         if (STAILQ_EMPTY(&sc->sc_grppollbuf)) {                         
470                                 DPRINTF(sc, ATH_DEBUG_XMIT, "%s: No more TxBufs left\n", __func__);
471 -                               ATH_TXBUF_UNLOCK_IRQ_EARLY(sc);
472 +                               ATH_TXBUF_UNLOCK_BH(sc);
473                                 return;
474                         }                                       
475 -                       ATH_TXBUF_UNLOCK_IRQ(sc);
476 +                       ATH_TXBUF_UNLOCK_BH(sc);
477  
478                         bf->bf_skbaddr = bus_map_single(sc->sc_bdev,
479                                 skb->data, skb->len, BUS_DMA_TODEVICE);
480 @@ -6599,9 +6664,9 @@
481                 dev_kfree_skb(lastbuf->bf_skb);
482                 lastbuf->bf_skb = NULL;
483                 ieee80211_unref_node(&lastbuf->bf_node);
484 -               ATH_TXBUF_LOCK_IRQ(sc);
485 +               ATH_TXBUF_LOCK_BH(sc);
486                 STAILQ_INSERT_TAIL(&sc->sc_txbuf, lastbuf, bf_list);
487 -               ATH_TXBUF_UNLOCK_IRQ(sc);
488 +               ATH_TXBUF_UNLOCK_BH(sc);
489                 
490                 /*
491                  *  move oldest from overflow to delivery
492 @@ -7426,9 +7491,6 @@
493                         if (sc->sc_reapcount > ATH_TXBUF_FREE_THRESHOLD) {
494                                 if (!sc->sc_dfswait)
495                                         netif_start_queue(sc->sc_dev);
496 -                               DPRINTF(sc, ATH_DEBUG_TX_PROC,
497 -                                       "%s: tx tasklet restart the queue\n",
498 -                                       __func__);
499                                 sc->sc_reapcount = 0;
500                                 sc->sc_devstopped = 0;
501                         } else
502 @@ -7463,11 +7525,22 @@
503         struct net_device *dev = (struct net_device *)data;
504         struct ath_softc *sc = dev->priv;
505  
506 +process_tx_again:
507         if (txqactive(sc->sc_ah, 0))
508                 ath_tx_processq(sc, &sc->sc_txq[0]);
509         if (txqactive(sc->sc_ah, sc->sc_cabq->axq_qnum))
510                 ath_tx_processq(sc, sc->sc_cabq);
511  
512 +       ATH_DISABLE_INTR();
513 +       if (sc->sc_isr & HAL_INT_TX) {
514 +               sc->sc_isr &= ~HAL_INT_TX;
515 +               ATH_ENABLE_INTR();
516 +               goto process_tx_again;
517 +       }
518 +       sc->sc_imask |= HAL_INT_TX;
519 +       ath_hal_intrset(sc->sc_ah, sc->sc_imask);
520 +       ATH_ENABLE_INTR();
521 +
522         netif_wake_queue(dev);
523  
524         if (sc->sc_softled)
525 @@ -7484,6 +7557,7 @@
526         struct net_device *dev = (struct net_device *)data;
527         struct ath_softc *sc = dev->priv;
528  
529 +process_tx_again:
530         /*
531          * Process each active queue.
532          */
533 @@ -7504,6 +7578,16 @@
534         if (sc->sc_uapsdq && txqactive(sc->sc_ah, sc->sc_uapsdq->axq_qnum))
535                 ath_tx_processq(sc, sc->sc_uapsdq);
536  
537 +       ATH_DISABLE_INTR();
538 +       if (sc->sc_isr & HAL_INT_TX) {
539 +               sc->sc_isr &= ~HAL_INT_TX;
540 +               ATH_ENABLE_INTR();
541 +               goto process_tx_again;
542 +       }
543 +       sc->sc_imask |= HAL_INT_TX;
544 +       ath_hal_intrset(sc->sc_ah, sc->sc_imask);
545 +       ATH_ENABLE_INTR();
546 +
547         netif_wake_queue(dev);
548  
549         if (sc->sc_softled)
550 @@ -7521,6 +7605,7 @@
551         unsigned int i;
552  
553         /* Process each active queue. */
554 +process_tx_again:
555         for (i = 0; i < HAL_NUM_TX_QUEUES; i++)
556                 if (ATH_TXQ_SETUP(sc, i) && txqactive(sc->sc_ah, i))
557                         ath_tx_processq(sc, &sc->sc_txq[i]);
558 @@ -7529,6 +7614,16 @@
559                 ath_tx_processq(sc, sc->sc_xrtxq);
560  #endif
561  
562 +       ATH_DISABLE_INTR();
563 +       if (sc->sc_isr & HAL_INT_TX) {
564 +               sc->sc_isr &= ~HAL_INT_TX;
565 +               ATH_ENABLE_INTR();
566 +               goto process_tx_again;
567 +       }
568 +       sc->sc_imask |= HAL_INT_TX;
569 +       ath_hal_intrset(sc->sc_ah, sc->sc_imask);
570 +       ATH_ENABLE_INTR();
571 +
572         netif_wake_queue(dev);
573  
574         if (sc->sc_softled)
575 @@ -7627,6 +7722,7 @@
576  ath_draintxq(struct ath_softc *sc)
577  {
578         struct ath_hal *ah = sc->sc_ah;
579 +       int npend = 0;
580         unsigned int i;
581  
582         /* XXX return value */
583 @@ -9160,9 +9256,9 @@
584         dev->mtu = mtu;
585         if ((dev->flags & IFF_RUNNING) && !sc->sc_invalid) {
586                 /* NB: the rx buffers may need to be reallocated */
587 -               tasklet_disable(&sc->sc_rxtq);
588 +               netif_poll_disable(dev);
589                 error = ath_reset(dev);
590 -               tasklet_enable(&sc->sc_rxtq);
591 +               netif_poll_enable(dev);
592         }
593         ATH_UNLOCK(sc);
594  
595 Only in madwifi.dev/ath: if_ath.c.orig
596 diff -ur madwifi.old/ath/if_athvar.h madwifi.dev/ath/if_athvar.h
597 --- madwifi.old/ath/if_athvar.h 2007-05-31 06:48:28.555526400 +0200
598 +++ madwifi.dev/ath/if_athvar.h 2007-05-31 06:48:57.847073408 +0200
599 @@ -48,6 +48,10 @@
600  #include "if_athioctl.h"
601  #include "net80211/ieee80211.h"                /* XXX for WME_NUM_AC */
602  
603 +#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0)
604 +#define irqs_disabled()                        0
605 +#endif
606 +
607  /*
608   * Deduce if tasklets are available.  If not then
609   * fall back to using the immediate work queue.
610 @@ -478,8 +482,12 @@
611  #define        ATH_TXQ_LOCK_DESTROY(_tq)       
612  #define        ATH_TXQ_LOCK(_tq)               spin_lock(&(_tq)->axq_lock)
613  #define        ATH_TXQ_UNLOCK(_tq)             spin_unlock(&(_tq)->axq_lock)
614 -#define        ATH_TXQ_LOCK_BH(_tq)            spin_lock_bh(&(_tq)->axq_lock)
615 -#define        ATH_TXQ_UNLOCK_BH(_tq)          spin_unlock_bh(&(_tq)->axq_lock)
616 +#define        ATH_TXQ_LOCK_BH(_tq) \
617 +       if (!irqs_disabled()) \
618 +               spin_lock_bh(&(_tq)->axq_lock)
619 +#define        ATH_TXQ_UNLOCK_BH(_tq) \
620 +       if (!irqs_disabled()) \
621 +               spin_unlock_bh(&(_tq)->axq_lock)
622  #define ATH_TXQ_LOCK_IRQ(_tq)          do {    \
623         unsigned long __axq_lockflags;          \
624         spin_lock_irqsave(&(_tq)->axq_lock, __axq_lockflags);
625 @@ -624,7 +632,6 @@
626         struct ath_buf *sc_rxbufcur;            /* current rx buffer */
627         u_int32_t *sc_rxlink;                   /* link ptr in last RX desc */
628         spinlock_t sc_rxbuflock; 
629 -       struct ATH_TQ_STRUCT sc_rxtq;           /* rx intr tasklet */
630         struct ATH_TQ_STRUCT sc_rxorntq;        /* rxorn intr tasklet */
631         u_int8_t sc_defant;                     /* current default antenna */
632         u_int8_t sc_rxotherant;                 /* rx's on non-default antenna*/
633 @@ -637,6 +644,7 @@
634         u_int sc_txintrperiod;                  /* tx interrupt batching */
635         struct ath_txq sc_txq[HAL_NUM_TX_QUEUES];
636         struct ath_txq *sc_ac2q[WME_NUM_AC];    /* WME AC -> h/w qnum */ 
637 +       HAL_INT sc_isr;                         /* unmasked ISR state */
638         struct ATH_TQ_STRUCT sc_txtq;           /* tx intr tasklet */
639         u_int8_t sc_grppoll_str[GRPPOLL_RATE_STR_LEN];  
640         struct ath_descdma sc_bdma;             /* beacon descriptors */
641 @@ -703,8 +711,12 @@
642  #define        ATH_TXBUF_LOCK_DESTROY(_sc)
643  #define        ATH_TXBUF_LOCK(_sc)             spin_lock(&(_sc)->sc_txbuflock)
644  #define        ATH_TXBUF_UNLOCK(_sc)           spin_unlock(&(_sc)->sc_txbuflock)
645 -#define        ATH_TXBUF_LOCK_BH(_sc)          spin_lock_bh(&(_sc)->sc_txbuflock)
646 -#define        ATH_TXBUF_UNLOCK_BH(_sc)        spin_unlock_bh(&(_sc)->sc_txbuflock)
647 +#define        ATH_TXBUF_LOCK_BH(_sc) \
648 +       if (!irqs_disabled()) \
649 +               spin_lock_bh(&(_sc)->sc_txbuflock)
650 +#define        ATH_TXBUF_UNLOCK_BH(_sc) \
651 +       if (!irqs_disabled()) \
652 +               spin_unlock_bh(&(_sc)->sc_txbuflock)
653  #define        ATH_TXBUF_LOCK_IRQ(_sc)         do {    \
654         unsigned long __txbuflockflags;         \
655         spin_lock_irqsave(&(_sc)->sc_txbuflock, __txbuflockflags);
656 @@ -722,8 +734,12 @@
657  #define        ATH_RXBUF_LOCK_DESTROY(_sc)
658  #define        ATH_RXBUF_LOCK(_sc)             spin_lock(&(_sc)->sc_rxbuflock)
659  #define        ATH_RXBUF_UNLOCK(_sc)           spin_unlock(&(_sc)->sc_rxbuflock)
660 -#define        ATH_RXBUF_LOCK_BH(_sc)          spin_lock_bh(&(_sc)->sc_rxbuflock)
661 -#define        ATH_RXBUF_UNLOCK_BH(_sc)        spin_unlock_bh(&(_sc)->sc_rxbuflock)
662 +#define        ATH_RXBUF_LOCK_BH(_sc) \
663 +       if (!irqs_disabled()) \
664 +               spin_lock_bh(&(_sc)->sc_rxbuflock)
665 +#define        ATH_RXBUF_UNLOCK_BH(_sc) \
666 +       if (!irqs_disabled()) \
667 +               spin_unlock_bh(&(_sc)->sc_rxbuflock)
668  #define        ATH_RXBUF_LOCK_IRQ(_sc)         do {    \
669         unsigned long __rxbuflockflags;         \
670         spin_lock_irqsave(&(_sc)->sc_rxbuflock, __rxbuflockflags);
671 @@ -733,6 +749,8 @@
672  #define        ATH_RXBUF_UNLOCK_IRQ_EARLY(_sc)         \
673         spin_unlock_irqrestore(&(_sc)->sc_rxbuflock, __rxbuflockflags);
674  
675 +#define ATH_DISABLE_INTR               local_irq_disable
676 +#define ATH_ENABLE_INTR                local_irq_enable
677  
678  /* Protects the device from concurrent accesses */
679  #define        ATH_LOCK_INIT(_sc)              init_MUTEX(&(_sc)->sc_lock)
680 Only in madwifi.dev/ath: if_athvar.h.orig
681 diff -ur madwifi.old/net80211/ieee80211_beacon.c madwifi.dev/net80211/ieee80211_beacon.c
682 --- madwifi.old/net80211/ieee80211_beacon.c     2007-05-31 06:48:28.555526400 +0200
683 +++ madwifi.dev/net80211/ieee80211_beacon.c     2007-05-31 06:48:57.847073408 +0200
684 @@ -286,7 +286,7 @@
685         int len_changed = 0;
686         u_int16_t capinfo;
687  
688 -       IEEE80211_LOCK(ic);
689 +       IEEE80211_BEACON_LOCK(ic);
690  
691         if ((ic->ic_flags & IEEE80211_F_DOTH) &&
692             (vap->iv_flags & IEEE80211_F_CHANSWITCH) &&
693 @@ -547,7 +547,7 @@
694                 vap->iv_flags_ext &= ~IEEE80211_FEXT_APPIE_UPDATE;
695         }
696  
697 -       IEEE80211_UNLOCK(ic);
698 +       IEEE80211_BEACON_UNLOCK(ic);
699  
700         return len_changed;
701  }
702 diff -ur madwifi.old/net80211/ieee80211_input.c madwifi.dev/net80211/ieee80211_input.c
703 --- madwifi.old/net80211/ieee80211_input.c      2007-05-31 06:48:28.557526096 +0200
704 +++ madwifi.dev/net80211/ieee80211_input.c      2007-05-31 06:48:57.849073104 +0200
705 @@ -1169,8 +1169,9 @@
706                 if (ni->ni_vlan != 0 && vap->iv_vlgrp != NULL) {
707                         /* attach vlan tag */
708                         vlan_hwaccel_receive_skb(skb, vap->iv_vlgrp, ni->ni_vlan);
709 -               } else
710 -                       netif_rx(skb);
711 +               } else {
712 +                       netif_receive_skb(skb);
713 +               }
714                 dev->last_rx = jiffies;
715         }
716  }
717 @@ -3675,9 +3676,9 @@
718         }
719  
720         /* Okay, take the first queued packet and put it out... */
721 -       IEEE80211_NODE_SAVEQ_LOCK(ni);
722 +       IEEE80211_NODE_SAVEQ_LOCK_IRQ(ni);
723         IEEE80211_NODE_SAVEQ_DEQUEUE(ni, skb, qlen);
724 -       IEEE80211_NODE_SAVEQ_UNLOCK(ni);
725 +       IEEE80211_NODE_SAVEQ_UNLOCK_IRQ(ni);
726         if (skb == NULL) {
727                 IEEE80211_NOTE_MAC(vap, IEEE80211_MSG_POWER, wh->i_addr2,
728                         "%s", "recv ps-poll, but queue empty");
729 Only in madwifi.dev/net80211: ieee80211_input.c.orig
730 diff -ur madwifi.old/net80211/ieee80211_linux.h madwifi.dev/net80211/ieee80211_linux.h
731 --- madwifi.old/net80211/ieee80211_linux.h      2007-05-24 19:31:37.000000000 +0200
732 +++ madwifi.dev/net80211/ieee80211_linux.h      2007-05-31 06:48:57.849073104 +0200
733 @@ -31,6 +31,10 @@
734  
735  #include <linux/wireless.h>
736  
737 +#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0)
738 +#define irqs_disabled()                        0
739 +#endif
740 +
741  /*
742   * Task deferral
743   *
744 @@ -86,8 +90,12 @@
745  } while (0)
746  #define        IEEE80211_UNLOCK_IRQ_EARLY(_ic)                                 \
747         spin_unlock_irqrestore(&(_ic)->ic_comlock, __ilockflags);
748 -#define IEEE80211_LOCK_BH(_ic) spin_lock_bh(&(_ic)->ic_comlock)
749 -#define IEEE80211_UNLOCK_BH(_ic) spin_unlock_bh(&(_ic)->ic_comlock)
750 +#define IEEE80211_LOCK_BH(_ic) \
751 +       if (!irqs_disabled()) \
752 +               spin_lock_bh(&(_ic)->ic_comlock)
753 +#define IEEE80211_UNLOCK_BH(_ic) \
754 +       if (!irqs_disabled()) \
755 +               spin_unlock_bh(&(_ic)->ic_comlock)
756  #define IEEE80211_LOCK(_ic)    spin_lock(&(_ic)->ic_comlock)
757  #define IEEE80211_UNLOCK(_ic)  spin_unlock(&(_ic)->ic_comlock)
758  
759 @@ -104,15 +112,22 @@
760  #define IEEE80211_VAPS_LOCK_DESTROY(_ic)
761  #define IEEE80211_VAPS_LOCK(_ic)       spin_lock(&(_ic)->ic_vapslock);
762  #define IEEE80211_VAPS_UNLOCK(_ic)     spin_unlock(&(_ic)->ic_vapslock);
763 -#define IEEE80211_VAPS_LOCK_BH(_ic)    spin_lock_bh(&(_ic)->ic_vapslock);
764 -#define IEEE80211_VAPS_UNLOCK_BH(_ic)  spin_unlock_bh(&(_ic)->ic_vapslock);
765 -#define IEEE80211_VAPS_LOCK_IRQ(_ic)   do {    \
766 -       int _vaps_lockflags;                    \
767 -       spin_lock_irqsave(&(_ic)->ic_vapslock, _vaps_lockflags);
768 -#define IEEE80211_VAPS_UNLOCK_IRQ(_ic) \
769 -       spin_unlock_irqrestore(&(_ic)->ic_vapslock, _vaps_lockflags); \
770 -} while (0)
771 -#define IEEE80211_VAPS_UNLOCK_IRQ_EARLY(_ic)   spin_unlock_irqrestore(&(_ic)->ic_vapslock, _vaps_lockflags)
772 +#define IEEE80211_VAPS_LOCK_BH(_ic) \
773 +       if (!irqs_disabled()) \
774 +               spin_lock_bh(&(_ic)->ic_vapslock);
775 +#define IEEE80211_VAPS_UNLOCK_BH(_ic) \
776 +       if (!irqs_disabled()) \
777 +               spin_unlock_bh(&(_ic)->ic_vapslock);
778 +#define IEEE80211_VAPS_LOCK_IRQ(_ic) do {                              \
779 +       unsigned long __vlockflags=0;                           \
780 +       unsigned int __vlocked=0;                               \
781 +       __vlocked=spin_is_locked(&(_ic)->ic_vapslock);          \
782 +       if(!__vlocked) spin_lock_irqsave(&(_ic)->ic_vapslock, __vlockflags);
783 +#define IEEE80211_VAPS_UNLOCK_IRQ(_ic)                                 \
784 +       if(!__vlocked) spin_unlock_irqrestore(&(_ic)->ic_vapslock, __vlockflags);       \
785 +} while (0);
786 +#define IEEE80211_VAPS_UNLOCK_IRQ_EARLY(_ic) \
787 +       if (!__vlocked) spin_unlock_irqrestore(&(_ic)->ic_vapslock, _vaps_lockflags)
788  
789  #if (defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK)) && defined(spin_is_locked)
790  #define IEEE80211_VAPS_LOCK_ASSERT(_ic) \
791 @@ -122,6 +137,11 @@
792  #define IEEE80211_VAPS_LOCK_ASSERT(_ic)
793  #endif
794  
795 +/*
796 + * Beacon locking definitions; piggyback on com lock.
797 + */
798 +#define        IEEE80211_BEACON_LOCK(_ic)         IEEE80211_LOCK_IRQ(_ic)
799 +#define        IEEE80211_BEACON_UNLOCK(_ic)       IEEE80211_UNLOCK_IRQ(_ic)
800  
801  /*
802   * Node locking definitions.
803 @@ -191,8 +211,12 @@
804  typedef spinlock_t ieee80211_scan_lock_t;
805  #define        IEEE80211_SCAN_LOCK_INIT(_nt, _name) spin_lock_init(&(_nt)->nt_scanlock)
806  #define        IEEE80211_SCAN_LOCK_DESTROY(_nt)
807 -#define        IEEE80211_SCAN_LOCK_BH(_nt)     spin_lock_bh(&(_nt)->nt_scanlock)
808 -#define        IEEE80211_SCAN_UNLOCK_BH(_nt)   spin_unlock_bh(&(_nt)->nt_scanlock)
809 +#define        IEEE80211_SCAN_LOCK_BH(_nt) \
810 +       if (!irqs_disabled()) \
811 +               spin_lock_bh(&(_nt)->nt_scanlock)
812 +#define        IEEE80211_SCAN_UNLOCK_BH(_nt) \
813 +       if (!irqs_disabled()) \
814 +               spin_unlock_bh(&(_nt)->nt_scanlock)
815  #define        IEEE80211_SCAN_LOCK_IRQ(_nt)    do {    \
816         unsigned long __scan_lockflags;         \
817         spin_lock_irqsave(&(_nt)->nt_scanlock, __scan_lockflags);
818 @@ -217,8 +241,12 @@
819  #define        ACL_LOCK_DESTROY(_as)
820  #define        ACL_LOCK(_as)                   spin_lock(&(_as)->as_lock)
821  #define        ACL_UNLOCK(_as)                 spin_unlock(&(_as)->as_lock)
822 -#define        ACL_LOCK_BH(_as)                spin_lock_bh(&(_as)->as_lock)
823 -#define        ACL_UNLOCK_BH(_as)              spin_unlock_bh(&(_as)->as_lock)
824 +#define        ACL_LOCK_BH(_as) \
825 +       if (!irqs_disabled()) \
826 +               spin_lock_bh(&(_as)->as_lock)
827 +#define        ACL_UNLOCK_BH(_as) \
828 +       if (!irqs_disabled()) \
829 +               spin_unlock_bh(&(_as)->as_lock)
830  
831  #if (defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK)) && defined(spin_is_locked)
832  #define        ACL_LOCK_ASSERT(_as) \
833 diff -ur madwifi.old/net80211/ieee80211_node.c madwifi.dev/net80211/ieee80211_node.c
834 --- madwifi.old/net80211/ieee80211_node.c       2007-05-31 06:48:28.558525944 +0200
835 +++ madwifi.dev/net80211/ieee80211_node.c       2007-05-31 06:48:57.868070216 +0200
836 @@ -1584,7 +1584,7 @@
837         struct ieee80211_node *ni;
838         u_int gen;
839  
840 -       IEEE80211_SCAN_LOCK_IRQ(nt);
841 +       IEEE80211_SCAN_LOCK_BH(nt);
842         gen = ++nt->nt_scangen;
843         
844  restart:
845 @@ -1604,7 +1604,7 @@
846         }
847         IEEE80211_NODE_TABLE_UNLOCK_IRQ(nt);
848  
849 -       IEEE80211_SCAN_UNLOCK_IRQ(nt);
850 +       IEEE80211_SCAN_UNLOCK_BH(nt);
851  }
852  EXPORT_SYMBOL(ieee80211_iterate_dev_nodes);
853  
854 Only in madwifi.dev/net80211: ieee80211_node.c.orig
855 diff -ur madwifi.old/net80211/ieee80211_power.c madwifi.dev/net80211/ieee80211_power.c
856 --- madwifi.old/net80211/ieee80211_power.c      2007-05-24 19:31:37.000000000 +0200
857 +++ madwifi.dev/net80211/ieee80211_power.c      2007-05-31 06:48:57.868070216 +0200
858 @@ -147,7 +147,7 @@
859  #endif
860                 struct sk_buff *skb;
861  
862 -               IEEE80211_NODE_SAVEQ_LOCK(ni);
863 +               IEEE80211_NODE_SAVEQ_LOCK_IRQ(ni);
864                 while ((skb = skb_peek(&ni->ni_savedq)) != NULL &&
865                      M_AGE_GET(skb) < IEEE80211_INACT_WAIT) {
866                         IEEE80211_NOTE(vap, IEEE80211_MSG_POWER, ni,
867 @@ -159,7 +159,7 @@
868                 }
869                 if (skb != NULL)
870                         M_AGE_SUB(skb, IEEE80211_INACT_WAIT);
871 -               IEEE80211_NODE_SAVEQ_UNLOCK(ni);
872 +               IEEE80211_NODE_SAVEQ_UNLOCK_IRQ(ni);
873  
874                 IEEE80211_NOTE(vap, IEEE80211_MSG_POWER, ni,
875                         "discard %u frames for age", discard);
876 @@ -185,7 +185,7 @@
877         KASSERT(aid < vap->iv_max_aid,
878                 ("bogus aid %u, max %u", aid, vap->iv_max_aid));
879  
880 -       IEEE80211_LOCK(ni->ni_ic);
881 +       IEEE80211_BEACON_LOCK(ni->ni_ic);
882         if (set != (isset(vap->iv_tim_bitmap, aid) != 0)) {
883                 if (set) {
884                         setbit(vap->iv_tim_bitmap, aid);
885 @@ -196,7 +196,7 @@
886                 }
887                 vap->iv_flags |= IEEE80211_F_TIMUPDATE;
888         }
889 -       IEEE80211_UNLOCK(ni->ni_ic);
890 +       IEEE80211_BEACON_UNLOCK(ni->ni_ic);
891  }
892  
893  /*
894 @@ -295,9 +295,9 @@
895                         struct sk_buff *skb;
896                         int qlen;
897  
898 -                       IEEE80211_NODE_SAVEQ_LOCK(ni);
899 +                       IEEE80211_NODE_SAVEQ_LOCK_IRQ(ni);
900                         IEEE80211_NODE_SAVEQ_DEQUEUE(ni, skb, qlen);
901 -                       IEEE80211_NODE_SAVEQ_UNLOCK(ni);
902 +                       IEEE80211_NODE_SAVEQ_UNLOCK_IRQ(ni);
903                         if (skb == NULL)
904                                 break;
905                         /* 
906 @@ -362,9 +362,9 @@
907                         for (;;) {
908                                 struct sk_buff *skb;
909  
910 -                               IEEE80211_NODE_SAVEQ_LOCK(ni);
911 +                               IEEE80211_NODE_SAVEQ_LOCK_IRQ(ni);
912                                 skb = __skb_dequeue(&ni->ni_savedq);
913 -                               IEEE80211_NODE_SAVEQ_UNLOCK(ni);
914 +                               IEEE80211_NODE_SAVEQ_UNLOCK_IRQ(ni);
915                                 if (skb == NULL)
916                                         break;
917                                 ieee80211_parent_queue_xmit(skb);
918 diff -ur madwifi.old/net80211/ieee80211_proto.c madwifi.dev/net80211/ieee80211_proto.c
919 --- madwifi.old/net80211/ieee80211_proto.c      2007-05-31 06:48:28.564525032 +0200
920 +++ madwifi.dev/net80211/ieee80211_proto.c      2007-05-31 06:48:57.869070064 +0200
921 @@ -635,9 +635,9 @@
922  {
923         struct ieee80211com *ic = vap->iv_ic;
924  
925 -       IEEE80211_LOCK(ic);
926 +       IEEE80211_BEACON_LOCK(ic);
927         ieee80211_wme_initparams_locked(vap);
928 -       IEEE80211_UNLOCK(ic);
929 +       IEEE80211_BEACON_UNLOCK(ic);
930  }
931  
932  void
933 @@ -920,9 +920,9 @@
934         struct ieee80211com *ic = vap->iv_ic;
935  
936         if (ic->ic_caps & IEEE80211_C_WME) {
937 -               IEEE80211_LOCK(ic);
938 +               IEEE80211_BEACON_LOCK(ic);
939                 ieee80211_wme_updateparams_locked(vap);
940 -               IEEE80211_UNLOCK(ic);
941 +               IEEE80211_BEACON_UNLOCK(ic);
942         }
943  }
944  
945 diff -ur madwifi.old/net80211/ieee80211_scan_sta.c madwifi.dev/net80211/ieee80211_scan_sta.c
946 --- madwifi.old/net80211/ieee80211_scan_sta.c   2007-05-21 17:53:39.000000000 +0200
947 +++ madwifi.dev/net80211/ieee80211_scan_sta.c   2007-05-31 06:48:57.870069912 +0200
948 @@ -163,9 +163,11 @@
949  {
950         struct sta_table *st = ss->ss_priv;
951  
952 -       spin_lock(&st->st_lock);
953 +       if (!irqs_disabled())
954 +               spin_lock_bh(&st->st_lock);
955         sta_flush_table(st);
956 -       spin_unlock(&st->st_lock);
957 +       if (!irqs_disabled())
958 +               spin_unlock_bh(&st->st_lock);
959         ss->ss_last = 0;
960         return 0;
961  }
962 @@ -215,7 +217,8 @@
963         int hash;
964  
965         hash = STA_HASH(macaddr);
966 -       spin_lock(&st->st_lock);  
967 +       if (!irqs_disabled())
968 +               spin_lock_bh(&st->st_lock);  
969         LIST_FOREACH(se, &st->st_hash[hash], se_hash)
970                 if (IEEE80211_ADDR_EQ(se->base.se_macaddr, macaddr) &&
971                     sp->ssid[1] == se->base.se_ssid[1] && 
972 @@ -225,7 +228,7 @@
973         MALLOC(se, struct sta_entry *, sizeof(struct sta_entry),
974                 M_80211_SCAN, M_NOWAIT | M_ZERO);
975         if (se == NULL) {
976 -               spin_unlock(&st->st_lock);
977 +               spin_unlock_bh(&st->st_lock);
978                 return 0;
979         }
980         se->se_scangen = st->st_scangen-1;
981 @@ -287,7 +290,8 @@
982         se->se_seen = 1;
983         se->se_notseen = 0;
984  
985 -       spin_unlock(&st->st_lock);
986 +       if (!irqs_disabled())
987 +               spin_unlock_bh(&st->st_lock);
988  
989         /*
990          * If looking for a quick choice and nothing's
991 @@ -1063,7 +1067,8 @@
992         u_int gen;
993         int res = 0;
994  
995 -       spin_lock(&st->st_scanlock);
996 +       if (!irqs_disabled())
997 +               spin_lock_bh(&st->st_scanlock);
998         gen = st->st_scangen++;
999  restart:
1000         spin_lock(&st->st_lock);
1001 @@ -1086,7 +1091,8 @@
1002         spin_unlock(&st->st_lock);
1003  
1004   done:
1005 -       spin_unlock(&st->st_scanlock);
1006 +       if (!irqs_disabled())
1007 +               spin_unlock_bh(&st->st_scanlock);
1008  
1009         return res;
1010  }
1011 @@ -1235,7 +1241,8 @@
1012         bestchan = NULL;
1013         bestrssi = -1;
1014  
1015 -       spin_lock(&st->st_lock);
1016 +       if (!irqs_disabled())
1017 +               spin_lock_bh(&st->st_lock);
1018         for (i = 0; i < ss->ss_last; i++) {
1019                 c = ss->ss_chans[i];
1020                 maxrssi = 0;
1021 @@ -1248,7 +1255,8 @@
1022                 if (bestchan == NULL || maxrssi < bestrssi)
1023                         bestchan = c;
1024         }
1025 -       spin_unlock(&st->st_lock);
1026 +       if (!irqs_disabled())
1027 +               spin_unlock_bh(&st->st_lock);
1028  
1029         return bestchan;
1030  }