2 * Copyright (c) 2010 Broadcom Corporation
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
17 #include <linux/kernel.h>
18 #include <linux/etherdevice.h>
19 #include <linux/module.h>
20 #include <net/cfg80211.h>
21 #include <net/rtnetlink.h>
22 #include <brcmu_utils.h>
23 #include <brcmu_wifi.h>
28 #include "fwil_types.h"
30 #include "wl_cfg80211.h"
35 MODULE_AUTHOR("Broadcom Corporation");
36 MODULE_DESCRIPTION("Broadcom 802.11 wireless LAN fullmac driver.");
37 MODULE_LICENSE("Dual BSD/GPL");
39 #define MAX_WAIT_FOR_8021X_TX 50 /* msecs */
41 /* AMPDU rx reordering definitions */
42 #define BRCMF_RXREORDER_FLOWID_OFFSET 0
43 #define BRCMF_RXREORDER_MAXIDX_OFFSET 2
44 #define BRCMF_RXREORDER_FLAGS_OFFSET 4
45 #define BRCMF_RXREORDER_CURIDX_OFFSET 6
46 #define BRCMF_RXREORDER_EXPIDX_OFFSET 8
48 #define BRCMF_RXREORDER_DEL_FLOW 0x01
49 #define BRCMF_RXREORDER_FLUSH_ALL 0x02
50 #define BRCMF_RXREORDER_CURIDX_VALID 0x04
51 #define BRCMF_RXREORDER_EXPIDX_VALID 0x08
52 #define BRCMF_RXREORDER_NEW_HOLE 0x10
56 module_param_named(debug, brcmf_msg_level, int, S_IRUSR | S_IWUSR);
57 MODULE_PARM_DESC(debug, "level of debug output");
60 static int brcmf_p2p_enable;
62 module_param_named(p2pon, brcmf_p2p_enable, int, 0);
63 MODULE_PARM_DESC(p2pon, "enable p2p management functionality");
66 char *brcmf_ifname(struct brcmf_pub *drvr, int ifidx)
68 if (ifidx < 0 || ifidx >= BRCMF_MAX_IFS) {
69 brcmf_err("ifidx %d out of range\n", ifidx);
73 if (drvr->iflist[ifidx] == NULL) {
74 brcmf_err("null i/f %d\n", ifidx);
78 if (drvr->iflist[ifidx]->ndev)
79 return drvr->iflist[ifidx]->ndev->name;
84 static void _brcmf_set_multicast_list(struct work_struct *work)
87 struct net_device *ndev;
88 struct netdev_hw_addr *ha;
95 ifp = container_of(work, struct brcmf_if, multicast_work);
97 brcmf_dbg(TRACE, "Enter, idx=%d\n", ifp->bssidx);
101 /* Determine initial value of allmulti flag */
102 cmd_value = (ndev->flags & IFF_ALLMULTI) ? true : false;
104 /* Send down the multicast list first. */
105 cnt = netdev_mc_count(ndev);
106 buflen = sizeof(cnt) + (cnt * ETH_ALEN);
107 buf = kmalloc(buflen, GFP_ATOMIC);
112 cnt_le = cpu_to_le32(cnt);
113 memcpy(bufp, &cnt_le, sizeof(cnt_le));
114 bufp += sizeof(cnt_le);
116 netdev_for_each_mc_addr(ha, ndev) {
119 memcpy(bufp, ha->addr, ETH_ALEN);
124 err = brcmf_fil_iovar_data_set(ifp, "mcast_list", buf, buflen);
126 brcmf_err("Setting mcast_list failed, %d\n", err);
127 cmd_value = cnt ? true : cmd_value;
133 * Now send the allmulti setting. This is based on the setting in the
134 * net_device flags, but might be modified above to be turned on if we
135 * were trying to set some addresses and dongle rejected it...
137 err = brcmf_fil_iovar_int_set(ifp, "allmulti", cmd_value);
139 brcmf_err("Setting allmulti failed, %d\n", err);
141 /*Finally, pick up the PROMISC flag */
142 cmd_value = (ndev->flags & IFF_PROMISC) ? true : false;
143 err = brcmf_fil_cmd_int_set(ifp, BRCMF_C_SET_PROMISC, cmd_value);
145 brcmf_err("Setting BRCMF_C_SET_PROMISC failed, %d\n",
150 _brcmf_set_mac_address(struct work_struct *work)
152 struct brcmf_if *ifp;
155 ifp = container_of(work, struct brcmf_if, setmacaddr_work);
157 brcmf_dbg(TRACE, "Enter, idx=%d\n", ifp->bssidx);
159 err = brcmf_fil_iovar_data_set(ifp, "cur_etheraddr", ifp->mac_addr,
162 brcmf_err("Setting cur_etheraddr failed, %d\n", err);
164 brcmf_dbg(TRACE, "MAC address updated to %pM\n",
166 memcpy(ifp->ndev->dev_addr, ifp->mac_addr, ETH_ALEN);
170 static int brcmf_netdev_set_mac_address(struct net_device *ndev, void *addr)
172 struct brcmf_if *ifp = netdev_priv(ndev);
173 struct sockaddr *sa = (struct sockaddr *)addr;
175 memcpy(&ifp->mac_addr, sa->sa_data, ETH_ALEN);
176 schedule_work(&ifp->setmacaddr_work);
180 static void brcmf_netdev_set_multicast_list(struct net_device *ndev)
182 struct brcmf_if *ifp = netdev_priv(ndev);
184 schedule_work(&ifp->multicast_work);
187 static netdev_tx_t brcmf_netdev_start_xmit(struct sk_buff *skb,
188 struct net_device *ndev)
191 struct brcmf_if *ifp = netdev_priv(ndev);
192 struct brcmf_pub *drvr = ifp->drvr;
193 struct ethhdr *eh = (struct ethhdr *)(skb->data);
195 brcmf_dbg(DATA, "Enter, idx=%d\n", ifp->bssidx);
197 /* Can the device send data? */
198 if (drvr->bus_if->state != BRCMF_BUS_DATA) {
199 brcmf_err("xmit rejected state=%d\n", drvr->bus_if->state);
200 netif_stop_queue(ndev);
206 if (!drvr->iflist[ifp->bssidx]) {
207 brcmf_err("bad ifidx %d\n", ifp->bssidx);
208 netif_stop_queue(ndev);
214 /* Make sure there's enough room for any header */
215 if (skb_headroom(skb) < drvr->hdrlen) {
216 struct sk_buff *skb2;
218 brcmf_dbg(INFO, "%s: insufficient headroom\n",
219 brcmf_ifname(drvr, ifp->bssidx));
220 drvr->bus_if->tx_realloc++;
221 skb2 = skb_realloc_headroom(skb, drvr->hdrlen);
225 brcmf_err("%s: skb_realloc_headroom failed\n",
226 brcmf_ifname(drvr, ifp->bssidx));
232 /* validate length for ether packet */
233 if (skb->len < sizeof(*eh)) {
239 if (eh->h_proto == htons(ETH_P_PAE))
240 atomic_inc(&ifp->pend_8021x_cnt);
242 ret = brcmf_fws_process_skb(ifp, skb);
246 ifp->stats.tx_dropped++;
248 ifp->stats.tx_packets++;
249 ifp->stats.tx_bytes += skb->len;
252 /* Return ok: we always eat the packet */
256 void brcmf_txflowblock_if(struct brcmf_if *ifp,
257 enum brcmf_netif_stop_reason reason, bool state)
261 if (!ifp || !ifp->ndev)
264 brcmf_dbg(TRACE, "enter: idx=%d stop=0x%X reason=%d state=%d\n",
265 ifp->bssidx, ifp->netif_stop, reason, state);
267 spin_lock_irqsave(&ifp->netif_stop_lock, flags);
269 if (!ifp->netif_stop)
270 netif_stop_queue(ifp->ndev);
271 ifp->netif_stop |= reason;
273 ifp->netif_stop &= ~reason;
274 if (!ifp->netif_stop)
275 netif_wake_queue(ifp->ndev);
277 spin_unlock_irqrestore(&ifp->netif_stop_lock, flags);
280 void brcmf_txflowblock(struct device *dev, bool state)
282 struct brcmf_bus *bus_if = dev_get_drvdata(dev);
283 struct brcmf_pub *drvr = bus_if->drvr;
285 brcmf_dbg(TRACE, "Enter\n");
287 brcmf_fws_bus_blocked(drvr, state);
290 static void brcmf_netif_rx(struct brcmf_if *ifp, struct sk_buff *skb)
292 skb->dev = ifp->ndev;
293 skb->protocol = eth_type_trans(skb, skb->dev);
295 if (skb->pkt_type == PACKET_MULTICAST)
296 ifp->stats.multicast++;
298 /* Process special event packets */
299 brcmf_fweh_process_skb(ifp->drvr, skb);
301 if (!(ifp->ndev->flags & IFF_UP)) {
302 brcmu_pkt_buf_free_skb(skb);
306 ifp->stats.rx_bytes += skb->len;
307 ifp->stats.rx_packets++;
309 brcmf_dbg(DATA, "rx proto=0x%X\n", ntohs(skb->protocol));
313 /* If the receive is not processed inside an ISR,
314 * the softirqd must be woken explicitly to service
315 * the NET_RX_SOFTIRQ. This is handled by netif_rx_ni().
320 static void brcmf_rxreorder_get_skb_list(struct brcmf_ampdu_rx_reorder *rfi,
322 struct sk_buff_head *skb_list)
324 /* initialize return list */
325 __skb_queue_head_init(skb_list);
327 if (rfi->pend_pkts == 0) {
328 brcmf_dbg(INFO, "no packets in reorder queue\n");
333 if (rfi->pktslots[start]) {
334 __skb_queue_tail(skb_list, rfi->pktslots[start]);
335 rfi->pktslots[start] = NULL;
338 if (start > rfi->max_idx)
340 } while (start != end);
341 rfi->pend_pkts -= skb_queue_len(skb_list);
344 static void brcmf_rxreorder_process_info(struct brcmf_if *ifp, u8 *reorder_data,
347 u8 flow_id, max_idx, cur_idx, exp_idx, end_idx;
348 struct brcmf_ampdu_rx_reorder *rfi;
349 struct sk_buff_head reorder_list;
350 struct sk_buff *pnext;
354 flow_id = reorder_data[BRCMF_RXREORDER_FLOWID_OFFSET];
355 flags = reorder_data[BRCMF_RXREORDER_FLAGS_OFFSET];
357 /* validate flags and flow id */
359 brcmf_err("invalid flags...so ignore this packet\n");
360 brcmf_netif_rx(ifp, pkt);
364 rfi = ifp->drvr->reorder_flows[flow_id];
365 if (flags & BRCMF_RXREORDER_DEL_FLOW) {
366 brcmf_dbg(INFO, "flow-%d: delete\n",
370 brcmf_dbg(INFO, "received flags to cleanup, but no flow (%d) yet\n",
372 brcmf_netif_rx(ifp, pkt);
376 brcmf_rxreorder_get_skb_list(rfi, rfi->exp_idx, rfi->exp_idx,
378 /* add the last packet */
379 __skb_queue_tail(&reorder_list, pkt);
381 ifp->drvr->reorder_flows[flow_id] = NULL;
384 /* from here on we need a flow reorder instance */
386 buf_size = sizeof(*rfi);
387 max_idx = reorder_data[BRCMF_RXREORDER_MAXIDX_OFFSET];
389 buf_size += (max_idx + 1) * sizeof(pkt);
391 /* allocate space for flow reorder info */
392 brcmf_dbg(INFO, "flow-%d: start, maxidx %d\n",
394 rfi = kzalloc(buf_size, GFP_ATOMIC);
396 brcmf_err("failed to alloc buffer\n");
397 brcmf_netif_rx(ifp, pkt);
401 ifp->drvr->reorder_flows[flow_id] = rfi;
402 rfi->pktslots = (struct sk_buff **)(rfi+1);
403 rfi->max_idx = max_idx;
405 if (flags & BRCMF_RXREORDER_NEW_HOLE) {
406 if (rfi->pend_pkts) {
407 brcmf_rxreorder_get_skb_list(rfi, rfi->exp_idx,
410 WARN_ON(rfi->pend_pkts);
412 __skb_queue_head_init(&reorder_list);
414 rfi->cur_idx = reorder_data[BRCMF_RXREORDER_CURIDX_OFFSET];
415 rfi->exp_idx = reorder_data[BRCMF_RXREORDER_EXPIDX_OFFSET];
416 rfi->max_idx = reorder_data[BRCMF_RXREORDER_MAXIDX_OFFSET];
417 rfi->pktslots[rfi->cur_idx] = pkt;
419 brcmf_dbg(DATA, "flow-%d: new hole %d (%d), pending %d\n",
420 flow_id, rfi->cur_idx, rfi->exp_idx, rfi->pend_pkts);
421 } else if (flags & BRCMF_RXREORDER_CURIDX_VALID) {
422 cur_idx = reorder_data[BRCMF_RXREORDER_CURIDX_OFFSET];
423 exp_idx = reorder_data[BRCMF_RXREORDER_EXPIDX_OFFSET];
425 if ((exp_idx == rfi->exp_idx) && (cur_idx != rfi->exp_idx)) {
426 /* still in the current hole */
427 /* enqueue the current on the buffer chain */
428 if (rfi->pktslots[cur_idx] != NULL) {
429 brcmf_dbg(INFO, "HOLE: ERROR buffer pending..free it\n");
430 brcmu_pkt_buf_free_skb(rfi->pktslots[cur_idx]);
431 rfi->pktslots[cur_idx] = NULL;
433 rfi->pktslots[cur_idx] = pkt;
435 rfi->cur_idx = cur_idx;
436 brcmf_dbg(DATA, "flow-%d: store pkt %d (%d), pending %d\n",
437 flow_id, cur_idx, exp_idx, rfi->pend_pkts);
439 /* can return now as there is no reorder
444 if (rfi->exp_idx == cur_idx) {
445 if (rfi->pktslots[cur_idx] != NULL) {
446 brcmf_dbg(INFO, "error buffer pending..free it\n");
447 brcmu_pkt_buf_free_skb(rfi->pktslots[cur_idx]);
448 rfi->pktslots[cur_idx] = NULL;
450 rfi->pktslots[cur_idx] = pkt;
453 /* got the expected one. flush from current to expected
454 * and update expected
456 brcmf_dbg(DATA, "flow-%d: expected %d (%d), pending %d\n",
457 flow_id, cur_idx, exp_idx, rfi->pend_pkts);
459 rfi->cur_idx = cur_idx;
460 rfi->exp_idx = exp_idx;
462 brcmf_rxreorder_get_skb_list(rfi, cur_idx, exp_idx,
464 brcmf_dbg(DATA, "flow-%d: freeing buffers %d, pending %d\n",
465 flow_id, skb_queue_len(&reorder_list),
470 brcmf_dbg(DATA, "flow-%d (0x%x): both moved, old %d/%d, new %d/%d\n",
471 flow_id, flags, rfi->cur_idx, rfi->exp_idx,
473 if (flags & BRCMF_RXREORDER_FLUSH_ALL)
474 end_idx = rfi->exp_idx;
478 /* flush pkts first */
479 brcmf_rxreorder_get_skb_list(rfi, rfi->exp_idx, end_idx,
482 if (exp_idx == ((cur_idx + 1) % (rfi->max_idx + 1))) {
483 __skb_queue_tail(&reorder_list, pkt);
485 rfi->pktslots[cur_idx] = pkt;
488 rfi->exp_idx = exp_idx;
489 rfi->cur_idx = cur_idx;
492 /* explicity window move updating the expected index */
493 exp_idx = reorder_data[BRCMF_RXREORDER_EXPIDX_OFFSET];
495 brcmf_dbg(DATA, "flow-%d (0x%x): change expected: %d -> %d\n",
496 flow_id, flags, rfi->exp_idx, exp_idx);
497 if (flags & BRCMF_RXREORDER_FLUSH_ALL)
498 end_idx = rfi->exp_idx;
502 brcmf_rxreorder_get_skb_list(rfi, rfi->exp_idx, end_idx,
504 __skb_queue_tail(&reorder_list, pkt);
505 /* set the new expected idx */
506 rfi->exp_idx = exp_idx;
509 skb_queue_walk_safe(&reorder_list, pkt, pnext) {
510 __skb_unlink(pkt, &reorder_list);
511 brcmf_netif_rx(ifp, pkt);
515 void brcmf_rx_frame(struct device *dev, struct sk_buff *skb)
517 struct brcmf_if *ifp;
518 struct brcmf_bus *bus_if = dev_get_drvdata(dev);
519 struct brcmf_pub *drvr = bus_if->drvr;
520 struct brcmf_skb_reorder_data *rd;
524 brcmf_dbg(DATA, "Enter: %s: rxp=%p\n", dev_name(dev), skb);
526 /* process and remove protocol-specific header */
527 ret = brcmf_proto_hdrpull(drvr, true, &ifidx, skb);
528 ifp = drvr->iflist[ifidx];
530 if (ret || !ifp || !ifp->ndev) {
531 if ((ret != -ENODATA) && ifp)
532 ifp->stats.rx_errors++;
533 brcmu_pkt_buf_free_skb(skb);
537 rd = (struct brcmf_skb_reorder_data *)skb->cb;
539 brcmf_rxreorder_process_info(ifp, rd->reorder, skb);
541 brcmf_netif_rx(ifp, skb);
544 void brcmf_txfinalize(struct brcmf_pub *drvr, struct sk_buff *txp, u8 ifidx,
547 struct brcmf_if *ifp;
551 ifp = drvr->iflist[ifidx];
555 eh = (struct ethhdr *)(txp->data);
556 type = ntohs(eh->h_proto);
558 if (type == ETH_P_PAE) {
559 atomic_dec(&ifp->pend_8021x_cnt);
560 if (waitqueue_active(&ifp->pend_8021x_wait))
561 wake_up(&ifp->pend_8021x_wait);
565 ifp->stats.tx_errors++;
567 brcmu_pkt_buf_free_skb(txp);
570 void brcmf_txcomplete(struct device *dev, struct sk_buff *txp, bool success)
572 struct brcmf_bus *bus_if = dev_get_drvdata(dev);
573 struct brcmf_pub *drvr = bus_if->drvr;
576 /* await txstatus signal for firmware if active */
577 if (brcmf_fws_fc_active(drvr->fws)) {
579 brcmf_fws_bustxfail(drvr->fws, txp);
581 if (brcmf_proto_hdrpull(drvr, false, &ifidx, txp))
582 brcmu_pkt_buf_free_skb(txp);
584 brcmf_txfinalize(drvr, txp, ifidx, success);
588 static struct net_device_stats *brcmf_netdev_get_stats(struct net_device *ndev)
590 struct brcmf_if *ifp = netdev_priv(ndev);
592 brcmf_dbg(TRACE, "Enter, idx=%d\n", ifp->bssidx);
597 static void brcmf_ethtool_get_drvinfo(struct net_device *ndev,
598 struct ethtool_drvinfo *info)
600 struct brcmf_if *ifp = netdev_priv(ndev);
601 struct brcmf_pub *drvr = ifp->drvr;
603 strlcpy(info->driver, KBUILD_MODNAME, sizeof(info->driver));
604 snprintf(info->version, sizeof(info->version), "n/a");
605 strlcpy(info->fw_version, drvr->fwver, sizeof(info->fw_version));
606 strlcpy(info->bus_info, dev_name(drvr->bus_if->dev),
607 sizeof(info->bus_info));
610 static const struct ethtool_ops brcmf_ethtool_ops = {
611 .get_drvinfo = brcmf_ethtool_get_drvinfo,
614 static int brcmf_netdev_stop(struct net_device *ndev)
616 struct brcmf_if *ifp = netdev_priv(ndev);
618 brcmf_dbg(TRACE, "Enter, idx=%d\n", ifp->bssidx);
620 brcmf_cfg80211_down(ndev);
622 /* Set state and stop OS transmissions */
623 netif_stop_queue(ndev);
628 static int brcmf_netdev_open(struct net_device *ndev)
630 struct brcmf_if *ifp = netdev_priv(ndev);
631 struct brcmf_pub *drvr = ifp->drvr;
632 struct brcmf_bus *bus_if = drvr->bus_if;
635 brcmf_dbg(TRACE, "Enter, idx=%d\n", ifp->bssidx);
637 /* If bus is not ready, can't continue */
638 if (bus_if->state != BRCMF_BUS_DATA) {
639 brcmf_err("failed bus is not ready\n");
643 atomic_set(&ifp->pend_8021x_cnt, 0);
645 /* Get current TOE mode from dongle */
646 if (brcmf_fil_iovar_int_get(ifp, "toe_ol", &toe_ol) >= 0
647 && (toe_ol & TOE_TX_CSUM_OL) != 0)
648 ndev->features |= NETIF_F_IP_CSUM;
650 ndev->features &= ~NETIF_F_IP_CSUM;
652 if (brcmf_cfg80211_up(ndev)) {
653 brcmf_err("failed to bring up cfg80211\n");
657 /* Allow transmit calls */
658 netif_start_queue(ndev);
662 static const struct net_device_ops brcmf_netdev_ops_pri = {
663 .ndo_open = brcmf_netdev_open,
664 .ndo_stop = brcmf_netdev_stop,
665 .ndo_get_stats = brcmf_netdev_get_stats,
666 .ndo_start_xmit = brcmf_netdev_start_xmit,
667 .ndo_set_mac_address = brcmf_netdev_set_mac_address,
668 .ndo_set_rx_mode = brcmf_netdev_set_multicast_list
671 int brcmf_net_attach(struct brcmf_if *ifp, bool rtnl_locked)
673 struct brcmf_pub *drvr = ifp->drvr;
674 struct net_device *ndev;
677 brcmf_dbg(TRACE, "Enter, idx=%d mac=%pM\n", ifp->bssidx,
681 /* set appropriate operations */
682 ndev->netdev_ops = &brcmf_netdev_ops_pri;
684 ndev->hard_header_len += drvr->hdrlen;
685 ndev->ethtool_ops = &brcmf_ethtool_ops;
687 drvr->rxsz = ndev->mtu + ndev->hard_header_len +
690 /* set the mac address */
691 memcpy(ndev->dev_addr, ifp->mac_addr, ETH_ALEN);
693 INIT_WORK(&ifp->setmacaddr_work, _brcmf_set_mac_address);
694 INIT_WORK(&ifp->multicast_work, _brcmf_set_multicast_list);
697 err = register_netdevice(ndev);
699 err = register_netdev(ndev);
701 brcmf_err("couldn't register the net device\n");
705 brcmf_dbg(INFO, "%s: Broadcom Dongle Host Driver\n", ndev->name);
707 ndev->destructor = brcmf_cfg80211_free_netdev;
711 drvr->iflist[ifp->bssidx] = NULL;
712 ndev->netdev_ops = NULL;
717 static int brcmf_net_p2p_open(struct net_device *ndev)
719 brcmf_dbg(TRACE, "Enter\n");
721 return brcmf_cfg80211_up(ndev);
724 static int brcmf_net_p2p_stop(struct net_device *ndev)
726 brcmf_dbg(TRACE, "Enter\n");
728 return brcmf_cfg80211_down(ndev);
731 static netdev_tx_t brcmf_net_p2p_start_xmit(struct sk_buff *skb,
732 struct net_device *ndev)
735 dev_kfree_skb_any(skb);
740 static const struct net_device_ops brcmf_netdev_ops_p2p = {
741 .ndo_open = brcmf_net_p2p_open,
742 .ndo_stop = brcmf_net_p2p_stop,
743 .ndo_start_xmit = brcmf_net_p2p_start_xmit
746 static int brcmf_net_p2p_attach(struct brcmf_if *ifp)
748 struct net_device *ndev;
750 brcmf_dbg(TRACE, "Enter, idx=%d mac=%pM\n", ifp->bssidx,
754 ndev->netdev_ops = &brcmf_netdev_ops_p2p;
756 /* set the mac address */
757 memcpy(ndev->dev_addr, ifp->mac_addr, ETH_ALEN);
759 if (register_netdev(ndev) != 0) {
760 brcmf_err("couldn't register the p2p net device\n");
764 brcmf_dbg(INFO, "%s: Broadcom Dongle Host Driver\n", ndev->name);
769 ifp->drvr->iflist[ifp->bssidx] = NULL;
770 ndev->netdev_ops = NULL;
775 struct brcmf_if *brcmf_add_if(struct brcmf_pub *drvr, s32 bssidx, s32 ifidx,
776 char *name, u8 *mac_addr)
778 struct brcmf_if *ifp;
779 struct net_device *ndev;
781 brcmf_dbg(TRACE, "Enter, idx=%d, ifidx=%d\n", bssidx, ifidx);
783 ifp = drvr->iflist[bssidx];
785 * Delete the existing interface before overwriting it
786 * in case we missed the BRCMF_E_IF_DEL event.
789 brcmf_err("ERROR: netdev:%s already exists\n",
792 netif_stop_queue(ifp->ndev);
793 unregister_netdev(ifp->ndev);
794 free_netdev(ifp->ndev);
795 drvr->iflist[bssidx] = NULL;
797 brcmf_err("ignore IF event\n");
798 return ERR_PTR(-EINVAL);
802 if (!brcmf_p2p_enable && bssidx == 1) {
803 /* this is P2P_DEVICE interface */
804 brcmf_dbg(INFO, "allocate non-netdev interface\n");
805 ifp = kzalloc(sizeof(*ifp), GFP_KERNEL);
807 return ERR_PTR(-ENOMEM);
809 brcmf_dbg(INFO, "allocate netdev interface\n");
810 /* Allocate netdev, including space for private structure */
811 ndev = alloc_netdev(sizeof(*ifp), name, ether_setup);
813 return ERR_PTR(-ENOMEM);
815 ifp = netdev_priv(ndev);
820 drvr->iflist[bssidx] = ifp;
822 ifp->bssidx = bssidx;
824 init_waitqueue_head(&ifp->pend_8021x_wait);
825 spin_lock_init(&ifp->netif_stop_lock);
827 if (mac_addr != NULL)
828 memcpy(ifp->mac_addr, mac_addr, ETH_ALEN);
830 brcmf_dbg(TRACE, " ==== pid:%x, if:%s (%pM) created ===\n",
831 current->pid, name, ifp->mac_addr);
836 void brcmf_del_if(struct brcmf_pub *drvr, s32 bssidx)
838 struct brcmf_if *ifp;
840 ifp = drvr->iflist[bssidx];
841 drvr->iflist[bssidx] = NULL;
843 brcmf_err("Null interface, idx=%d\n", bssidx);
846 brcmf_dbg(TRACE, "Enter, idx=%d, ifidx=%d\n", bssidx, ifp->ifidx);
849 if (ifp->ndev->netdev_ops == &brcmf_netdev_ops_pri) {
851 brcmf_netdev_stop(ifp->ndev);
855 netif_stop_queue(ifp->ndev);
858 if (ifp->ndev->netdev_ops == &brcmf_netdev_ops_pri) {
859 cancel_work_sync(&ifp->setmacaddr_work);
860 cancel_work_sync(&ifp->multicast_work);
862 /* unregister will take care of freeing it */
863 unregister_netdev(ifp->ndev);
869 int brcmf_attach(struct device *dev)
871 struct brcmf_pub *drvr = NULL;
874 brcmf_dbg(TRACE, "Enter\n");
876 /* Allocate primary brcmf_info */
877 drvr = kzalloc(sizeof(struct brcmf_pub), GFP_ATOMIC);
881 mutex_init(&drvr->proto_block);
883 /* Link to bus module */
885 drvr->bus_if = dev_get_drvdata(dev);
886 drvr->bus_if->drvr = drvr;
888 /* create device debugfs folder */
889 brcmf_debugfs_attach(drvr);
891 /* Attach and link in the protocol */
892 ret = brcmf_proto_attach(drvr);
894 brcmf_err("brcmf_prot_attach failed\n");
898 /* attach firmware event handler */
899 brcmf_fweh_attach(drvr);
909 int brcmf_bus_start(struct device *dev)
912 struct brcmf_bus *bus_if = dev_get_drvdata(dev);
913 struct brcmf_pub *drvr = bus_if->drvr;
914 struct brcmf_if *ifp;
915 struct brcmf_if *p2p_ifp;
917 brcmf_dbg(TRACE, "\n");
919 /* add primary networking interface */
920 ifp = brcmf_add_if(drvr, 0, 0, "wlan%d", NULL);
924 if (brcmf_p2p_enable)
925 p2p_ifp = brcmf_add_if(drvr, 1, 0, "p2p%d", NULL);
931 /* signal bus ready */
932 brcmf_bus_change_state(bus_if, BRCMF_BUS_DATA);
934 /* Bus is ready, do any initialization */
935 ret = brcmf_c_preinit_dcmds(ifp);
939 ret = brcmf_fws_init(drvr);
943 brcmf_fws_add_interface(ifp);
945 drvr->config = brcmf_cfg80211_attach(drvr, bus_if->dev);
946 if (drvr->config == NULL) {
951 ret = brcmf_fweh_activate_events(ifp);
955 ret = brcmf_net_attach(ifp, false);
958 brcmf_err("failed: %d\n", ret);
959 brcmf_cfg80211_detach(drvr->config);
961 brcmf_fws_del_interface(ifp);
962 brcmf_fws_deinit(drvr);
964 if (drvr->iflist[0]) {
965 free_netdev(ifp->ndev);
966 drvr->iflist[0] = NULL;
969 free_netdev(p2p_ifp->ndev);
970 drvr->iflist[1] = NULL;
974 if ((brcmf_p2p_enable) && (p2p_ifp))
975 if (brcmf_net_p2p_attach(p2p_ifp) < 0)
976 brcmf_p2p_enable = 0;
981 void brcmf_bus_add_txhdrlen(struct device *dev, uint len)
983 struct brcmf_bus *bus_if = dev_get_drvdata(dev);
984 struct brcmf_pub *drvr = bus_if->drvr;
991 static void brcmf_bus_detach(struct brcmf_pub *drvr)
993 brcmf_dbg(TRACE, "Enter\n");
996 /* Stop the bus module */
997 brcmf_bus_stop(drvr->bus_if);
1001 void brcmf_dev_reset(struct device *dev)
1003 struct brcmf_bus *bus_if = dev_get_drvdata(dev);
1004 struct brcmf_pub *drvr = bus_if->drvr;
1009 if (drvr->iflist[0])
1010 brcmf_fil_cmd_int_set(drvr->iflist[0], BRCMF_C_TERMINATED, 1);
1013 void brcmf_detach(struct device *dev)
1016 struct brcmf_bus *bus_if = dev_get_drvdata(dev);
1017 struct brcmf_pub *drvr = bus_if->drvr;
1019 brcmf_dbg(TRACE, "Enter\n");
1024 /* stop firmware event handling */
1025 brcmf_fweh_detach(drvr);
1027 brcmf_bus_change_state(bus_if, BRCMF_BUS_DOWN);
1029 /* make sure primary interface removed last */
1030 for (i = BRCMF_MAX_IFS-1; i > -1; i--)
1031 if (drvr->iflist[i]) {
1032 brcmf_fws_del_interface(drvr->iflist[i]);
1033 brcmf_del_if(drvr, i);
1036 brcmf_cfg80211_detach(drvr->config);
1038 brcmf_fws_deinit(drvr);
1040 brcmf_bus_detach(drvr);
1042 brcmf_proto_detach(drvr);
1044 brcmf_debugfs_detach(drvr);
1045 bus_if->drvr = NULL;
1049 s32 brcmf_iovar_data_set(struct device *dev, char *name, void *data, u32 len)
1051 struct brcmf_bus *bus_if = dev_get_drvdata(dev);
1052 struct brcmf_if *ifp = bus_if->drvr->iflist[0];
1054 return brcmf_fil_iovar_data_set(ifp, name, data, len);
1057 static int brcmf_get_pend_8021x_cnt(struct brcmf_if *ifp)
1059 return atomic_read(&ifp->pend_8021x_cnt);
1062 int brcmf_netdev_wait_pend8021x(struct net_device *ndev)
1064 struct brcmf_if *ifp = netdev_priv(ndev);
1067 err = wait_event_timeout(ifp->pend_8021x_wait,
1068 !brcmf_get_pend_8021x_cnt(ifp),
1069 msecs_to_jiffies(MAX_WAIT_FOR_8021X_TX));
1077 * return chip id and rev of the device encoded in u32.
1079 u32 brcmf_get_chip_info(struct brcmf_if *ifp)
1081 struct brcmf_bus *bus = ifp->drvr->bus_if;
1083 return bus->chip << 4 | bus->chiprev;
1086 static void brcmf_driver_register(struct work_struct *work)
1088 #ifdef CONFIG_BRCMFMAC_SDIO
1089 brcmf_sdio_register();
1091 #ifdef CONFIG_BRCMFMAC_USB
1092 brcmf_usb_register();
1095 static DECLARE_WORK(brcmf_driver_work, brcmf_driver_register);
1097 static int __init brcmfmac_module_init(void)
1099 brcmf_debugfs_init();
1100 #ifdef CONFIG_BRCMFMAC_SDIO
1103 if (!schedule_work(&brcmf_driver_work))
1109 static void __exit brcmfmac_module_exit(void)
1111 cancel_work_sync(&brcmf_driver_work);
1113 #ifdef CONFIG_BRCMFMAC_SDIO
1116 #ifdef CONFIG_BRCMFMAC_USB
1119 brcmf_debugfs_exit();
1122 module_init(brcmfmac_module_init);
1123 module_exit(brcmfmac_module_exit);