| /* Copyright (c) 2014 Broadcom Corporation |
| * |
| * Permission to use, copy, modify, and/or distribute this software for any |
| * purpose with or without fee is hereby granted, provided that the above |
| * copyright notice and this permission notice appear in all copies. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES |
| * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF |
| * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY |
| * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES |
| * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION |
| * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN |
| * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. |
| */ |
| |
| |
| #include <linux/types.h> |
| #include <linux/netdevice.h> |
| #include <linux/etherdevice.h> |
| #include <brcmu_utils.h> |
| |
| #include "core.h" |
| #include "debug.h" |
| #include "bus.h" |
| #include "proto.h" |
| #include "flowring.h" |
| #include "msgbuf.h" |
| #include "common.h" |
| |
| |
| #define BRCMF_FLOWRING_HIGH 1024 |
| #define BRCMF_FLOWRING_LOW (BRCMF_FLOWRING_HIGH - 256) |
| #define BRCMF_FLOWRING_INVALID_IFIDX 0xff |
| |
| #define BRCMF_FLOWRING_HASH_AP(da, fifo, ifidx) (da[5] + fifo + ifidx * 16) |
| #define BRCMF_FLOWRING_HASH_STA(fifo, ifidx) (fifo + ifidx * 16) |
| |
| static const u8 brcmf_flowring_prio2fifo[] = { |
| 1, |
| 0, |
| 0, |
| 1, |
| 2, |
| 2, |
| 3, |
| 3 |
| }; |
| |
| |
| static bool |
| brcmf_flowring_is_tdls_mac(struct brcmf_flowring *flow, u8 mac[ETH_ALEN]) |
| { |
| struct brcmf_flowring_tdls_entry *search; |
| |
| search = flow->tdls_entry; |
| |
| while (search) { |
| if (memcmp(search->mac, mac, ETH_ALEN) == 0) |
| return true; |
| search = search->next; |
| } |
| |
| return false; |
| } |
| |
| |
| u32 brcmf_flowring_lookup(struct brcmf_flowring *flow, u8 da[ETH_ALEN], |
| u8 prio, u8 ifidx) |
| { |
| struct brcmf_flowring_hash *hash; |
| u8 hash_idx; |
| u32 i; |
| bool found; |
| bool sta; |
| u8 fifo; |
| u8 *mac; |
| |
| fifo = brcmf_flowring_prio2fifo[prio]; |
| sta = (flow->addr_mode[ifidx] == ADDR_INDIRECT); |
| mac = da; |
| if ((!sta) && (is_multicast_ether_addr(da))) { |
| mac = (u8 *)ALLFFMAC; |
| fifo = 0; |
| } |
| if ((sta) && (flow->tdls_active) && |
| (brcmf_flowring_is_tdls_mac(flow, da))) { |
| sta = false; |
| } |
| hash_idx = sta ? BRCMF_FLOWRING_HASH_STA(fifo, ifidx) : |
| BRCMF_FLOWRING_HASH_AP(mac, fifo, ifidx); |
| found = false; |
| hash = flow->hash; |
| for (i = 0; i < BRCMF_FLOWRING_HASHSIZE; i++) { |
| if ((sta || (memcmp(hash[hash_idx].mac, mac, ETH_ALEN) == 0)) && |
| (hash[hash_idx].fifo == fifo) && |
| (hash[hash_idx].ifidx == ifidx)) { |
| found = true; |
| break; |
| } |
| hash_idx++; |
| } |
| if (found) |
| return hash[hash_idx].flowid; |
| |
| return BRCMF_FLOWRING_INVALID_ID; |
| } |
| |
| |
| u32 brcmf_flowring_create(struct brcmf_flowring *flow, u8 da[ETH_ALEN], |
| u8 prio, u8 ifidx) |
| { |
| struct brcmf_flowring_ring *ring; |
| struct brcmf_flowring_hash *hash; |
| u8 hash_idx; |
| u32 i; |
| bool found; |
| u8 fifo; |
| bool sta; |
| u8 *mac; |
| |
| fifo = brcmf_flowring_prio2fifo[prio]; |
| sta = (flow->addr_mode[ifidx] == ADDR_INDIRECT); |
| mac = da; |
| if ((!sta) && (is_multicast_ether_addr(da))) { |
| mac = (u8 *)ALLFFMAC; |
| fifo = 0; |
| } |
| if ((sta) && (flow->tdls_active) && |
| (brcmf_flowring_is_tdls_mac(flow, da))) { |
| sta = false; |
| } |
| hash_idx = sta ? BRCMF_FLOWRING_HASH_STA(fifo, ifidx) : |
| BRCMF_FLOWRING_HASH_AP(mac, fifo, ifidx); |
| found = false; |
| hash = flow->hash; |
| for (i = 0; i < BRCMF_FLOWRING_HASHSIZE; i++) { |
| if ((hash[hash_idx].ifidx == BRCMF_FLOWRING_INVALID_IFIDX) && |
| (is_zero_ether_addr(hash[hash_idx].mac))) { |
| found = true; |
| break; |
| } |
| hash_idx++; |
| } |
| if (found) { |
| for (i = 0; i < flow->nrofrings; i++) { |
| if (flow->rings[i] == NULL) |
| break; |
| } |
| if (i == flow->nrofrings) |
| return -ENOMEM; |
| |
| ring = kzalloc(sizeof(*ring), GFP_ATOMIC); |
| if (!ring) |
| return -ENOMEM; |
| |
| memcpy(hash[hash_idx].mac, mac, ETH_ALEN); |
| hash[hash_idx].fifo = fifo; |
| hash[hash_idx].ifidx = ifidx; |
| hash[hash_idx].flowid = i; |
| |
| ring->hash_id = hash_idx; |
| ring->status = RING_CLOSED; |
| skb_queue_head_init(&ring->skblist); |
| flow->rings[i] = ring; |
| |
| return i; |
| } |
| return BRCMF_FLOWRING_INVALID_ID; |
| } |
| |
| |
| u8 brcmf_flowring_tid(struct brcmf_flowring *flow, u8 flowid) |
| { |
| struct brcmf_flowring_ring *ring; |
| |
| ring = flow->rings[flowid]; |
| |
| return flow->hash[ring->hash_id].fifo; |
| } |
| |
| |
| static void brcmf_flowring_block(struct brcmf_flowring *flow, u8 flowid, |
| bool blocked) |
| { |
| struct brcmf_flowring_ring *ring; |
| struct brcmf_bus *bus_if; |
| struct brcmf_pub *drvr; |
| struct brcmf_if *ifp; |
| bool currently_blocked; |
| int i; |
| u8 ifidx; |
| unsigned long flags; |
| |
| spin_lock_irqsave(&flow->block_lock, flags); |
| |
| ring = flow->rings[flowid]; |
| ifidx = brcmf_flowring_ifidx_get(flow, flowid); |
| |
| currently_blocked = false; |
| for (i = 0; i < flow->nrofrings; i++) { |
| if (flow->rings[i]) { |
| ring = flow->rings[i]; |
| if ((ring->status == RING_OPEN) && |
| (brcmf_flowring_ifidx_get(flow, i) == ifidx)) { |
| if (ring->blocked) { |
| currently_blocked = true; |
| break; |
| } |
| } |
| } |
| } |
| ring->blocked = blocked; |
| if (currently_blocked == blocked) { |
| spin_unlock_irqrestore(&flow->block_lock, flags); |
| return; |
| } |
| |
| bus_if = dev_get_drvdata(flow->dev); |
| drvr = bus_if->drvr; |
| ifp = drvr->iflist[ifidx]; |
| brcmf_txflowblock_if(ifp, BRCMF_NETIF_STOP_REASON_FLOW, blocked); |
| |
| spin_unlock_irqrestore(&flow->block_lock, flags); |
| } |
| |
| |
| void brcmf_flowring_delete(struct brcmf_flowring *flow, u8 flowid) |
| { |
| struct brcmf_flowring_ring *ring; |
| u8 hash_idx; |
| struct sk_buff *skb; |
| |
| ring = flow->rings[flowid]; |
| if (!ring) |
| return; |
| brcmf_flowring_block(flow, flowid, false); |
| hash_idx = ring->hash_id; |
| flow->hash[hash_idx].ifidx = BRCMF_FLOWRING_INVALID_IFIDX; |
| eth_zero_addr(flow->hash[hash_idx].mac); |
| flow->rings[flowid] = NULL; |
| |
| skb = skb_dequeue(&ring->skblist); |
| while (skb) { |
| brcmu_pkt_buf_free_skb(skb); |
| skb = skb_dequeue(&ring->skblist); |
| } |
| |
| kfree(ring); |
| } |
| |
| |
| u32 brcmf_flowring_enqueue(struct brcmf_flowring *flow, u8 flowid, |
| struct sk_buff *skb) |
| { |
| struct brcmf_flowring_ring *ring; |
| |
| ring = flow->rings[flowid]; |
| |
| skb_queue_tail(&ring->skblist, skb); |
| |
| if (!ring->blocked && |
| (skb_queue_len(&ring->skblist) > BRCMF_FLOWRING_HIGH)) { |
| brcmf_flowring_block(flow, flowid, true); |
| brcmf_dbg(MSGBUF, "Flowcontrol: BLOCK for ring %d\n", flowid); |
| /* To prevent (work around) possible race condition, check |
| * queue len again. It is also possible to use locking to |
| * protect, but that is undesirable for every enqueue and |
| * dequeue. This simple check will solve a possible race |
| * condition if it occurs. |
| */ |
| if (skb_queue_len(&ring->skblist) < BRCMF_FLOWRING_LOW) |
| brcmf_flowring_block(flow, flowid, false); |
| } |
| return skb_queue_len(&ring->skblist); |
| } |
| |
| |
| struct sk_buff *brcmf_flowring_dequeue(struct brcmf_flowring *flow, u8 flowid) |
| { |
| struct brcmf_flowring_ring *ring; |
| struct sk_buff *skb; |
| |
| ring = flow->rings[flowid]; |
| if (ring->status != RING_OPEN) |
| return NULL; |
| |
| skb = skb_dequeue(&ring->skblist); |
| |
| if (ring->blocked && |
| (skb_queue_len(&ring->skblist) < BRCMF_FLOWRING_LOW)) { |
| brcmf_flowring_block(flow, flowid, false); |
| brcmf_dbg(MSGBUF, "Flowcontrol: OPEN for ring %d\n", flowid); |
| } |
| |
| return skb; |
| } |
| |
| |
| void brcmf_flowring_reinsert(struct brcmf_flowring *flow, u8 flowid, |
| struct sk_buff *skb) |
| { |
| struct brcmf_flowring_ring *ring; |
| |
| ring = flow->rings[flowid]; |
| |
| skb_queue_head(&ring->skblist, skb); |
| } |
| |
| |
| u32 brcmf_flowring_qlen(struct brcmf_flowring *flow, u8 flowid) |
| { |
| struct brcmf_flowring_ring *ring; |
| |
| ring = flow->rings[flowid]; |
| if (!ring) |
| return 0; |
| |
| if (ring->status != RING_OPEN) |
| return 0; |
| |
| return skb_queue_len(&ring->skblist); |
| } |
| |
| |
| void brcmf_flowring_open(struct brcmf_flowring *flow, u8 flowid) |
| { |
| struct brcmf_flowring_ring *ring; |
| |
| ring = flow->rings[flowid]; |
| if (!ring) { |
| brcmf_err("Ring NULL, for flowid %d\n", flowid); |
| return; |
| } |
| |
| ring->status = RING_OPEN; |
| } |
| |
| |
| u8 brcmf_flowring_ifidx_get(struct brcmf_flowring *flow, u8 flowid) |
| { |
| struct brcmf_flowring_ring *ring; |
| u8 hash_idx; |
| |
| ring = flow->rings[flowid]; |
| hash_idx = ring->hash_id; |
| |
| return flow->hash[hash_idx].ifidx; |
| } |
| |
| |
| struct brcmf_flowring *brcmf_flowring_attach(struct device *dev, u16 nrofrings) |
| { |
| struct brcmf_flowring *flow; |
| u32 i; |
| |
| flow = kzalloc(sizeof(*flow), GFP_KERNEL); |
| if (flow) { |
| flow->dev = dev; |
| flow->nrofrings = nrofrings; |
| spin_lock_init(&flow->block_lock); |
| for (i = 0; i < ARRAY_SIZE(flow->addr_mode); i++) |
| flow->addr_mode[i] = ADDR_INDIRECT; |
| for (i = 0; i < ARRAY_SIZE(flow->hash); i++) |
| flow->hash[i].ifidx = BRCMF_FLOWRING_INVALID_IFIDX; |
| flow->rings = kcalloc(nrofrings, sizeof(*flow->rings), |
| GFP_KERNEL); |
| if (!flow->rings) { |
| kfree(flow); |
| flow = NULL; |
| } |
| } |
| |
| return flow; |
| } |
| |
| |
| void brcmf_flowring_detach(struct brcmf_flowring *flow) |
| { |
| struct brcmf_bus *bus_if = dev_get_drvdata(flow->dev); |
| struct brcmf_pub *drvr = bus_if->drvr; |
| struct brcmf_flowring_tdls_entry *search; |
| struct brcmf_flowring_tdls_entry *remove; |
| u8 flowid; |
| |
| for (flowid = 0; flowid < flow->nrofrings; flowid++) { |
| if (flow->rings[flowid]) |
| brcmf_msgbuf_delete_flowring(drvr, flowid); |
| } |
| |
| search = flow->tdls_entry; |
| while (search) { |
| remove = search; |
| search = search->next; |
| kfree(remove); |
| } |
| kfree(flow->rings); |
| kfree(flow); |
| } |
| |
| |
| void brcmf_flowring_configure_addr_mode(struct brcmf_flowring *flow, int ifidx, |
| enum proto_addr_mode addr_mode) |
| { |
| struct brcmf_bus *bus_if = dev_get_drvdata(flow->dev); |
| struct brcmf_pub *drvr = bus_if->drvr; |
| u32 i; |
| u8 flowid; |
| |
| if (flow->addr_mode[ifidx] != addr_mode) { |
| for (i = 0; i < ARRAY_SIZE(flow->hash); i++) { |
| if (flow->hash[i].ifidx == ifidx) { |
| flowid = flow->hash[i].flowid; |
| if (flow->rings[flowid]->status != RING_OPEN) |
| continue; |
| flow->rings[flowid]->status = RING_CLOSING; |
| brcmf_msgbuf_delete_flowring(drvr, flowid); |
| } |
| } |
| flow->addr_mode[ifidx] = addr_mode; |
| } |
| } |
| |
| |
| void brcmf_flowring_delete_peer(struct brcmf_flowring *flow, int ifidx, |
| u8 peer[ETH_ALEN]) |
| { |
| struct brcmf_bus *bus_if = dev_get_drvdata(flow->dev); |
| struct brcmf_pub *drvr = bus_if->drvr; |
| struct brcmf_flowring_hash *hash; |
| struct brcmf_flowring_tdls_entry *prev; |
| struct brcmf_flowring_tdls_entry *search; |
| u32 i; |
| u8 flowid; |
| bool sta; |
| |
| sta = (flow->addr_mode[ifidx] == ADDR_INDIRECT); |
| |
| search = flow->tdls_entry; |
| prev = NULL; |
| while (search) { |
| if (memcmp(search->mac, peer, ETH_ALEN) == 0) { |
| sta = false; |
| break; |
| } |
| prev = search; |
| search = search->next; |
| } |
| |
| hash = flow->hash; |
| for (i = 0; i < BRCMF_FLOWRING_HASHSIZE; i++) { |
| if ((sta || (memcmp(hash[i].mac, peer, ETH_ALEN) == 0)) && |
| (hash[i].ifidx == ifidx)) { |
| flowid = flow->hash[i].flowid; |
| if (flow->rings[flowid]->status == RING_OPEN) { |
| flow->rings[flowid]->status = RING_CLOSING; |
| brcmf_msgbuf_delete_flowring(drvr, flowid); |
| } |
| } |
| } |
| |
| if (search) { |
| if (prev) |
| prev->next = search->next; |
| else |
| flow->tdls_entry = search->next; |
| kfree(search); |
| if (flow->tdls_entry == NULL) |
| flow->tdls_active = false; |
| } |
| } |
| |
| |
| void brcmf_flowring_add_tdls_peer(struct brcmf_flowring *flow, int ifidx, |
| u8 peer[ETH_ALEN]) |
| { |
| struct brcmf_flowring_tdls_entry *tdls_entry; |
| struct brcmf_flowring_tdls_entry *search; |
| |
| tdls_entry = kzalloc(sizeof(*tdls_entry), GFP_ATOMIC); |
| if (tdls_entry == NULL) |
| return; |
| |
| memcpy(tdls_entry->mac, peer, ETH_ALEN); |
| tdls_entry->next = NULL; |
| if (flow->tdls_entry == NULL) { |
| flow->tdls_entry = tdls_entry; |
| } else { |
| search = flow->tdls_entry; |
| if (memcmp(search->mac, peer, ETH_ALEN) == 0) |
| return; |
| while (search->next) { |
| search = search->next; |
| if (memcmp(search->mac, peer, ETH_ALEN) == 0) |
| return; |
| } |
| search->next = tdls_entry; |
| } |
| |
| flow->tdls_active = true; |
| } |