/linux-2.4.37.9/net/sched/ |
D | sch_generic.c | 80 struct Qdisc *q = dev->qdisc; in qdisc_restart() 109 q = dev->qdisc; in qdisc_restart() 149 if (dev->qdisc != &noop_qdisc) { in dev_watchdog() 205 noop_enqueue(struct sk_buff *skb, struct Qdisc * qdisc) in noop_enqueue() argument 212 noop_dequeue(struct Qdisc * qdisc) in noop_dequeue() argument 218 noop_requeue(struct sk_buff *skb, struct Qdisc* qdisc) in noop_requeue() argument 277 pfifo_fast_enqueue(struct sk_buff *skb, struct Qdisc* qdisc) in pfifo_fast_enqueue() argument 281 list = ((struct sk_buff_head*)qdisc->data) + in pfifo_fast_enqueue() 284 if (list->qlen < qdisc->dev->tx_queue_len) { in pfifo_fast_enqueue() 286 qdisc->q.qlen++; in pfifo_fast_enqueue() [all …]
|
D | sch_tbf.c | 132 struct Qdisc *qdisc; /* Inner qdisc, default - bfifo queue */ member 153 if ((ret = q->qdisc->enqueue(skb, q->qdisc)) != 0) { in tbf_enqueue() 169 if ((ret = q->qdisc->ops->requeue(skb, q->qdisc)) == 0) in tbf_requeue() 180 if ((len = q->qdisc->ops->drop(q->qdisc)) != 0) { in tbf_drop() 200 skb = q->qdisc->dequeue(q->qdisc); in tbf_dequeue() 250 if (q->qdisc->ops->requeue(skb, q->qdisc) != NET_XMIT_SUCCESS) { in tbf_dequeue() 266 qdisc_reset(q->qdisc); in tbf_reset() 342 if (q->qdisc == &noop_qdisc) { in tbf_change() 348 if (child) q->qdisc = child; in tbf_change() 382 q->qdisc = &noop_qdisc; in tbf_init() [all …]
|
D | sch_prio.c | 73 struct Qdisc *qdisc; in prio_enqueue() local 76 qdisc = q->queues[prio_classify(skb, sch)]; in prio_enqueue() 78 if ((ret = qdisc->enqueue(skb, qdisc)) == 0) { in prio_enqueue() 93 struct Qdisc *qdisc; in prio_requeue() local 96 qdisc = q->queues[prio_classify(skb, sch)]; in prio_requeue() 98 if ((ret = qdisc->ops->requeue(skb, qdisc)) == 0) { in prio_requeue() 113 struct Qdisc *qdisc; in prio_dequeue() local 116 qdisc = q->queues[prio]; in prio_dequeue() 117 skb = qdisc->dequeue(qdisc); in prio_dequeue() 132 struct Qdisc *qdisc; in prio_drop() local [all …]
|
D | sch_netem.c | 58 struct Qdisc *qdisc; member 205 ret = q->qdisc->enqueue(skb, q->qdisc); in netem_enqueue() 226 if ((ret = q->qdisc->ops->requeue(skb, q->qdisc)) == 0) in netem_requeue() 237 if ((len = q->qdisc->ops->drop(q->qdisc)) != 0) { in netem_drop() 253 skb = q->qdisc->dequeue(q->qdisc); in netem_dequeue() 287 if (q->qdisc->enqueue(skb, q->qdisc)) in netem_watchdog() 300 qdisc_reset(q->qdisc); in netem_reset() 379 ret = set_fifo_limit(q->qdisc, qopt->limit); in netem_change() 434 q->qdisc = qdisc_create_dflt(sch->dev, &pfifo_qdisc_ops); in netem_init() 435 if (!q->qdisc) { in netem_init() [all …]
|
D | sch_hfsc.c | 133 struct Qdisc *qdisc; /* leaf qdisc */ member 819 if (cl->qdisc->q.qlen == 0 && cl->cl_flags & HFSC_FSC) in update_vf() 954 unsigned int len = cl->qdisc->q.qlen; in hfsc_purge_queue() 956 qdisc_reset(cl->qdisc); in hfsc_purge_queue() 1090 if (cl->qdisc->q.qlen != 0) { in hfsc_change_class() 1092 update_ed(cl, qdisc_peek_len(cl->qdisc)); in hfsc_change_class() 1141 cl->qdisc = qdisc_create_dflt(sch->dev, &pfifo_qdisc_ops); in hfsc_change_class() 1142 if (cl->qdisc == NULL) in hfsc_change_class() 1143 cl->qdisc = &noop_qdisc; in hfsc_change_class() 1183 qdisc_destroy(cl->qdisc); in hfsc_destroy_class() [all …]
|
D | sch_cbq.c | 124 struct Qdisc *qdisc; /* Ptr to CBQ discipline */ member 321 struct cbq_sched_data *q = (struct cbq_sched_data*)cl->qdisc->data; in cbq_activate_class() 345 struct cbq_sched_data *q = (struct cbq_sched_data*)this->qdisc->data; in cbq_deactivate_class() 466 struct cbq_sched_data *q = (struct cbq_sched_data *)cl->qdisc->data; in cbq_ovl_classic() 520 struct cbq_sched_data *q = (struct cbq_sched_data *)cl->qdisc->data; in cbq_ovl_rclassic() 539 struct cbq_sched_data *q = (struct cbq_sched_data *)cl->qdisc->data; in cbq_ovl_delay() 575 struct cbq_sched_data *q = (struct cbq_sched_data*)cl->qdisc->data; in cbq_ovl_lowprio() 593 cl->qdisc->q.qlen--; in cbq_ovl_drop() 829 struct cbq_sched_data *q = (struct cbq_sched_data*)cl->qdisc->data; in cbq_under_limit() 1106 if (cl->quantum <= 0 || cl->quantum>32*cl->qdisc->dev->mtu) { in cbq_normalize_quanta() [all …]
|
D | sch_teql.c | 125 struct net_device *m = dat->m->dev.qdisc->dev; in teql_dequeue() 131 sch->q.qlen = dat->q.qlen + dat->m->dev.qdisc->q.qlen; in teql_dequeue() 169 qdisc_reset(master->dev.qdisc); in teql_destroy() 235 struct teql_sched_data *q = (void*)dev->qdisc->data; in __teql_resolve()
|
D | sch_api.c | 301 dev_graft_qdisc(struct net_device *dev, struct Qdisc *qdisc) in dev_graft_qdisc() argument 310 if (qdisc && qdisc->flags&TCQ_F_INGRESS) { in dev_graft_qdisc() 318 dev->qdisc_ingress = qdisc; in dev_graft_qdisc() 330 if (qdisc == NULL) in dev_graft_qdisc() 331 qdisc = &noop_qdisc; in dev_graft_qdisc() 332 dev->qdisc_sleeping = qdisc; in dev_graft_qdisc() 333 dev->qdisc = &noop_qdisc; in dev_graft_qdisc()
|
D | sch_atm.c | 93 static int find_flow(struct atm_qdisc_data *qdisc,struct atm_flow_data *flow) in find_flow() argument 97 DPRINTK("find_flow(qdisc %p,flow %p)\n",qdisc,flow); in find_flow() 98 for (walk = qdisc->flows; walk; walk = walk->next) in find_flow()
|
/linux-2.4.37.9/include/net/ |
D | pkt_sched.h | 432 void qdisc_reset(struct Qdisc *qdisc); 433 void qdisc_destroy(struct Qdisc *qdisc);
|
/linux-2.4.37.9/net/irda/ |
D | irda_device.c | 472 if (skb_queue_len(&dev->qdisc->q)) in irda_device_txqueue_empty()
|
/linux-2.4.37.9/include/linux/ |
D | netdevice.h | 353 struct Qdisc *qdisc; member
|
/linux-2.4.37.9/net/core/ |
D | dev.c | 1071 q = dev->qdisc; in dev_queue_xmit()
|
/linux-2.4.37.9/drivers/isdn/ |
D | isdn_net.c | 307 qdisc_reset(lp->netdev->dev.qdisc); in isdn_net_unbind_channel()
|
/linux-2.4.37.9/ |
D | CREDITS | 716 D: HTB qdisc and random networking hacks
|