/linux-3.4.99/crypto/ |
D | gf128mul.c | 56 #define gf128mul_dat(q) { \ argument 57 q(0x00), q(0x01), q(0x02), q(0x03), q(0x04), q(0x05), q(0x06), q(0x07),\ 58 q(0x08), q(0x09), q(0x0a), q(0x0b), q(0x0c), q(0x0d), q(0x0e), q(0x0f),\ 59 q(0x10), q(0x11), q(0x12), q(0x13), q(0x14), q(0x15), q(0x16), q(0x17),\ 60 q(0x18), q(0x19), q(0x1a), q(0x1b), q(0x1c), q(0x1d), q(0x1e), q(0x1f),\ 61 q(0x20), q(0x21), q(0x22), q(0x23), q(0x24), q(0x25), q(0x26), q(0x27),\ 62 q(0x28), q(0x29), q(0x2a), q(0x2b), q(0x2c), q(0x2d), q(0x2e), q(0x2f),\ 63 q(0x30), q(0x31), q(0x32), q(0x33), q(0x34), q(0x35), q(0x36), q(0x37),\ 64 q(0x38), q(0x39), q(0x3a), q(0x3b), q(0x3c), q(0x3d), q(0x3e), q(0x3f),\ 65 q(0x40), q(0x41), q(0x42), q(0x43), q(0x44), q(0x45), q(0x46), q(0x47),\ [all …]
|
/linux-3.4.99/drivers/media/video/ |
D | videobuf-core.c | 52 #define CALL(q, f, arg...) \ argument 53 ((q->int_ops->f) ? q->int_ops->f(arg) : 0) 55 struct videobuf_buffer *videobuf_alloc_vb(struct videobuf_queue *q) in videobuf_alloc_vb() argument 59 BUG_ON(q->msize < sizeof(*vb)); in videobuf_alloc_vb() 61 if (!q->int_ops || !q->int_ops->alloc_vb) { in videobuf_alloc_vb() 66 vb = q->int_ops->alloc_vb(q->msize); in videobuf_alloc_vb() 76 static int is_state_active_or_queued(struct videobuf_queue *q, struct videobuf_buffer *vb) in is_state_active_or_queued() argument 81 spin_lock_irqsave(q->irqlock, flags); in is_state_active_or_queued() 83 spin_unlock_irqrestore(q->irqlock, flags); in is_state_active_or_queued() 87 int videobuf_waiton(struct videobuf_queue *q, struct videobuf_buffer *vb, in videobuf_waiton() argument [all …]
|
D | videobuf2-core.c | 33 #define call_memop(q, op, args...) \ argument 34 (((q)->mem_ops->op) ? \ 35 ((q)->mem_ops->op(args)) : 0) 37 #define call_qop(q, op, args...) \ argument 38 (((q)->ops->op) ? ((q)->ops->op(args)) : 0) 49 struct vb2_queue *q = vb->vb2_queue; in __vb2_buf_mem_alloc() local 55 mem_priv = call_memop(q, alloc, q->alloc_ctx[plane], in __vb2_buf_mem_alloc() 56 q->plane_sizes[plane]); in __vb2_buf_mem_alloc() 62 vb->v4l2_planes[plane].length = q->plane_sizes[plane]; in __vb2_buf_mem_alloc() 69 call_memop(q, put, vb->planes[plane - 1].mem_priv); in __vb2_buf_mem_alloc() [all …]
|
D | bw-qcam.c | 138 static inline int read_lpstatus(struct qcam *q) in read_lpstatus() argument 140 return parport_read_status(q->pport); in read_lpstatus() 143 static inline int read_lpdata(struct qcam *q) in read_lpdata() argument 145 return parport_read_data(q->pport); in read_lpdata() 148 static inline void write_lpdata(struct qcam *q, int d) in write_lpdata() argument 150 parport_write_data(q->pport, d); in write_lpdata() 153 static void write_lpcontrol(struct qcam *q, int d) in write_lpcontrol() argument 157 parport_data_reverse(q->pport); in write_lpcontrol() 160 parport_data_forward(q->pport); in write_lpcontrol() 166 parport_write_control(q->pport, d); in write_lpcontrol() [all …]
|
/linux-3.4.99/drivers/isdn/hardware/eicon/ |
D | dqueue.c | 17 diva_data_q_init(diva_um_idi_data_queue_t *q, in diva_data_q_init() argument 22 q->max_length = max_length; in diva_data_q_init() 23 q->segments = max_segments; in diva_data_q_init() 25 for (i = 0; i < q->segments; i++) { in diva_data_q_init() 26 q->data[i] = NULL; in diva_data_q_init() 27 q->length[i] = 0; in diva_data_q_init() 29 q->read = q->write = q->count = q->segment_pending = 0; in diva_data_q_init() 31 for (i = 0; i < q->segments; i++) { in diva_data_q_init() 32 if (!(q->data[i] = diva_os_malloc(0, q->max_length))) { in diva_data_q_init() 33 diva_data_q_finit(q); in diva_data_q_init() [all …]
|
/linux-3.4.99/sound/core/seq/ |
D | seq_queue.c | 63 static int queue_list_add(struct snd_seq_queue *q) in queue_list_add() argument 71 queue_list[i] = q; in queue_list_add() 72 q->queue = i; in queue_list_add() 84 struct snd_seq_queue *q; in queue_list_remove() local 88 q = queue_list[id]; in queue_list_remove() 89 if (q) { in queue_list_remove() 90 spin_lock(&q->owner_lock); in queue_list_remove() 91 if (q->owner == client) { in queue_list_remove() 93 q->klocked = 1; in queue_list_remove() 94 spin_unlock(&q->owner_lock); in queue_list_remove() [all …]
|
/linux-3.4.99/sound/core/seq/oss/ |
D | seq_oss_readq.c | 48 struct seq_oss_readq *q; in snd_seq_oss_readq_new() local 50 if ((q = kzalloc(sizeof(*q), GFP_KERNEL)) == NULL) { in snd_seq_oss_readq_new() 55 if ((q->q = kcalloc(maxlen, sizeof(union evrec), GFP_KERNEL)) == NULL) { in snd_seq_oss_readq_new() 57 kfree(q); in snd_seq_oss_readq_new() 61 q->maxlen = maxlen; in snd_seq_oss_readq_new() 62 q->qlen = 0; in snd_seq_oss_readq_new() 63 q->head = q->tail = 0; in snd_seq_oss_readq_new() 64 init_waitqueue_head(&q->midi_sleep); in snd_seq_oss_readq_new() 65 spin_lock_init(&q->lock); in snd_seq_oss_readq_new() 66 q->pre_event_timeout = SNDRV_SEQ_OSS_MAX_TIMEOUT; in snd_seq_oss_readq_new() [all …]
|
D | seq_oss_writeq.c | 39 struct seq_oss_writeq *q; in snd_seq_oss_writeq_new() local 42 if ((q = kzalloc(sizeof(*q), GFP_KERNEL)) == NULL) in snd_seq_oss_writeq_new() 44 q->dp = dp; in snd_seq_oss_writeq_new() 45 q->maxlen = maxlen; in snd_seq_oss_writeq_new() 46 spin_lock_init(&q->sync_lock); in snd_seq_oss_writeq_new() 47 q->sync_event_put = 0; in snd_seq_oss_writeq_new() 48 q->sync_time = 0; in snd_seq_oss_writeq_new() 49 init_waitqueue_head(&q->sync_sleep); in snd_seq_oss_writeq_new() 58 return q; in snd_seq_oss_writeq_new() 65 snd_seq_oss_writeq_delete(struct seq_oss_writeq *q) in snd_seq_oss_writeq_delete() argument [all …]
|
/linux-3.4.99/drivers/infiniband/hw/amso1100/ |
D | c2_mq.c | 36 void *c2_mq_alloc(struct c2_mq *q) in c2_mq_alloc() argument 38 BUG_ON(q->magic != C2_MQ_MAGIC); in c2_mq_alloc() 39 BUG_ON(q->type != C2_MQ_ADAPTER_TARGET); in c2_mq_alloc() 41 if (c2_mq_full(q)) { in c2_mq_alloc() 46 (struct c2wr_hdr *) (q->msg_pool.host + q->priv * q->msg_size); in c2_mq_alloc() 53 return q->msg_pool.host + q->priv * q->msg_size; in c2_mq_alloc() 58 void c2_mq_produce(struct c2_mq *q) in c2_mq_produce() argument 60 BUG_ON(q->magic != C2_MQ_MAGIC); in c2_mq_produce() 61 BUG_ON(q->type != C2_MQ_ADAPTER_TARGET); in c2_mq_produce() 63 if (!c2_mq_full(q)) { in c2_mq_produce() [all …]
|
/linux-3.4.99/drivers/s390/cio/ |
D | qdio_main.c | 103 static inline int qdio_check_ccq(struct qdio_q *q, unsigned int ccq) in qdio_check_ccq() argument 115 DBF_ERROR("%4x ccq:%3d", SCH_NO(q), ccq); in qdio_check_ccq() 130 static int qdio_do_eqbs(struct qdio_q *q, unsigned char *state, in qdio_do_eqbs() argument 133 int rc, tmp_count = count, tmp_start = start, nr = q->nr, retried = 0; in qdio_do_eqbs() 136 BUG_ON(!q->irq_ptr->sch_token); in qdio_do_eqbs() 137 qperf_inc(q, eqbs); in qdio_do_eqbs() 139 if (!q->is_input_q) in qdio_do_eqbs() 140 nr += q->irq_ptr->nr_input_qs; in qdio_do_eqbs() 142 ccq = do_eqbs(q->irq_ptr->sch_token, state, nr, &tmp_start, &tmp_count, in qdio_do_eqbs() 144 rc = qdio_check_ccq(q, ccq); in qdio_do_eqbs() [all …]
|
/linux-3.4.99/net/sched/ |
D | sch_choke.c | 87 static unsigned int choke_len(const struct choke_sched_data *q) in choke_len() argument 89 return (q->tail - q->head) & q->tab_mask; in choke_len() 93 static int use_ecn(const struct choke_sched_data *q) in use_ecn() argument 95 return q->flags & TC_RED_ECN; in use_ecn() 99 static int use_harddrop(const struct choke_sched_data *q) in use_harddrop() argument 101 return q->flags & TC_RED_HARDDROP; in use_harddrop() 105 static void choke_zap_head_holes(struct choke_sched_data *q) in choke_zap_head_holes() argument 108 q->head = (q->head + 1) & q->tab_mask; in choke_zap_head_holes() 109 if (q->head == q->tail) in choke_zap_head_holes() 111 } while (q->tab[q->head] == NULL); in choke_zap_head_holes() [all …]
|
D | sch_sfq.c | 152 static inline struct sfq_head *sfq_dep_head(struct sfq_sched_data *q, sfq_index val) in sfq_dep_head() argument 155 return &q->slots[val].dep; in sfq_dep_head() 156 return &q->dep[val - SFQ_MAX_FLOWS]; in sfq_dep_head() 173 static unsigned int sfq_hash(const struct sfq_sched_data *q, in sfq_hash() argument 181 (__force u32)keys->ports, q->perturbation); in sfq_hash() 182 return hash & (q->divisor - 1); in sfq_hash() 188 struct sfq_sched_data *q = qdisc_priv(sch); in sfq_classify() local 194 TC_H_MIN(skb->priority) <= q->divisor) in sfq_classify() 197 if (!q->filter_list) { in sfq_classify() 199 return sfq_hash(q, skb) + 1; in sfq_classify() [all …]
|
D | sch_sfb.c | 126 static void increment_one_qlen(u32 sfbhash, u32 slot, struct sfb_sched_data *q) in increment_one_qlen() argument 129 struct sfb_bucket *b = &q->bins[slot].bins[0][0]; in increment_one_qlen() 141 static void increment_qlen(const struct sk_buff *skb, struct sfb_sched_data *q) in increment_qlen() argument 147 increment_one_qlen(sfbhash, 0, q); in increment_qlen() 151 increment_one_qlen(sfbhash, 1, q); in increment_qlen() 155 struct sfb_sched_data *q) in decrement_one_qlen() argument 158 struct sfb_bucket *b = &q->bins[slot].bins[0][0]; in decrement_one_qlen() 170 static void decrement_qlen(const struct sk_buff *skb, struct sfb_sched_data *q) in decrement_qlen() argument 176 decrement_one_qlen(sfbhash, 0, q); in decrement_qlen() 180 decrement_one_qlen(sfbhash, 1, q); in decrement_qlen() [all …]
|
D | sch_red.c | 49 static inline int red_use_ecn(struct red_sched_data *q) in red_use_ecn() argument 51 return q->flags & TC_RED_ECN; in red_use_ecn() 54 static inline int red_use_harddrop(struct red_sched_data *q) in red_use_harddrop() argument 56 return q->flags & TC_RED_HARDDROP; in red_use_harddrop() 61 struct red_sched_data *q = qdisc_priv(sch); in red_enqueue() local 62 struct Qdisc *child = q->qdisc; in red_enqueue() 65 q->vars.qavg = red_calc_qavg(&q->parms, in red_enqueue() 66 &q->vars, in red_enqueue() 69 if (red_is_idling(&q->vars)) in red_enqueue() 70 red_end_of_idle_period(&q->vars); in red_enqueue() [all …]
|
D | sch_tbf.c | 117 #define L2T(q, L) qdisc_l2t((q)->R_tab, L) argument 118 #define L2T_P(q, L) qdisc_l2t((q)->P_tab, L) argument 122 struct tbf_sched_data *q = qdisc_priv(sch); in tbf_enqueue() local 125 if (qdisc_pkt_len(skb) > q->max_size) in tbf_enqueue() 128 ret = qdisc_enqueue(skb, q->qdisc); in tbf_enqueue() 135 sch->q.qlen++; in tbf_enqueue() 141 struct tbf_sched_data *q = qdisc_priv(sch); in tbf_drop() local 144 if (q->qdisc->ops->drop && (len = q->qdisc->ops->drop(q->qdisc)) != 0) { in tbf_drop() 145 sch->q.qlen--; in tbf_drop() 153 struct tbf_sched_data *q = qdisc_priv(sch); in tbf_dequeue() local [all …]
|
D | sch_netem.c | 169 static bool loss_4state(struct netem_sched_data *q) in loss_4state() argument 171 struct clgstate *clg = &q->clg; in loss_4state() 233 static bool loss_gilb_ell(struct netem_sched_data *q) in loss_gilb_ell() argument 235 struct clgstate *clg = &q->clg; in loss_gilb_ell() 253 static bool loss_event(struct netem_sched_data *q) in loss_event() argument 255 switch (q->loss_model) { in loss_event() 258 return q->loss && q->loss >= get_crandom(&q->loss_cor); in loss_event() 266 return loss_4state(q); in loss_event() 274 return loss_gilb_ell(q); in loss_event() 312 static psched_time_t packet_len_2_sched_time(unsigned int len, struct netem_sched_data *q) in packet_len_2_sched_time() argument [all …]
|
/linux-3.4.99/block/ |
D | blk-sysfs.c | 34 static ssize_t queue_requests_show(struct request_queue *q, char *page) in queue_requests_show() argument 36 return queue_var_show(q->nr_requests, (page)); in queue_requests_show() 40 queue_requests_store(struct request_queue *q, const char *page, size_t count) in queue_requests_store() argument 42 struct request_list *rl = &q->rq; in queue_requests_store() 46 if (!q->request_fn) in queue_requests_store() 53 spin_lock_irq(q->queue_lock); in queue_requests_store() 54 q->nr_requests = nr; in queue_requests_store() 55 blk_queue_congestion_threshold(q); in queue_requests_store() 57 if (rl->count[BLK_RW_SYNC] >= queue_congestion_on_threshold(q)) in queue_requests_store() 58 blk_set_queue_congested(q, BLK_RW_SYNC); in queue_requests_store() [all …]
|
D | blk-core.c | 95 void blk_queue_congestion_threshold(struct request_queue *q) in blk_queue_congestion_threshold() argument 99 nr = q->nr_requests - (q->nr_requests / 8) + 1; in blk_queue_congestion_threshold() 100 if (nr > q->nr_requests) in blk_queue_congestion_threshold() 101 nr = q->nr_requests; in blk_queue_congestion_threshold() 102 q->nr_congestion_on = nr; in blk_queue_congestion_threshold() 104 nr = q->nr_requests - (q->nr_requests / 8) - (q->nr_requests / 16) - 1; in blk_queue_congestion_threshold() 107 q->nr_congestion_off = nr; in blk_queue_congestion_threshold() 122 struct request_queue *q = bdev_get_queue(bdev); in blk_get_backing_dev_info() local 124 if (q) in blk_get_backing_dev_info() 125 ret = &q->backing_dev_info; in blk_get_backing_dev_info() [all …]
|
D | elevator.c | 61 struct request_queue *q = rq->q; in elv_iosched_allow_merge() local 62 struct elevator_queue *e = q->elevator; in elv_iosched_allow_merge() 65 return e->type->ops.elevator_allow_merge_fn(q, rq, bio); in elv_iosched_allow_merge() 124 static int elevator_init_queue(struct request_queue *q, in elevator_init_queue() argument 127 eq->elevator_data = eq->type->ops.elevator_init_fn(q); in elevator_init_queue() 149 static struct elevator_queue *elevator_alloc(struct request_queue *q, in elevator_alloc() argument 155 eq = kmalloc_node(sizeof(*eq), GFP_KERNEL | __GFP_ZERO, q->node); in elevator_alloc() 164 GFP_KERNEL, q->node); in elevator_alloc() 188 int elevator_init(struct request_queue *q, char *name) in elevator_init() argument 194 if (unlikely(q->elevator)) in elevator_init() [all …]
|
D | blk.h | 16 static inline void __blk_get_queue(struct request_queue *q) in __blk_get_queue() argument 18 kobject_get(&q->kobj); in __blk_get_queue() 22 void blk_rq_bio_prep(struct request_queue *q, struct request *rq, 24 int blk_rq_append_bio(struct request_queue *q, struct request *rq, 26 void blk_drain_queue(struct request_queue *q, bool drain_all); 28 void __blk_queue_free_tags(struct request_queue *q); 64 void blk_abort_flushes(struct request_queue *q); 66 static inline struct request *__elv_next_request(struct request_queue *q) in __elv_next_request() argument 71 if (!list_empty(&q->queue_head)) { in __elv_next_request() 72 rq = list_entry_rq(q->queue_head.next); in __elv_next_request() [all …]
|
/linux-3.4.99/drivers/net/wireless/b43/ |
D | pio.c | 37 static u16 generate_cookie(struct b43_pio_txqueue *q, in generate_cookie() argument 50 cookie = (((u16)q->index + 1) << 12); in generate_cookie() 62 struct b43_pio_txqueue *q = NULL; in parse_cookie() local 67 q = pio->tx_queue_AC_BK; in parse_cookie() 70 q = pio->tx_queue_AC_BE; in parse_cookie() 73 q = pio->tx_queue_AC_VI; in parse_cookie() 76 q = pio->tx_queue_AC_VO; in parse_cookie() 79 q = pio->tx_queue_mcast; in parse_cookie() 82 if (B43_WARN_ON(!q)) in parse_cookie() 85 if (B43_WARN_ON(pack_index >= ARRAY_SIZE(q->packets))) in parse_cookie() [all …]
|
/linux-3.4.99/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 182 static inline unsigned int txq_avail(const struct sge_txq *q) in txq_avail() argument 184 return q->size - 1 - q->in_use; in txq_avail() 265 const struct ulptx_sgl *sgl, const struct sge_txq *q) in unmap_sgl() argument 284 if (likely((u8 *)(p + 1) <= (u8 *)q->stat)) { in unmap_sgl() 290 } else if ((u8 *)p == (u8 *)q->stat) { in unmap_sgl() 291 p = (const struct ulptx_sge_pair *)q->desc; in unmap_sgl() 293 } else if ((u8 *)p + 8 == (u8 *)q->stat) { in unmap_sgl() 294 const __be64 *addr = (const __be64 *)q->desc; in unmap_sgl() 302 const __be64 *addr = (const __be64 *)q->desc; in unmap_sgl() 314 if ((u8 *)p == (u8 *)q->stat) in unmap_sgl() [all …]
|
/linux-3.4.99/arch/mips/include/asm/ |
D | smtc_ipi.h | 51 static inline void smtc_ipi_nq(struct smtc_ipi_q *q, struct smtc_ipi *p) in smtc_ipi_nq() argument 55 spin_lock_irqsave(&q->lock, flags); in smtc_ipi_nq() 56 if (q->head == NULL) in smtc_ipi_nq() 57 q->head = q->tail = p; in smtc_ipi_nq() 59 q->tail->flink = p; in smtc_ipi_nq() 61 q->tail = p; in smtc_ipi_nq() 62 q->depth++; in smtc_ipi_nq() 67 spin_unlock_irqrestore(&q->lock, flags); in smtc_ipi_nq() 70 static inline struct smtc_ipi *__smtc_ipi_dq(struct smtc_ipi_q *q) in __smtc_ipi_dq() argument 74 if (q->head == NULL) in __smtc_ipi_dq() [all …]
|
/linux-3.4.99/arch/sh/kernel/cpu/sh5/ |
D | switchto.S | 66 st.q r0, ( 9*8), r9 67 st.q r0, (10*8), r10 68 st.q r0, (11*8), r11 69 st.q r0, (12*8), r12 70 st.q r0, (13*8), r13 71 st.q r0, (14*8), r14 ! for unwind, want to look as though we took a trap at 74 st.q r0, (16*8), r16 76 st.q r0, (24*8), r24 77 st.q r0, (25*8), r25 78 st.q r0, (26*8), r26 [all …]
|
/linux-3.4.99/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 167 static inline struct sge_qset *fl_to_qset(const struct sge_fl *q, int qidx) in fl_to_qset() argument 169 return container_of(q, struct sge_qset, fl[qidx]); in fl_to_qset() 172 static inline struct sge_qset *rspq_to_qset(const struct sge_rspq *q) in rspq_to_qset() argument 174 return container_of(q, struct sge_qset, rspq); in rspq_to_qset() 177 static inline struct sge_qset *txq_to_qset(const struct sge_txq *q, int qidx) in txq_to_qset() argument 179 return container_of(q, struct sge_qset, txq[qidx]); in txq_to_qset() 192 const struct sge_rspq *q, unsigned int credits) in refill_rspq() argument 196 V_RSPQ(q->cntxt_id) | V_CREDITS(credits)); in refill_rspq() 236 static inline void unmap_skb(struct sk_buff *skb, struct sge_txq *q, in unmap_skb() argument 240 struct tx_sw_desc *d = &q->sdesc[cidx]; in unmap_skb() [all …]
|