Home
last modified time | relevance | path

Searched refs:q (Results 1 – 25 of 806) sorted by relevance

12345678910>>...33

/linux-3.4.99/crypto/
Dgf128mul.c56 #define gf128mul_dat(q) { \ argument
57 q(0x00), q(0x01), q(0x02), q(0x03), q(0x04), q(0x05), q(0x06), q(0x07),\
58 q(0x08), q(0x09), q(0x0a), q(0x0b), q(0x0c), q(0x0d), q(0x0e), q(0x0f),\
59 q(0x10), q(0x11), q(0x12), q(0x13), q(0x14), q(0x15), q(0x16), q(0x17),\
60 q(0x18), q(0x19), q(0x1a), q(0x1b), q(0x1c), q(0x1d), q(0x1e), q(0x1f),\
61 q(0x20), q(0x21), q(0x22), q(0x23), q(0x24), q(0x25), q(0x26), q(0x27),\
62 q(0x28), q(0x29), q(0x2a), q(0x2b), q(0x2c), q(0x2d), q(0x2e), q(0x2f),\
63 q(0x30), q(0x31), q(0x32), q(0x33), q(0x34), q(0x35), q(0x36), q(0x37),\
64 q(0x38), q(0x39), q(0x3a), q(0x3b), q(0x3c), q(0x3d), q(0x3e), q(0x3f),\
65 q(0x40), q(0x41), q(0x42), q(0x43), q(0x44), q(0x45), q(0x46), q(0x47),\
[all …]
/linux-3.4.99/drivers/media/video/
Dvideobuf-core.c52 #define CALL(q, f, arg...) \ argument
53 ((q->int_ops->f) ? q->int_ops->f(arg) : 0)
55 struct videobuf_buffer *videobuf_alloc_vb(struct videobuf_queue *q) in videobuf_alloc_vb() argument
59 BUG_ON(q->msize < sizeof(*vb)); in videobuf_alloc_vb()
61 if (!q->int_ops || !q->int_ops->alloc_vb) { in videobuf_alloc_vb()
66 vb = q->int_ops->alloc_vb(q->msize); in videobuf_alloc_vb()
76 static int is_state_active_or_queued(struct videobuf_queue *q, struct videobuf_buffer *vb) in is_state_active_or_queued() argument
81 spin_lock_irqsave(q->irqlock, flags); in is_state_active_or_queued()
83 spin_unlock_irqrestore(q->irqlock, flags); in is_state_active_or_queued()
87 int videobuf_waiton(struct videobuf_queue *q, struct videobuf_buffer *vb, in videobuf_waiton() argument
[all …]
Dvideobuf2-core.c33 #define call_memop(q, op, args...) \ argument
34 (((q)->mem_ops->op) ? \
35 ((q)->mem_ops->op(args)) : 0)
37 #define call_qop(q, op, args...) \ argument
38 (((q)->ops->op) ? ((q)->ops->op(args)) : 0)
49 struct vb2_queue *q = vb->vb2_queue; in __vb2_buf_mem_alloc() local
55 mem_priv = call_memop(q, alloc, q->alloc_ctx[plane], in __vb2_buf_mem_alloc()
56 q->plane_sizes[plane]); in __vb2_buf_mem_alloc()
62 vb->v4l2_planes[plane].length = q->plane_sizes[plane]; in __vb2_buf_mem_alloc()
69 call_memop(q, put, vb->planes[plane - 1].mem_priv); in __vb2_buf_mem_alloc()
[all …]
Dbw-qcam.c138 static inline int read_lpstatus(struct qcam *q) in read_lpstatus() argument
140 return parport_read_status(q->pport); in read_lpstatus()
143 static inline int read_lpdata(struct qcam *q) in read_lpdata() argument
145 return parport_read_data(q->pport); in read_lpdata()
148 static inline void write_lpdata(struct qcam *q, int d) in write_lpdata() argument
150 parport_write_data(q->pport, d); in write_lpdata()
153 static void write_lpcontrol(struct qcam *q, int d) in write_lpcontrol() argument
157 parport_data_reverse(q->pport); in write_lpcontrol()
160 parport_data_forward(q->pport); in write_lpcontrol()
166 parport_write_control(q->pport, d); in write_lpcontrol()
[all …]
/linux-3.4.99/drivers/isdn/hardware/eicon/
Ddqueue.c17 diva_data_q_init(diva_um_idi_data_queue_t *q, in diva_data_q_init() argument
22 q->max_length = max_length; in diva_data_q_init()
23 q->segments = max_segments; in diva_data_q_init()
25 for (i = 0; i < q->segments; i++) { in diva_data_q_init()
26 q->data[i] = NULL; in diva_data_q_init()
27 q->length[i] = 0; in diva_data_q_init()
29 q->read = q->write = q->count = q->segment_pending = 0; in diva_data_q_init()
31 for (i = 0; i < q->segments; i++) { in diva_data_q_init()
32 if (!(q->data[i] = diva_os_malloc(0, q->max_length))) { in diva_data_q_init()
33 diva_data_q_finit(q); in diva_data_q_init()
[all …]
/linux-3.4.99/sound/core/seq/
Dseq_queue.c63 static int queue_list_add(struct snd_seq_queue *q) in queue_list_add() argument
71 queue_list[i] = q; in queue_list_add()
72 q->queue = i; in queue_list_add()
84 struct snd_seq_queue *q; in queue_list_remove() local
88 q = queue_list[id]; in queue_list_remove()
89 if (q) { in queue_list_remove()
90 spin_lock(&q->owner_lock); in queue_list_remove()
91 if (q->owner == client) { in queue_list_remove()
93 q->klocked = 1; in queue_list_remove()
94 spin_unlock(&q->owner_lock); in queue_list_remove()
[all …]
/linux-3.4.99/sound/core/seq/oss/
Dseq_oss_readq.c48 struct seq_oss_readq *q; in snd_seq_oss_readq_new() local
50 if ((q = kzalloc(sizeof(*q), GFP_KERNEL)) == NULL) { in snd_seq_oss_readq_new()
55 if ((q->q = kcalloc(maxlen, sizeof(union evrec), GFP_KERNEL)) == NULL) { in snd_seq_oss_readq_new()
57 kfree(q); in snd_seq_oss_readq_new()
61 q->maxlen = maxlen; in snd_seq_oss_readq_new()
62 q->qlen = 0; in snd_seq_oss_readq_new()
63 q->head = q->tail = 0; in snd_seq_oss_readq_new()
64 init_waitqueue_head(&q->midi_sleep); in snd_seq_oss_readq_new()
65 spin_lock_init(&q->lock); in snd_seq_oss_readq_new()
66 q->pre_event_timeout = SNDRV_SEQ_OSS_MAX_TIMEOUT; in snd_seq_oss_readq_new()
[all …]
Dseq_oss_writeq.c39 struct seq_oss_writeq *q; in snd_seq_oss_writeq_new() local
42 if ((q = kzalloc(sizeof(*q), GFP_KERNEL)) == NULL) in snd_seq_oss_writeq_new()
44 q->dp = dp; in snd_seq_oss_writeq_new()
45 q->maxlen = maxlen; in snd_seq_oss_writeq_new()
46 spin_lock_init(&q->sync_lock); in snd_seq_oss_writeq_new()
47 q->sync_event_put = 0; in snd_seq_oss_writeq_new()
48 q->sync_time = 0; in snd_seq_oss_writeq_new()
49 init_waitqueue_head(&q->sync_sleep); in snd_seq_oss_writeq_new()
58 return q; in snd_seq_oss_writeq_new()
65 snd_seq_oss_writeq_delete(struct seq_oss_writeq *q) in snd_seq_oss_writeq_delete() argument
[all …]
/linux-3.4.99/drivers/infiniband/hw/amso1100/
Dc2_mq.c36 void *c2_mq_alloc(struct c2_mq *q) in c2_mq_alloc() argument
38 BUG_ON(q->magic != C2_MQ_MAGIC); in c2_mq_alloc()
39 BUG_ON(q->type != C2_MQ_ADAPTER_TARGET); in c2_mq_alloc()
41 if (c2_mq_full(q)) { in c2_mq_alloc()
46 (struct c2wr_hdr *) (q->msg_pool.host + q->priv * q->msg_size); in c2_mq_alloc()
53 return q->msg_pool.host + q->priv * q->msg_size; in c2_mq_alloc()
58 void c2_mq_produce(struct c2_mq *q) in c2_mq_produce() argument
60 BUG_ON(q->magic != C2_MQ_MAGIC); in c2_mq_produce()
61 BUG_ON(q->type != C2_MQ_ADAPTER_TARGET); in c2_mq_produce()
63 if (!c2_mq_full(q)) { in c2_mq_produce()
[all …]
/linux-3.4.99/drivers/s390/cio/
Dqdio_main.c103 static inline int qdio_check_ccq(struct qdio_q *q, unsigned int ccq) in qdio_check_ccq() argument
115 DBF_ERROR("%4x ccq:%3d", SCH_NO(q), ccq); in qdio_check_ccq()
130 static int qdio_do_eqbs(struct qdio_q *q, unsigned char *state, in qdio_do_eqbs() argument
133 int rc, tmp_count = count, tmp_start = start, nr = q->nr, retried = 0; in qdio_do_eqbs()
136 BUG_ON(!q->irq_ptr->sch_token); in qdio_do_eqbs()
137 qperf_inc(q, eqbs); in qdio_do_eqbs()
139 if (!q->is_input_q) in qdio_do_eqbs()
140 nr += q->irq_ptr->nr_input_qs; in qdio_do_eqbs()
142 ccq = do_eqbs(q->irq_ptr->sch_token, state, nr, &tmp_start, &tmp_count, in qdio_do_eqbs()
144 rc = qdio_check_ccq(q, ccq); in qdio_do_eqbs()
[all …]
/linux-3.4.99/net/sched/
Dsch_choke.c87 static unsigned int choke_len(const struct choke_sched_data *q) in choke_len() argument
89 return (q->tail - q->head) & q->tab_mask; in choke_len()
93 static int use_ecn(const struct choke_sched_data *q) in use_ecn() argument
95 return q->flags & TC_RED_ECN; in use_ecn()
99 static int use_harddrop(const struct choke_sched_data *q) in use_harddrop() argument
101 return q->flags & TC_RED_HARDDROP; in use_harddrop()
105 static void choke_zap_head_holes(struct choke_sched_data *q) in choke_zap_head_holes() argument
108 q->head = (q->head + 1) & q->tab_mask; in choke_zap_head_holes()
109 if (q->head == q->tail) in choke_zap_head_holes()
111 } while (q->tab[q->head] == NULL); in choke_zap_head_holes()
[all …]
Dsch_sfq.c152 static inline struct sfq_head *sfq_dep_head(struct sfq_sched_data *q, sfq_index val) in sfq_dep_head() argument
155 return &q->slots[val].dep; in sfq_dep_head()
156 return &q->dep[val - SFQ_MAX_FLOWS]; in sfq_dep_head()
173 static unsigned int sfq_hash(const struct sfq_sched_data *q, in sfq_hash() argument
181 (__force u32)keys->ports, q->perturbation); in sfq_hash()
182 return hash & (q->divisor - 1); in sfq_hash()
188 struct sfq_sched_data *q = qdisc_priv(sch); in sfq_classify() local
194 TC_H_MIN(skb->priority) <= q->divisor) in sfq_classify()
197 if (!q->filter_list) { in sfq_classify()
199 return sfq_hash(q, skb) + 1; in sfq_classify()
[all …]
Dsch_sfb.c126 static void increment_one_qlen(u32 sfbhash, u32 slot, struct sfb_sched_data *q) in increment_one_qlen() argument
129 struct sfb_bucket *b = &q->bins[slot].bins[0][0]; in increment_one_qlen()
141 static void increment_qlen(const struct sk_buff *skb, struct sfb_sched_data *q) in increment_qlen() argument
147 increment_one_qlen(sfbhash, 0, q); in increment_qlen()
151 increment_one_qlen(sfbhash, 1, q); in increment_qlen()
155 struct sfb_sched_data *q) in decrement_one_qlen() argument
158 struct sfb_bucket *b = &q->bins[slot].bins[0][0]; in decrement_one_qlen()
170 static void decrement_qlen(const struct sk_buff *skb, struct sfb_sched_data *q) in decrement_qlen() argument
176 decrement_one_qlen(sfbhash, 0, q); in decrement_qlen()
180 decrement_one_qlen(sfbhash, 1, q); in decrement_qlen()
[all …]
Dsch_red.c49 static inline int red_use_ecn(struct red_sched_data *q) in red_use_ecn() argument
51 return q->flags & TC_RED_ECN; in red_use_ecn()
54 static inline int red_use_harddrop(struct red_sched_data *q) in red_use_harddrop() argument
56 return q->flags & TC_RED_HARDDROP; in red_use_harddrop()
61 struct red_sched_data *q = qdisc_priv(sch); in red_enqueue() local
62 struct Qdisc *child = q->qdisc; in red_enqueue()
65 q->vars.qavg = red_calc_qavg(&q->parms, in red_enqueue()
66 &q->vars, in red_enqueue()
69 if (red_is_idling(&q->vars)) in red_enqueue()
70 red_end_of_idle_period(&q->vars); in red_enqueue()
[all …]
Dsch_tbf.c117 #define L2T(q, L) qdisc_l2t((q)->R_tab, L) argument
118 #define L2T_P(q, L) qdisc_l2t((q)->P_tab, L) argument
122 struct tbf_sched_data *q = qdisc_priv(sch); in tbf_enqueue() local
125 if (qdisc_pkt_len(skb) > q->max_size) in tbf_enqueue()
128 ret = qdisc_enqueue(skb, q->qdisc); in tbf_enqueue()
135 sch->q.qlen++; in tbf_enqueue()
141 struct tbf_sched_data *q = qdisc_priv(sch); in tbf_drop() local
144 if (q->qdisc->ops->drop && (len = q->qdisc->ops->drop(q->qdisc)) != 0) { in tbf_drop()
145 sch->q.qlen--; in tbf_drop()
153 struct tbf_sched_data *q = qdisc_priv(sch); in tbf_dequeue() local
[all …]
Dsch_netem.c169 static bool loss_4state(struct netem_sched_data *q) in loss_4state() argument
171 struct clgstate *clg = &q->clg; in loss_4state()
233 static bool loss_gilb_ell(struct netem_sched_data *q) in loss_gilb_ell() argument
235 struct clgstate *clg = &q->clg; in loss_gilb_ell()
253 static bool loss_event(struct netem_sched_data *q) in loss_event() argument
255 switch (q->loss_model) { in loss_event()
258 return q->loss && q->loss >= get_crandom(&q->loss_cor); in loss_event()
266 return loss_4state(q); in loss_event()
274 return loss_gilb_ell(q); in loss_event()
312 static psched_time_t packet_len_2_sched_time(unsigned int len, struct netem_sched_data *q) in packet_len_2_sched_time() argument
[all …]
/linux-3.4.99/block/
Dblk-sysfs.c34 static ssize_t queue_requests_show(struct request_queue *q, char *page) in queue_requests_show() argument
36 return queue_var_show(q->nr_requests, (page)); in queue_requests_show()
40 queue_requests_store(struct request_queue *q, const char *page, size_t count) in queue_requests_store() argument
42 struct request_list *rl = &q->rq; in queue_requests_store()
46 if (!q->request_fn) in queue_requests_store()
53 spin_lock_irq(q->queue_lock); in queue_requests_store()
54 q->nr_requests = nr; in queue_requests_store()
55 blk_queue_congestion_threshold(q); in queue_requests_store()
57 if (rl->count[BLK_RW_SYNC] >= queue_congestion_on_threshold(q)) in queue_requests_store()
58 blk_set_queue_congested(q, BLK_RW_SYNC); in queue_requests_store()
[all …]
Dblk-core.c95 void blk_queue_congestion_threshold(struct request_queue *q) in blk_queue_congestion_threshold() argument
99 nr = q->nr_requests - (q->nr_requests / 8) + 1; in blk_queue_congestion_threshold()
100 if (nr > q->nr_requests) in blk_queue_congestion_threshold()
101 nr = q->nr_requests; in blk_queue_congestion_threshold()
102 q->nr_congestion_on = nr; in blk_queue_congestion_threshold()
104 nr = q->nr_requests - (q->nr_requests / 8) - (q->nr_requests / 16) - 1; in blk_queue_congestion_threshold()
107 q->nr_congestion_off = nr; in blk_queue_congestion_threshold()
122 struct request_queue *q = bdev_get_queue(bdev); in blk_get_backing_dev_info() local
124 if (q) in blk_get_backing_dev_info()
125 ret = &q->backing_dev_info; in blk_get_backing_dev_info()
[all …]
Delevator.c61 struct request_queue *q = rq->q; in elv_iosched_allow_merge() local
62 struct elevator_queue *e = q->elevator; in elv_iosched_allow_merge()
65 return e->type->ops.elevator_allow_merge_fn(q, rq, bio); in elv_iosched_allow_merge()
124 static int elevator_init_queue(struct request_queue *q, in elevator_init_queue() argument
127 eq->elevator_data = eq->type->ops.elevator_init_fn(q); in elevator_init_queue()
149 static struct elevator_queue *elevator_alloc(struct request_queue *q, in elevator_alloc() argument
155 eq = kmalloc_node(sizeof(*eq), GFP_KERNEL | __GFP_ZERO, q->node); in elevator_alloc()
164 GFP_KERNEL, q->node); in elevator_alloc()
188 int elevator_init(struct request_queue *q, char *name) in elevator_init() argument
194 if (unlikely(q->elevator)) in elevator_init()
[all …]
Dblk.h16 static inline void __blk_get_queue(struct request_queue *q) in __blk_get_queue() argument
18 kobject_get(&q->kobj); in __blk_get_queue()
22 void blk_rq_bio_prep(struct request_queue *q, struct request *rq,
24 int blk_rq_append_bio(struct request_queue *q, struct request *rq,
26 void blk_drain_queue(struct request_queue *q, bool drain_all);
28 void __blk_queue_free_tags(struct request_queue *q);
64 void blk_abort_flushes(struct request_queue *q);
66 static inline struct request *__elv_next_request(struct request_queue *q) in __elv_next_request() argument
71 if (!list_empty(&q->queue_head)) { in __elv_next_request()
72 rq = list_entry_rq(q->queue_head.next); in __elv_next_request()
[all …]
/linux-3.4.99/drivers/net/wireless/b43/
Dpio.c37 static u16 generate_cookie(struct b43_pio_txqueue *q, in generate_cookie() argument
50 cookie = (((u16)q->index + 1) << 12); in generate_cookie()
62 struct b43_pio_txqueue *q = NULL; in parse_cookie() local
67 q = pio->tx_queue_AC_BK; in parse_cookie()
70 q = pio->tx_queue_AC_BE; in parse_cookie()
73 q = pio->tx_queue_AC_VI; in parse_cookie()
76 q = pio->tx_queue_AC_VO; in parse_cookie()
79 q = pio->tx_queue_mcast; in parse_cookie()
82 if (B43_WARN_ON(!q)) in parse_cookie()
85 if (B43_WARN_ON(pack_index >= ARRAY_SIZE(q->packets))) in parse_cookie()
[all …]
/linux-3.4.99/drivers/net/ethernet/chelsio/cxgb4/
Dsge.c182 static inline unsigned int txq_avail(const struct sge_txq *q) in txq_avail() argument
184 return q->size - 1 - q->in_use; in txq_avail()
265 const struct ulptx_sgl *sgl, const struct sge_txq *q) in unmap_sgl() argument
284 if (likely((u8 *)(p + 1) <= (u8 *)q->stat)) { in unmap_sgl()
290 } else if ((u8 *)p == (u8 *)q->stat) { in unmap_sgl()
291 p = (const struct ulptx_sge_pair *)q->desc; in unmap_sgl()
293 } else if ((u8 *)p + 8 == (u8 *)q->stat) { in unmap_sgl()
294 const __be64 *addr = (const __be64 *)q->desc; in unmap_sgl()
302 const __be64 *addr = (const __be64 *)q->desc; in unmap_sgl()
314 if ((u8 *)p == (u8 *)q->stat) in unmap_sgl()
[all …]
/linux-3.4.99/arch/mips/include/asm/
Dsmtc_ipi.h51 static inline void smtc_ipi_nq(struct smtc_ipi_q *q, struct smtc_ipi *p) in smtc_ipi_nq() argument
55 spin_lock_irqsave(&q->lock, flags); in smtc_ipi_nq()
56 if (q->head == NULL) in smtc_ipi_nq()
57 q->head = q->tail = p; in smtc_ipi_nq()
59 q->tail->flink = p; in smtc_ipi_nq()
61 q->tail = p; in smtc_ipi_nq()
62 q->depth++; in smtc_ipi_nq()
67 spin_unlock_irqrestore(&q->lock, flags); in smtc_ipi_nq()
70 static inline struct smtc_ipi *__smtc_ipi_dq(struct smtc_ipi_q *q) in __smtc_ipi_dq() argument
74 if (q->head == NULL) in __smtc_ipi_dq()
[all …]
/linux-3.4.99/arch/sh/kernel/cpu/sh5/
Dswitchto.S66 st.q r0, ( 9*8), r9
67 st.q r0, (10*8), r10
68 st.q r0, (11*8), r11
69 st.q r0, (12*8), r12
70 st.q r0, (13*8), r13
71 st.q r0, (14*8), r14 ! for unwind, want to look as though we took a trap at
74 st.q r0, (16*8), r16
76 st.q r0, (24*8), r24
77 st.q r0, (25*8), r25
78 st.q r0, (26*8), r26
[all …]
/linux-3.4.99/drivers/net/ethernet/chelsio/cxgb3/
Dsge.c167 static inline struct sge_qset *fl_to_qset(const struct sge_fl *q, int qidx) in fl_to_qset() argument
169 return container_of(q, struct sge_qset, fl[qidx]); in fl_to_qset()
172 static inline struct sge_qset *rspq_to_qset(const struct sge_rspq *q) in rspq_to_qset() argument
174 return container_of(q, struct sge_qset, rspq); in rspq_to_qset()
177 static inline struct sge_qset *txq_to_qset(const struct sge_txq *q, int qidx) in txq_to_qset() argument
179 return container_of(q, struct sge_qset, txq[qidx]); in txq_to_qset()
192 const struct sge_rspq *q, unsigned int credits) in refill_rspq() argument
196 V_RSPQ(q->cntxt_id) | V_CREDITS(credits)); in refill_rspq()
236 static inline void unmap_skb(struct sk_buff *skb, struct sge_txq *q, in unmap_skb() argument
240 struct tx_sw_desc *d = &q->sdesc[cidx]; in unmap_skb()
[all …]

12345678910>>...33