/linux-6.6.21/tools/net/ynl/generated/ |
D | devlink-user.h | 89 void devlink_get_req_free(struct devlink_get_req *req); 92 devlink_get_req_set_bus_name(struct devlink_get_req *req, const char *bus_name) in devlink_get_req_set_bus_name() argument 94 free(req->bus_name); in devlink_get_req_set_bus_name() 95 req->_present.bus_name_len = strlen(bus_name); in devlink_get_req_set_bus_name() 96 req->bus_name = malloc(req->_present.bus_name_len + 1); in devlink_get_req_set_bus_name() 97 memcpy(req->bus_name, bus_name, req->_present.bus_name_len); in devlink_get_req_set_bus_name() 98 req->bus_name[req->_present.bus_name_len] = 0; in devlink_get_req_set_bus_name() 101 devlink_get_req_set_dev_name(struct devlink_get_req *req, const char *dev_name) in devlink_get_req_set_dev_name() argument 103 free(req->dev_name); in devlink_get_req_set_dev_name() 104 req->_present.dev_name_len = strlen(dev_name); in devlink_get_req_set_dev_name() [all …]
|
D | ethtool-user.h | 227 void ethtool_strset_get_req_free(struct ethtool_strset_get_req *req); 230 ethtool_strset_get_req_set_header_dev_index(struct ethtool_strset_get_req *req, in ethtool_strset_get_req_set_header_dev_index() argument 233 req->_present.header = 1; in ethtool_strset_get_req_set_header_dev_index() 234 req->header._present.dev_index = 1; in ethtool_strset_get_req_set_header_dev_index() 235 req->header.dev_index = dev_index; in ethtool_strset_get_req_set_header_dev_index() 238 ethtool_strset_get_req_set_header_dev_name(struct ethtool_strset_get_req *req, in ethtool_strset_get_req_set_header_dev_name() argument 241 free(req->header.dev_name); in ethtool_strset_get_req_set_header_dev_name() 242 req->header._present.dev_name_len = strlen(dev_name); in ethtool_strset_get_req_set_header_dev_name() 243 req->header.dev_name = malloc(req->header._present.dev_name_len + 1); in ethtool_strset_get_req_set_header_dev_name() 244 memcpy(req->header.dev_name, dev_name, req->header._present.dev_name_len); in ethtool_strset_get_req_set_header_dev_name() [all …]
|
D | fou-user.h | 54 void fou_add_req_free(struct fou_add_req *req); 57 fou_add_req_set_port(struct fou_add_req *req, __u16 port /* big-endian */) in fou_add_req_set_port() argument 59 req->_present.port = 1; in fou_add_req_set_port() 60 req->port = port; in fou_add_req_set_port() 63 fou_add_req_set_ipproto(struct fou_add_req *req, __u8 ipproto) in fou_add_req_set_ipproto() argument 65 req->_present.ipproto = 1; in fou_add_req_set_ipproto() 66 req->ipproto = ipproto; in fou_add_req_set_ipproto() 68 static inline void fou_add_req_set_type(struct fou_add_req *req, __u8 type) in fou_add_req_set_type() argument 70 req->_present.type = 1; in fou_add_req_set_type() 71 req->type = type; in fou_add_req_set_type() [all …]
|
/linux-6.6.21/drivers/media/mc/ |
D | mc-request.c | 39 static void media_request_clean(struct media_request *req) in media_request_clean() argument 44 WARN_ON(req->state != MEDIA_REQUEST_STATE_CLEANING); in media_request_clean() 45 WARN_ON(req->updating_count); in media_request_clean() 46 WARN_ON(req->access_count); in media_request_clean() 48 list_for_each_entry_safe(obj, obj_safe, &req->objects, list) { in media_request_clean() 53 req->updating_count = 0; in media_request_clean() 54 req->access_count = 0; in media_request_clean() 55 WARN_ON(req->num_incomplete_objects); in media_request_clean() 56 req->num_incomplete_objects = 0; in media_request_clean() 57 wake_up_interruptible_all(&req->poll_wait); in media_request_clean() [all …]
|
/linux-6.6.21/drivers/s390/scsi/ |
D | zfcp_fsf.c | 80 static void zfcp_fsf_class_not_supp(struct zfcp_fsf_req *req) in zfcp_fsf_class_not_supp() argument 82 dev_err(&req->adapter->ccw_device->dev, "FCP device not " in zfcp_fsf_class_not_supp() 84 zfcp_erp_adapter_shutdown(req->adapter, 0, "fscns_1"); in zfcp_fsf_class_not_supp() 85 req->status |= ZFCP_STATUS_FSFREQ_ERROR; in zfcp_fsf_class_not_supp() 92 void zfcp_fsf_req_free(struct zfcp_fsf_req *req) in zfcp_fsf_req_free() argument 94 if (likely(req->pool)) { in zfcp_fsf_req_free() 95 if (likely(!zfcp_fsf_req_is_status_read_buffer(req))) in zfcp_fsf_req_free() 96 mempool_free(req->qtcb, req->adapter->pool.qtcb_pool); in zfcp_fsf_req_free() 97 mempool_free(req, req->pool); in zfcp_fsf_req_free() 101 if (likely(!zfcp_fsf_req_is_status_read_buffer(req))) in zfcp_fsf_req_free() [all …]
|
/linux-6.6.21/drivers/staging/greybus/ |
D | audio_apbridgea.c | 16 struct audio_apbridgea_set_config_request req; in gb_audio_apbridgea_set_config() local 18 req.hdr.type = AUDIO_APBRIDGEA_TYPE_SET_CONFIG; in gb_audio_apbridgea_set_config() 19 req.hdr.i2s_port = cpu_to_le16(i2s_port); in gb_audio_apbridgea_set_config() 20 req.format = cpu_to_le32(format); in gb_audio_apbridgea_set_config() 21 req.rate = cpu_to_le32(rate); in gb_audio_apbridgea_set_config() 22 req.mclk_freq = cpu_to_le32(mclk_freq); in gb_audio_apbridgea_set_config() 24 return gb_hd_output(connection->hd, &req, sizeof(req), in gb_audio_apbridgea_set_config() 33 struct audio_apbridgea_register_cport_request req; in gb_audio_apbridgea_register_cport() local 36 req.hdr.type = AUDIO_APBRIDGEA_TYPE_REGISTER_CPORT; in gb_audio_apbridgea_register_cport() 37 req.hdr.i2s_port = cpu_to_le16(i2s_port); in gb_audio_apbridgea_register_cport() [all …]
|
D | audio_gb.c | 50 struct gb_audio_get_control_request req; in gb_audio_gb_get_control() local 54 req.control_id = control_id; in gb_audio_gb_get_control() 55 req.index = index; in gb_audio_gb_get_control() 58 &req, sizeof(req), &resp, sizeof(resp)); in gb_audio_gb_get_control() 72 struct gb_audio_set_control_request req; in gb_audio_gb_set_control() local 74 req.control_id = control_id; in gb_audio_gb_set_control() 75 req.index = index; in gb_audio_gb_set_control() 76 memcpy(&req.value, value, sizeof(req.value)); in gb_audio_gb_set_control() 79 &req, sizeof(req), NULL, 0); in gb_audio_gb_set_control() 86 struct gb_audio_enable_widget_request req; in gb_audio_gb_enable_widget() local [all …]
|
/linux-6.6.21/drivers/nvme/target/ |
D | io-cmd-file.c | 76 static ssize_t nvmet_file_submit_bvec(struct nvmet_req *req, loff_t pos, in nvmet_file_submit_bvec() argument 79 struct kiocb *iocb = &req->f.iocb; in nvmet_file_submit_bvec() 84 if (req->cmd->rw.opcode == nvme_cmd_write) { in nvmet_file_submit_bvec() 85 if (req->cmd->rw.control & cpu_to_le16(NVME_RW_FUA)) in nvmet_file_submit_bvec() 87 call_iter = req->ns->file->f_op->write_iter; in nvmet_file_submit_bvec() 90 call_iter = req->ns->file->f_op->read_iter; in nvmet_file_submit_bvec() 94 iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count); in nvmet_file_submit_bvec() 97 iocb->ki_filp = req->ns->file; in nvmet_file_submit_bvec() 105 struct nvmet_req *req = container_of(iocb, struct nvmet_req, f.iocb); in nvmet_file_io_done() local 108 if (req->f.bvec != req->inline_bvec) { in nvmet_file_io_done() [all …]
|
D | admin-cmd.c | 28 static u32 nvmet_feat_data_len(struct nvmet_req *req, u32 cdw10) in nvmet_feat_data_len() argument 32 return sizeof(req->sq->ctrl->hostid); in nvmet_feat_data_len() 43 static void nvmet_execute_get_log_page_noop(struct nvmet_req *req) in nvmet_execute_get_log_page_noop() argument 45 nvmet_req_complete(req, nvmet_zero_sgl(req, 0, req->transfer_len)); in nvmet_execute_get_log_page_noop() 48 static void nvmet_execute_get_log_page_error(struct nvmet_req *req) in nvmet_execute_get_log_page_error() argument 50 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_execute_get_log_page_error() 60 if (nvmet_copy_to_sgl(req, offset, &ctrl->slots[slot], in nvmet_execute_get_log_page_error() 71 nvmet_req_complete(req, 0); in nvmet_execute_get_log_page_error() 74 static u16 nvmet_get_smart_log_nsid(struct nvmet_req *req, in nvmet_get_smart_log_nsid() argument 80 status = nvmet_req_find_ns(req); in nvmet_get_smart_log_nsid() [all …]
|
D | fabrics-cmd-auth.c | 34 static u16 nvmet_auth_negotiate(struct nvmet_req *req, void *d) in nvmet_auth_negotiate() argument 36 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_auth_negotiate() 41 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate() 45 req->sq->dhchap_tid = le16_to_cpu(data->t_id); in nvmet_auth_negotiate() 70 __func__, ctrl->cntlid, req->sq->qid); in nvmet_auth_negotiate() 74 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate() 98 __func__, ctrl->cntlid, req->sq->qid); in nvmet_auth_negotiate() 102 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate() 107 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate() 112 static u16 nvmet_auth_reply(struct nvmet_req *req, void *d) in nvmet_auth_reply() argument [all …]
|
D | zns.c | 73 void nvmet_execute_identify_ctrl_zns(struct nvmet_req *req) in nvmet_execute_identify_ctrl_zns() argument 75 u8 zasl = req->sq->ctrl->subsys->zasl; in nvmet_execute_identify_ctrl_zns() 76 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_execute_identify_ctrl_zns() 91 status = nvmet_copy_to_sgl(req, 0, id, sizeof(*id)); in nvmet_execute_identify_ctrl_zns() 95 nvmet_req_complete(req, status); in nvmet_execute_identify_ctrl_zns() 98 void nvmet_execute_identify_ns_zns(struct nvmet_req *req) in nvmet_execute_identify_ns_zns() argument 105 if (le32_to_cpu(req->cmd->identify.nsid) == NVME_NSID_ALL) { in nvmet_execute_identify_ns_zns() 106 req->error_loc = offsetof(struct nvme_identify, nsid); in nvmet_execute_identify_ns_zns() 117 status = nvmet_req_find_ns(req); in nvmet_execute_identify_ns_zns() 121 if (nvmet_ns_revalidate(req->ns)) { in nvmet_execute_identify_ns_zns() [all …]
|
D | io-cmd-bdev.c | 122 u16 blk_to_nvme_status(struct nvmet_req *req, blk_status_t blk_sts) in blk_to_nvme_status() argument 137 req->error_loc = offsetof(struct nvme_rw_command, length); in blk_to_nvme_status() 141 req->error_loc = offsetof(struct nvme_rw_command, slba); in blk_to_nvme_status() 144 req->error_loc = offsetof(struct nvme_common_command, opcode); in blk_to_nvme_status() 145 switch (req->cmd->common.opcode) { in blk_to_nvme_status() 156 req->error_loc = offsetof(struct nvme_rw_command, nsid); in blk_to_nvme_status() 161 req->error_loc = offsetof(struct nvme_common_command, opcode); in blk_to_nvme_status() 164 switch (req->cmd->common.opcode) { in blk_to_nvme_status() 167 req->error_slba = le64_to_cpu(req->cmd->rw.slba); in blk_to_nvme_status() 170 req->error_slba = in blk_to_nvme_status() [all …]
|
/linux-6.6.21/drivers/block/drbd/ |
D | drbd_req.c | 26 struct drbd_request *req; in drbd_req_new() local 28 req = mempool_alloc(&drbd_request_mempool, GFP_NOIO); in drbd_req_new() 29 if (!req) in drbd_req_new() 31 memset(req, 0, sizeof(*req)); in drbd_req_new() 33 req->rq_state = (bio_data_dir(bio_src) == WRITE ? RQ_WRITE : 0) in drbd_req_new() 36 req->device = device; in drbd_req_new() 37 req->master_bio = bio_src; in drbd_req_new() 38 req->epoch = 0; in drbd_req_new() 40 drbd_clear_interval(&req->i); in drbd_req_new() 41 req->i.sector = bio_src->bi_iter.bi_sector; in drbd_req_new() [all …]
|
/linux-6.6.21/io_uring/ |
D | poll.c | 34 struct io_kiocb *req; member 71 static bool io_poll_get_ownership_slowpath(struct io_kiocb *req) in io_poll_get_ownership_slowpath() argument 80 v = atomic_fetch_or(IO_POLL_RETRY_FLAG, &req->poll_refs); in io_poll_get_ownership_slowpath() 83 return !(atomic_fetch_inc(&req->poll_refs) & IO_POLL_REF_MASK); in io_poll_get_ownership_slowpath() 92 static inline bool io_poll_get_ownership(struct io_kiocb *req) in io_poll_get_ownership() argument 94 if (unlikely(atomic_read(&req->poll_refs) >= IO_POLL_REF_BIAS)) in io_poll_get_ownership() 95 return io_poll_get_ownership_slowpath(req); in io_poll_get_ownership() 96 return !(atomic_fetch_inc(&req->poll_refs) & IO_POLL_REF_MASK); in io_poll_get_ownership() 99 static void io_poll_mark_cancelled(struct io_kiocb *req) in io_poll_mark_cancelled() argument 101 atomic_or(IO_POLL_CANCEL_FLAG, &req->poll_refs); in io_poll_mark_cancelled() [all …]
|
D | rw.c | 31 static inline bool io_file_supports_nowait(struct io_kiocb *req) in io_file_supports_nowait() argument 33 return req->flags & REQ_F_SUPPORT_NOWAIT; in io_file_supports_nowait() 55 static int io_iov_buffer_select_prep(struct io_kiocb *req) in io_iov_buffer_select_prep() argument 59 struct io_rw *rw = io_kiocb_to_cmd(req, struct io_rw); in io_iov_buffer_select_prep() 65 if (req->ctx->compat) in io_iov_buffer_select_prep() 76 int io_prep_rw(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_prep_rw() argument 78 struct io_rw *rw = io_kiocb_to_cmd(req, struct io_rw); in io_prep_rw() 84 req->buf_index = READ_ONCE(sqe->buf_index); in io_prep_rw() 86 if (req->opcode == IORING_OP_READ_FIXED || in io_prep_rw() 87 req->opcode == IORING_OP_WRITE_FIXED) { in io_prep_rw() [all …]
|
D | kbuf.h | 44 void __user *io_buffer_select(struct io_kiocb *req, size_t *len, 48 int io_remove_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 49 int io_remove_buffers(struct io_kiocb *req, unsigned int issue_flags); 51 int io_provide_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 52 int io_provide_buffers(struct io_kiocb *req, unsigned int issue_flags); 59 unsigned int __io_put_kbuf(struct io_kiocb *req, unsigned issue_flags); 61 void io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags); 65 static inline void io_kbuf_recycle_ring(struct io_kiocb *req) in io_kbuf_recycle_ring() argument 74 if (req->buf_list) { in io_kbuf_recycle_ring() 75 if (req->flags & REQ_F_PARTIAL_IO) { in io_kbuf_recycle_ring() [all …]
|
D | timeout.c | 38 static inline bool io_is_timeout_noseq(struct io_kiocb *req) in io_is_timeout_noseq() argument 40 struct io_timeout *timeout = io_kiocb_to_cmd(req, struct io_timeout); in io_is_timeout_noseq() 41 struct io_timeout_data *data = req->async_data; in io_is_timeout_noseq() 46 static inline void io_put_req(struct io_kiocb *req) in io_put_req() argument 48 if (req_ref_put_and_test(req)) { in io_put_req() 49 io_queue_next(req); in io_put_req() 50 io_free_req(req); in io_put_req() 68 static void io_timeout_complete(struct io_kiocb *req, struct io_tw_state *ts) in io_timeout_complete() argument 70 struct io_timeout *timeout = io_kiocb_to_cmd(req, struct io_timeout); in io_timeout_complete() 71 struct io_timeout_data *data = req->async_data; in io_timeout_complete() [all …]
|
/linux-6.6.21/net/handshake/ |
D | request.c | 67 static bool handshake_req_hash_add(struct handshake_req *req) in handshake_req_hash_add() argument 72 &req->hr_rhash, in handshake_req_hash_add() 77 static void handshake_req_destroy(struct handshake_req *req) in handshake_req_destroy() argument 79 if (req->hr_proto->hp_destroy) in handshake_req_destroy() 80 req->hr_proto->hp_destroy(req); in handshake_req_destroy() 81 rhashtable_remove_fast(&handshake_rhashtbl, &req->hr_rhash, in handshake_req_destroy() 83 kfree(req); in handshake_req_destroy() 89 struct handshake_req *req; in handshake_sk_destruct() local 91 req = handshake_req_hash_lookup(sk); in handshake_sk_destruct() 92 if (!req) in handshake_sk_destruct() [all …]
|
/linux-6.6.21/drivers/net/ethernet/marvell/octeontx2/af/ |
D | mcs_rvu_if.c | 21 struct _req_type *req; \ 23 req = (struct _req_type *)otx2_mbox_alloc_msg_rsp( \ 26 if (!req) \ 28 req->hdr.sig = OTX2_MBOX_REQ_SIG; \ 29 req->hdr.id = _id; \ 30 return req; \ 73 struct mcs_set_lmac_mode *req, in rvu_mbox_handler_mcs_set_lmac_mode() argument 78 if (req->mcs_id >= rvu->mcs_blk_cnt) in rvu_mbox_handler_mcs_set_lmac_mode() 81 mcs = mcs_get_pdata(req->mcs_id); in rvu_mbox_handler_mcs_set_lmac_mode() 83 if (BIT_ULL(req->lmac_id) & mcs->hw->lmac_bmap) in rvu_mbox_handler_mcs_set_lmac_mode() [all …]
|
/linux-6.6.21/drivers/infiniband/hw/hfi1/ |
D | user_sdma.c | 38 static int user_sdma_send_pkts(struct user_sdma_request *req, u16 maxpkts); 41 static void user_sdma_free_request(struct user_sdma_request *req); 42 static int check_header_template(struct user_sdma_request *req, 45 static int set_txreq_header(struct user_sdma_request *req, 47 static int set_txreq_header_ahg(struct user_sdma_request *req, 290 struct user_sdma_request *req; in hfi1_user_sdma_process_request() local 297 if (iovec[idx].iov_len < sizeof(info) + sizeof(req->hdr)) { in hfi1_user_sdma_process_request() 302 iovec[idx].iov_len, sizeof(info) + sizeof(req->hdr)); in hfi1_user_sdma_process_request() 352 req = pq->reqs + info.comp_idx; in hfi1_user_sdma_process_request() 353 req->data_iovs = req_iovcnt(info.ctrl) - 1; /* subtract header vector */ in hfi1_user_sdma_process_request() [all …]
|
/linux-6.6.21/net/sunrpc/ |
D | backchannel_rqst.c | 40 static void xprt_free_allocation(struct rpc_rqst *req) in xprt_free_allocation() argument 44 dprintk("RPC: free allocations for req= %p\n", req); in xprt_free_allocation() 45 WARN_ON_ONCE(test_bit(RPC_BC_PA_IN_USE, &req->rq_bc_pa_state)); in xprt_free_allocation() 46 xbufp = &req->rq_rcv_buf; in xprt_free_allocation() 48 xbufp = &req->rq_snd_buf; in xprt_free_allocation() 50 kfree(req); in xprt_free_allocation() 78 struct rpc_rqst *req; in xprt_alloc_bc_req() local 81 req = kzalloc(sizeof(*req), gfp_flags); in xprt_alloc_bc_req() 82 if (req == NULL) in xprt_alloc_bc_req() 85 req->rq_xprt = xprt; in xprt_alloc_bc_req() [all …]
|
/linux-6.6.21/drivers/peci/ |
D | request.c | 88 static u8 peci_request_data_cc(struct peci_request *req) in peci_request_data_cc() argument 90 return req->rx.buf[0]; in peci_request_data_cc() 102 int peci_request_status(struct peci_request *req) in peci_request_status() argument 104 u8 cc = peci_request_data_cc(req); in peci_request_status() 107 dev_dbg(&req->device->dev, "ret: %#02x\n", cc); in peci_request_status() 133 static int peci_request_xfer(struct peci_request *req) in peci_request_xfer() argument 135 struct peci_device *device = req->device; in peci_request_xfer() 140 ret = controller->ops->xfer(controller, device->addr, req); in peci_request_xfer() 146 static int peci_request_xfer_retry(struct peci_request *req) in peci_request_xfer_retry() argument 149 struct peci_device *device = req->device; in peci_request_xfer_retry() [all …]
|
/linux-6.6.21/crypto/ |
D | chacha20poly1305.c | 42 struct ahash_request req; /* must be last member */ member 48 struct skcipher_request req; /* must be last member */ member 70 static inline void async_done_continue(struct aead_request *req, int err, in async_done_continue() argument 74 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); in async_done_continue() 77 err = cont(req); in async_done_continue() 81 aead_request_complete(req, err); in async_done_continue() 84 static void chacha_iv(u8 *iv, struct aead_request *req, u32 icb) in chacha_iv() argument 86 struct chachapoly_ctx *ctx = crypto_aead_ctx(crypto_aead_reqtfm(req)); in chacha_iv() 91 memcpy(iv + sizeof(leicb) + ctx->saltlen, req->iv, in chacha_iv() 95 static int poly_verify_tag(struct aead_request *req) in poly_verify_tag() argument [all …]
|
D | seqiv.c | 21 static void seqiv_aead_encrypt_complete2(struct aead_request *req, int err) in seqiv_aead_encrypt_complete2() argument 23 struct aead_request *subreq = aead_request_ctx(req); in seqiv_aead_encrypt_complete2() 32 geniv = crypto_aead_reqtfm(req); in seqiv_aead_encrypt_complete2() 33 memcpy(req->iv, subreq->iv, crypto_aead_ivsize(geniv)); in seqiv_aead_encrypt_complete2() 41 struct aead_request *req = data; in seqiv_aead_encrypt_complete() local 43 seqiv_aead_encrypt_complete2(req, err); in seqiv_aead_encrypt_complete() 44 aead_request_complete(req, err); in seqiv_aead_encrypt_complete() 47 static int seqiv_aead_encrypt(struct aead_request *req) in seqiv_aead_encrypt() argument 49 struct crypto_aead *geniv = crypto_aead_reqtfm(req); in seqiv_aead_encrypt() 51 struct aead_request *subreq = aead_request_ctx(req); in seqiv_aead_encrypt() [all …]
|
/linux-6.6.21/drivers/clk/sunxi/ |
D | clk-sunxi.c | 33 static void sun4i_get_pll1_factors(struct factors_request *req) in sun4i_get_pll1_factors() argument 38 div = req->rate / 6000000; in sun4i_get_pll1_factors() 39 req->rate = 6000000 * div; in sun4i_get_pll1_factors() 42 req->m = 0; in sun4i_get_pll1_factors() 45 if (req->rate >= 768000000 || req->rate == 42000000 || in sun4i_get_pll1_factors() 46 req->rate == 54000000) in sun4i_get_pll1_factors() 47 req->k = 1; in sun4i_get_pll1_factors() 49 req->k = 0; in sun4i_get_pll1_factors() 53 req->p = 3; in sun4i_get_pll1_factors() 57 req->p = 2; in sun4i_get_pll1_factors() [all …]
|