Symbol: rq
function parameter
Defined...
-
block/bfq-cgroup.c:350:59-350:75: void bfqg_stats_update_legacy_io(struct request_queue *q, struct request *rq)
-
block/bfq-iosched.c:1069:41-1069:57: static unsigned long bfq_serv_to_charge(struct request *rq,
-
block/bfq-iosched.c:1820:11-1820:27: struct request *rq,
-
block/bfq-iosched.c:2200:29-2200:45: static void bfq_add_request(struct request *rq)
-
block/bfq-iosched.c:2368:46-2368:62: static sector_t get_sdist(sector_t last_pos, struct request *rq)
-
block/bfq-iosched.c:2393:11-2393:27: struct request *rq)
-
block/bfq-iosched.c:2571:58-2571:74: static void bfq_requests_merged(struct request_queue *q, struct request *rq,
-
block/bfq-iosched.c:3207:58-3207:74: static bool bfq_allow_bio_merge(struct request_queue *q, struct request *rq,
-
block/bfq-iosched.c:3413:12-3413:28: struct request *rq)
-
block/bfq-iosched.c:3430:58-3430:74: static void bfq_update_rate_reset(struct bfq_data *bfqd, struct request *rq)
-
block/bfq-iosched.c:3571:57-3571:73: static void bfq_update_peak_rate(struct bfq_data *bfqd, struct request *rq)
-
block/bfq-iosched.c:3633:58-3633:74: static void bfq_dispatch_remove(struct request_queue *q, struct request *rq)
-
block/bfq-iosched.c:5169:11-5169:27: struct request *rq,
-
block/bfq-iosched.c:5816:10-5816:26: struct request *rq)
-
block/bfq-iosched.c:5978:8-5978:24: struct request *rq)
-
block/bfq-iosched.c:6050:57-6050:73: static bool __bfq_insert_request(struct bfq_data *bfqd, struct request *rq)
-
block/bfq-iosched.c:6137:60-6137:76: static void bfq_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq,
-
block/bfq-iosched.c:6544:40-6544:56: static void bfq_finish_requeue_request(struct request *rq)
-
block/bfq-iosched.c:6599:32-6599:48: static void bfq_finish_request(struct request *rq)
-
block/bfq-iosched.c:6704:33-6704:49: static void bfq_prepare_request(struct request *rq)
-
block/bfq-iosched.c:6739:38-6739:54: static struct bfq_queue *bfq_init_rq(struct request *rq)
-
block/blk-cgroup.h:491:41-491:57: static inline bool blk_cgroup_mergeable(struct request *rq, struct bio *bio)
-
block/blk-crypto-internal.h:57:47-57:63: static inline void blk_crypto_rq_set_defaults(struct request *rq)
-
block/blk-crypto-internal.h:63:47-63:63: static inline bool blk_crypto_rq_is_encrypted(struct request *rq)
-
block/blk-crypto-internal.h:124:45-124:61: static inline void bio_crypt_do_front_merge(struct request *rq,
-
block/blk-crypto-internal.h:143:52-143:68: static inline blk_status_t blk_crypto_init_request(struct request *rq)
-
block/blk-crypto-internal.h:151:44-151:60: static inline void blk_crypto_free_request(struct request *rq)
-
block/blk-crypto-internal.h:169:42-169:58: static inline int blk_crypto_rq_bio_prep(struct request *rq, struct bio *bio,
-
block/blk-crypto-internal.h:184:61-184:77: static inline blk_status_t blk_crypto_insert_cloned_request(struct request *rq)
-
block/blk-crypto.c:186:34-186:50: bool bio_crypt_rq_ctx_compatible(struct request *rq, struct bio *bio)
-
block/blk-crypto.c:221:40-221:56: blk_status_t __blk_crypto_init_request(struct request *rq)
-
block/blk-crypto.c:237:32-237:48: void __blk_crypto_free_request(struct request *rq)
-
block/blk-crypto.c:298:30-298:46: int __blk_crypto_rq_bio_prep(struct request *rq, struct bio *bio,
-
block/blk-flush.c:105:60-105:76: static unsigned int blk_flush_policy(unsigned long fflags, struct request *rq)
-
block/blk-flush.c:122:39-122:55: static unsigned int blk_flush_cur_seq(struct request *rq)
-
block/blk-flush.c:127:39-127:55: static void blk_flush_restore_request(struct request *rq)
-
block/blk-flush.c:141:32-141:48: static void blk_flush_queue_rq(struct request *rq, bool add_front)
-
block/blk-flush.c:146:34-146:50: static void blk_account_io_flush(struct request *rq)
-
block/blk-flush.c:170:36-170:52: static void blk_flush_complete_seq(struct request *rq,
-
block/blk-flush.c:274:18-274:34: bool is_flush_rq(struct request *rq)
-
block/blk-flush.c:357:34-357:50: static void mq_flush_data_end_io(struct request *rq, blk_status_t error)
-
block/blk-flush.c:390:23-390:39: void blk_insert_flush(struct request *rq)
-
block/blk-integrity.c:369:39-369:55: static void blk_integrity_nop_prepare(struct request *rq)
-
block/blk-integrity.c:373:40-373:56: static void blk_integrity_nop_complete(struct request *rq,
-
block/blk-iocost.c:2535:42-2535:58: static void calc_size_vtime_cost_builtin(struct request *rq, struct ioc *ioc,
-
block/blk-iocost.c:2552:33-2552:49: static u64 calc_size_vtime_cost(struct request *rq, struct ioc *ioc)
-
block/blk-iocost.c:2695:49-2695:65: static void ioc_rqos_merge(struct rq_qos *rqos, struct request *rq,
-
block/blk-iocost.c:2762:48-2762:64: static void ioc_rqos_done(struct rq_qos *rqos, struct request *rq)
-
block/blk-ioprio.c:190:53-190:69: static void blkcg_ioprio_track(struct rq_qos *rqos, struct request *rq,
-
block/blk-map.c:129:30-129:46: static int bio_copy_user_iov(struct request *rq, struct rq_map_data *map_data,
-
block/blk-map.c:234:29-234:45: static int bio_map_user_iov(struct request *rq, struct iov_iter *iter,
-
block/blk-map.c:490:23-490:39: int blk_rq_append_bio(struct request *rq, struct bio *bio)
-
block/blk-map.c:529:50-529:66: int blk_rq_map_user_iov(struct request_queue *q, struct request *rq,
-
block/blk-map.c:573:46-573:62: int blk_rq_map_user(struct request_queue *q, struct request *rq,
-
block/blk-map.c:634:46-634:62: int blk_rq_map_kern(struct request_queue *q, struct request *rq, void *kbuf,
-
block/blk-merge.c:375:37-375:53: unsigned int blk_recalc_rq_segments(struct request *rq)
-
block/blk-merge.c:528:46-528:62: int __blk_rq_map_sg(struct request_queue *q, struct request *rq,
-
block/blk-merge.c:551:52-551:68: static inline unsigned int blk_rq_get_max_segments(struct request *rq)
-
block/blk-merge.c:558:51-558:67: static inline unsigned int blk_rq_get_max_sectors(struct request *rq,
-
block/blk-merge.c:700:29-700:45: void blk_rq_set_mixed_merge(struct request *rq)
-
block/blk-merge.c:826:3-826:19: struct request *rq)
-
block/blk-merge.c:837:3-837:19: struct request *rq)
-
block/blk-merge.c:852:53-852:69: bool blk_attempt_req_merge(struct request_queue *q, struct request *rq,
-
block/blk-merge.c:858:22-858:38: bool blk_rq_merge_ok(struct request *rq, struct bio *bio)
-
block/blk-merge.c:888:30-888:46: enum elv_merge blk_try_merge(struct request *rq, struct bio *bio)
-
block/blk-merge.c:991:10-991:26: struct request *rq,
-
block/blk-mq-debugfs.c:306:50-306:66: int __blk_mq_debugfs_rq_show(struct seq_file *m, struct request *rq)
-
block/blk-mq-debugfs.c:380:31-380:47: static bool hctx_show_busy_rq(struct request *rq, void *data, bool reserved)
-
block/blk-mq-sched.c:381:61-381:77: bool blk_mq_sched_try_insert_merge(struct request_queue *q, struct request *rq,
-
block/blk-mq-sched.c:389:12-389:28: struct request *rq)
-
block/blk-mq-sched.c:408:34-408:50: void blk_mq_sched_insert_request(struct request *rq, bool at_head,
-
block/blk-mq-sched.h:44:51-44:67: blk_mq_sched_allow_merge(struct request_queue *q, struct request *rq,
-
block/blk-mq-sched.h:56:51-56:67: static inline void blk_mq_sched_completed_request(struct request *rq, u64 now)
-
block/blk-mq-sched.h:66:49-66:65: static inline void blk_mq_sched_requeue_request(struct request *rq)
-
block/blk-mq-tag.c:451:47-451:63: static bool blk_mq_tagset_count_completed_rqs(struct request *rq,
-
block/blk-mq-tag.c:678:23-678:39: u32 blk_mq_unique_tag(struct request *rq)
-
block/blk-mq.c:51:34-51:56: static int blk_mq_poll_stats_bkt(const struct request *rq)
-
block/blk-mq.c:88:37-88:53: static inline blk_qc_t blk_rq_to_qc(struct request *rq)
-
block/blk-mq.c:131:35-131:51: static bool blk_mq_check_inflight(struct request *rq, void *priv,
-
block/blk-mq.c:324:43-324:59: void blk_rq_init(struct request_queue *q, struct request *rq)
-
block/blk-mq.c:604:35-604:51: static void __blk_mq_free_request(struct request *rq)
-
block/blk-mq.c:622:26-622:42: void blk_mq_free_request(struct request *rq)
-
block/blk-mq.c:653:24-653:40: void blk_dump_rq_flags(struct request *rq, char *msg)
-
block/blk-mq.c:667:27-667:43: static void req_bio_endio(struct request *rq, struct bio *bio,
-
block/blk-mq.c:893:36-893:52: static void __blk_account_io_start(struct request *rq)
-
block/blk-mq.c:917:46-917:62: static inline void __blk_mq_end_request_acct(struct request *rq, u64 now)
-
block/blk-mq.c:928:34-928:50: inline void __blk_mq_end_request(struct request *rq, blk_status_t error)
-
block/blk-mq.c:942:25-942:41: void blk_mq_end_request(struct request *rq, blk_status_t error)
-
block/blk-mq.c:1034:45-1034:61: static inline bool blk_mq_complete_need_ipi(struct request *rq)
-
block/blk-mq.c:1060:38-1060:54: static void blk_mq_complete_send_ipi(struct request *rq)
-
block/blk-mq.c:1073:34-1073:50: static void blk_mq_raise_softirq(struct request *rq)
-
block/blk-mq.c:1084:37-1084:53: bool blk_mq_complete_request_remote(struct request *rq)
-
block/blk-mq.c:1115:30-1115:46: void blk_mq_complete_request(struct request *rq)
-
block/blk-mq.c:1130:27-1130:43: void blk_mq_start_request(struct request *rq)
-
block/blk-mq.c:1169:55-1169:71: static void blk_add_rq_to_plug(struct blk_plug *plug, struct request *rq)
-
block/blk-mq.c:1203:28-1203:44: void blk_execute_rq_nowait(struct request *rq, bool at_head)
-
block/blk-mq.c:1221:29-1221:45: static void blk_end_sync_rq(struct request *rq, blk_status_t ret)
-
block/blk-mq.c:1229:28-1229:44: static bool blk_rq_is_poll(struct request *rq)
-
block/blk-mq.c:1240:36-1240:52: static void blk_rq_poll_completion(struct request *rq, struct completion *wait)
-
block/blk-mq.c:1258:29-1258:45: blk_status_t blk_execute_rq(struct request *rq, bool at_head)
-
block/blk-mq.c:1294:38-1294:54: static void __blk_mq_requeue_request(struct request *rq)
-
block/blk-mq.c:1309:29-1309:45: void blk_mq_requeue_request(struct request *rq, bool kick_requeue_list)
-
block/blk-mq.c:1357:33-1357:49: void blk_mq_add_to_requeue_list(struct request *rq, bool at_head,
-
block/blk-mq.c:1396:32-1396:48: static bool blk_mq_rq_inflight(struct request *rq, void *priv,
-
block/blk-mq.c:1438:32-1438:48: static bool blk_mq_req_expired(struct request *rq, unsigned long *next)
-
block/blk-mq.c:1458:24-1458:40: void blk_mq_put_rq_ref(struct request *rq)
-
block/blk-mq.c:1466:34-1466:50: static bool blk_mq_check_expired(struct request *rq, void *priv, bool reserved)
-
block/blk-mq.c:1600:39-1600:55: static bool __blk_mq_alloc_driver_tag(struct request *rq)
-
block/blk-mq.c:1624:58-1624:74: bool __blk_mq_get_driver_tag(struct blk_mq_hw_ctx *hctx, struct request *rq)
-
block/blk-mq.c:1666:6-1666:22: struct request *rq)
-
block/blk-mq.c:1757:40-1757:56: static void blk_mq_handle_dev_resource(struct request *rq,
-
block/blk-mq.c:1774:41-1774:57: static void blk_mq_handle_zone_resource(struct request *rq,
-
block/blk-mq.c:1793:51-1793:67: static enum prep_dispatch blk_mq_prep_dispatch_rq(struct request *rq,
-
block/blk-mq.c:2340:10-2340:26: struct request *rq,
-
block/blk-mq.c:2356:58-2356:74: void __blk_mq_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq,
-
block/blk-mq.c:2376:35-2376:51: void blk_mq_request_bypass_insert(struct request *rq, bool at_head,
-
block/blk-mq.c:2424:35-2424:51: static void blk_mq_bio_to_request(struct request *rq, struct bio *bio,
-
block/blk-mq.c:2443:10-2443:26: struct request *rq, bool last)
-
block/blk-mq.c:2476:7-2476:23: struct request *rq,
-
block/blk-mq.c:2531:3-2531:19: struct request *rq)
-
block/blk-mq.c:2542:51-2542:67: static blk_status_t blk_mq_request_issue_directly(struct request *rq, bool last)
-
block/blk-mq.c:2859:40-2859:56: blk_status_t blk_insert_cloned_request(struct request *rq)
-
block/blk-mq.c:2923:26-2923:42: void blk_rq_unprep_clone(struct request *rq)
-
block/blk-mq.c:2952:23-2952:39: int blk_rq_prep_clone(struct request *rq, struct request *rq_src,
-
block/blk-mq.c:3009:44-3009:60: void blk_steal_bios(struct bio_list *list, struct request *rq)
-
block/blk-mq.c:3182:60-3182:76: static int blk_mq_init_request(struct blk_mq_tag_set *set, struct request *rq,
-
block/blk-mq.c:3279:32-3279:48: static bool blk_mq_has_request(struct request *rq, void *data, bool reserved)
-
block/blk-mq.c:4600:12-4600:28: struct request *rq)
-
block/blk-mq.c:4736:28-4736:44: unsigned int blk_mq_rq_cpu(struct request *rq)
-
block/blk-mq.h:207:47-207:63: static inline void blk_mq_set_rq_budget_token(struct request *rq, int token)
-
block/blk-mq.h:216:46-216:62: static inline int blk_mq_get_rq_budget_token(struct request *rq)
-
block/blk-mq.h:252:9-252:25: struct request *rq)
-
block/blk-mq.h:263:42-263:58: static inline void blk_mq_put_driver_tag(struct request *rq)
-
block/blk-mq.h:273:42-273:58: static inline bool blk_mq_get_driver_tag(struct request *rq)
-
block/blk-pm.h:19:42-19:58: static inline void blk_pm_mark_last_busy(struct request *rq)
-
block/blk-rq-qos.c:41:41-41:57: void __rq_qos_done(struct rq_qos *rqos, struct request *rq)
-
block/blk-rq-qos.c:50:42-50:58: void __rq_qos_issue(struct rq_qos *rqos, struct request *rq)
-
block/blk-rq-qos.c:59:44-59:60: void __rq_qos_requeue(struct rq_qos *rqos, struct request *rq)
-
block/blk-rq-qos.c:77:42-77:58: void __rq_qos_track(struct rq_qos *rqos, struct request *rq, struct bio *bio)
-
block/blk-rq-qos.c:86:42-86:58: void __rq_qos_merge(struct rq_qos *rqos, struct request *rq, struct bio *bio)
-
block/blk-rq-qos.h:167:57-167:73: static inline void rq_qos_done(struct request_queue *q, struct request *rq)
-
block/blk-rq-qos.h:173:58-173:74: static inline void rq_qos_issue(struct request_queue *q, struct request *rq)
-
block/blk-rq-qos.h:179:60-179:76: static inline void rq_qos_requeue(struct request_queue *q, struct request *rq)
-
block/blk-rq-qos.h:203:58-203:74: static inline void rq_qos_track(struct request_queue *q, struct request *rq,
-
block/blk-rq-qos.h:210:58-210:74: static inline void rq_qos_merge(struct request_queue *q, struct request *rq,
-
block/blk-stat.c:51:19-51:35: void blk_stat_add(struct request *rq, u64 now)
-
block/blk-throttle.c:2224:26-2224:42: void blk_throtl_stat_add(struct request *rq, u64 time_ns)
-
block/blk-wbt.c:34:36-34:52: static inline void wbt_clear_state(struct request *rq)
-
block/blk-wbt.c:39:40-39:56: static inline enum wbt_flags wbt_flags(struct request *rq)
-
block/blk-wbt.c:44:35-44:51: static inline bool wbt_is_tracked(struct request *rq)
-
block/blk-wbt.c:49:32-49:48: static inline bool wbt_is_read(struct request *rq)
-
block/blk-wbt.c:186:43-186:59: static void wbt_done(struct rq_qos *rqos, struct request *rq)
-
block/blk-wbt.c:592:44-592:60: static void wbt_track(struct rq_qos *rqos, struct request *rq, struct bio *bio)
-
block/blk-wbt.c:598:44-598:60: static void wbt_issue(struct rq_qos *rqos, struct request *rq)
-
block/blk-wbt.c:618:46-618:62: static void wbt_requeue(struct rq_qos *rqos, struct request *rq)
-
block/blk-wbt.c:671:25-671:47: static int wbt_data_dir(const struct request *rq)
-
block/blk-zoned.c:58:36-58:52: bool blk_req_needs_zone_write_lock(struct request *rq)
-
block/blk-zoned.c:76:33-76:49: bool blk_req_zone_write_trylock(struct request *rq)
-
block/blk-zoned.c:90:32-90:48: void __blk_req_zone_write_lock(struct request *rq)
-
block/blk-zoned.c:101:34-101:50: void __blk_req_zone_write_unlock(struct request *rq)
-
block/blk.h:124:33-124:49: static inline bool rq_mergeable(struct request *rq)
-
block/blk.h:325:35-325:51: static inline bool blk_do_io_stat(struct request *rq)
-
block/bsg-lib.c:205:26-205:42: static void bsg_complete(struct request *rq)
-
block/bsg-lib.c:334:45-334:61: static enum blk_eh_timer_return bsg_timeout(struct request *rq, bool reserved)
-
block/elevator.c:60:40-60:56: static int elv_iosched_allow_bio_merge(struct request *rq, struct bio *bio)
-
block/elevator.c:74:23-74:39: bool elv_bio_merge_ok(struct request *rq, struct bio *bio)
-
block/elevator.c:205:37-205:53: static inline void __elv_rqhash_del(struct request *rq)
-
block/elevator.c:211:46-211:62: void elv_rqhash_del(struct request_queue *q, struct request *rq)
-
block/elevator.c:218:46-218:62: void elv_rqhash_add(struct request_queue *q, struct request *rq)
-
block/elevator.c:228:53-228:69: void elv_rqhash_reposition(struct request_queue *q, struct request *rq)
-
block/elevator.c:259:39-259:55: void elv_rb_add(struct rb_root *root, struct request *rq)
-
block/elevator.c:280:39-280:55: void elv_rb_del(struct rb_root *root, struct request *rq)
-
block/elevator.c:364:56-364:72: bool elv_attempt_insert_merge(struct request_queue *q, struct request *rq,
-
block/elevator.c:402:50-402:66: void elv_merged_request(struct request_queue *q, struct request *rq,
-
block/elevator.c:416:50-416:66: void elv_merge_requests(struct request_queue *q, struct request *rq,
-
block/elevator.c:428:61-428:77: struct request *elv_latter_request(struct request_queue *q, struct request *rq)
-
block/elevator.c:438:61-438:77: struct request *elv_former_request(struct request_queue *q, struct request *rq)
-
block/elevator.c:829:11-829:27: struct request *rq)
-
block/elevator.c:841:11-841:27: struct request *rq)
-
block/kyber-iosched.c:532:32-532:48: static int rq_get_domain_token(struct request *rq)
-
block/kyber-iosched.c:537:33-537:49: static void rq_set_domain_token(struct request *rq, int token)
-
block/kyber-iosched.c:543:7-543:23: struct request *rq)
-
block/kyber-iosched.c:587:35-587:51: static void kyber_prepare_request(struct request *rq)
-
block/kyber-iosched.c:615:34-615:50: static void kyber_finish_request(struct request *rq)
-
block/kyber-iosched.c:640:37-640:53: static void kyber_completed_request(struct request *rq, u64 now)
-
block/mq-deadline.c:119:48-119:64: deadline_rb_root(struct dd_per_prio *per_prio, struct request *rq)
-
block/mq-deadline.c:128:25-128:41: static u8 dd_rq_ioclass(struct request *rq)
-
block/mq-deadline.c:137:25-137:41: deadline_latter_request(struct request *rq)
-
block/mq-deadline.c:148:50-148:66: deadline_add_rq_rb(struct dd_per_prio *per_prio, struct request *rq)
-
block/mq-deadline.c:156:50-156:66: deadline_del_rq_rb(struct dd_per_prio *per_prio, struct request *rq)
-
block/mq-deadline.c:171:9-171:25: struct request *rq)
-
block/mq-deadline.c:240:9-240:25: struct request *rq)
-
block/mq-deadline.c:351:53-351:69: static bool started_after(struct deadline_data *dd, struct request *rq,
-
block/mq-deadline.c:660:54-660:71: static int dd_request_merge(struct request_queue *q, struct request **rq,
-
block/mq-deadline.c:712:59-712:75: static void dd_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq,
-
block/mq-deadline.c:787:32-787:48: static void dd_prepare_request(struct request *rq)
-
block/mq-deadline.c:808:31-808:47: static void dd_finish_request(struct request *rq)
-
block/t10-pi.c:135:34-135:50: static void t10_pi_type1_prepare(struct request *rq)
-
block/t10-pi.c:184:35-184:51: static void t10_pi_type1_complete(struct request *rq, unsigned int nr_bytes)
-
block/t10-pi.c:238:34-238:50: static void t10_pi_type3_prepare(struct request *rq)
-
block/t10-pi.c:243:35-243:51: static void t10_pi_type3_complete(struct request *rq, unsigned int nr_bytes)
-
block/t10-pi.c:374:34-374:50: static void ext_pi_type1_prepare(struct request *rq)
-
block/t10-pi.c:412:35-412:51: static void ext_pi_type1_complete(struct request *rq, unsigned int nr_bytes)
-
drivers/ata/libata-scsi.c:1047:30-1047:46: bool ata_scsi_dma_need_drain(struct request *rq)
-
drivers/block/aoe/aoecmd.c:825:26-825:42: bufinit(struct buf *buf, struct request *rq, struct bio *bio)
-
drivers/block/aoe/aoecmd.c:1029:35-1029:51: aoe_end_request(struct aoedev *d, struct request *rq, int fastfail)
-
drivers/block/loop.c:263:52-263:68: static int lo_write_simple(struct loop_device *lo, struct request *rq,
-
drivers/block/loop.c:280:51-280:67: static int lo_read_simple(struct loop_device *lo, struct request *rq,
-
drivers/block/loop.c:309:49-309:65: static int lo_fallocate(struct loop_device *lo, struct request *rq, loff_t pos,
-
drivers/block/loop.c:330:49-330:65: static int lo_req_flush(struct loop_device *lo, struct request *rq)
-
drivers/block/loop.c:339:28-339:44: static void lo_complete_rq(struct request *rq)
-
drivers/block/loop.c:463:54-463:70: static int do_req_filebacked(struct loop_device *lo, struct request *rq)
-
drivers/block/mtip32xx/mtip32xx.c:2053:55-2053:71: static void mtip_hw_submit_io(struct driver_data *dd, struct request *rq,
-
drivers/block/mtip32xx/mtip32xx.c:2544:34-2544:50: static void mtip_softirq_done_fn(struct request *rq)
-
drivers/block/mtip32xx/mtip32xx.c:3359:55-3359:71: static inline bool is_stopped(struct driver_data *dd, struct request *rq)
-
drivers/block/mtip32xx/mtip32xx.c:3380:7-3380:23: struct request *rq)
-
drivers/block/mtip32xx/mtip32xx.c:3404:3-3404:19: struct request *rq)
-
drivers/block/mtip32xx/mtip32xx.c:3462:55-3462:71: static void mtip_free_cmd(struct blk_mq_tag_set *set, struct request *rq,
-
drivers/block/mtip32xx/mtip32xx.c:3475:54-3475:70: static int mtip_init_cmd(struct blk_mq_tag_set *set, struct request *rq,
-
drivers/block/mtip32xx/mtip32xx.c:3676:33-3676:49: static bool mtip_no_dev_cleanup(struct request *rq, void *data, bool reserv)
-
drivers/block/nbd.c:1722:57-1722:73: static int nbd_init_request(struct blk_mq_tag_set *set, struct request *rq,
-
drivers/block/null_blk/main.c:776:30-776:46: static void null_complete_rq(struct request *rq)
-
drivers/block/null_blk/main.c:1482:36-1482:52: static bool should_timeout_request(struct request *rq)
-
drivers/block/null_blk/main.c:1491:36-1491:52: static bool should_requeue_request(struct request *rq)
-
drivers/block/null_blk/main.c:1581:49-1581:65: static enum blk_eh_timer_return null_timeout_rq(struct request *rq, bool res)
-
drivers/block/rnbd/rnbd-clt.c:390:34-390:50: static void rnbd_softirq_done_fn(struct request *rq)
-
drivers/block/rnbd/rnbd-clt.c:1006:10-1006:26: struct request *rq,
-
drivers/block/rnbd/rnbd-proto.h:267:36-267:52: static inline u32 rq_to_rnbd_flags(struct request *rq)
-
drivers/block/sx8.c:698:51-698:67: static inline enum dma_data_direction carm_rq_dir(struct request *rq)
-
drivers/block/xen-blkfront.c:122:43-122:59: static inline struct blkif_req *blkif_req(struct request *rq)
-
drivers/block/xen-blkfront.c:925:31-925:47: static void blkif_complete_rq(struct request *rq)
-
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c:2170:40-2170:61: static int i915_reset_gen7_sol_offsets(struct i915_request *rq)
-
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c:2334:9-2334:30: struct i915_request *rq,
-
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c:2945:5-2945:26: struct i915_request *rq)
-
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c:3017:55-3017:76: static int eb_request_add(struct i915_execbuffer *eb, struct i915_request *rq,
-
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c:3180:43-3180:64: eb_fences_add(struct i915_execbuffer *eb, struct i915_request *rq,
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c:304:6-304:27: struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/gen2_engine_cs.c:16:21-16:42: int gen2_emit_flush(struct i915_request *rq, u32 mode)
-
drivers/gpu/drm/i915/gt/gen2_engine_cs.c:43:25-43:46: int gen4_emit_flush_rcs(struct i915_request *rq, u32 mode)
-
drivers/gpu/drm/i915/gt/gen2_engine_cs.c:129:25-129:46: int gen4_emit_flush_vcs(struct i915_request *rq, u32 mode)
-
drivers/gpu/drm/i915/gt/gen2_engine_cs.c:144:36-144:57: static u32 *__gen2_emit_breadcrumb(struct i915_request *rq, u32 *cs,
-
drivers/gpu/drm/i915/gt/gen2_engine_cs.c:172:27-172:48: u32 *gen3_emit_breadcrumb(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen2_engine_cs.c:177:27-177:48: u32 *gen5_emit_breadcrumb(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen2_engine_cs.c:186:24-186:45: int i830_emit_bb_start(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/gen2_engine_cs.c:251:24-251:45: int gen3_emit_bb_start(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/gen2_engine_cs.c:271:24-271:45: int gen4_emit_bb_start(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:55:35-55:56: gen6_emit_post_sync_nonzero_flush(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:89:25-89:46: int gen6_emit_flush_rcs(struct i915_request *rq, u32 mode)
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:143:31-143:52: u32 *gen6_emit_breadcrumb_rcs(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:178:24-178:45: static int mi_flush_dw(struct i915_request *rq, u32 flags)
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:214:26-214:47: static int gen6_flush_dw(struct i915_request *rq, u32 mode, u32 invflags)
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:219:25-219:46: int gen6_emit_flush_xcs(struct i915_request *rq, u32 mode)
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:224:25-224:46: int gen6_emit_flush_vcs(struct i915_request *rq, u32 mode)
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:229:24-229:45: int gen6_emit_bb_start(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:251:19-251:40: hsw_emit_bb_start(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:272:26-272:47: static int gen7_stall_cs(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:289:25-289:46: int gen7_emit_flush_rcs(struct i915_request *rq, u32 mode)
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:353:31-353:52: u32 *gen7_emit_breadcrumb_rcs(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:375:31-375:52: u32 *gen6_emit_breadcrumb_xcs(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen6_engine_cs.c:393:31-393:52: u32 *gen7_emit_breadcrumb_xcs(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:13:25-13:46: int gen8_emit_flush_rcs(struct i915_request *rq, u32 mode)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:79:25-79:46: int gen8_emit_flush_xcs(struct i915_request *rq, u32 mode)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:112:26-112:47: int gen11_emit_flush_rcs(struct i915_request *rq, u32 mode)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:178:26-178:47: int gen12_emit_flush_rcs(struct i915_request *rq, u32 mode)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:263:26-263:47: int gen12_emit_flush_xcs(struct i915_request *rq, u32 mode)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:330:24-330:51: static u32 hwsp_offset(const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:342:31-342:52: int gen8_emit_init_breadcrumb(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:389:35-389:56: static int __gen125_emit_bb_start(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:428:32-428:53: int gen125_emit_bb_start_noarb(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:435:26-435:47: int gen125_emit_bb_start(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:442:30-442:51: int gen8_emit_bb_start_noarb(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:478:24-478:45: int gen8_emit_bb_start(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:506:34-506:55: static void assert_request_valid(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:519:31-519:52: static u32 *gen8_emit_wa_tail(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:532:35-532:56: static u32 *emit_preempt_busywait(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:548:32-548:53: gen8_emit_fini_breadcrumb_tail(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:563:33-563:54: static u32 *emit_xcs_breadcrumb(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:568:36-568:57: u32 *gen8_emit_fini_breadcrumb_xcs(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:573:36-573:57: u32 *gen8_emit_fini_breadcrumb_rcs(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:591:37-591:58: u32 *gen11_emit_fini_breadcrumb_rcs(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:625:41-625:62: static u32 *gen12_emit_preempt_busywait(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:642:33-642:54: static u32 ccs_semaphore_offset(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:649:34-649:55: static u32 *ccs_emit_wa_busywait(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:678:33-678:54: gen12_emit_fini_breadcrumb_tail(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:697:37-697:58: u32 *gen12_emit_fini_breadcrumb_xcs(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/gen8_engine_cs.c:704:37-704:58: u32 *gen12_emit_fini_breadcrumb_rcs(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c:106:46-106:67: check_signal_order(struct intel_context *ce, struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c:333:32-333:53: static void irq_signal_request(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c:344:31-344:52: static void insert_breadcrumb(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c:405:37-405:58: bool i915_request_enable_breadcrumb(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c:430:37-430:58: void i915_request_cancel_breadcrumb(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_context.c:467:7-467:28: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_context.h:125:7-125:28: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_context.h:313:10-313:31: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_context_sseu.c:16:34-16:55: static int gen8_emit_rpcs_config(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_engine_cs.c:1578:44-1578:65: static struct intel_timeline *get_timeline(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_engine_cs.c:1599:42-1599:63: static int print_ring(char *buf, int sz, struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_engine_cs.c:1802:55-1802:76: static void print_request_ring(struct drm_printer *m, struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_engine_cs.c:1882:33-1882:54: static void engine_dump_request(struct i915_request *rq, struct drm_printer *m, const char *msg)
-
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c:51:56-51:77: static void idle_pulse(struct intel_engine_cs *engine, struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c:59:30-59:51: static void heartbeat_commit(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c:68:28-68:55: static void show_heartbeat(const struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c:88:46-88:67: reset_engine(struct intel_engine_cs *engine, struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_engine_pm.c:93:24-93:45: __queue_and_release_pm(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:212:4-212:25: struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:232:56-232:77: active_request(const struct intel_timeline * const tl, struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:255:20-255:47: static int rq_prio(const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:260:27-260:54: static int effective_prio(const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:297:5-297:32: const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:417:33-417:54: execlists_context_status_change(struct i915_request *rq, unsigned long status)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:430:26-430:47: static void reset_active(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:468:25-468:52: static bool bad_request(const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:474:25-474:46: __execlists_schedule_in(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:529:35-529:56: static void execlists_schedule_in(struct i915_request *rq, int idx)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:546:26-546:47: resubmit_virtual_request(struct i915_request *rq, struct virtual_engine *ve)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:559:27-559:48: static void kick_siblings(struct i915_request *rq, struct intel_context *ce)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:588:38-588:66: static void __execlists_schedule_out(struct i915_request * const rq,
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:650:43-650:64: static inline void execlists_schedule_out(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:664:37-664:58: static u64 execlists_update_context(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:727:54-727:75: dump_port(char *buf, int buflen, const char *prefix, struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:951:41-951:68: static unsigned long i915_request_flags(const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:986:8-986:35: const struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:1064:27-1064:48: static void defer_request(struct i915_request *rq, struct list_head * const pl)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:1126:3-1126:30: const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:1144:8-1144:35: const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:1179:51-1179:78: timeslice_expired(struct intel_engine_cs *engine, const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:1229:10-1229:37: const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:1242:5-1242:32: const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:1251:23-1251:50: static bool completed(const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:2034:30-2034:51: static void __execlists_hold(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:2078:7-2078:28: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:2106:26-2106:53: static bool hold_request(const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:2132:32-2132:53: static void __execlists_unhold(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:2176:9-2176:30: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:2512:6-2512:27: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:2522:5-2522:32: const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:2534:9-2534:36: const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:2604:11-2604:32: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:2675:22-2675:43: static int emit_pdps(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:3231:27-3231:48: static void add_to_engine(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:3237:32-3237:53: static void remove_from_engine(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:3276:28-3276:55: static void kick_execlists(const struct i915_request *rq, int prio)
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:3847:36-3847:57: static void virtual_submit_request(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_gt_buffer_pool.h:25:6-25:27: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_migrate.c:328:32-328:53: static int emit_no_arbitration(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_migrate.c:344:21-344:42: static int emit_pte(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_migrate.c:539:26-539:47: static int emit_copy_ccs(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_migrate.c:594:22-594:43: static int emit_copy(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_migrate.c:890:23-890:44: static int emit_clear(struct i915_request *rq, u32 offset, int size,
-
drivers/gpu/drm/i915/gt/intel_renderstate.c:210:7-210:28: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_reset.c:75:25-75:46: static bool mark_guilty(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_reset.c:126:27-126:48: static void mark_innocent(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_reset.c:137:27-137:48: void __i915_request_reset(struct i915_request *rq, bool guilty)
-
drivers/gpu/drm/i915/gt/intel_ring.c:229:23-229:44: u32 *intel_ring_begin(struct i915_request *rq, unsigned int num_dwords)
-
drivers/gpu/drm/i915/gt/intel_ring.c:310:32-310:53: int intel_ring_cacheline_align(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_ring.h:41:39-41:60: static inline void intel_ring_advance(struct i915_request *rq, u32 *cs)
-
drivers/gpu/drm/i915/gt/intel_ring.h:81:37-81:64: static inline u32 intel_ring_offset(const struct i915_request *rq, void *addr)
-
drivers/gpu/drm/i915/gt/intel_ring_submission.c:600:9-600:30: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_ring_submission.c:618:6-618:27: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_ring_submission.c:649:24-649:45: static int load_pd_dir(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_ring_submission.c:683:27-683:48: static int mi_set_context(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_ring_submission.c:800:27-800:48: static int remap_l3_slice(struct i915_request *rq, int slice)
-
drivers/gpu/drm/i915/gt/intel_ring_submission.c:830:21-830:42: static int remap_l3(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_ring_submission.c:851:22-851:43: static int switch_mm(struct i915_request *rq, struct i915_address_space *vm)
-
drivers/gpu/drm/i915/gt/intel_ring_submission.c:877:28-877:49: static int clear_residuals(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_ring_submission.c:908:27-908:48: static int switch_context(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_ring_submission.c:1095:27-1095:48: static void add_to_engine(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_ring_submission.c:1101:32-1101:53: static void remove_from_engine(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_rps.c:1005:22-1005:43: void intel_rps_boost(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_timeline.c:327:9-327:30: struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/intel_timeline.h:96:10-96:37: const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_workarounds.c:821:30-821:51: int intel_engine_emit_ctx_wa(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/intel_workarounds.c:2740:13-2740:34: wa_list_srm(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/mock_engine.c:256:32-256:53: static void mock_add_to_engine(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/mock_engine.c:262:37-262:58: static void mock_remove_from_engine(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/selftest_context.c:15:25-15:46: static int request_sync(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/selftest_engine_cs.c:42:28-42:49: static int write_timestamp(struct i915_request *rq, int slot)
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:28:23-28:44: static bool is_active(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:43:7-43:28: struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:69:6-69:27: struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:822:22-822:43: emit_semaphore_chain(struct i915_request *rq, struct i915_vma *vma, int idx)
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:97:10-97:37: const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:103:6-103:27: struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:268:44-268:71: static u32 hws_seqno(const struct hang *h, const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:289:48-289:69: static bool wait_until_running(struct hang *h, struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:880:31-880:52: static int active_request_put(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:35:23-35:44: static bool is_active(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:50:7-50:28: struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:1026:27-1026:48: static int move_to_active(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:1660:6-1660:27: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/selftest_mocs.c:37:29-37:50: static int request_add_sync(struct i915_request *rq, int err)
-
drivers/gpu/drm/i915/gt/selftest_mocs.c:48:29-48:50: static int request_add_spin(struct i915_request *rq, struct igt_spinner *spin)
-
drivers/gpu/drm/i915/gt/selftest_mocs.c:101:22-101:43: static int read_regs(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/selftest_mocs.c:129:28-129:49: static int read_mocs_table(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/selftest_mocs.c:146:28-146:49: static int read_l3cc_table(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:454:31-454:52: static int emit_ggtt_store_dw(struct i915_request *rq, u32 addr, u32 value)
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:778:27-778:48: static int emit_read_hwsp(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:952:43-952:64: static struct i915_request *wrap_timeline(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/selftest_workarounds.c:37:29-37:50: static int request_add_sync(struct i915_request *rq, int err)
-
drivers/gpu/drm/i915/gt/selftest_workarounds.c:48:29-48:50: static int request_add_spin(struct i915_request *rq, struct igt_spinner *spin)
-
drivers/gpu/drm/i915/gt/selftest_workarounds.c:275:6-275:28: struct i915_request **rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:366:31-366:52: request_to_scheduling_context(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:617:53-617:74: static int __guc_add_request(struct intel_guc *guc, struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:700:51-700:72: static int guc_add_request(struct intel_guc *guc, struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:712:37-712:58: static inline void guc_set_lrc_tail(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:718:27-718:54: static inline int rq_prio(const struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:723:29-723:50: static bool is_multi_lrc_rq(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:728:26-728:47: static bool can_merge_rq(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:774:33-774:54: static int __guc_wq_item_append(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:819:10-819:31: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:836:30-836:51: static bool multi_lrc_submit(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:1899:6-1899:27: struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:1910:10-1910:31: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:1935:49-1935:70: static bool need_tasklet(struct intel_guc *guc, struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:1945:32-1945:53: static void guc_submit_request(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:2738:12-2738:33: struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:2771:55-2771:76: static void guc_context_ban(struct intel_context *ce, struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:3103:28-3103:49: static void add_to_context(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:3127:27-3127:48: static void guc_prio_fini(struct i915_request *rq, struct intel_context *ce)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:3139:33-3139:54: static void remove_from_context(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:3251:30-3251:51: static int guc_request_alloc(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:3683:44-3683:65: static void guc_bump_inflight_request_prio(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:3707:46-3707:67: static void guc_retire_inflight_request_prio(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:4583:54-4583:75: static int emit_bb_start_parent_no_preempt_mid_batch(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:4630:53-4630:74: static int emit_bb_start_child_no_preempt_mid_batch(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:4675:52-4675:73: __emit_fini_breadcrumb_parent_no_preempt_mid_batch(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:4718:35-4718:56: static inline bool skip_handshake(struct i915_request *rq)
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:4725:50-4725:71: emit_fini_breadcrumb_parent_no_preempt_mid_batch(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:4769:51-4769:72: __emit_fini_breadcrumb_child_no_preempt_mid_batch(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:4801:49-4801:70: emit_fini_breadcrumb_child_no_preempt_mid_batch(struct i915_request *rq,
-
drivers/gpu/drm/i915/gt/uc/selftest_guc.c:9:29-9:50: static int request_add_spin(struct i915_request *rq, struct igt_spinner *spin)
-
drivers/gpu/drm/i915/gvt/scheduler.c:262:35-262:56: static inline bool is_gvt_request(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_active.c:429:54-429:75: int i915_active_add_request(struct i915_active *ref, struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_active.c:737:31-737:52: int i915_request_await_active(struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_active.c:986:39-986:60: void i915_request_add_active_barriers(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_active.c:1072:6-1072:27: struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_active.h:220:50-220:71: static inline int __i915_request_await_exclusive(struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_gpu_error.c:1333:7-1333:34: const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_gpu_error.c:1425:7-1425:34: const struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_gpu_error.c:1497:7-1497:28: struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_perf.c:2152:17-2152:38: gen8_store_flex(struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_perf.c:2177:16-2177:37: gen8_load_flex(struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_request.c:170:21-170:42: __notify_execute_cb(struct i915_request *rq, bool (*fn)(struct irq_work *wrk))
-
drivers/gpu/drm/i915/i915_request.c:183:37-183:58: static void __notify_execute_cb_irq(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:194:41-194:62: void i915_request_notify_execute_cb_imm(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:199:33-199:54: static void __i915_request_fill(struct i915_request *rq, u8 val)
-
drivers/gpu/drm/i915/i915_request.c:223:28-223:49: i915_request_active_engine(struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_request.c:255:32-255:53: static void __rq_init_watchdog(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:276:31-276:52: static void __rq_arm_watchdog(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:295:34-295:55: static void __rq_cancel_watchdog(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:335:26-335:47: bool i915_request_retire(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:397:31-397:52: void i915_request_retire_upto(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:480:19-480:40: __await_execution(struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_request.c:531:26-531:47: void __i915_request_skip(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:549:34-549:55: bool i915_request_set_error_once(struct i915_request *rq, int error)
-
drivers/gpu/drm/i915/i915_request.c:567:44-567:65: struct i915_request *i915_request_mark_eio(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:740:26-740:47: void i915_request_cancel(struct i915_request *rq, int error)
-
drivers/gpu/drm/i915/i915_request.c:1050:26-1050:47: i915_request_await_start(struct i915_request *rq, struct i915_request *signal)
-
drivers/gpu/drm/i915/i915_request.c:1116:21-1116:42: already_busywaiting(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:1325:27-1325:48: static void mark_external(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:1339:31-1339:52: __i915_request_await_external(struct i915_request *rq, struct dma_fence *fence)
-
drivers/gpu/drm/i915/i915_request.c:1349:29-1349:50: i915_request_await_external(struct i915_request *rq, struct dma_fence *fence)
-
drivers/gpu/drm/i915/i915_request.c:1374:35-1374:56: static inline bool is_parallel_rq(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:1379:55-1379:76: static inline struct intel_context *request_to_parent(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:1394:30-1394:51: i915_request_await_execution(struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_request.c:1491:30-1491:51: i915_request_await_dma_fence(struct i915_request *rq, struct dma_fence *fence)
-
drivers/gpu/drm/i915/i915_request.c:1559:29-1559:50: int i915_request_await_deps(struct i915_request *rq, const struct i915_deps *deps)
-
drivers/gpu/drm/i915/i915_request.c:1612:41-1612:62: __i915_request_ensure_parallel_ordering(struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_request.c:1642:32-1642:53: __i915_request_ensure_ordering(struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_request.c:1687:32-1687:53: __i915_request_add_to_timeline(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:1742:44-1742:65: struct i915_request *__i915_request_commit(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:1772:30-1772:51: void __i915_request_queue_bh(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:1778:27-1778:48: void __i915_request_queue(struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_request.c:1800:23-1800:44: void i915_request_add(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:1857:33-1857:61: static bool __i915_spin_request(struct i915_request * const rq, int state)
-
drivers/gpu/drm/i915/i915_request.c:1936:32-1936:53: long i915_request_wait_timeout(struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_request.c:2075:24-2075:45: long i915_request_wait(struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_request.c:2102:26-2102:53: static char queue_status(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:2113:31-2113:58: static const char *run_status(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:2127:33-2127:60: static const char *fence_status(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:2139:10-2139:37: const struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_request.c:2190:63-2190:84: static bool engine_match_ring(struct intel_engine_cs *engine, struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:2197:24-2197:45: static bool match_ring(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.c:2217:49-2217:70: enum i915_request_state i915_test_request_state(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:393:18-393:39: i915_request_get(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:399:22-399:43: i915_request_get_rcu(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:405:18-405:39: i915_request_put(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:447:42-447:69: static inline bool i915_request_signaled(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:453:43-453:70: static inline bool i915_request_is_active(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:458:51-458:78: static inline bool i915_request_in_priority_queue(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:464:37-464:64: i915_request_has_initial_breadcrumb(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:477:32-477:59: static inline u32 __hwsp_seqno(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:497:30-497:57: static inline u32 hwsp_seqno(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:508:47-508:74: static inline bool __i915_request_has_started(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:539:41-539:68: static inline bool i915_request_started(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:564:44-564:71: static inline bool i915_request_is_running(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:594:42-594:69: static inline bool i915_request_is_ready(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:599:47-599:74: static inline bool __i915_request_is_complete(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:604:43-604:70: static inline bool i915_request_completed(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:620:47-620:68: static inline void i915_request_mark_complete(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:626:47-626:74: static inline bool i915_request_has_waitboost(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:631:47-631:74: static inline bool i915_request_has_nopreempt(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:637:46-637:73: static inline bool i915_request_has_sentinel(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:642:41-642:68: static inline bool i915_request_on_hold(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:647:42-647:63: static inline void i915_request_set_hold(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:652:44-652:65: static inline void i915_request_clear_hold(struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:658:23-658:50: i915_request_timeline(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:667:26-667:53: i915_request_gem_context(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:674:30-674:57: i915_request_active_timeline(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_request.h:686:27-686:54: i915_request_active_seqno(const struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_scheduler.c:289:20-289:41: void i915_schedule(struct i915_request *rq, const struct i915_sched_attr *attr)
-
drivers/gpu/drm/i915/i915_scheduler.c:410:10-410:37: const struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_trace.h:267:1-267:1: TRACE_EVENT(i915_request_queue,
-
drivers/gpu/drm/i915/i915_trace.h:321:1-321:1: DEFINE_EVENT(i915_request, i915_request_add,
-
drivers/gpu/drm/i915/i915_trace.h:327:1-327:1: DEFINE_EVENT(i915_request, i915_request_guc_submit,
-
drivers/gpu/drm/i915/i915_trace.h:332:1-332:1: DEFINE_EVENT(i915_request, i915_request_submit,
-
drivers/gpu/drm/i915/i915_trace.h:337:1-337:1: DEFINE_EVENT(i915_request, i915_request_execute,
-
drivers/gpu/drm/i915/i915_trace.h:342:1-342:1: TRACE_EVENT(i915_request_in,
-
drivers/gpu/drm/i915/i915_trace.h:372:1-372:1: TRACE_EVENT(i915_request_out,
-
drivers/gpu/drm/i915/i915_trace.h:602:1-602:1: DEFINE_EVENT(i915_request, i915_request_retire,
-
drivers/gpu/drm/i915/i915_trace.h:607:1-607:1: TRACE_EVENT(i915_request_wait_begin,
-
drivers/gpu/drm/i915/i915_trace.h:641:1-641:1: DEFINE_EVENT(i915_request, i915_request_wait_end,
-
drivers/gpu/drm/i915/i915_vma.c:1794:27-1794:48: __i915_request_await_bind(struct i915_request *rq, struct i915_vma *vma)
-
drivers/gpu/drm/i915/i915_vma.c:1799:60-1799:81: static int __i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq)
-
drivers/gpu/drm/i915/i915_vma.c:1812:9-1812:30: struct i915_request *rq,
-
drivers/gpu/drm/i915/i915_vma.h:64:47-64:68: i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq,
-
drivers/gpu/drm/i915/pxp/intel_pxp_cmd.c:81:32-81:53: static void pxp_request_commit(struct i915_request *rq)
-
drivers/gpu/drm/i915/selftests/i915_perf.c:155:28-155:49: static int write_timestamp(struct i915_request *rq, int slot)
-
drivers/gpu/drm/i915/selftests/i915_perf.c:182:28-182:49: static ktime_t poll_status(struct i915_request *rq, int slot)
-
drivers/gpu/drm/i915/selftests/igt_spinner.c:117:10-117:37: const struct i915_request *rq)
-
drivers/gpu/drm/i915/selftests/igt_spinner.c:123:6-123:27: struct i915_request *rq,
-
drivers/gpu/drm/i915/selftests/igt_spinner.c:233:43-233:70: hws_seqno(const struct igt_spinner *spin, const struct i915_request *rq)
-
drivers/gpu/drm/i915/selftests/igt_spinner.c:266:53-266:74: bool igt_wait_for_spinner(struct igt_spinner *spin, struct i915_request *rq)
-
drivers/gpu/drm/i915/selftests/intel_scheduler_helpers.c:91:32-91:53: int intel_selftest_wait_for_rq(struct i915_request *rq)
-
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/g94.c:28:54-28:59: g94_aux_stat(struct nvkm_i2c *i2c, u32 *hi, u32 *lo, u32 *rq, u32 *tx)
-
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/gk104.c:28:56-28:61: gk104_aux_stat(struct nvkm_i2c *i2c, u32 *hi, u32 *lo, u32 *rq, u32 *tx)
-
drivers/gpu/drm/scheduler/sched_main.c:74:10-74:31: struct drm_sched_rq *rq)
-
drivers/gpu/drm/scheduler/sched_main.c:90:30-90:51: void drm_sched_rq_add_entity(struct drm_sched_rq *rq,
-
drivers/gpu/drm/scheduler/sched_main.c:109:33-109:54: void drm_sched_rq_remove_entity(struct drm_sched_rq *rq,
-
drivers/gpu/drm/scheduler/sched_main.c:130:28-130:49: drm_sched_rq_select_entity(struct drm_sched_rq *rq)
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:2165:23-2165:44: static int __flush_rq(struct bnxt_qplib_q *rq, struct bnxt_qplib_qp *qp,
-
drivers/infiniband/hw/mlx5/qp.c:1349:8-1349:27: struct mlx5_ib_rq *rq, void *qpin,
-
drivers/infiniband/hw/mlx5/qp.c:1419:10-1419:29: struct mlx5_ib_rq *rq)
-
drivers/infiniband/hw/mlx5/qp.c:1425:11-1425:30: struct mlx5_ib_rq *rq,
-
drivers/infiniband/hw/mlx5/qp.c:1436:9-1436:28: struct mlx5_ib_rq *rq, u32 tdn,
-
drivers/infiniband/hw/mlx5/qp.c:3693:27-3693:46: struct mlx5_ib_dev *dev, struct mlx5_ib_rq *rq, int new_state,
-
drivers/infiniband/hw/mlx5/qp.c:4699:6-4699:25: struct mlx5_ib_rq *rq,
-
drivers/infiniband/hw/mlx5/qpc.c:562:5-562:26: struct mlx5_core_qp *rq)
-
drivers/infiniband/hw/mlx5/qpc.c:586:6-586:27: struct mlx5_core_qp *rq)
-
drivers/infiniband/sw/rdmavt/qp.c:785:18-785:33: int rvt_alloc_rq(struct rvt_rq *rq, u32 size, int node,
-
drivers/infiniband/sw/rdmavt/qp.c:2323:32-2323:47: static inline u32 get_rvt_head(struct rvt_rq *rq, void *ip)
-
drivers/infiniband/sw/rxe/rxe_verbs.c:235:26-235:41: static int post_one_recv(struct rxe_rq *rq, const struct ib_recv_wr *ibwr)
-
drivers/input/misc/xen-kbdfront.c:181:34-181:38: static irqreturn_t input_handler(int rq, void *dev_id)
-
drivers/isdn/hardware/mISDN/avmfritz.c:895:37-895:57: open_bchannel(struct fritzcard *fc, struct channel_req *rq)
-
drivers/isdn/hardware/mISDN/hfcmulti.c:4040:8-4040:28: struct channel_req *rq)
-
drivers/isdn/hardware/mISDN/hfcmulti.c:4081:8-4081:28: struct channel_req *rq)
-
drivers/isdn/hardware/mISDN/hfcpci.c:1874:8-1874:28: struct channel_req *rq)
-
drivers/isdn/hardware/mISDN/hfcpci.c:1925:35-1925:55: open_bchannel(struct hfc_pci *hc, struct channel_req *rq)
-
drivers/isdn/hardware/mISDN/hfcsusb.c:412:8-412:28: struct channel_req *rq)
-
drivers/isdn/hardware/mISDN/hfcsusb.c:465:35-465:55: open_bchannel(struct hfcsusb *hw, struct channel_req *rq)
-
drivers/isdn/hardware/mISDN/mISDNipac.c:743:44-743:64: open_dchannel_caller(struct isac_hw *isac, struct channel_req *rq, void *caller)
-
drivers/isdn/hardware/mISDN/mISDNipac.c:761:37-761:57: open_dchannel(struct isac_hw *isac, struct channel_req *rq)
-
drivers/isdn/hardware/mISDN/mISDNipac.c:1479:37-1479:57: open_bchannel(struct ipac_hw *ipac, struct channel_req *rq)
-
drivers/isdn/hardware/mISDN/mISDNisar.c:1635:33-1635:53: isar_open(struct isar_hw *isar, struct channel_req *rq)
-
drivers/isdn/hardware/mISDN/netjet.c:849:38-849:58: open_bchannel(struct tiger_hw *card, struct channel_req *rq)
-
drivers/isdn/hardware/mISDN/w6692.c:987:38-987:58: open_bchannel(struct w6692_hw *card, struct channel_req *rq)
-
drivers/isdn/hardware/mISDN/w6692.c:1158:38-1158:58: open_dchannel(struct w6692_hw *card, struct channel_req *rq, void *caller)
-
drivers/isdn/mISDN/l1oip_core.c:976:55-976:75: open_dchannel(struct l1oip *hc, struct dchannel *dch, struct channel_req *rq)
-
drivers/isdn/mISDN/l1oip_core.c:1003:55-1003:75: open_bchannel(struct l1oip *hc, struct dchannel *dch, struct channel_req *rq)
-
drivers/md/dm-integrity.c:365:34-365:50: static void dm_integrity_prepare(struct request *rq)
-
drivers/md/dm-integrity.c:369:35-369:51: static void dm_integrity_complete(struct request *rq, unsigned int nr_bytes)
-
drivers/md/dm-mpath.c:502:58-502:74: static int multipath_clone_and_map(struct dm_target *ti, struct request *rq,
-
drivers/md/dm-rq.c:122:49-122:65: static struct dm_rq_target_io *tio_from_request(struct request *rq)
-
drivers/md/dm-rq.c:181:41-181:57: static void dm_mq_delay_requeue_request(struct request *rq, unsigned long msecs)
-
drivers/md/dm-rq.c:250:29-250:45: static void dm_softirq_done(struct request *rq)
-
drivers/md/dm-rq.c:275:33-275:49: static void dm_complete_request(struct request *rq, blk_status_t error)
-
drivers/md/dm-rq.c:290:38-290:54: static void dm_kill_unmapped_request(struct request *rq, blk_status_t error)
-
drivers/md/dm-rq.c:317:47-317:63: static int setup_clone(struct request *clone, struct request *rq,
-
drivers/md/dm-rq.c:335:51-335:67: static void init_tio(struct dm_rq_target_io *tio, struct request *rq,
-
drivers/md/dm-rq.c:453:59-453:75: static int dm_mq_init_request(struct blk_mq_tag_set *set, struct request *rq,
-
drivers/md/dm-target.c:133:58-133:74: static int io_err_clone_and_map_rq(struct dm_target *ti, struct request *rq,
-
drivers/mmc/core/mmc_test.c:766:32-766:53: static void mmc_test_req_reset(struct mmc_test_req *rq)
-
drivers/mmc/core/queue.h:25:58-25:74: static inline struct mmc_queue_req *req_to_mmc_queue_req(struct request *rq)
-
drivers/net/dsa/hirschmann/hellcreek_ptp.c:222:5-222:31: struct ptp_clock_request *rq, int on)
-
drivers/net/dsa/mv88e6xxx/ptp.c:269:11-269:37: struct ptp_clock_request *rq, int on)
-
drivers/net/dsa/mv88e6xxx/ptp.c:322:5-322:31: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/3com/3c574_cs.c:1034:46-1034:60: static int el3_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/3com/3c59x.c:3029:49-3029:63: static int vortex_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/8390/axnet_cs.c:608:48-608:62: static int axnet_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/8390/pcnet_cs.c:1109:45-1109:59: static int ei_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/adaptec/starfire.c:1906:49-1906:63: static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/amd/pcnet32.c:2777:50-2777:64: static int pcnet32_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c:390:12-390:38: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c:427:9-427:35: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c:461:11-461:37: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c:488:11-488:37: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c:13803:8-13803:34: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/broadcom/bnxt/bnxt_ptp.c:378:11-378:37: struct ptp_clock_request *rq)
-
drivers/net/ethernet/broadcom/bnxt/bnxt_ptp.c:414:7-414:33: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/broadcom/tg3.c:6262:6-6262:32: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/cadence/macb_main.c:3674:47-3674:61: static int macb_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/cadence/macb_ptp.c:184:6-184:32: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/cadence/macb_ptp.c:422:43-422:57: int gem_get_hwtst(struct net_device *dev, struct ifreq *rq)
-
drivers/net/ethernet/cavium/common/cavium_ptp.c:207:9-207:35: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/cavium/liquidio/lio_main.c:1627:7-1627:48: struct ptp_clock_request __maybe_unused *rq,
-
drivers/net/ethernet/cavium/thunder/nicvf_main.c:530:5-530:23: struct rcv_queue *rq, struct sk_buff **skb)
-
drivers/net/ethernet/cavium/thunder/nicvf_main.c:773:29-773:47: struct snd_queue *sq, struct rcv_queue *rq)
-
drivers/net/ethernet/chelsio/cxgb3/sge.c:1962:51-1962:68: static inline int rx_offload(struct t3cdev *tdev, struct sge_rspq *rq,
-
drivers/net/ethernet/chelsio/cxgb3/sge.c:2089:42-2089:59: static void rx_eth(struct adapter *adap, struct sge_rspq *rq,
-
drivers/net/ethernet/chelsio/cxgb3/sge.c:2720:6-2720:23: struct sge_rspq *rq)
-
drivers/net/ethernet/chelsio/cxgb4/sge.c:4864:41-4864:58: void free_rspq_fl(struct adapter *adap, struct sge_rspq *rq,
-
drivers/net/ethernet/cisco/enic/enic.h:232:58-232:71: static inline unsigned int enic_cq_rq(struct enic *enic, unsigned int rq)
-
drivers/net/ethernet/cisco/enic/enic.h:258:2-258:15: unsigned int rq)
-
drivers/net/ethernet/cisco/enic/enic_clsf.c:21:64-21:68: int enic_addfltr_5t(struct enic *enic, struct flow_keys *keys, u16 rq)
-
drivers/net/ethernet/cisco/enic/enic_main.c:1223:30-1223:46: static void enic_free_rq_buf(struct vnic_rq *rq, struct vnic_rq_buf *buf)
-
drivers/net/ethernet/cisco/enic/enic_main.c:1236:30-1236:46: static int enic_rq_alloc_buf(struct vnic_rq *rq)
-
drivers/net/ethernet/cisco/enic/enic_main.c:1297:34-1297:50: static void enic_rq_indicate_buf(struct vnic_rq *rq,
-
drivers/net/ethernet/cisco/enic/enic_main.c:1447:56-1447:72: static void enic_set_int_moderation(struct enic *enic, struct vnic_rq *rq)
-
drivers/net/ethernet/cisco/enic/enic_main.c:1459:57-1459:73: static void enic_calc_int_moderation(struct enic *enic, struct vnic_rq *rq)
-
drivers/net/ethernet/cisco/enic/enic_res.h:122:39-122:55: static inline void enic_queue_rq_desc(struct vnic_rq *rq,
-
drivers/net/ethernet/cisco/enic/vnic_rq.c:31:31-31:47: static int vnic_rq_alloc_bufs(struct vnic_rq *rq)
-
drivers/net/ethernet/cisco/enic/vnic_rq.c:66:19-66:35: void vnic_rq_free(struct vnic_rq *rq)
-
drivers/net/ethernet/cisco/enic/vnic_rq.c:85:42-85:58: int vnic_rq_alloc(struct vnic_dev *vdev, struct vnic_rq *rq, unsigned int index,
-
drivers/net/ethernet/cisco/enic/vnic_rq.c:114:32-114:48: static void vnic_rq_init_start(struct vnic_rq *rq, unsigned int cq_index,
-
drivers/net/ethernet/cisco/enic/vnic_rq.c:138:19-138:35: void vnic_rq_init(struct vnic_rq *rq, unsigned int cq_index,
-
drivers/net/ethernet/cisco/enic/vnic_rq.c:146:35-146:51: unsigned int vnic_rq_error_status(struct vnic_rq *rq)
-
drivers/net/ethernet/cisco/enic/vnic_rq.c:151:21-151:37: void vnic_rq_enable(struct vnic_rq *rq)
-
drivers/net/ethernet/cisco/enic/vnic_rq.c:156:21-156:37: int vnic_rq_disable(struct vnic_rq *rq)
-
drivers/net/ethernet/cisco/enic/vnic_rq.c:184:20-184:36: void vnic_rq_clean(struct vnic_rq *rq,
-
drivers/net/ethernet/cisco/enic/vnic_rq.h:97:47-97:63: static inline unsigned int vnic_rq_desc_avail(struct vnic_rq *rq)
-
drivers/net/ethernet/cisco/enic/vnic_rq.h:103:46-103:62: static inline unsigned int vnic_rq_desc_used(struct vnic_rq *rq)
-
drivers/net/ethernet/cisco/enic/vnic_rq.h:109:39-109:55: static inline void *vnic_rq_next_desc(struct vnic_rq *rq)
-
drivers/net/ethernet/cisco/enic/vnic_rq.h:114:47-114:63: static inline unsigned int vnic_rq_next_index(struct vnic_rq *rq)
-
drivers/net/ethernet/cisco/enic/vnic_rq.h:119:33-119:49: static inline void vnic_rq_post(struct vnic_rq *rq,
-
drivers/net/ethernet/cisco/enic/vnic_rq.h:154:41-154:57: static inline void vnic_rq_return_descs(struct vnic_rq *rq, unsigned int count)
-
drivers/net/ethernet/cisco/enic/vnic_rq.h:164:36-164:52: static inline void vnic_rq_service(struct vnic_rq *rq,
-
drivers/net/ethernet/cisco/enic/vnic_rq.h:192:32-192:48: static inline int vnic_rq_fill(struct vnic_rq *rq,
-
drivers/net/ethernet/dec/tulip/tulip_core.c:897:51-897:65: static int private_ioctl (struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/dec/tulip/winbond-840.c:1438:49-1438:63: static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/dlink/dl2k.c:1342:36-1342:50: rio_ioctl (struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/dlink/sundance.c:1802:49-1802:63: static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/fealnx.c:1873:46-1873:60: static int mii_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c:2452:55-2452:69: static int dpaa2_eth_ts_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c:2488:52-2488:66: static int dpaa2_eth_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/freescale/dpaa2/dpaa2-ptp.c:17:8-17:34: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/freescale/enetc/enetc.c:2734:42-2734:56: int enetc_ioctl(struct net_device *ndev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/freescale/fec_main.c:2930:52-2930:66: static int fec_enet_ioctl(struct net_device *ndev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/freescale/fec_ptp.c:440:6-440:32: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/freescale/gianfar.c:2128:47-2128:61: static int gfar_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/huawei/hinic/hinic_debugfs.c:53:61-53:78: static u64 hinic_dbg_get_rq_info(struct hinic_dev *nic_dev, struct hinic_rq *rq, int idx)
-
drivers/net/ethernet/huawei/hinic/hinic_debugfs.c:244:25-244:42: void hinic_rq_debug_rem(struct hinic_rq *rq)
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:155:7-155:24: struct hinic_rq *rq, u16 global_qid)
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:247:29-247:46: static int alloc_rq_skb_arr(struct hinic_rq *rq)
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:264:29-264:46: static void free_rq_skb_arr(struct hinic_rq *rq)
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:316:25-316:42: static int alloc_rq_cqe(struct hinic_rq *rq)
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:360:25-360:42: static void free_rq_cqe(struct hinic_rq *rq)
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:384:19-384:36: int hinic_init_rq(struct hinic_rq *rq, struct hinic_hwif *hwif,
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:435:21-435:38: void hinic_clean_rq(struct hinic_rq *rq)
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:468:30-468:47: int hinic_get_rq_free_wqebbs(struct hinic_rq *rq)
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:799:39-799:56: struct hinic_rq_wqe *hinic_rq_get_wqe(struct hinic_rq *rq,
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:818:25-818:42: void hinic_rq_write_wqe(struct hinic_rq *rq, u16 prod_idx,
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:840:40-840:57: struct hinic_rq_wqe *hinic_rq_read_wqe(struct hinic_rq *rq,
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:875:45-875:62: struct hinic_rq_wqe *hinic_rq_read_next_wqe(struct hinic_rq *rq,
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:902:23-902:40: void hinic_rq_put_wqe(struct hinic_rq *rq, u16 cons_idx,
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:925:23-925:40: void hinic_rq_get_sge(struct hinic_rq *rq, struct hinic_rq_wqe *rq_wqe,
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:943:27-943:44: void hinic_rq_prepare_wqe(struct hinic_rq *rq, u16 prod_idx,
-
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c:971:22-971:39: void hinic_rq_update(struct hinic_rq *rq, u16 prod_idx)
-
drivers/net/ethernet/huawei/hinic/hinic_rx.c:582:43-582:60: int hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq,
-
drivers/net/ethernet/intel/i40e/i40e_ptp.c:509:10-509:36: struct ptp_clock_request *rq,
-
drivers/net/ethernet/intel/i40e/i40e_ptp.c:614:8-614:34: struct ptp_clock_request *rq,
-
drivers/net/ethernet/intel/ice/ice_ptp.c:1393:5-1393:31: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/intel/igb/igb_ptp.c:510:6-510:32: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/intel/igb/igb_ptp.c:662:12-662:38: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/intel/igb/igb_ptp.c:808:7-808:33: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/intel/igc/igc_ptp.c:247:12-247:38: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c:637:9-637:35: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/jme.c:2614:38-2614:52: jme_ioctl(struct net_device *netdev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/korina.c:919:49-919:63: static int korina_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ptp.c:195:7-195:33: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/mellanox/mlx5/core/en/ptp.c:494:9-494:26: struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c:726:51-726:68: void mlx5e_rep_tc_receive(struct mlx5_cqe64 *cqe, struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:137:30-137:47: static int mlx5e_rq_to_ready(struct mlx5e_rq *rq, int curr_state)
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:267:51-267:68: mlx5e_rx_reporter_build_diagnose_output_rq_common(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:332:52-332:69: static int mlx5e_rx_reporter_build_diagnose_output(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:352:50-352:67: static int mlx5e_rx_reporter_diagnose_generic_rq(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:440:59-440:76: static int mlx5e_rx_reporter_build_diagnose_output_ptp_rq(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:689:32-689:49: void mlx5e_reporter_rx_timeout(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:710:32-710:49: void mlx5e_reporter_rq_cqe_err(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c:41:11-41:28: struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c:93:33-93:50: static void mlx5e_close_trap_rq(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h:153:53-153:70: static inline u16 mlx5e_shampo_get_cqe_header_index(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
-
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h:299:37-299:54: static inline void mlx5e_rqwq_reset(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h:325:39-325:56: static inline u32 mlx5e_rqwq_get_size(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h:335:41-335:58: static inline u32 mlx5e_rqwq_get_cur_sz(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h:345:39-345:56: static inline u16 mlx5e_rqwq_get_head(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h:355:46-355:63: static inline u16 mlx5e_rqwq_get_wqe_counter(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c:59:45-59:62: mlx5e_xmit_xdp_buff(struct mlx5e_xdpsq *sq, struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c:160:23-160:40: bool mlx5e_xdp_handle(struct mlx5e_rq *rq, struct page *page,
-
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c:670:33-670:50: void mlx5e_xdp_rx_poll_complete(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c:11:48-11:65: static struct sk_buff *mlx5e_xsk_construct_skb(struct mlx5e_rq *rq, void *data,
-
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c:27:53-27:70: struct sk_buff *mlx5e_xsk_skb_from_cqe_mpwrq_linear(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c:82:47-82:64: struct sk_buff *mlx5e_xsk_skb_from_cqe_linear(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.h:21:45-21:62: static inline int mlx5e_xsk_page_alloc_pool(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.h:38:47-38:64: static inline bool mlx5e_xsk_update_rx_wakeup(struct mlx5e_rq *rq, bool alloc_err)
-
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/setup.c:53:9-53:26: struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_rx.c:461:30-461:47: static void resync_update_sn(struct mlx5e_rq *rq, struct sk_buff *skb)
-
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_rx.c:540:31-540:48: void mlx5e_ktls_handle_rx_skb(struct mlx5e_rq *rq, struct sk_buff *skb,
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:202:40-202:57: static inline void mlx5e_build_umr_wqe(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:220:37-220:54: static int mlx5e_rq_shampo_hd_alloc(struct mlx5e_rq *rq, int node)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:229:37-229:54: static void mlx5e_rq_shampo_hd_free(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:234:42-234:59: static int mlx5e_rq_shampo_hd_info_alloc(struct mlx5e_rq *rq, int node)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:253:42-253:59: static void mlx5e_rq_shampo_hd_info_free(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:259:38-259:55: static int mlx5e_rq_alloc_mpwqe_info(struct mlx5e_rq *rq, int node)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:357:65-357:82: static int mlx5e_create_rq_umr_mkey(struct mlx5_core_dev *mdev, struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:366:12-366:29: struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:384:40-384:57: static void mlx5e_init_frags_partition(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:417:24-417:41: int mlx5e_init_di_list(struct mlx5e_rq *rq, int wq_sz, int node)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:430:25-430:42: void mlx5e_free_di_list(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:442:43-442:60: static int mlx5e_alloc_mpwqe_rq_drop_page(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:457:43-457:60: static void mlx5e_free_mpwqe_rq_drop_page(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:465:9-465:26: struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:494:5-494:22: struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:539:34-539:51: static void mlx5e_rq_free_shampo(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:553:16-553:33: int node, struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:749:27-749:44: static void mlx5e_free_rq(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:789:21-789:38: int mlx5e_create_rq(struct mlx5e_rq *rq, struct mlx5e_rq_param *param)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:836:27-836:44: int mlx5e_modify_rq_state(struct mlx5e_rq *rq, int curr_state, int next_state)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:865:40-865:57: static int mlx5e_modify_rq_scatter_fcs(struct mlx5e_rq *rq, bool enable)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:894:32-894:49: static int mlx5e_modify_rq_vsd(struct mlx5e_rq *rq, bool vsd)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:922:23-922:40: void mlx5e_destroy_rq(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:927:32-927:49: int mlx5e_wait_for_min_rx_wqes(struct mlx5e_rq *rq, int wait_time)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:947:38-947:55: void mlx5e_free_rx_in_progress_descs(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:979:26-979:43: void mlx5e_free_rx_descs(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:1017:5-1017:22: struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:1067:24-1067:41: void mlx5e_activate_rq(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:1072:26-1072:43: void mlx5e_deactivate_rq(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:1078:21-1078:38: void mlx5e_close_rq(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:2974:32-2974:49: static void mlx5e_free_drop_rq(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:2980:11-2980:28: struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:4510:39-4510:56: static void mlx5e_rq_replace_xdp_prog(struct mlx5e_rq *rq, struct bpf_prog *prog)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:90:42-90:59: static inline void mlx5e_read_title_slot(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:135:41-135:58: static inline void mlx5e_decompress_cqe(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:165:49-165:66: static inline void mlx5e_decompress_cqe_no_hash(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:176:46-176:63: static inline u32 mlx5e_decompress_cqes_cont(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:206:47-206:64: static inline u32 mlx5e_decompress_cqes_start(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:224:39-224:56: static inline bool mlx5e_rx_cache_put(struct mlx5e_rq *rq, struct page *page)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:246:39-246:56: static inline bool mlx5e_rx_cache_get(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:272:41-272:58: static inline int mlx5e_page_alloc_pool(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:294:36-294:53: static inline int mlx5e_page_alloc(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:303:27-303:44: void mlx5e_page_dma_unmap(struct mlx5e_rq *rq, struct page *page)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:312:33-312:50: void mlx5e_page_release_dynamic(struct mlx5e_rq *rq, struct page *page, bool recycle)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:327:39-327:56: static inline void mlx5e_page_release(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:341:37-341:54: static inline int mlx5e_get_rx_frag(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:357:38-357:55: static inline void mlx5e_put_rx_frag(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:365:52-365:69: static inline struct mlx5e_wqe_frag_info *get_frag(struct mlx5e_rq *rq, u16 ix)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:370:31-370:48: static int mlx5e_alloc_rx_wqe(struct mlx5e_rq *rq, struct mlx5e_rx_wqe_cyc *wqe,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:398:38-398:55: static inline void mlx5e_free_rx_wqe(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:408:34-408:51: static void mlx5e_dealloc_rx_wqe(struct mlx5e_rq *rq, u16 ix)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:415:32-415:49: static int mlx5e_alloc_rx_wqes(struct mlx5e_rq *rq, u16 ix, u8 wqe_bulk)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:450:20-450:37: mlx5e_add_skb_frag(struct mlx5e_rq *rq, struct sk_buff *skb,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:477:21-477:38: mlx5e_free_rx_mpwqe(struct mlx5e_rq *rq, struct mlx5e_mpw_info *wi, bool recycle)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:495:33-495:50: static void mlx5e_post_rx_mpwqe(struct mlx5e_rq *rq, u8 n)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:548:38-548:55: static int mlx5e_build_shampo_hd_umr(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:621:36-621:53: static int mlx5e_alloc_rx_hd_mpwqe(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:660:33-660:50: static int mlx5e_alloc_rx_mpwqe(struct mlx5e_rq *rq, u16 ix)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:736:30-736:47: void mlx5e_shampo_dealloc_hd(struct mlx5e_rq *rq, u16 len, u16 start, bool close)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:768:36-768:53: static void mlx5e_dealloc_rx_mpwqe(struct mlx5e_rq *rq, u16 ix)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:775:49-775:66: INDIRECT_CALLABLE_SCOPE bool mlx5e_post_rx_wqes(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:945:51-945:68: INDIRECT_CALLABLE_SCOPE bool mlx5e_post_rx_mpwqes(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1075:41-1075:58: static void *mlx5e_shampo_get_packet_hd(struct mlx5e_rq *rq, u16 header_index)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1083:46-1083:63: static void mlx5e_shampo_update_ipv4_udp_hdr(struct mlx5e_rq *rq, struct iphdr *ipv4)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1102:46-1102:63: static void mlx5e_shampo_update_ipv6_udp_hdr(struct mlx5e_rq *rq, struct ipv6hdr *ipv6)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1121:47-1121:64: static void mlx5e_shampo_update_fin_psh_flags(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1133:46-1133:63: static void mlx5e_shampo_update_ipv4_tcp_hdr(struct mlx5e_rq *rq, struct iphdr *ipv4,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1157:46-1157:63: static void mlx5e_shampo_update_ipv6_tcp_hdr(struct mlx5e_rq *rq, struct ipv6hdr *ipv6,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1178:37-1178:54: static void mlx5e_shampo_update_hdr(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe, bool match)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1236:37-1236:54: static inline void mlx5e_enable_ecn(struct mlx5e_rq *rq, struct sk_buff *skb)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1338:10-1338:27: struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1409:11-1409:28: struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1462:42-1462:59: static void mlx5e_shampo_complete_rx_cqe(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1483:42-1483:59: static inline void mlx5e_complete_rx_cqe(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1496:40-1496:57: struct sk_buff *mlx5e_build_linear_skb(struct mlx5e_rq *rq, void *va,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1516:33-1516:50: static void mlx5e_fill_xdp_buff(struct mlx5e_rq *rq, void *va, u16 headroom,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1524:27-1524:44: mlx5e_skb_from_cqe_linear(struct mlx5e_rq *rq, struct mlx5e_wqe_frag_info *wi,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1568:30-1568:47: mlx5e_skb_from_cqe_nonlinear(struct mlx5e_rq *rq, struct mlx5e_wqe_frag_info *wi,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1674:28-1674:45: static void trigger_report(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1686:37-1686:54: static void mlx5e_handle_rx_err_cqe(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1692:33-1692:50: static void mlx5e_handle_rx_cqe(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1741:37-1741:54: static void mlx5e_handle_rx_cqe_rep(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1790:43-1790:60: static void mlx5e_handle_rx_cqe_mpwrq_rep(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1849:42-1849:59: mlx5e_fill_skb_data(struct sk_buff *skb, struct mlx5e_rq *rq, struct mlx5e_dma_info *di,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1873:36-1873:53: mlx5e_skb_from_cqe_mpwrq_nonlinear(struct mlx5e_rq *rq, struct mlx5e_mpw_info *wi,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1908:33-1908:50: mlx5e_skb_from_cqe_mpwrq_linear(struct mlx5e_rq *rq, struct mlx5e_mpw_info *wi,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:1961:27-1961:44: mlx5e_skb_from_cqe_shampo(struct mlx5e_rq *rq, struct mlx5e_mpw_info *wi,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:2022:24-2022:41: mlx5e_shampo_flush_skb(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe, bool match)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:2045:31-2045:48: mlx5e_free_rx_shampo_hd_entry(struct mlx5e_rq *rq, u16 header_index)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:2057:46-2057:63: static void mlx5e_handle_rx_cqe_mpwrq_shampo(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:2142:39-2142:56: static void mlx5e_handle_rx_cqe_mpwrq(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:2259:42-2259:59: static inline void mlx5i_complete_rx_cqe(struct mlx5e_rq *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:2344:33-2344:50: static void mlx5i_handle_rx_cqe(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:2386:27-2386:44: int mlx5e_rq_set_handlers(struct mlx5e_rq *rq, struct mlx5e_params *params, bool xsk)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:2435:38-2435:55: static void mlx5e_trap_handle_rx_cqe(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:2472:33-2472:50: void mlx5e_rq_set_trap_handlers(struct mlx5e_rq *rq, struct mlx5e_params *params)
-
drivers/net/ethernet/mellanox/mlx5/core/en_txrx.c:60:33-60:50: static void mlx5e_handle_rx_dim(struct mlx5e_rq *rq)
-
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c:369:5-369:31: struct ptp_clock_request *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c:468:6-468:32: struct ptp_clock_request *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c:547:10-547:36: struct ptp_clock_request *rq,
-
drivers/net/ethernet/mellanox/mlx5/core/lib/clock.c:558:7-558:33: struct ptp_clock_request *rq,
-
drivers/net/ethernet/microchip/lan966x/lan966x_ptp.c:594:10-594:36: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/microchip/lan966x/lan966x_ptp.c:695:9-695:35: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/microchip/lan966x/lan966x_ptp.c:740:10-740:36: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/mscc/ocelot_ptp.c:186:9-186:35: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/natsemi/natsemi.c:3072:49-3072:63: static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/neterion/s2io.c:6625:47-6625:61: static int s2io_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/neterion/vxge/vxge-main.c:3236:47-3236:61: static int vxge_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/packetengines/hamachi.c:1875:59-1875:73: static int hamachi_siocdevprivate(struct net_device *dev, struct ifreq *rq,
-
drivers/net/ethernet/packetengines/hamachi.c:1903:50-1903:64: static int hamachi_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/packetengines/yellowfin.c:1352:49-1352:63: static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/qlogic/qede/qede_ptp.c:120:11-120:37: struct ptp_clock_request *rq,
-
drivers/net/ethernet/realtek/8139cp.c:1606:46-1606:60: static int cp_ioctl (struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/realtek/8139too.c:2501:49-2501:63: static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c:1941:48-1941:62: static int sxgbe_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/sis/sis900.c:2228:50-2228:64: static int mii_ioctl(struct net_device *net_dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/smsc/epic100.c:1486:49-1486:63: static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/smsc/smc91c92_cs.c:1997:47-1997:61: static int smc_ioctl (struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c:5787:49-5787:63: static int stmmac_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/stmicro/stmmac/stmmac_ptp.c:173:5-173:31: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/ti/cpts.c:312:7-312:33: struct ptp_clock_request *rq, int on)
-
drivers/net/ethernet/ti/tlan.c:936:47-936:61: static int tlan_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/via/via-rhine.c:2387:49-2387:63: static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/via/via-velocity.c:2427:51-2427:65: static int velocity_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/xilinx/xilinx_axienet_main.c:1285:50-1285:64: static int axienet_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/xilinx/xilinx_emaclite.c:1220:52-1220:66: static int xemaclite_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/ethernet/xircom/xirc2ps_cs.c:1415:34-1415:48: do_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/fddi/skfp/skfddi.c:960:56-960:70: static int skfp_siocdevprivate(struct net_device *dev, struct ifreq *rq, void __user *data, int cmd)
-
drivers/net/hippi/rrunner.c:1575:54-1575:68: static int rr_siocdevprivate(struct net_device *dev, struct ifreq *rq,
-
drivers/net/phy/dp83640.c:471:10-471:36: struct ptp_clock_request *rq, int on)
-
drivers/net/plip/plip.c:1216:45-1216:59: plip_siocdevprivate(struct net_device *dev, struct ifreq *rq,
-
drivers/net/slip/slip.c:1186:54-1186:68: static int sl_siocdevprivate(struct net_device *dev, struct ifreq *rq,
-
drivers/net/usb/asix_devices.c:107:48-107:62: static int asix_ioctl (struct net_device *net, struct ifreq *rq, int cmd)
-
drivers/net/usb/ax88179_178a.c:887:50-887:64: static int ax88179_ioctl(struct net_device *net, struct ifreq *rq, int cmd)
-
drivers/net/usb/dm9601.c:271:49-271:63: static int dm9601_ioctl(struct net_device *net, struct ifreq *rq, int cmd)
-
drivers/net/usb/mcs7830.c:328:50-328:64: static int mcs7830_ioctl(struct net_device *net, struct ifreq *rq, int cmd)
-
drivers/net/usb/pegasus.c:1004:59-1004:73: static int pegasus_siocdevprivate(struct net_device *net, struct ifreq *rq,
-
drivers/net/usb/r8152.c:9157:53-9157:67: static int rtl8152_ioctl(struct net_device *netdev, struct ifreq *rq, int cmd)
-
drivers/net/usb/rtl8150.c:825:62-825:76: static int rtl8150_siocdevprivate(struct net_device *netdev, struct ifreq *rq,
-
drivers/net/usb/smsc75xx.c:748:54-748:68: static int smsc75xx_ioctl(struct net_device *netdev, struct ifreq *rq, int cmd)
-
drivers/net/usb/smsc95xx.c:797:54-797:68: static int smsc95xx_ioctl(struct net_device *netdev, struct ifreq *rq, int cmd)
-
drivers/net/usb/sr9700.c:239:52-239:66: static int sr9700_ioctl(struct net_device *netdev, struct ifreq *rq, int cmd)
-
drivers/net/usb/sr9800.c:488:45-488:59: static int sr_ioctl(struct net_device *net, struct ifreq *rq, int cmd)
-
drivers/net/veth.c:264:30-264:46: static void __veth_xdp_flush(struct veth_rq *rq)
-
drivers/net/veth.c:275:24-275:40: static int veth_xdp_rx(struct veth_rq *rq, struct sk_buff *skb)
-
drivers/net/veth.c:286:8-286:24: struct veth_rq *rq, bool xdp)
-
drivers/net/veth.c:525:31-525:47: static void veth_xdp_flush_bq(struct veth_rq *rq, struct veth_xdp_tx_bq *bq)
-
drivers/net/veth.c:549:28-549:44: static void veth_xdp_flush(struct veth_rq *rq, struct veth_xdp_tx_bq *bq)
-
drivers/net/veth.c:572:24-572:40: static int veth_xdp_tx(struct veth_rq *rq, struct xdp_buff *xdp,
-
drivers/net/veth.c:588:43-588:59: static struct xdp_frame *veth_xdp_rcv_one(struct veth_rq *rq,
-
drivers/net/veth.c:657:35-657:51: static void veth_xdp_rcv_bulk_skb(struct veth_rq *rq, void **frames,
-
drivers/net/veth.c:700:41-700:57: static int veth_convert_skb_to_xdp_buff(struct veth_rq *rq,
-
drivers/net/veth.c:801:41-801:57: static struct sk_buff *veth_xdp_rcv_skb(struct veth_rq *rq,
-
drivers/net/veth.c:910:25-910:41: static int veth_xdp_rcv(struct veth_rq *rq, int budget,
-
drivers/net/virtio_net.c:328:24-328:46: static void give_pages(struct receive_queue *rq, struct page *page)
-
drivers/net/virtio_net.c:338:32-338:54: static struct page *get_a_page(struct receive_queue *rq, gfp_t gfp_mask)
-
drivers/net/virtio_net.c:408:8-408:30: struct receive_queue *rq,
-
drivers/net/virtio_net.c:685:40-685:62: static struct page *xdp_linearize_page(struct receive_queue *rq,
-
drivers/net/virtio_net.c:737:10-737:32: struct receive_queue *rq,
-
drivers/net/virtio_net.c:881:8-881:30: struct receive_queue *rq,
-
drivers/net/virtio_net.c:904:7-904:29: struct receive_queue *rq,
-
drivers/net/virtio_net.c:1200:50-1200:72: static void receive_buf(struct virtnet_info *vi, struct receive_queue *rq,
-
drivers/net/virtio_net.c:1266:55-1266:77: static int add_recvbuf_small(struct virtnet_info *vi, struct receive_queue *rq,
-
drivers/net/virtio_net.c:1292:53-1292:75: static int add_recvbuf_big(struct virtnet_info *vi, struct receive_queue *rq,
-
drivers/net/virtio_net.c:1341:43-1341:65: static unsigned int get_mergeable_buf_len(struct receive_queue *rq,
-
drivers/net/virtio_net.c:1359:6-1359:28: struct receive_queue *rq, gfp_t gfp)
-
drivers/net/virtio_net.c:1408:52-1408:74: static bool try_fill_recv(struct virtnet_info *vi, struct receive_queue *rq,
-
drivers/net/virtio_net.c:1504:28-1504:50: static int virtnet_receive(struct receive_queue *rq, int budget,
-
drivers/net/virtio_net.c:1593:34-1593:56: static void virtnet_poll_cleantx(struct receive_queue *rq)
-
drivers/net/vmxnet3/vmxnet3_drv.c:560:25-560:50: vmxnet3_rq_alloc_rx_buf(struct vmxnet3_rx_queue *rq, u32 ring_idx,
-
drivers/net/vmxnet3/vmxnet3_drv.c:1263:18-1263:43: vmxnet3_rx_error(struct vmxnet3_rx_queue *rq, struct Vmxnet3_RxCompDesc *rcd,
-
drivers/net/vmxnet3/vmxnet3_drv.c:1345:24-1345:49: vmxnet3_rq_rx_complete(struct vmxnet3_rx_queue *rq,
-
drivers/net/vmxnet3/vmxnet3_drv.c:1665:20-1665:45: vmxnet3_rq_cleanup(struct vmxnet3_rx_queue *rq,
-
drivers/net/vmxnet3/vmxnet3_drv.c:1718:32-1718:57: static void vmxnet3_rq_destroy(struct vmxnet3_rx_queue *rq,
-
drivers/net/vmxnet3/vmxnet3_drv.c:1784:17-1784:42: vmxnet3_rq_init(struct vmxnet3_rx_queue *rq,
-
drivers/net/vmxnet3/vmxnet3_drv.c:1855:19-1855:44: vmxnet3_rq_create(struct vmxnet3_rx_queue *rq, struct vmxnet3_adapter *adapter)
-
drivers/net/wireless/atmel/atmel.c:2623:48-2623:62: static int atmel_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/net/wireless/cisco/airo.c:7663:56-7663:70: static int airo_siocdevprivate(struct net_device *dev, struct ifreq *rq,
-
drivers/nvme/host/core.c:668:3-668:19: struct request *rq)
-
drivers/nvme/host/core.c:680:49-680:65: bool __nvme_check_ready(struct nvme_ctrl *ctrl, struct request *rq,
-
drivers/nvme/host/core.c:975:28-975:44: static int nvme_execute_rq(struct request *rq, bool at_head)
-
drivers/nvme/host/core.c:1130:30-1130:46: int nvme_execute_passthru_rq(struct request *rq)
-
drivers/nvme/host/core.c:1158:36-1158:52: static void nvme_keep_alive_end_io(struct request *rq, blk_status_t status)
-
drivers/nvme/host/fabrics.h:190:52-190:68: static inline void nvmf_complete_timed_out_request(struct request *rq)
-
drivers/nvme/host/fc.c:1827:50-1827:66: nvme_fc_exit_request(struct blk_mq_tag_set *set, struct request *rq,
-
drivers/nvme/host/fc.c:2084:3-2084:19: struct request *rq, u32 rqno)
-
drivers/nvme/host/fc.c:2135:50-2135:66: nvme_fc_init_request(struct blk_mq_tag_set *set, struct request *rq,
-
drivers/nvme/host/fc.c:2569:17-2569:33: nvme_fc_timeout(struct request *rq, bool reserved)
-
drivers/nvme/host/fc.c:2597:45-2597:61: nvme_fc_map_data(struct nvme_fc_ctrl *ctrl, struct request *rq,
-
drivers/nvme/host/fc.c:2632:47-2632:63: nvme_fc_unmap_data(struct nvme_fc_ctrl *ctrl, struct request *rq,
-
drivers/nvme/host/fc.c:2850:21-2850:37: nvme_fc_complete_rq(struct request *rq)
-
drivers/nvme/host/nvme.h:519:28-519:44: static inline u16 nvme_cid(struct request *rq)
-
drivers/nvme/host/nvme.h:750:61-750:77: static inline bool nvme_check_ready(struct nvme_ctrl *ctrl, struct request *rq,
-
drivers/nvme/host/rdma.c:290:3-290:19: struct request *rq, unsigned int hctx_idx)
-
drivers/nvme/host/rdma.c:298:3-298:19: struct request *rq, unsigned int hctx_idx,
-
drivers/nvme/host/rdma.c:1284:62-1284:78: static void nvme_rdma_dma_unmap_req(struct ib_device *ibdev, struct request *rq)
-
drivers/nvme/host/rdma.c:1301:3-1301:19: struct request *rq)
-
drivers/nvme/host/rdma.c:1530:59-1530:75: static int nvme_rdma_dma_map_req(struct ib_device *ibdev, struct request *rq,
-
drivers/nvme/host/rdma.c:1591:3-1591:19: struct request *rq, struct nvme_command *c)
-
drivers/nvme/host/rdma.c:2007:42-2007:58: static void nvme_rdma_complete_timed_out(struct request *rq)
-
drivers/nvme/host/rdma.c:2017:19-2017:35: nvme_rdma_timeout(struct request *rq, bool reserved)
-
drivers/nvme/host/rdma.c:2171:35-2171:51: static void nvme_rdma_complete_rq(struct request *rq)
-
drivers/nvme/host/tcp.c:456:3-456:19: struct request *rq, unsigned int hctx_idx)
-
drivers/nvme/host/tcp.c:464:3-464:19: struct request *rq, unsigned int hctx_idx,
-
drivers/nvme/host/tcp.c:749:41-749:57: static inline void nvme_tcp_end_request(struct request *rq, u16 status)
-
drivers/nvme/host/tcp.c:2315:41-2315:57: static void nvme_tcp_complete_timed_out(struct request *rq)
-
drivers/nvme/host/tcp.c:2325:18-2325:34: nvme_tcp_timeout(struct request *rq, bool reserved)
-
drivers/nvme/host/tcp.c:2362:4-2362:20: struct request *rq)
-
drivers/nvme/host/tcp.c:2382:3-2382:19: struct request *rq)
-
drivers/nvme/target/passthru.c:188:37-188:53: static void nvmet_passthru_req_done(struct request *rq,
-
drivers/nvme/target/passthru.c:198:57-198:73: static int nvmet_passthru_map_sg(struct nvmet_req *req, struct request *rq)
-
drivers/pci/pci.c:6051:42-6051:46: int pcie_set_readrq(struct pci_dev *dev, int rq)
-
drivers/platform/chrome/wilco_ec/mailbox.c:92:9-92:34: struct wilco_ec_request *rq)
-
drivers/platform/chrome/wilco_ec/mailbox.c:117:9-117:34: struct wilco_ec_request *rq)
-
drivers/platform/chrome/wilco_ec/properties.c:35:9-35:37: struct ec_property_request *rq,
-
drivers/platform/chrome/wilco_ec/sysfs.c:159:5-159:32: struct usb_charge_request *rq,
-
drivers/platform/chrome/wilco_ec/telemetry.c:154:32-154:63: static int check_telem_request(struct wilco_ec_telem_request *rq,
-
drivers/ptp/ptp_clockmatrix.c:274:10-274:36: struct ptp_clock_request *rq, int on)
-
drivers/ptp/ptp_clockmatrix.c:1936:4-1936:30: struct ptp_clock_request *rq, int on)
-
drivers/ptp/ptp_dte.c:220:8-220:34: struct ptp_clock_request *rq, int on)
-
drivers/ptp/ptp_idt82p33.c:558:7-558:33: struct ptp_clock_request *rq, int on)
-
drivers/ptp/ptp_kvm_common.c:106:6-106:32: struct ptp_clock_request *rq, int on)
-
drivers/ptp/ptp_ocp.c:945:49-945:75: ptp_ocp_enable(struct ptp_clock_info *ptp_info, struct ptp_clock_request *rq,
-
drivers/ptp/ptp_pch.c:412:6-412:32: struct ptp_clock_request *rq, int on)
-
drivers/ptp/ptp_qoriq.c:266:8-266:34: struct ptp_clock_request *rq, int on)
-
drivers/scsi/elx/efct/efct_hw_queues.c:455:16-455:30: efct_hw_del_rq(struct hw_rq *rq)
-
drivers/scsi/esas2r/esas2r.h:1169:8-1169:31: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r.h:1200:43-1200:66: static inline void esas2r_rq_init_request(struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r.h:1264:44-1264:67: static inline void esas2r_rq_free_sg_lists(struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r.h:1277:46-1277:69: static inline void esas2r_rq_destroy_request(struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r.h:1304:6-1304:29: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r.h:1394:9-1394:32: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:387:6-387:29: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:460:11-460:34: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:489:10-489:33: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:503:10-503:33: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:518:12-518:35: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:549:8-549:31: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:578:11-578:34: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:625:7-625:30: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:688:7-688:30: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:740:10-740:33: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:789:8-789:31: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:826:11-826:34: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:882:8-882:31: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:940:11-940:34: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:1045:8-1045:31: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_disc.c:1084:5-1084:28: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_flash.c:134:7-134:30: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_flash.c:177:8-177:31: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_flash.c:227:50-227:73: static bool load_image(struct esas2r_adapter *a, struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_flash.c:303:11-303:34: struct esas2r_request *rq, u8 fi_stat)
-
drivers/scsi/esas2r/esas2r_flash.c:324:9-324:32: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_flash.c:828:10-828:33: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_flash.c:847:9-847:32: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_flash.c:1213:7-1213:30: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_flash.c:1258:51-1258:74: bool esas2r_nvram_write(struct esas2r_adapter *a, struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_flash.c:1391:6-1391:29: struct esas2r_request *rq, struct esas2r_sg_context *sgc)
-
drivers/scsi/esas2r/esas2r_init.c:104:6-104:29: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_init.c:1146:8-1146:31: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_int.c:173:9-173:32: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_int.c:749:51-749:74: void esas2r_ae_complete(struct esas2r_adapter *a, struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_int.c:876:54-876:77: void esas2r_dummy_complete(struct esas2r_adapter *a, struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_int.c:880:12-880:35: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_int.c:920:9-920:32: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_io.c:46:53-46:76: void esas2r_start_request(struct esas2r_adapter *a, struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_io.c:120:5-120:28: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_io.c:138:10-138:33: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_io.c:858:53-858:76: bool esas2r_ioreq_aborted(struct esas2r_adapter *a, struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_ioctl.c:83:5-83:28: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_ioctl.c:182:9-182:32: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_ioctl.c:199:6-199:29: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_ioctl.c:294:10-294:33: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_ioctl.c:331:11-331:34: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_ioctl.c:343:10-343:33: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_ioctl.c:391:11-391:34: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_ioctl.c:607:10-607:33: struct esas2r_request *rq, void *context)
-
drivers/scsi/esas2r/esas2r_ioctl.c:669:9-669:32: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_ioctl.c:688:7-688:30: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_ioctl.c:745:10-745:33: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_ioctl.c:1204:9-1204:32: struct esas2r_request *rq, void *context)
-
drivers/scsi/esas2r/esas2r_ioctl.c:1249:51-1249:74: int esas2r_write_params(struct esas2r_adapter *a, struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_ioctl.c:1807:9-1807:32: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_ioctl.c:1925:5-1925:28: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_main.c:890:11-890:34: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_main.c:1196:5-1196:28: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_main.c:1233:52-1233:75: void esas2r_wait_request(struct esas2r_adapter *a, struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_main.c:1481:52-1481:75: void esas2r_free_request(struct esas2r_adapter *a, struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_main.c:1514:5-1514:28: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_vda.c:67:10-67:33: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_vda.c:270:11-270:34: struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_vda.c:347:8-347:31: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_vda.c:373:6-373:29: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_vda.c:420:52-420:75: void esas2r_build_ae_req(struct esas2r_adapter *a, struct esas2r_request *rq)
-
drivers/scsi/esas2r/esas2r_vda.c:449:6-449:29: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_vda.c:466:8-466:31: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_vda.c:483:6-483:29: struct esas2r_request *rq,
-
drivers/scsi/esas2r/esas2r_vda.c:504:31-504:54: static void clear_vda_request(struct esas2r_request *rq)
-
drivers/scsi/fnic/fnic_fcs.c:827:37-827:53: static void fnic_rq_cmpl_frame_recv(struct vnic_rq *rq, struct cq_desc
-
drivers/scsi/fnic/fnic_fcs.c:970:25-970:41: int fnic_alloc_rq_frame(struct vnic_rq *rq)
-
drivers/scsi/fnic/fnic_fcs.c:1004:23-1004:39: void fnic_free_rq_buf(struct vnic_rq *rq, struct vnic_rq_buf *buf)
-
drivers/scsi/fnic/fnic_res.h:223:39-223:55: static inline void fnic_queue_rq_desc(struct vnic_rq *rq,
-
drivers/scsi/fnic/vnic_rq.c:27:31-27:47: static int vnic_rq_alloc_bufs(struct vnic_rq *rq)
-
drivers/scsi/fnic/vnic_rq.c:65:19-65:35: void vnic_rq_free(struct vnic_rq *rq)
-
drivers/scsi/fnic/vnic_rq.c:82:42-82:58: int vnic_rq_alloc(struct vnic_dev *vdev, struct vnic_rq *rq, unsigned int index,
-
drivers/scsi/fnic/vnic_rq.c:111:19-111:35: void vnic_rq_init(struct vnic_rq *rq, unsigned int cq_index,
-
drivers/scsi/fnic/vnic_rq.c:137:35-137:51: unsigned int vnic_rq_error_status(struct vnic_rq *rq)
-
drivers/scsi/fnic/vnic_rq.c:142:21-142:37: void vnic_rq_enable(struct vnic_rq *rq)
-
drivers/scsi/fnic/vnic_rq.c:147:21-147:37: int vnic_rq_disable(struct vnic_rq *rq)
-
drivers/scsi/fnic/vnic_rq.c:165:20-165:36: void vnic_rq_clean(struct vnic_rq *rq,
-
drivers/scsi/fnic/vnic_rq.h:105:47-105:63: static inline unsigned int vnic_rq_desc_avail(struct vnic_rq *rq)
-
drivers/scsi/fnic/vnic_rq.h:111:46-111:62: static inline unsigned int vnic_rq_desc_used(struct vnic_rq *rq)
-
drivers/scsi/fnic/vnic_rq.h:117:39-117:55: static inline void *vnic_rq_next_desc(struct vnic_rq *rq)
-
drivers/scsi/fnic/vnic_rq.h:122:47-122:63: static inline unsigned int vnic_rq_next_index(struct vnic_rq *rq)
-
drivers/scsi/fnic/vnic_rq.h:127:51-127:67: static inline unsigned int vnic_rq_next_buf_index(struct vnic_rq *rq)
-
drivers/scsi/fnic/vnic_rq.h:132:33-132:49: static inline void vnic_rq_post(struct vnic_rq *rq,
-
drivers/scsi/fnic/vnic_rq.h:165:40-165:56: static inline int vnic_rq_posting_soon(struct vnic_rq *rq)
-
drivers/scsi/fnic/vnic_rq.h:170:41-170:57: static inline void vnic_rq_return_descs(struct vnic_rq *rq, unsigned int count)
-
drivers/scsi/fnic/vnic_rq.h:180:36-180:52: static inline void vnic_rq_service(struct vnic_rq *rq,
-
drivers/scsi/fnic/vnic_rq.h:208:32-208:48: static inline int vnic_rq_fill(struct vnic_rq *rq,
-
drivers/scsi/hosts.c:569:39-569:55: static bool scsi_host_check_in_flight(struct request *rq, void *data,
-
drivers/scsi/hosts.c:665:36-665:52: static bool complete_all_cmds_iter(struct request *rq, void *data, bool rsvd)
-
drivers/scsi/lpfc/lpfc_init.c:10940:44-10940:63: lpfc_free_rq_buffer(struct lpfc_hba *phba, struct lpfc_queue *rq)
-
drivers/scsi/mpi3mr/mpi3mr_os.c:390:31-390:47: static bool mpi3mr_print_scmd(struct request *rq,
-
drivers/scsi/mpi3mr/mpi3mr_os.c:423:31-423:47: static bool mpi3mr_flush_scmd(struct request *rq,
-
drivers/scsi/mpi3mr/mpi3mr_os.c:464:38-464:54: static bool mpi3mr_count_dev_pending(struct request *rq,
-
drivers/scsi/mpi3mr/mpi3mr_os.c:498:38-498:54: static bool mpi3mr_count_tgt_pending(struct request *rq,
-
drivers/scsi/scsi_debugfs.c:34:39-34:55: void scsi_show_rq(struct seq_file *m, struct request *rq)
-
drivers/scsi/scsi_ioctl.c:345:57-345:73: static int scsi_fill_sghdr_rq(struct scsi_device *sdev, struct request *rq,
-
drivers/scsi/scsi_ioctl.c:369:35-369:51: static int scsi_complete_sghdr_rq(struct request *rq, struct sg_io_hdr *hdr,
-
drivers/scsi/scsi_lib.c:634:39-634:61: static unsigned int scsi_rq_err_bytes(const struct request *rq)
-
drivers/scsi/scsi_lib.c:995:3-995:19: struct request *rq)
-
drivers/scsi/scsi_lib.c:1116:32-1116:48: static void scsi_initialize_rq(struct request *rq)
-
drivers/scsi/scsi_lib.c:1144:29-1144:45: static void scsi_cleanup_rq(struct request *rq)
-
drivers/scsi/scsi_lib.c:1420:27-1420:43: static void scsi_complete(struct request *rq)
-
drivers/scsi/scsi_lib.c:1801:61-1801:77: static int scsi_mq_init_request(struct blk_mq_tag_set *set, struct request *rq,
-
drivers/scsi/scsi_lib.c:1829:62-1829:78: static void scsi_mq_exit_request(struct blk_mq_tag_set *set, struct request *rq,
-
drivers/scsi/sd_zbc.c:501:40-501:56: static bool sd_zbc_need_zone_wp_update(struct request *rq)
-
drivers/scsi/sg.c:1320:14-1320:30: sg_rq_end_io(struct request *rq, blk_status_t status)
-
drivers/staging/ks7010/ks_wlan_net.c:2456:57-2456:71: static int ks_wlan_netdev_ioctl(struct net_device *dev, struct ifreq *rq,
-
drivers/staging/octeon/ethernet-mdio.c:56:43-56:57: int cvm_oct_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/staging/rtl8192u/r8192U_core.c:3462:50-3462:64: static int rtl8192_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/staging/rtl8712/rtl871x_ioctl_linux.c:2178:41-2178:55: int r871x_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/staging/rtl8723bs/os_dep/ioctl_linux.c:1326:39-1326:53: int rtw_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
-
drivers/ufs/core/ufshcd-crypto.h:16:47-16:63: static inline void ufshcd_prepare_lrbp_crypto(struct request *rq,
-
drivers/ufs/core/ufshpb.c:471:51-471:70: static void ufshpb_put_req(struct ufshpb_lu *hpb, struct ufshpb_req *rq)
-
drivers/video/fbdev/xen-fbfront.c:342:40-342:44: static irqreturn_t xenfb_event_handler(int rq, void *dev_id)
-
fs/dlm/lock.c:2681:51-2681:67: static int modes_require_bast(struct dlm_lkb *gr, struct dlm_lkb *rq)
-
fs/erofs/decompressor.c:190:26-190:57: int z_erofs_fixup_insize(struct z_erofs_decompress_req *rq, const char *padbuf,
-
fs/erofs/decompressor.c:272:35-272:66: static int z_erofs_lz4_decompress(struct z_erofs_decompress_req *rq,
-
fs/erofs/decompressor.c:316:38-316:69: static int z_erofs_shifted_transform(struct z_erofs_decompress_req *rq,
-
fs/erofs/decompressor.c:374:24-374:55: int z_erofs_decompress(struct z_erofs_decompress_req *rq,
-
fs/erofs/decompressor_lzma.c:152:29-152:60: int z_erofs_lzma_decompress(struct z_erofs_decompress_req *rq,
-
fs/nfsd/nfsd.h:140:33-140:50: static inline int nfsd_v4client(struct svc_rqst *rq)
-
include/linux/blk-integrity.h:102:37-102:53: static inline bool blk_integrity_rq(struct request *rq)
-
include/linux/blk-integrity.h:111:48-111:64: static inline struct bio_vec *rq_integrity_vec(struct request *rq)
-
include/linux/blk-mq.h:203:42-203:58: static inline bool blk_rq_is_passthrough(struct request *rq)
-
include/linux/blk-mq.h:259:5-259:21: struct request *rq, struct request *prev)
-
include/linux/blk-mq.h:774:48-774:64: static inline enum mq_rq_state blk_mq_rq_state(struct request *rq)
-
include/linux/blk-mq.h:779:42-779:58: static inline int blk_mq_request_started(struct request *rq)
-
include/linux/blk-mq.h:784:44-784:60: static inline int blk_mq_request_completed(struct request *rq)
-
include/linux/blk-mq.h:796:48-796:64: static inline void blk_mq_set_request_complete(struct request *rq)
-
include/linux/blk-mq.h:805:51-805:67: static inline void blk_mq_complete_request_direct(struct request *rq,
-
include/linux/blk-mq.h:821:43-821:59: static inline bool blk_mq_need_time_stamp(struct request *rq)
-
include/linux/blk-mq.h:913:38-913:54: static inline void *blk_mq_rq_to_pdu(struct request *rq)
-
include/linux/blk-mq.h:925:38-925:54: static inline void blk_mq_cleanup_rq(struct request *rq)
-
include/linux/blk-mq.h:931:36-931:52: static inline void blk_rq_bio_prep(struct request *rq, struct bio *bio,
-
include/linux/blk-mq.h:943:31-943:47: static inline bool rq_is_sync(struct request *rq)
-
include/linux/blk-mq.h:1004:35-1004:57: static inline sector_t blk_rq_pos(const struct request *rq)
-
include/linux/blk-mq.h:1009:41-1009:63: static inline unsigned int blk_rq_bytes(const struct request *rq)
-
include/linux/blk-mq.h:1014:36-1014:58: static inline int blk_rq_cur_bytes(const struct request *rq)
-
include/linux/blk-mq.h:1023:43-1023:65: static inline unsigned int blk_rq_sectors(const struct request *rq)
-
include/linux/blk-mq.h:1028:47-1028:69: static inline unsigned int blk_rq_cur_sectors(const struct request *rq)
-
include/linux/blk-mq.h:1033:49-1033:71: static inline unsigned int blk_rq_stats_sectors(const struct request *rq)
-
include/linux/blk-mq.h:1044:49-1044:65: static inline unsigned int blk_rq_payload_bytes(struct request *rq)
-
include/linux/blk-mq.h:1055:39-1055:55: static inline struct bio_vec req_bvec(struct request *rq)
-
include/linux/blk-mq.h:1062:46-1062:62: static inline unsigned int blk_rq_count_bios(struct request *rq)
-
include/linux/blk-mq.h:1094:54-1094:70: static inline unsigned short blk_rq_nr_phys_segments(struct request *rq)
-
include/linux/blk-mq.h:1105:57-1105:73: static inline unsigned short blk_rq_nr_discard_segments(struct request *rq)
-
include/linux/blk-mq.h:1112:58-1112:74: static inline int blk_rq_map_sg(struct request_queue *q, struct request *rq,
-
include/linux/blk-mq.h:1122:43-1122:59: static inline unsigned int blk_rq_zone_no(struct request *rq)
-
include/linux/blk-mq.h:1127:47-1127:63: static inline unsigned int blk_rq_zone_is_seq(struct request *rq)
-
include/linux/blk-mq.h:1137:44-1137:60: static inline void blk_req_zone_write_lock(struct request *rq)
-
include/linux/blk-mq.h:1143:46-1143:62: static inline void blk_req_zone_write_unlock(struct request *rq)
-
include/linux/blk-mq.h:1149:49-1149:65: static inline bool blk_req_zone_is_write_locked(struct request *rq)
-
include/linux/blk-mq.h:1155:49-1155:65: static inline bool blk_req_can_dispatch_to_zone(struct request *rq)
-
include/linux/blktrace_api.h:120:44-120:60: static inline sector_t blk_rq_trace_sector(struct request *rq)
-
include/linux/blktrace_api.h:131:52-131:68: static inline unsigned int blk_rq_trace_nr_sectors(struct request *rq)
-
include/linux/mii.h:51:45-51:59: static inline struct mii_ioctl_data *if_mii(struct ifreq *rq)
-
include/linux/t10-pi.h:40:34-40:50: static inline u32 t10_pi_ref_tag(struct request *rq)
-
include/linux/t10-pi.h:71:34-71:50: static inline u64 ext_pi_ref_tag(struct request *rq)
-
include/rdma/rdmavt_qp.h:276:36-276:51: static inline u32 rvt_get_rq_count(struct rvt_rq *rq, u32 head, u32 tail)
-
include/rdma/rdmavt_qp.h:544:49-544:64: static inline struct rvt_rwqe *rvt_get_rwqe_ptr(struct rvt_rq *rq, unsigned n)
-
include/rdma/rdmavt_qp.h:952:32-952:47: static inline void rvt_free_rq(struct rvt_rq *rq)
-
include/trace/events/block.h:73:1-73:1: TRACE_EVENT(block_rq_requeue,
-
include/trace/events/block.h:147:1-147:1: DEFINE_EVENT(block_rq_completion, block_rq_complete,
-
include/trace/events/block.h:163:1-163:1: DEFINE_EVENT(block_rq_completion, block_rq_error,
-
include/trace/events/block.h:213:1-213:1: DEFINE_EVENT(block_rq, block_rq_insert,
-
include/trace/events/block.h:227:1-227:1: DEFINE_EVENT(block_rq, block_rq_issue,
-
include/trace/events/block.h:241:1-241:1: DEFINE_EVENT(block_rq, block_rq_merge,
-
include/trace/events/block.h:521:1-521:1: TRACE_EVENT(block_rq_remap,
-
include/trace/events/nbd.h:61:1-61:1: DECLARE_EVENT_CLASS(nbd_send_request,
-
include/trace/events/nbd.h:94:1-94:1: NBD_DEFINE_EVENT(nbd_send_request, nbd_send_request,
-
include/trace/events/sched.h:698:1-698:1: DECLARE_TRACE(pelt_rt_tp,
-
include/trace/events/sched.h:702:1-702:1: DECLARE_TRACE(pelt_dl_tp,
-
include/trace/events/sched.h:706:1-706:1: DECLARE_TRACE(pelt_thermal_tp,
-
include/trace/events/sched.h:710:1-710:1: DECLARE_TRACE(pelt_irq_tp,
-
include/trace/events/sched.h:718:1-718:1: DECLARE_TRACE(sched_cpu_capacity_tp,
-
include/trace/events/sched.h:734:1-734:1: DECLARE_TRACE(sched_update_nr_running_tp,
-
kernel/sched/core.c:236:25-236:36: void sched_core_enqueue(struct rq *rq, struct task_struct *p)
-
kernel/sched/core.c:246:25-246:36: void sched_core_dequeue(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/core.c:268:44-268:55: static struct task_struct *sched_core_find(struct rq *rq, unsigned long cookie)
-
kernel/sched/core.c:537:30-537:41: void raw_spin_rq_lock_nested(struct rq *rq, int subclass)
-
kernel/sched/core.c:562:26-562:37: bool raw_spin_rq_trylock(struct rq *rq)
-
kernel/sched/core.c:586:25-586:36: void raw_spin_rq_unlock(struct rq *rq)
-
kernel/sched/core.c:680:34-680:45: static void update_rq_clock_task(struct rq *rq, s64 delta)
-
kernel/sched/core.c:734:22-734:33: void update_rq_clock(struct rq *rq)
-
kernel/sched/core.c:761:26-761:37: static void hrtick_clear(struct rq *rq)
-
kernel/sched/core.c:788:30-788:41: static void __hrtick_restart(struct rq *rq)
-
kernel/sched/core.c:814:19-814:30: void hrtick_start(struct rq *rq, u64 delay)
-
kernel/sched/core.c:851:28-851:39: static void hrtick_rq_init(struct rq *rq)
-
kernel/sched/core.c:1031:19-1031:30: void resched_curr(struct rq *rq)
-
kernel/sched/core.c:1384:19-1384:30: uclamp_idle_value(struct rq *rq, enum uclamp_id clamp_id,
-
kernel/sched/core.c:1400:38-1400:49: static inline void uclamp_idle_reset(struct rq *rq, enum uclamp_id clamp_id,
-
kernel/sched/core.c:1411:34-1411:45: unsigned int uclamp_rq_max_value(struct rq *rq, enum uclamp_id clamp_id,
-
kernel/sched/core.c:1533:37-1533:48: static inline void uclamp_rq_inc_id(struct rq *rq, struct task_struct *p,
-
kernel/sched/core.c:1571:37-1571:48: static inline void uclamp_rq_dec_id(struct rq *rq, struct task_struct *p,
-
kernel/sched/core.c:1637:34-1637:45: static inline void uclamp_rq_inc(struct rq *rq, struct task_struct *p)
-
kernel/sched/core.c:1661:34-1661:45: static inline void uclamp_rq_dec(struct rq *rq, struct task_struct *p)
-
kernel/sched/core.c:1681:39-1681:50: static inline void uclamp_rq_reinc_id(struct rq *rq, struct task_struct *p,
-
kernel/sched/core.c:1983:35-1983:46: static void __init init_uclamp_rq(struct rq *rq)
-
kernel/sched/core.c:2061:33-2061:44: static inline void enqueue_task(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/core.c:2078:33-2078:44: static inline void dequeue_task(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/core.c:2095:20-2095:31: void activate_task(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/core.c:2102:22-2102:33: void deactivate_task(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/core.c:2173:40-2173:51: static inline void check_class_changed(struct rq *rq, struct task_struct *p,
-
kernel/sched/core.c:2186:25-2186:36: void check_preempt_curr(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/core.c:2210:36-2210:47: static void migrate_disable_switch(struct rq *rq, struct task_struct *p)
-
kernel/sched/core.c:2271:40-2271:51: static inline bool rq_has_pinned_tasks(struct rq *rq)
-
kernel/sched/core.c:2325:36-2325:47: static struct rq *move_queued_task(struct rq *rq, struct rq_flags *rf,
-
kernel/sched/core.c:2371:34-2371:45: static struct rq *__migrate_task(struct rq *rq, struct rq_flags *rf,
-
kernel/sched/core.c:2703:29-2703:40: static int affine_move_task(struct rq *rq, struct task_struct *p, struct rq_flags *rf,
-
kernel/sched/core.c:2849:7-2849:18: struct rq *rq,
-
kernel/sched/core.c:3620:28-3620:39: static void ttwu_do_wakeup(struct rq *rq, struct task_struct *p, int wake_flags,
-
kernel/sched/core.c:3656:18-3656:29: ttwu_do_activate(struct rq *rq, struct task_struct *p, int wake_flags,
-
kernel/sched/core.c:4782:34-4782:45: static void do_balance_callbacks(struct rq *rq, struct callback_head *head)
-
kernel/sched/core.c:4818:28-4818:39: __splice_balance_callbacks(struct rq *rq, bool split)
-
kernel/sched/core.c:4842:62-4842:73: static inline struct callback_head *splice_balance_callbacks(struct rq *rq)
-
kernel/sched/core.c:4847:33-4847:44: static void __balance_callbacks(struct rq *rq)
-
kernel/sched/core.c:4852:38-4852:49: static inline void balance_callbacks(struct rq *rq, struct callback_head *head)
-
kernel/sched/core.c:4881:21-4881:32: prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf)
-
kernel/sched/core.c:4897:39-4897:50: static inline void finish_lock_switch(struct rq *rq)
-
kernel/sched/core.c:4951:21-4951:32: prepare_task_switch(struct rq *rq, struct task_struct *prev,
-
kernel/sched/core.c:5095:16-5095:27: context_switch(struct rq *rq, struct task_struct *prev,
-
kernel/sched/core.c:5348:32-5348:43: static u64 cpu_resched_latency(struct rq *rq)
-
kernel/sched/core.c:5728:35-5728:46: static void put_prev_task_balance(struct rq *rq, struct task_struct *prev,
-
kernel/sched/core.c:5754:18-5754:29: __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
-
kernel/sched/core.c:5812:45-5812:56: static inline struct task_struct *pick_task(struct rq *rq)
-
kernel/sched/core.c:5831:16-5831:27: pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
-
kernel/sched/core.c:6126:32-6126:43: static void sched_core_balance(struct rq *rq)
-
kernel/sched/core.c:6148:32-6148:43: static void queue_core_balance(struct rq *rq)
-
kernel/sched/core.c:9082:26-9082:37: static void balance_push(struct rq *rq)
-
kernel/sched/core.c:9189:20-9189:31: void set_rq_online(struct rq *rq)
-
kernel/sched/core.c:9204:21-9204:32: void set_rq_offline(struct rq *rq)
-
kernel/sched/core.c:9425:31-9425:42: static void calc_load_migrate(struct rq *rq)
-
kernel/sched/core.c:9433:27-9433:38: static void dump_rq_tasks(struct rq *rq, const char *loglvl)
-
kernel/sched/core.c:11160:41-11160:52: void call_trace_sched_update_nr_running(struct rq *rq, int count)
-
kernel/sched/core_sched.c:239:37-239:48: void __sched_core_account_forceidle(struct rq *rq)
-
kernel/sched/core_sched.c:284:24-284:35: void __sched_core_tick(struct rq *rq)
-
kernel/sched/deadline.c:529:33-529:44: static inline int dl_overloaded(struct rq *rq)
-
kernel/sched/deadline.c:534:36-534:47: static inline void dl_set_overload(struct rq *rq)
-
kernel/sched/deadline.c:550:38-550:49: static inline void dl_clear_overload(struct rq *rq)
-
kernel/sched/deadline.c:604:38-604:49: static void enqueue_pushable_dl_task(struct rq *rq, struct task_struct *p)
-
kernel/sched/deadline.c:617:38-617:49: static void dequeue_pushable_dl_task(struct rq *rq, struct task_struct *p)
-
kernel/sched/deadline.c:633:41-633:52: static inline int has_pushable_dl_tasks(struct rq *rq)
-
kernel/sched/deadline.c:640:38-640:49: static inline bool need_pull_dl_task(struct rq *rq, struct task_struct *prev)
-
kernel/sched/deadline.c:651:46-651:57: static inline void deadline_queue_push_tasks(struct rq *rq)
-
kernel/sched/deadline.c:659:45-659:56: static inline void deadline_queue_pull_task(struct rq *rq)
-
kernel/sched/deadline.c:666:45-666:56: static struct rq *dl_task_offline_migration(struct rq *rq, struct task_struct *p)
-
kernel/sched/deadline.c:949:57-949:68: update_dl_revised_wakeup(struct sched_dl_entity *dl_se, struct rq *rq)
-
kernel/sched/deadline.c:1276:36-1276:47: static u64 grub_reclaim(u64 delta, struct rq *rq, struct sched_dl_entity *dl_se)
-
kernel/sched/deadline.c:1302:28-1302:39: static void update_curr_dl(struct rq *rq)
-
kernel/sched/deadline.c:1669:29-1669:40: static void enqueue_task_dl(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/deadline.c:1750:31-1750:42: static void __dequeue_task_dl(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/deadline.c:1757:29-1757:40: static void dequeue_task_dl(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/deadline.c:1790:27-1790:38: static void yield_task_dl(struct rq *rq)
-
kernel/sched/deadline.c:1898:36-1898:47: static void check_preempt_equal_dl(struct rq *rq, struct task_struct *p)
-
kernel/sched/deadline.c:1919:23-1919:34: static int balance_dl(struct rq *rq, struct task_struct *p, struct rq_flags *rf)
-
kernel/sched/deadline.c:1941:35-1941:46: static void check_preempt_curr_dl(struct rq *rq, struct task_struct *p,
-
kernel/sched/deadline.c:1961:29-1961:40: static void start_hrtick_dl(struct rq *rq, struct task_struct *p)
-
kernel/sched/deadline.c:1971:30-1971:41: static void set_next_task_dl(struct rq *rq, struct task_struct *p, bool first)
-
kernel/sched/deadline.c:2005:41-2005:52: static struct task_struct *pick_task_dl(struct rq *rq)
-
kernel/sched/deadline.c:2021:46-2021:57: static struct task_struct *pick_next_task_dl(struct rq *rq)
-
kernel/sched/deadline.c:2032:30-2032:41: static void put_prev_task_dl(struct rq *rq, struct task_struct *p)
-
kernel/sched/deadline.c:2055:26-2055:37: static void task_tick_dl(struct rq *rq, struct task_struct *p, int queued)
-
kernel/sched/deadline.c:2083:25-2083:36: static int pick_dl_task(struct rq *rq, struct task_struct *p, int cpu)
-
kernel/sched/deadline.c:2095:59-2095:70: static struct task_struct *pick_earliest_pushable_dl_task(struct rq *rq, int cpu)
-
kernel/sched/deadline.c:2209:64-2209:75: static struct rq *find_lock_later_rq(struct task_struct *task, struct rq *rq)
-
kernel/sched/deadline.c:2266:55-2266:66: static struct task_struct *pick_next_pushable_dl_task(struct rq *rq)
-
kernel/sched/deadline.c:2290:25-2290:36: static int push_dl_task(struct rq *rq)
-
kernel/sched/deadline.c:2368:27-2368:38: static void push_dl_tasks(struct rq *rq)
-
kernel/sched/deadline.c:2471:27-2471:38: static void task_woken_dl(struct rq *rq, struct task_struct *p)
-
kernel/sched/deadline.c:2518:26-2518:37: static void rq_online_dl(struct rq *rq)
-
kernel/sched/deadline.c:2529:27-2529:38: static void rq_offline_dl(struct rq *rq)
-
kernel/sched/deadline.c:2582:30-2582:41: static void switched_from_dl(struct rq *rq, struct task_struct *p)
-
kernel/sched/deadline.c:2630:28-2630:39: static void switched_to_dl(struct rq *rq, struct task_struct *p)
-
kernel/sched/deadline.c:2660:29-2660:40: static void prio_changed_dl(struct rq *rq, struct task_struct *p,
-
kernel/sched/debug.c:530:32-530:43: print_task(struct seq_file *m, struct rq *rq, struct task_struct *p)
-
kernel/sched/debug.c:559:42-559:53: static void print_rq(struct seq_file *m, struct rq *rq, int rq_cpu)
-
kernel/sched/fair.c:435:44-435:55: static inline void assert_list_leaf_cfs_rq(struct rq *rq)
-
kernel/sched/fair.c:928:30-928:41: static void update_curr_fair(struct rq *rq)
-
kernel/sched/fair.c:1196:34-1196:45: static void account_numa_enqueue(struct rq *rq, struct task_struct *p)
-
kernel/sched/fair.c:1202:34-1202:45: static void account_numa_dequeue(struct rq *rq, struct task_struct *p)
-
kernel/sched/fair.c:2922:28-2922:39: static void task_tick_numa(struct rq *rq, struct task_struct *curr)
-
kernel/sched/fair.c:4177:64-4177:75: static inline void update_misfit_status(struct task_struct *p, struct rq *rq)
-
kernel/sched/fair.c:5464:51-5464:62: static void __maybe_unused update_runtime_enabled(struct rq *rq)
-
kernel/sched/fair.c:5483:55-5483:66: static void __maybe_unused unthrottle_offline_cfs_rqs(struct rq *rq)
-
kernel/sched/fair.c:5563:31-5563:42: static void hrtick_start_fair(struct rq *rq, struct task_struct *p)
-
kernel/sched/fair.c:5589:27-5589:38: static void hrtick_update(struct rq *rq)
-
kernel/sched/fair.c:5616:47-5616:58: static inline void update_overutilized_status(struct rq *rq)
-
kernel/sched/fair.c:5628:26-5628:37: static int sched_idle_rq(struct rq *rq)
-
kernel/sched/fair.c:5658:19-5658:30: enqueue_task_fair(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/fair.c:5773:31-5773:42: static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/fair.c:5862:31-5862:42: static unsigned long cpu_load(struct rq *rq)
-
kernel/sched/fair.c:5880:39-5880:50: static unsigned long cpu_load_without(struct rq *rq, struct task_struct *p)
-
kernel/sched/fair.c:5898:35-5898:46: static unsigned long cpu_runnable(struct rq *rq)
-
kernel/sched/fair.c:5903:43-5903:54: static unsigned long cpu_runnable_without(struct rq *rq, struct task_struct *p)
-
kernel/sched/fair.c:6231:25-6231:36: void __update_idle_core(struct rq *rq)
-
kernel/sched/fair.c:7015:14-7015:25: balance_fair(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
-
kernel/sched/fair.c:7104:34-7104:45: static void check_preempt_wakeup(struct rq *rq, struct task_struct *p, int wake_flags)
-
kernel/sched/fair.c:7202:43-7202:54: static struct task_struct *pick_task_fair(struct rq *rq)
-
kernel/sched/fair.c:7235:21-7235:32: pick_next_task_fair(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
-
kernel/sched/fair.c:7378:50-7378:61: static struct task_struct *__pick_next_task_fair(struct rq *rq)
-
kernel/sched/fair.c:7386:32-7386:43: static void put_prev_task_fair(struct rq *rq, struct task_struct *prev)
-
kernel/sched/fair.c:7402:29-7402:40: static void yield_task_fair(struct rq *rq)
-
kernel/sched/fair.c:7433:32-7433:43: static bool yield_to_task_fair(struct rq *rq, struct task_struct *p)
-
kernel/sched/fair.c:8032:25-8032:36: static void attach_task(struct rq *rq, struct task_struct *p)
-
kernel/sched/fair.c:8045:29-8045:40: static void attach_one_task(struct rq *rq, struct task_struct *p)
-
kernel/sched/fair.c:8090:40-8090:51: static inline bool others_have_blocked(struct rq *rq)
-
kernel/sched/fair.c:8109:45-8109:56: static inline void update_blocked_load_tick(struct rq *rq)
-
kernel/sched/fair.c:8114:47-8114:58: static inline void update_blocked_load_status(struct rq *rq, bool has_blocked)
-
kernel/sched/fair.c:8126:37-8126:48: static bool __update_blocked_others(struct rq *rq, bool *done)
-
kernel/sched/fair.c:8154:35-8154:46: static bool __update_blocked_fair(struct rq *rq, bool *done)
-
kernel/sched/fair.c:8448:20-8448:31: check_cpu_capacity(struct rq *rq, struct sched_domain *sd)
-
kernel/sched/fair.c:8459:39-8459:50: static inline int check_misfit_status(struct rq *rq, struct sched_domain *sd)
-
kernel/sched/fair.c:8883:45-8883:56: static inline enum fbq_type fbq_classify_rq(struct rq *rq)
-
kernel/sched/fair.c:10285:31-10285:42: static void rebalance_domains(struct rq *rq, enum cpu_idle_type idle)
-
kernel/sched/fair.c:10367:34-10367:45: static inline int on_null_domain(struct rq *rq)
-
kernel/sched/fair.c:10441:32-10441:43: static void nohz_balancer_kick(struct rq *rq)
-
kernel/sched/fair.c:10571:29-10571:40: void nohz_balance_exit_idle(struct rq *rq)
-
kernel/sched/fair.c:10662:31-10662:42: static bool update_nohz_stats(struct rq *rq)
-
kernel/sched/fair.c:11022:27-11022:38: void trigger_load_balance(struct rq *rq)
-
kernel/sched/fair.c:11037:28-11037:39: static void rq_online_fair(struct rq *rq)
-
kernel/sched/fair.c:11044:29-11044:40: static void rq_offline_fair(struct rq *rq)
-
kernel/sched/fair.c:11065:35-11065:46: static inline void task_tick_core(struct rq *rq, struct task_struct *curr)
-
kernel/sched/fair.c:11107:27-11107:38: void task_vruntime_update(struct rq *rq, struct task_struct *p, bool in_fi)
-
kernel/sched/fair.c:11175:28-11175:39: static void task_tick_fair(struct rq *rq, struct task_struct *curr, int queued)
-
kernel/sched/fair.c:11235:19-11235:30: prio_changed_fair(struct rq *rq, struct task_struct *p, int oldprio)
-
kernel/sched/fair.c:11372:32-11372:43: static void switched_from_fair(struct rq *rq, struct task_struct *p)
-
kernel/sched/fair.c:11377:30-11377:41: static void switched_to_fair(struct rq *rq, struct task_struct *p)
-
kernel/sched/fair.c:11399:32-11399:43: static void set_next_task_fair(struct rq *rq, struct task_struct *p, bool first)
-
kernel/sched/fair.c:11741:42-11741:53: static unsigned int get_rr_interval_fair(struct rq *rq, struct task_struct *task)
-
kernel/sched/idle.c:415:14-415:25: balance_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
-
kernel/sched/idle.c:424:37-424:48: static void check_preempt_curr_idle(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/idle.c:429:32-429:43: static void put_prev_task_idle(struct rq *rq, struct task_struct *prev)
-
kernel/sched/idle.c:433:32-433:43: static void set_next_task_idle(struct rq *rq, struct task_struct *next, bool first)
-
kernel/sched/idle.c:440:43-440:54: static struct task_struct *pick_task_idle(struct rq *rq)
-
kernel/sched/idle.c:446:41-446:52: struct task_struct *pick_next_task_idle(struct rq *rq)
-
kernel/sched/idle.c:460:19-460:30: dequeue_task_idle(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/idle.c:476:28-476:39: static void task_tick_idle(struct rq *rq, struct task_struct *curr, int queued)
-
kernel/sched/idle.c:480:30-480:41: static void switched_to_idle(struct rq *rq, struct task_struct *p)
-
kernel/sched/idle.c:486:19-486:30: prio_changed_idle(struct rq *rq, struct task_struct *p, int oldprio)
-
kernel/sched/idle.c:491:30-491:41: static void update_curr_idle(struct rq *rq)
-
kernel/sched/loadavg.c:233:33-233:44: static void calc_load_nohz_fold(struct rq *rq)
-
kernel/sched/loadavg.c:258:28-258:39: void calc_load_nohz_remote(struct rq *rq)
-
kernel/sched/pelt.c:346:36-346:47: int update_rt_rq_load_avg(u64 now, struct rq *rq, int running)
-
kernel/sched/pelt.c:372:36-372:47: int update_dl_rq_load_avg(u64 now, struct rq *rq, int running)
-
kernel/sched/pelt.c:430:25-430:36: int update_irq_load_avg(struct rq *rq, u64 running)
-
kernel/sched/pelt.h:19:34-19:45: update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity)
-
kernel/sched/pelt.h:24:36-24:47: static inline u64 thermal_load_avg(struct rq *rq)
-
kernel/sched/pelt.h:76:41-76:52: static inline void update_rq_clock_pelt(struct rq *rq, s64 delta)
-
kernel/sched/pelt.h:115:46-115:57: static inline void update_idle_rq_clock_pelt(struct rq *rq)
-
kernel/sched/pelt.h:135:33-135:44: static inline u64 rq_clock_pelt(struct rq *rq)
-
kernel/sched/rt.c:321:38-321:49: static inline bool need_pull_rt_task(struct rq *rq, struct task_struct *prev)
-
kernel/sched/rt.c:327:33-327:44: static inline int rt_overloaded(struct rq *rq)
-
kernel/sched/rt.c:332:36-332:47: static inline void rt_set_overload(struct rq *rq)
-
kernel/sched/rt.c:351:38-351:49: static inline void rt_clear_overload(struct rq *rq)
-
kernel/sched/rt.c:408:38-408:49: static inline int has_pushable_tasks(struct rq *rq)
-
kernel/sched/rt.c:419:40-419:51: static inline void rt_queue_push_tasks(struct rq *rq)
-
kernel/sched/rt.c:427:39-427:50: static inline void rt_queue_pull_task(struct rq *rq)
-
kernel/sched/rt.c:432:35-432:46: static void enqueue_pushable_task(struct rq *rq, struct task_struct *p)
-
kernel/sched/rt.c:443:35-443:46: static void dequeue_pushable_task(struct rq *rq, struct task_struct *p)
-
kernel/sched/rt.c:780:31-780:42: static void __disable_runtime(struct rq *rq)
-
kernel/sched/rt.c:862:30-862:41: static void __enable_runtime(struct rq *rq)
-
kernel/sched/rt.c:1045:28-1045:39: static void update_curr_rt(struct rq *rq)
-
kernel/sched/rt.c:1536:17-1536:28: enqueue_task_rt(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/rt.c:1552:29-1552:40: static void dequeue_task_rt(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/rt.c:1580:29-1580:40: static void requeue_task_rt(struct rq *rq, struct task_struct *p, int head)
-
kernel/sched/rt.c:1591:27-1591:38: static void yield_task_rt(struct rq *rq)
-
kernel/sched/rt.c:1671:38-1671:49: static void check_preempt_equal_prio(struct rq *rq, struct task_struct *p)
-
kernel/sched/rt.c:1698:23-1698:34: static int balance_rt(struct rq *rq, struct task_struct *p, struct rq_flags *rf)
-
kernel/sched/rt.c:1719:35-1719:46: static void check_preempt_curr_rt(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/rt.c:1744:37-1744:48: static inline void set_next_task_rt(struct rq *rq, struct task_struct *p, bool first)
-
kernel/sched/rt.c:1786:47-1786:58: static struct task_struct *_pick_next_task_rt(struct rq *rq)
-
kernel/sched/rt.c:1800:41-1800:52: static struct task_struct *pick_task_rt(struct rq *rq)
-
kernel/sched/rt.c:1812:46-1812:57: static struct task_struct *pick_next_task_rt(struct rq *rq)
-
kernel/sched/rt.c:1822:30-1822:41: static void put_prev_task_rt(struct rq *rq, struct task_struct *p)
-
kernel/sched/rt.c:1847:25-1847:36: static int pick_rt_task(struct rq *rq, struct task_struct *p, int cpu)
-
kernel/sched/rt.c:1860:55-1860:66: static struct task_struct *pick_highest_pushable_task(struct rq *rq, int cpu)
-
kernel/sched/rt.c:1970:65-1970:76: static struct rq *find_lock_lowest_rq(struct task_struct *task, struct rq *rq)
-
kernel/sched/rt.c:2026:52-2026:63: static struct task_struct *pick_next_pushable_task(struct rq *rq)
-
kernel/sched/rt.c:2051:25-2051:36: static int push_rt_task(struct rq *rq, bool pull)
-
kernel/sched/rt.c:2169:27-2169:38: static void push_rt_tasks(struct rq *rq)
-
kernel/sched/rt.c:2276:30-2276:41: static void tell_cpu_to_push(struct rq *rq)
-
kernel/sched/rt.c:2460:27-2460:38: static void task_woken_rt(struct rq *rq, struct task_struct *p)
-
kernel/sched/rt.c:2474:26-2474:37: static void rq_online_rt(struct rq *rq)
-
kernel/sched/rt.c:2485:27-2485:38: static void rq_offline_rt(struct rq *rq)
-
kernel/sched/rt.c:2499:30-2499:41: static void switched_from_rt(struct rq *rq, struct task_struct *p)
-
kernel/sched/rt.c:2530:28-2530:39: static void switched_to_rt(struct rq *rq, struct task_struct *p)
-
kernel/sched/rt.c:2561:17-2561:28: prio_changed_rt(struct rq *rq, struct task_struct *p, int oldprio)
-
kernel/sched/rt.c:2598:22-2598:33: static void watchdog(struct rq *rq, struct task_struct *p)
-
kernel/sched/rt.c:2633:26-2633:37: static void task_tick_rt(struct rq *rq, struct task_struct *p, int queued)
-
kernel/sched/rt.c:2667:40-2667:51: static unsigned int get_rr_interval_rt(struct rq *rq, struct task_struct *task)
-
kernel/sched/sched.h:1129:26-1129:37: static inline int cpu_of(struct rq *rq)
-
kernel/sched/sched.h:1155:39-1155:50: static inline bool sched_core_enabled(struct rq *rq)
-
kernel/sched/sched.h:1169:40-1169:51: static inline raw_spinlock_t *rq_lockp(struct rq *rq)
-
kernel/sched/sched.h:1177:42-1177:53: static inline raw_spinlock_t *__rq_lockp(struct rq *rq)
-
kernel/sched/sched.h:1193:43-1193:54: static inline bool sched_cpu_cookie_match(struct rq *rq, struct task_struct *p)
-
kernel/sched/sched.h:1202:44-1202:55: static inline bool sched_core_cookie_match(struct rq *rq, struct task_struct *p)
-
kernel/sched/sched.h:1225:45-1225:56: static inline bool sched_group_cookie_match(struct rq *rq,
-
kernel/sched/sched.h:1293:43-1293:54: static inline void lockdep_assert_rq_held(struct rq *rq)
-
kernel/sched/sched.h:1302:37-1302:48: static inline void raw_spin_rq_lock(struct rq *rq)
-
kernel/sched/sched.h:1307:41-1307:52: static inline void raw_spin_rq_lock_irq(struct rq *rq)
-
kernel/sched/sched.h:1313:43-1313:54: static inline void raw_spin_rq_unlock_irq(struct rq *rq)
-
kernel/sched/sched.h:1319:55-1319:66: static inline unsigned long _raw_spin_rq_lock_irqsave(struct rq *rq)
-
kernel/sched/sched.h:1327:50-1327:61: static inline void raw_spin_rq_unlock_irqrestore(struct rq *rq, unsigned long flags)
-
kernel/sched/sched.h:1341:37-1341:48: static inline void update_idle_core(struct rq *rq)
-
kernel/sched/sched.h:1439:41-1439:52: static inline void assert_clock_updated(struct rq *rq)
-
kernel/sched/sched.h:1448:28-1448:39: static inline u64 rq_clock(struct rq *rq)
-
kernel/sched/sched.h:1456:33-1456:44: static inline u64 rq_clock_task(struct rq *rq)
-
kernel/sched/sched.h:1477:36-1477:47: static inline u64 rq_clock_thermal(struct rq *rq)
-
kernel/sched/sched.h:1482:41-1482:52: static inline void rq_clock_skip_update(struct rq *rq)
-
kernel/sched/sched.h:1492:47-1492:58: static inline void rq_clock_cancel_skipupdate(struct rq *rq)
-
kernel/sched/sched.h:1523:32-1523:43: static inline void rq_pin_lock(struct rq *rq, struct rq_flags *rf)
-
kernel/sched/sched.h:1536:34-1536:45: static inline void rq_unpin_lock(struct rq *rq, struct rq_flags *rf)
-
kernel/sched/sched.h:1546:34-1546:45: static inline void rq_repin_lock(struct rq *rq, struct rq_flags *rf)
-
kernel/sched/sched.h:1565:37-1565:48: static inline void __task_rq_unlock(struct rq *rq, struct rq_flags *rf)
-
kernel/sched/sched.h:1573:16-1573:27: task_rq_unlock(struct rq *rq, struct task_struct *p, struct rq_flags *rf)
-
kernel/sched/sched.h:1583:17-1583:28: rq_lock_irqsave(struct rq *rq, struct rq_flags *rf)
-
kernel/sched/sched.h:1591:13-1591:24: rq_lock_irq(struct rq *rq, struct rq_flags *rf)
-
kernel/sched/sched.h:1599:9-1599:20: rq_lock(struct rq *rq, struct rq_flags *rf)
-
kernel/sched/sched.h:1607:22-1607:33: rq_unlock_irqrestore(struct rq *rq, struct rq_flags *rf)
-
kernel/sched/sched.h:1615:15-1615:26: rq_unlock_irq(struct rq *rq, struct rq_flags *rf)
-
kernel/sched/sched.h:1623:11-1623:22: rq_unlock(struct rq *rq, struct rq_flags *rf)
-
kernel/sched/sched.h:1690:24-1690:35: queue_balance_callback(struct rq *rq,
-
kernel/sched/sched.h:1850:49-1850:60: static inline void sched_core_account_forceidle(struct rq *rq)
-
kernel/sched/sched.h:1858:36-1858:47: static inline void sched_core_tick(struct rq *rq)
-
kernel/sched/sched.h:2016:32-2016:43: static inline int task_current(struct rq *rq, struct task_struct *p)
-
kernel/sched/sched.h:2021:32-2021:43: static inline int task_running(struct rq *rq, struct task_struct *p)
-
kernel/sched/sched.h:2174:34-2174:45: static inline void put_prev_task(struct rq *rq, struct task_struct *prev)
-
kernel/sched/sched.h:2180:34-2180:45: static inline void set_next_task(struct rq *rq, struct task_struct *next)
-
kernel/sched/sched.h:2219:40-2219:51: static inline bool sched_stop_runnable(struct rq *rq)
-
kernel/sched/sched.h:2224:38-2224:49: static inline bool sched_dl_runnable(struct rq *rq)
-
kernel/sched/sched.h:2229:38-2229:49: static inline bool sched_rt_runnable(struct rq *rq)
-
kernel/sched/sched.h:2234:40-2234:51: static inline bool sched_fair_runnable(struct rq *rq)
-
kernel/sched/sched.h:2255:49-2255:60: static inline struct task_struct *get_push_task(struct rq *rq)
-
kernel/sched/sched.h:2279:35-2279:46: static inline void idle_set_state(struct rq *rq,
-
kernel/sched/sched.h:2285:52-2285:63: static inline struct cpuidle_state *idle_get_state(struct rq *rq)
-
kernel/sched/sched.h:2359:49-2359:60: static inline void sched_update_tick_dependency(struct rq *rq) { }
-
kernel/sched/sched.h:2362:35-2362:46: static inline void add_nr_running(struct rq *rq, unsigned count)
-
kernel/sched/sched.h:2381:35-2381:46: static inline void sub_nr_running(struct rq *rq, unsigned count)
-
kernel/sched/sched.h:2423:34-2423:45: static inline int hrtick_enabled(struct rq *rq)
-
kernel/sched/sched.h:2430:39-2430:50: static inline int hrtick_enabled_fair(struct rq *rq)
-
kernel/sched/sched.h:2437:37-2437:48: static inline int hrtick_enabled_dl(struct rq *rq)
-
kernel/sched/sched.h:2812:40-2812:51: static inline void cpufreq_update_util(struct rq *rq, unsigned int flags)
-
kernel/sched/sched.h:2858:39-2858:50: static inline unsigned long cpu_bw_dl(struct rq *rq)
-
kernel/sched/sched.h:2863:41-2863:52: static inline unsigned long cpu_util_dl(struct rq *rq)
-
kernel/sched/sched.h:2916:41-2916:52: static inline unsigned long cpu_util_rt(struct rq *rq)
-
kernel/sched/sched.h:2943:35-2943:46: unsigned long uclamp_rq_util_with(struct rq *rq, unsigned long util,
-
kernel/sched/sched.h:2979:40-2979:51: static inline bool uclamp_rq_is_capped(struct rq *rq)
-
kernel/sched/sched.h:3022:42-3022:53: static inline unsigned long cpu_util_irq(struct rq *rq)
-
kernel/sched/sched.h:3074:41-3074:52: static inline void membarrier_switch_mm(struct rq *rq,
-
kernel/sched/stats.c:6:32-6:43: void __update_stats_wait_start(struct rq *rq, struct task_struct *p,
-
kernel/sched/stats.c:20:30-20:41: void __update_stats_wait_end(struct rq *rq, struct task_struct *p,
-
kernel/sched/stats.c:47:37-47:48: void __update_stats_enqueue_sleeper(struct rq *rq, struct task_struct *p,
-
kernel/sched/stats.h:13:22-13:33: rq_sched_info_arrive(struct rq *rq, unsigned long long delta)
-
kernel/sched/stats.h:25:22-25:33: rq_sched_info_depart(struct rq *rq, unsigned long long delta)
-
kernel/sched/stats.h:32:23-32:34: rq_sched_info_dequeue(struct rq *rq, unsigned long long delta)
-
kernel/sched/stats.h:213:39-213:50: static inline void sched_info_dequeue(struct rq *rq, struct task_struct *t)
-
kernel/sched/stats.h:232:31-232:42: static void sched_info_arrive(struct rq *rq, struct task_struct *t)
-
kernel/sched/stats.h:254:39-254:50: static inline void sched_info_enqueue(struct rq *rq, struct task_struct *t)
-
kernel/sched/stats.h:268:38-268:49: static inline void sched_info_depart(struct rq *rq, struct task_struct *t)
-
kernel/sched/stats.h:284:19-284:30: sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next)
-
kernel/sched/stop_task.c:19:14-19:25: balance_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
-
kernel/sched/stop_task.c:26:25-26:36: check_preempt_curr_stop(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/stop_task.c:31:32-31:43: static void set_next_task_stop(struct rq *rq, struct task_struct *stop, bool first)
-
kernel/sched/stop_task.c:36:43-36:54: static struct task_struct *pick_task_stop(struct rq *rq)
-
kernel/sched/stop_task.c:44:48-44:59: static struct task_struct *pick_next_task_stop(struct rq *rq)
-
kernel/sched/stop_task.c:55:19-55:30: enqueue_task_stop(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/stop_task.c:61:19-61:30: dequeue_task_stop(struct rq *rq, struct task_struct *p, int flags)
-
kernel/sched/stop_task.c:66:29-66:40: static void yield_task_stop(struct rq *rq)
-
kernel/sched/stop_task.c:71:32-71:43: static void put_prev_task_stop(struct rq *rq, struct task_struct *prev)
-
kernel/sched/stop_task.c:98:28-98:39: static void task_tick_stop(struct rq *rq, struct task_struct *curr, int queued)
-
kernel/sched/stop_task.c:102:30-102:41: static void switched_to_stop(struct rq *rq, struct task_struct *p)
-
kernel/sched/stop_task.c:108:19-108:30: prio_changed_stop(struct rq *rq, struct task_struct *p, int oldprio)
-
kernel/sched/stop_task.c:113:30-113:41: static void update_curr_stop(struct rq *rq)
-
kernel/sched/topology.c:485:21-485:32: void rq_attach_root(struct rq *rq, struct root_domain *rd)
-
kernel/trace/blktrace.c:804:28-804:44: blk_trace_request_get_cgid(struct request *rq)
-
kernel/trace/blktrace.c:828:30-828:46: static void blk_add_trace_rq(struct request *rq, blk_status_t error,
-
kernel/trace/blktrace.c:851:51-851:67: static void blk_add_trace_rq_insert(void *ignore, struct request *rq)
-
kernel/trace/blktrace.c:857:50-857:66: static void blk_add_trace_rq_issue(void *ignore, struct request *rq)
-
kernel/trace/blktrace.c:863:50-863:66: static void blk_add_trace_rq_merge(void *ignore, struct request *rq)
-
kernel/trace/blktrace.c:869:52-869:68: static void blk_add_trace_rq_requeue(void *ignore, struct request *rq)
-
kernel/trace/blktrace.c:875:53-875:69: static void blk_add_trace_rq_complete(void *ignore, struct request *rq,
-
kernel/trace/blktrace.c:1043:50-1043:66: static void blk_add_trace_rq_remap(void *ignore, struct request *rq, dev_t dev,
-
kernel/trace/blktrace.c:1076:26-1076:42: void blk_add_driver_data(struct request *rq, void *data, size_t len)
-
net/bridge/br_ioctl.c:144:51-144:65: int br_dev_siocdevprivate(struct net_device *dev, struct ifreq *rq,
-
net/key/af_key.c:1940:44-1940:72: parse_ipsecrequest(struct xfrm_policy *xp, struct sadb_x_ipsecrequest *rq)
record
Declared as a prototype...
Defined...
variable
Defined...
-
block/bfq-iosched.c:1025:2-1025:18: struct request *rq;
-
block/bfq-iosched.c:5023:2-5023:29: struct request *rq = bfqq->next_rq;
-
block/bfq-iosched.c:5082:2-5082:23: struct request *rq = NULL;
-
block/bfq-iosched.c:5222:2-5222:18: struct request *rq;
-
block/bfq-iosched.c:6198:3-6198:19: struct request *rq;
-
block/blk-flush.c:225:2-225:18: struct request *rq, *n;
-
block/blk-merge.c:1041:2-1041:18: struct request *rq;
-
block/blk-merge.c:1072:2-1072:18: struct request *rq;
-
block/blk-merge.c:1097:2-1097:18: struct request *rq;
-
block/blk-mq-sched.c:63:2-63:18: struct request *rq;
-
block/blk-mq-sched.c:106:3-106:19: struct request *rq;
-
block/blk-mq-sched.c:224:2-224:18: struct request *rq;
-
block/blk-mq-tag.c:252:2-252:18: struct request *rq;
-
block/blk-mq-tag.c:271:2-271:18: struct request *rq;
-
block/blk-mq-tag.c:341:2-341:18: struct request *rq;
-
block/blk-mq.c:347:2-347:43: struct request *rq = tags->static_rqs[tag];
-
block/blk-mq.c:414:2-414:18: struct request *rq;
-
block/blk-mq.c:444:2-444:18: struct request *rq;
-
block/blk-mq.c:519:2-519:18: struct request *rq;
-
block/blk-mq.c:647:2-647:18: struct request *rq;
-
block/blk-mq.c:972:2-972:18: struct request *rq;
-
block/blk-mq.c:1012:2-1012:18: struct request *rq, *next;
-
block/blk-mq.c:1325:2-1325:18: struct request *rq, *next;
-
block/blk-mq.c:1834:2-1834:18: struct request *rq;
-
block/blk-mq.c:1852:2-1852:18: struct request *rq, *nxt;
-
block/blk-mq.c:2396:2-2396:18: struct request *rq;
-
block/blk-mq.c:2550:2-2550:18: struct request *rq;
-
block/blk-mq.c:2606:3-2606:24: struct request *rq = rq_list_pop(&plug->mq_list);
-
block/blk-mq.c:2626:2-2626:18: struct request *rq;
-
block/blk-mq.c:2675:3-2675:24: struct request *rq = list_first_entry(list, struct request,
-
block/blk-mq.c:2725:2-2725:18: struct request *rq;
-
block/blk-mq.c:2755:2-2755:18: struct request *rq;
-
block/blk-mq.c:2804:2-2804:18: struct request *rq;
-
block/blk-mq.c:3048:4-3048:40: struct request *rq = drv_tags->rqs[i];
-
block/blk-mq.c:3086:4-3086:43: struct request *rq = tags->static_rqs[i];
-
block/blk-mq.c:3255:4-3255:25: struct request *rq = p;
-
block/blk-mq.c:4634:2-4634:44: struct request *rq = blk_qc_to_rq(hctx, qc);
-
block/blk-mq.h:330:3-330:24: struct request *rq = list_entry_rq(list->next);
-
block/bsg-lib.c:32:2-32:18: struct request *rq;
-
block/bsg-lib.c:159:2-159:45: struct request *rq = blk_mq_rq_from_pdu(job);
-
block/bsg-lib.c:192:2-192:45: struct request *rq = blk_mq_rq_from_pdu(job);
-
block/elevator.c:238:2-238:18: struct request *rq;
-
block/elevator.c:291:2-291:18: struct request *rq;
-
block/kyber-iosched.c:596:2-596:18: struct request *rq, *next;
-
block/kyber-iosched.c:760:2-760:18: struct request *rq;
-
block/kyber-iosched.c:808:2-808:18: struct request *rq;
-
block/mq-deadline.c:269:2-269:23: struct request *rq = rq_entry_fifo(per_prio->fifo_list[data_dir].next);
-
block/mq-deadline.c:288:2-288:18: struct request *rq;
-
block/mq-deadline.c:322:2-322:18: struct request *rq;
-
block/mq-deadline.c:369:2-369:18: struct request *rq, *next_rq;
-
block/mq-deadline.c:485:2-485:18: struct request *rq;
-
block/mq-deadline.c:518:2-518:18: struct request *rq;
-
block/mq-deadline.c:777:3-777:19: struct request *rq;
-
block/mq-deadline.c:975:1-975:1: DEADLINE_DEBUGFS_DDIR_ATTRS(DD_RT_PRIO, DD_READ, read0);
-
block/mq-deadline.c:976:1-976:1: DEADLINE_DEBUGFS_DDIR_ATTRS(DD_RT_PRIO, DD_WRITE, write0);
-
block/mq-deadline.c:977:1-977:1: DEADLINE_DEBUGFS_DDIR_ATTRS(DD_BE_PRIO, DD_READ, read1);
-
block/mq-deadline.c:978:1-978:1: DEADLINE_DEBUGFS_DDIR_ATTRS(DD_BE_PRIO, DD_WRITE, write1);
-
block/mq-deadline.c:979:1-979:1: DEADLINE_DEBUGFS_DDIR_ATTRS(DD_IDLE_PRIO, DD_READ, read2);
-
block/mq-deadline.c:980:1-980:1: DEADLINE_DEBUGFS_DDIR_ATTRS(DD_IDLE_PRIO, DD_WRITE, write2);
-
drivers/ata/libata-scsi.c:1495:2-1495:42: struct request *rq = scsi_cmd_to_rq(scmd);
-
drivers/ata/libata-scsi.c:1530:2-1530:42: struct request *rq = scsi_cmd_to_rq(scmd);
-
drivers/block/aoe/aoecmd.c:836:2-836:18: struct request *rq;
-
drivers/block/aoe/aoecmd.c:1056:2-1056:28: struct request *rq = buf->rq;
-
drivers/block/aoe/aoedev.c:162:2-162:18: struct request *rq;
-
drivers/block/loop.c:376:2-376:45: struct request *rq = blk_mq_rq_from_pdu(cmd);
-
drivers/block/loop.c:400:2-400:45: struct request *rq = blk_mq_rq_from_pdu(cmd);
-
drivers/block/loop.c:1800:2-1800:27: struct request *rq = bd->rq;
-
drivers/block/loop.c:1842:2-1842:45: struct request *rq = blk_mq_rq_from_pdu(cmd);
-
drivers/block/mtip32xx/mtip32xx.c:964:2-964:18: struct request *rq;
-
drivers/block/mtip32xx/mtip32xx.c:3444:2-3444:27: struct request *rq = bd->rq;
-
drivers/block/nbd.c:832:2-832:18: struct request *rq;
-
drivers/block/null_blk/main.c:1210:2-1210:28: struct request *rq = cmd->rq;
-
drivers/block/null_blk/main.c:1282:2-1282:28: struct request *rq = cmd->rq;
-
drivers/block/paride/pd.c:775:2-775:18: struct request *rq;
-
drivers/block/pktcdvd.c:688:2-688:18: struct request *rq;
-
drivers/block/rbd.c:3586:3-3586:50: struct request *rq = blk_mq_rq_from_pdu(img_req);
-
drivers/block/rbd.c:4640:2-4640:53: struct request *rq = blk_mq_rq_from_pdu(img_request);
-
drivers/block/rnbd/rnbd-clt.c:406:2-406:27: struct request *rq = iu->rq;
-
drivers/block/rnbd/rnbd-clt.c:1131:2-1131:27: struct request *rq = bd->rq;
-
drivers/block/sx8.c:509:2-509:18: struct request *rq;
-
drivers/block/sx8.c:558:2-558:18: struct request *rq;
-
drivers/block/sx8.c:707:2-707:27: struct request *rq = bd->rq;
-
drivers/block/sx8.c:911:2-911:18: struct request *rq;
-
drivers/block/xen-blkfront.c:937:2-937:35: struct request_queue *rq = info->rq;
-
drivers/block/xen-blkfront.c:1593:5-1593:38: struct request_queue *rq = info->rq;
-
drivers/gpu/drm/amd/amdgpu/amdgpu_job.c:287:3-287:47: struct drm_sched_rq *rq = &sched->sched_rq[i];
-
drivers/gpu/drm/i915/display/intel_atomic_plane.c:915:2-915:34: struct i915_request *rq = wait->request;
-
drivers/gpu/drm/i915/display/intel_overlay.c:233:2-233:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/display/intel_overlay.c:255:2-255:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/display/intel_overlay.c:320:2-320:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/display/intel_overlay.c:401:2-401:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/display/intel_overlay.c:454:2-454:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/i915_gem_busy.c:43:2-43:29: const struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/i915_gem_context.c:1335:2-1335:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c:1987:3-1987:43: struct i915_request *rq = eb->requests[j];
-
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c:2438:2-2438:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c:2472:2-2472:28: struct i915_request *rq = NULL;
-
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c:3010:2-3010:23: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c:3072:3-3072:43: struct i915_request *rq = eb->requests[i];
-
drivers/gpu/drm/i915/gem/i915_gem_throttle.c:64:4-64:25: struct i915_request *rq, *target = NULL;
-
drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c:175:2-175:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/i915_gem_wait.c:61:2-61:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_client_blt.c:340:2-340:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_coherency.c:198:2-198:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c:74:3-74:29: struct i915_request *rq = NULL;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c:194:3-194:29: struct i915_request *rq = NULL;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c:231:2-231:28: struct i915_request *rq = NULL;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c:921:2-921:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c:1028:2-1028:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c:1074:2-1074:28: struct i915_request *rq = NULL;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c:1494:2-1494:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c:1586:2-1586:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_migrate.c:188:2-188:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_migrate.c:341:3-341:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c:542:3-542:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c:1170:2-1170:28: struct i915_request *rq = NULL;
-
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c:1509:3-1509:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/selftests/igt_gem_utils.c:24:2-24:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gem/selftests/igt_gem_utils.c:115:2-115:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c:207:3-207:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c:245:3-246:4: struct i915_request *rq =
-
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c:457:2-457:23: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/gt/intel_breadcrumbs.c:490:2-490:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_context.c:496:2-496:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_context.c:534:2-534:23: struct i915_request *rq, *active = NULL;
-
drivers/gpu/drm/i915/gt/intel_context_sseu.c:43:2-43:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_engine_cs.c:1727:3-1727:39: struct i915_request * const *port, *rq;
-
drivers/gpu/drm/i915/gt/intel_engine_cs.c:1917:2-1917:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_engine_cs.c:1979:2-1979:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c:42:2-42:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c:113:2-113:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c:250:2-250:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_engine_heartbeat.c:341:2-341:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_engine_pm.c:85:2-85:44: struct i915_request *rq = to_request(fence);
-
drivers/gpu/drm/i915/gt/intel_engine_pm.c:128:2-128:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_engine_pm.h:90:2-90:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:368:2-368:23: struct i915_request *rq, *rn, *active = NULL;
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:773:2-773:38: struct i915_request * const *port, *rq, *prev = NULL;
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:921:3-921:49: struct i915_request *rq = execlists->pending[n];
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:1022:3-1022:29: struct i915_request *rq = READ_ONCE(ve->request);
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:1114:2-1114:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:1400:3-1400:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:1479:3-1479:24: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:1954:5-1954:43: struct i915_request *rq = *execlists->active;
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:2271:2-2271:38: struct i915_request * const *port, *rq;
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:2984:2-2984:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:3105:2-3105:23: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:3740:2-3740:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:4036:2-4036:23: struct i915_request *rq, *last;
-
drivers/gpu/drm/i915/gt/intel_execlists_submission.c:4090:3-4090:29: struct i915_request *rq = READ_ONCE(ve->request);
-
drivers/gpu/drm/i915/gt/intel_gt.c:537:3-537:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_gt.c:585:3-585:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_gt.c:621:3-621:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_gt_requests.c:19:2-19:23: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/gt/intel_gt_requests.c:243:2-243:23: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/gt/intel_migrate.c:709:2-709:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_migrate.c:967:2-967:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/intel_ring_submission.c:350:2-350:29: struct i915_request *pos, *rq;
-
drivers/gpu/drm/i915/gt/intel_timeline.c:424:3-424:24: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/gt/intel_workarounds.c:2783:2-2783:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/mock_engine.c:297:2-297:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_context.c:51:3-51:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_context.c:78:2-78:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_context.c:234:3-234:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_context.c:327:2-327:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_engine_cs.c:155:4-155:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_engine_cs.c:295:4-295:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_engine_heartbeat.c:207:2-207:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_engine_pm.c:78:2-78:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_engine_pm.c:260:3-260:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:127:3-127:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:184:3-184:28: struct i915_request *rq[2];
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:346:3-346:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:495:3-495:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:603:3-603:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:722:5-722:26: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:862:2-862:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:897:2-897:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:948:4-948:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:1058:2-1058:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:1132:3-1132:33: struct i915_request *rq[3] = {};
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:1261:2-1261:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:1332:3-1332:24: struct i915_request *rq, *nop;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:1432:3-1432:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:1719:2-1719:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:1757:3-1757:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:1850:3-1850:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:2056:2-2056:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:2100:2-2100:32: struct i915_request *rq[2] = {};
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:2171:2-2171:32: struct i915_request *rq[3] = {};
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:2260:2-2260:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:2317:2-2317:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:2568:3-2568:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:2698:2-2698:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:2797:2-2797:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:2971:3-2971:29: struct i915_request *rq = NULL;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:3149:2-3149:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:3216:2-3216:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:3301:4-3301:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:3400:3-3400:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:3490:2-3490:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:3762:6-3762:27: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:3779:6-3779:27: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:4035:2-4035:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:4102:2-4102:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:4234:3-4234:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_execlists.c:4345:2-4345:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:124:2-124:28: struct i915_request *rq = NULL;
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:302:2-302:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:390:5-390:26: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:489:5-489:26: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:610:5-610:26: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:745:4-745:30: struct i915_request *rq = NULL;
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:908:2-908:32: struct i915_request *rq[8] = {};
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:1053:4-1053:30: struct i915_request *rq = NULL;
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:1301:2-1301:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:1433:2-1433:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:1689:4-1689:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:1819:2-1819:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_hangcheck.c:1916:2-1916:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:80:2-80:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:107:2-107:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:398:2-398:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:528:2-528:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:563:2-563:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:618:2-618:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:740:2-740:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:798:2-798:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:1048:2-1048:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:1190:2-1190:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:1351:2-1351:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:1506:2-1506:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:1681:2-1681:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_lrc.c:1797:2-1797:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_migrate.c:44:2-44:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_migrate.c:145:2-145:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_migrate.c:263:2-263:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_migrate.c:696:3-696:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_migrate.c:774:3-774:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_mocs.c:219:2-219:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_mocs.c:326:2-326:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_rc6.c:127:2-127:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_reset.c:60:3-60:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_ring_submission.c:72:2-72:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_rps.c:244:3-244:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_rps.c:402:3-402:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_rps.c:630:3-630:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_rps.c:771:3-771:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_rps.c:900:2-900:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_rps.c:1155:3-1155:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_rps.c:1257:3-1257:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_slpc.c:75:3-75:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_slpc.c:205:3-205:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:487:2-487:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:554:4-554:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:624:4-624:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:696:3-696:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:895:2-895:28: struct i915_request *rq = fetch_and_zero(&w->rq);
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:941:2-941:23: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:1025:4-1025:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:1186:3-1186:33: struct i915_request *rq[3] = {};
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:1262:3-1262:33: struct i915_request *rq[3] = {};
-
drivers/gpu/drm/i915/gt/selftest_timeline.c:1362:4-1362:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_workarounds.c:104:2-104:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_workarounds.c:307:2-307:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_workarounds.c:527:3-527:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_workarounds.c:854:2-854:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_workarounds.c:901:2-901:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/selftest_workarounds.c:1259:2-1259:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:880:3-880:24: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:1624:2-1624:23: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:1660:2-1660:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:1747:2-1747:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:1762:2-1762:23: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:3189:2-3189:28: struct i915_request *rq = container_of(wrk, typeof(*rq), submit_work);
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:3197:2-3197:23: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:4363:2-4363:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c:4470:3-4470:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/uc/selftest_guc.c:24:2-24:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gt/uc/selftest_guc.c:52:2-52:54: struct i915_request *last[3] = {NULL, NULL, NULL}, *rq;
-
drivers/gpu/drm/i915/gt/uc/selftest_guc.c:148:2-148:40: struct i915_request *spin_rq = NULL, *rq, *last = NULL;
-
drivers/gpu/drm/i915/gt/uc/selftest_guc_multi_lrc.c:79:2-79:23: struct i915_request *rq, *child_rq;
-
drivers/gpu/drm/i915/gt/uc/selftest_guc_multi_lrc.c:113:2-113:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gvt/scheduler.c:289:2-289:28: struct i915_request *rq = data;
-
drivers/gpu/drm/i915/gvt/scheduler.c:463:2-463:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gvt/scheduler.c:595:2-595:38: struct i915_request *rq = workload->req;
-
drivers/gpu/drm/i915/gvt/scheduler.c:802:2-802:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/gvt/scheduler.c:937:2-937:38: struct i915_request *rq = workload->req;
-
drivers/gpu/drm/i915/gvt/scheduler.c:1072:2-1072:38: struct i915_request *rq = workload->req;
-
drivers/gpu/drm/i915/i915_gpu_error.c:1560:2-1560:28: struct i915_request *rq = NULL;
-
drivers/gpu/drm/i915/i915_perf.c:1978:2-1978:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/i915_perf.c:2204:2-2204:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/i915_perf.c:2225:2-2225:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/i915_request.c:115:2-115:44: struct i915_request *rq = to_request(fence);
-
drivers/gpu/drm/i915/i915_request.c:262:2-263:3: struct i915_request *rq =
-
drivers/gpu/drm/i915/i915_request.c:419:2-419:38: struct i915_request * const *port, *rq;
-
drivers/gpu/drm/i915/i915_request.c:789:2-789:28: struct i915_request *rq = container_of(fence, typeof(*rq), semaphore);
-
drivers/gpu/drm/i915/i915_request.c:805:2-805:23: struct i915_request *rq, *rn;
-
drivers/gpu/drm/i915/i915_request.c:817:2-817:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/i915_request.c:853:2-853:28: struct i915_request *rq = arg;
-
drivers/gpu/drm/i915/i915_request.c:876:2-876:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/i915_request.c:1021:2-1021:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/i915_scheduler.c:133:2-133:54: const struct i915_request *rq = node_to_request(node);
-
drivers/gpu/drm/i915/pxp/intel_pxp_cmd.c:97:2-97:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_active.c:101:3-101:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_gem.c:28:3-28:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_gem_evict.c:456:4-456:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c:1950:2-1950:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c:1969:2-1969:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c:2123:4-2123:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c:2224:5-2224:26: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_perf.c:195:2-195:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_perf.c:288:2-288:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:356:4-356:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:395:4-395:48: struct i915_request *rq = requests[count - 1];
-
drivers/gpu/drm/i915/selftests/i915_request.c:410:4-410:40: struct i915_request *rq = requests[n];
-
drivers/gpu/drm/i915/selftests/i915_request.c:627:2-627:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:680:2-680:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:741:2-741:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:802:2-802:23: struct i915_request *rq, *nop;
-
drivers/gpu/drm/i915/selftests/i915_request.c:1432:3-1432:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:1468:3-1468:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:1512:2-1512:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:1634:2-1634:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:1829:2-1829:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:1958:2-1958:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:2049:3-2049:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:2121:3-2121:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:2181:2-2181:23: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:2236:3-2236:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:2326:4-2326:25: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:2417:3-2417:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:2529:3-2529:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:2677:3-2677:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:2711:3-2711:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:2745:3-2745:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:2924:3-2924:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:2995:3-2995:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/i915_request.c:3067:3-3067:24: struct i915_request *rq;
-
drivers/gpu/drm/i915/selftests/igt_spinner.c:144:2-144:28: struct i915_request *rq = NULL;
-
drivers/gpu/drm/i915/selftests/intel_memory_region.c:1018:2-1018:23: struct i915_request *rq;
-
drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c:39:2-39:23: int rq = ram->freq < 1000000; /* XXX */
-
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c:133:2-133:14: u32 hi, lo, rq, tx;
-
drivers/gpu/drm/scheduler/sched_entity.c:453:2-453:23: struct drm_sched_rq *rq;
-
drivers/gpu/drm/scheduler/sched_main.c:1050:3-1050:47: struct drm_sched_rq *rq = &sched->sched_rq[i];
-
drivers/gpu/drm/scheduler/sched_main.c:1103:4-1103:48: struct drm_sched_rq *rq = &sched->sched_rq[i];
-
drivers/infiniband/hw/bnxt_re/ib_verbs.c:1095:2-1095:23: struct bnxt_qplib_q *rq;
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:177:2-177:33: struct bnxt_qplib_q *rq = &qp->rq;
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:199:2-199:33: struct bnxt_qplib_q *rq = &qp->rq;
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:815:2-815:33: struct bnxt_qplib_q *rq = &qp->rq;
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:956:2-956:33: struct bnxt_qplib_q *rq = &qp->rq;
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:1525:2-1525:33: struct bnxt_qplib_q *rq = &qp->rq;
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:1538:2-1538:33: struct bnxt_qplib_q *rq = &qp->rq;
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:1948:2-1948:33: struct bnxt_qplib_q *rq = &qp->rq;
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:1957:2-1957:33: struct bnxt_qplib_q *rq = &qp->rq;
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:2440:2-2440:23: struct bnxt_qplib_q *rq;
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:2521:2-2521:23: struct bnxt_qplib_q *rq;
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:2620:2-2620:23: struct bnxt_qplib_q *rq;
-
drivers/infiniband/hw/bnxt_re/qplib_fp.c:2714:2-2714:28: struct bnxt_qplib_q *sq, *rq;
-
drivers/infiniband/hw/mlx5/devx.c:638:4-638:44: struct mlx5_ib_rq *rq = &raw_packet_qp->rq;
-
drivers/infiniband/hw/mlx5/qp.c:1493:2-1493:42: struct mlx5_ib_rq *rq = &raw_packet_qp->rq;
-
drivers/infiniband/hw/mlx5/qp.c:1585:2-1585:42: struct mlx5_ib_rq *rq = &raw_packet_qp->rq;
-
drivers/infiniband/hw/mlx5/qp.c:1602:2-1602:42: struct mlx5_ib_rq *rq = &raw_packet_qp->rq;
-
drivers/infiniband/hw/mlx5/qp.c:3815:2-3815:42: struct mlx5_ib_rq *rq = &raw_packet_qp->rq;
-
drivers/infiniband/hw/mlx5/qp.c:4776:2-4776:42: struct mlx5_ib_rq *rq = &raw_packet_qp->rq;
-
drivers/infiniband/sw/rdmavt/qp.c:2348:2-2348:17: struct rvt_rq *rq;
-
drivers/infiniband/sw/rxe/rxe_verbs.c:729:2-729:27: struct rxe_rq *rq = &qp->rq;
-
drivers/infiniband/ulp/srp/ib_srp.c:2162:2-2162:43: struct request *rq = scsi_cmd_to_rq(scmnd);
-
drivers/isdn/hardware/mISDN/avmfritz.c:920:2-920:22: struct channel_req *rq;
-
drivers/isdn/hardware/mISDN/hfcmulti.c:4170:2-4170:22: struct channel_req *rq;
-
drivers/isdn/hardware/mISDN/hfcpci.c:1952:2-1952:22: struct channel_req *rq;
-
drivers/isdn/hardware/mISDN/hfcsusb.c:522:2-522:22: struct channel_req *rq;
-
drivers/isdn/hardware/mISDN/mISDNipac.c:1531:2-1531:22: struct channel_req *rq;
-
drivers/isdn/hardware/mISDN/netjet.c:875:2-875:22: struct channel_req *rq;
-
drivers/isdn/hardware/mISDN/speedfax.c:241:2-241:22: struct channel_req *rq;
-
drivers/isdn/hardware/mISDN/w6692.c:1181:2-1181:22: struct channel_req *rq;
-
drivers/isdn/mISDN/l1oip_core.c:1034:2-1034:22: struct channel_req *rq;
-
drivers/isdn/mISDN/layer2.c:2112:2-2112:21: struct channel_req rq;
-
drivers/isdn/mISDN/stack.c:421:2-421:21: struct channel_req rq;
-
drivers/isdn/mISDN/stack.c:458:2-458:21: struct channel_req rq, rq2;
-
drivers/isdn/mISDN/stack.c:515:2-515:21: struct channel_req rq;
-
drivers/isdn/mISDN/tei.c:788:2-788:21: struct channel_req rq;
-
drivers/md/dm-rq.c:160:2-160:28: struct request *rq = tio->orig;
-
drivers/md/dm-rq.c:190:2-190:28: struct request *rq = tio->orig;
-
drivers/md/dm-rq.c:364:2-364:28: struct request *rq = tio->orig;
-
drivers/md/dm-rq.c:476:2-476:27: struct request *rq = bd->rq;
-
drivers/mmc/core/mmc_test.c:777:2-777:59: struct mmc_test_req *rq = kmalloc(sizeof(*rq), GFP_KERNEL);
-
drivers/mmc/core/mmc_test.c:2363:2-2363:47: struct mmc_test_req *rq = mmc_test_req_alloc();
-
drivers/mtd/mtd_blkdevs.c:108:2-108:18: struct request *rq;
-
drivers/net/ethernet/cavium/thunder/nicvf_main.c:857:2-857:39: struct rcv_queue *rq = &qs->rq[cq_idx];
-
drivers/net/ethernet/cavium/thunder/nicvf_queues.c:749:2-749:20: struct rcv_queue *rq;
-
drivers/net/ethernet/cavium/thunder/nicvf_queues.c:1814:2-1814:20: struct rcv_queue *rq;
-
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c:1178:2-1178:64: const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq;
-
drivers/net/ethernet/cisco/enic/enic_main.c:1637:2-1637:42: unsigned int rq = (napi - &enic->napi[0]);
-
drivers/net/ethernet/freescale/gianfar.c:1779:2-1779:14: int i, rq = 0;
-
drivers/net/ethernet/huawei/hinic/hinic_debugfs.c:230:2-230:19: struct hinic_rq *rq;
-
drivers/net/ethernet/huawei/hinic/hinic_main.c:255:3-255:61: struct hinic_rq *rq = hinic_hwdev_get_rq(nic_dev->hwdev, i);
-
drivers/net/ethernet/huawei/hinic/hinic_rx.c:247:2-247:29: struct hinic_rq *rq = rxq->rq;
-
drivers/net/ethernet/huawei/hinic/hinic_rx.c:361:2-361:29: struct hinic_rq *rq = rxq->rq;
-
drivers/net/ethernet/huawei/hinic/hinic_rx.c:463:2-463:29: struct hinic_rq *rq = rxq->rq;
-
drivers/net/ethernet/huawei/hinic/hinic_rx.c:498:2-498:29: struct hinic_rq *rq = rxq->rq;
-
drivers/net/ethernet/huawei/hinic/hinic_rx.c:521:2-521:29: struct hinic_rq *rq = rxq->rq;
-
drivers/net/ethernet/huawei/hinic/hinic_rx.c:567:2-567:29: struct hinic_rq *rq = rxq->rq;
-
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c:76:2-76:49: struct otx2_rcv_queue *rq = &pfvf->qset.rq[qidx];
-
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c:1357:2-1357:28: int stack_pages, pool_id, rq;
-
drivers/net/ethernet/mellanox/mlx5/core/en/ptp.c:136:2-136:28: struct mlx5e_rq *rq = &c->rq;
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:69:2-69:19: struct mlx5e_rq *rq;
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:158:2-158:24: struct mlx5e_rq *rq = ctx;
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:184:2-184:19: struct mlx5e_rq *rq;
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:486:3-486:47: struct mlx5e_rq *rq = &priv->channels.c[i]->rq;
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:569:2-569:24: struct mlx5e_rq *rq = ctx;
-
drivers/net/ethernet/mellanox/mlx5/core/en/reporter_rx.c:655:3-655:47: struct mlx5e_rq *rq = &priv->channels.c[i]->rq;
-
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c:13:2-13:35: struct mlx5e_rq *rq = &trap_ctx->rq;
-
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c:66:2-66:28: struct mlx5e_rq *rq = &t->rq;
-
drivers/net/ethernet/mellanox/mlx5/core/en_dim.c:47:2-47:24: struct mlx5e_rq *rq = container_of(dim, struct mlx5e_rq, dim);
-
drivers/net/ethernet/mellanox/mlx5/core/en_main.c:437:2-437:24: struct mlx5e_rq *rq = container_of(recover_work, struct mlx5e_rq, recover_work);
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:845:2-845:28: struct mlx5e_rq *rq = &c->rq;
-
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c:2202:2-2202:24: struct mlx5e_rq *rq = container_of(cq, struct mlx5e_rq, cq);
-
drivers/net/ethernet/mellanox/mlx5/core/en_txrx.c:121:2-121:28: struct mlx5e_rq *rq = &c->rq;
-
drivers/net/ethernet/microsoft/mana/hw_channel.c:183:2-183:21: struct gdma_queue *rq;
-
drivers/net/ethernet/microsoft/mana/hw_channel.c:594:2-594:36: struct gdma_queue *rq = hwc->rxq->gdma_wq;
-
drivers/net/veth.c:316:2-316:23: struct veth_rq *rq = NULL;
-
drivers/net/veth.c:461:2-461:18: struct veth_rq *rq;
-
drivers/net/veth.c:970:2-971:3: struct veth_rq *rq =
-
drivers/net/veth.c:1007:3-1007:35: struct veth_rq *rq = &priv->rq[i];
-
drivers/net/veth.c:1015:3-1015:35: struct veth_rq *rq = &priv->rq[i];
-
drivers/net/veth.c:1041:3-1041:35: struct veth_rq *rq = &priv->rq[i];
-
drivers/net/veth.c:1050:3-1050:35: struct veth_rq *rq = &priv->rq[i];
-
drivers/net/veth.c:1074:3-1074:35: struct veth_rq *rq = &priv->rq[i];
-
drivers/net/veth.c:1097:3-1097:35: struct veth_rq *rq = &priv->rq[i];
-
drivers/net/veth.c:1114:3-1114:35: struct veth_rq *rq = &priv->rq[i];
-
drivers/net/veth.c:1189:3-1189:35: struct veth_rq *rq = &priv->rq[i];
-
drivers/net/veth.c:1197:4-1197:36: struct veth_rq *rq = &priv->rq[i];
-
drivers/net/virtio_net.c:598:2-598:33: struct receive_queue *rq = vi->rq;
-
drivers/net/virtio_net.c:1440:2-1440:48: struct receive_queue *rq = &vi->rq[vq2rxq(rvq)];
-
drivers/net/virtio_net.c:1490:3-1490:39: struct receive_queue *rq = &vi->rq[i];
-
drivers/net/virtio_net.c:1618:2-1619:3: struct receive_queue *rq =
-
drivers/net/virtio_net.c:1955:3-1955:39: struct receive_queue *rq = &vi->rq[i];
-
drivers/net/virtio_net.c:2528:3-2528:39: struct receive_queue *rq = &vi->rq[i];
-
drivers/net/vmxnet3/vmxnet3_drv.c:1769:3-1769:53: struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i];
-
drivers/net/vmxnet3/vmxnet3_drv.c:1986:2-1986:32: struct vmxnet3_rx_queue *rq = container_of(napi,
-
drivers/net/vmxnet3/vmxnet3_drv.c:2050:2-2050:32: struct vmxnet3_rx_queue *rq = data;
-
drivers/net/vmxnet3/vmxnet3_drv.c:2241:4-2241:54: struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i];
-
drivers/net/vmxnet3/vmxnet3_drv.c:2551:3-2551:53: struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i];
-
drivers/net/vmxnet3/vmxnet3_drv.c:2941:3-2941:53: struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i];
-
drivers/net/vmxnet3/vmxnet3_drv.c:2982:3-2982:53: struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i];
-
drivers/net/vmxnet3/vmxnet3_ethtool.c:491:3-491:53: struct vmxnet3_rx_queue *rq = &adapter->rx_queue[i];
-
drivers/nvme/host/core.c:1188:2-1188:18: struct request *rq;
-
drivers/nvme/host/fc.c:1912:2-1912:27: struct request *rq = op->rq;
-
drivers/nvme/host/fc.c:1924:2-1924:27: struct request *rq = op->rq;
-
drivers/nvme/host/fc.c:2794:2-2794:27: struct request *rq = bd->rq;
-
drivers/nvme/host/nvme.h:529:2-529:18: struct request *rq;
-
drivers/nvme/host/nvme.h:653:2-653:40: struct nvme_request *rq = nvme_req(req);
-
drivers/nvme/host/rdma.c:1228:2-1228:45: struct request *rq = blk_mq_rq_from_pdu(req);
-
drivers/nvme/host/rdma.c:1481:2-1481:45: struct request *rq = blk_mq_rq_from_pdu(req);
-
drivers/nvme/host/rdma.c:1758:2-1758:18: struct request *rq;
-
drivers/nvme/host/rdma.c:2057:2-2057:27: struct request *rq = bd->rq;
-
drivers/nvme/host/rdma.c:2142:2-2142:45: struct request *rq = blk_mq_rq_from_pdu(req);
-
drivers/nvme/host/tcp.c:224:2-224:18: struct request *rq;
-
drivers/nvme/host/tcp.c:266:2-266:45: struct request *rq = blk_mq_rq_from_pdu(req);
-
drivers/nvme/host/tcp.c:538:2-538:18: struct request *rq;
-
drivers/nvme/host/tcp.c:563:2-563:18: struct request *rq;
-
drivers/nvme/host/tcp.c:620:2-620:45: struct request *rq = blk_mq_rq_from_pdu(req);
-
drivers/nvme/host/tcp.c:654:2-654:18: struct request *rq;
-
drivers/nvme/host/tcp.c:761:2-762:57: struct request *rq =
-
drivers/nvme/host/tcp.c:848:3-849:21: struct request *rq = nvme_cid_to_rq(nvme_tcp_tagset(queue),
-
drivers/nvme/host/tcp.c:861:3-862:21: struct request *rq = nvme_cid_to_rq(nvme_tcp_tagset(queue),
-
drivers/nvme/host/tcp.c:2448:2-2448:27: struct request *rq = bd->rq;
-
drivers/nvme/target/loop.c:108:3-108:19: struct request *rq;
-
drivers/nvme/target/passthru.c:165:2-165:30: struct request *rq = req->p.rq;
-
drivers/nvme/target/passthru.c:235:2-235:23: struct request *rq = NULL;
-
drivers/platform/chrome/wilco_ec/debugfs.c:178:2-178:20: struct ec_request rq;
-
drivers/platform/chrome/wilco_ec/mailbox.c:198:2-198:27: struct wilco_ec_request *rq;
-
drivers/platform/chrome/wilco_ec/properties.c:62:2-62:29: struct ec_property_request rq;
-
drivers/platform/chrome/wilco_ec/properties.c:84:2-84:29: struct ec_property_request rq;
-
drivers/platform/chrome/wilco_ec/sysfs.c:74:2-74:28: struct boot_on_ac_request rq;
-
drivers/platform/chrome/wilco_ec/sysfs.c:184:2-184:28: struct usb_charge_request rq;
-
drivers/platform/chrome/wilco_ec/sysfs.c:204:2-204:28: struct usb_charge_request rq;
-
drivers/ptp/ptp_chardev.c:21:2-21:27: struct ptp_clock_request rq;
-
drivers/scsi/elx/efct/efct_hw.c:748:3-748:33: struct hw_rq *rq = hw->hw_rq[i];
-
drivers/scsi/elx/efct/efct_hw.c:781:2-781:16: struct hw_rq *rq;
-
drivers/scsi/elx/efct/efct_hw.c:1099:3-1099:33: struct hw_rq *rq = hw->hw_rq[i];
-
drivers/scsi/elx/efct/efct_hw.c:1239:3-1239:33: struct hw_rq *rq = hw->hw_rq[i];
-
drivers/scsi/elx/efct/efct_hw.c:1300:3-1300:38: struct hw_rq *rq = hw->hw_rq[rq_idx];
-
drivers/scsi/elx/efct/efct_hw.c:1330:3-1330:33: struct hw_rq *rq = hw->hw_rq[i];
-
drivers/scsi/elx/efct/efct_hw_queues.c:314:2-314:21: struct hw_rq *rq = NULL;
-
drivers/scsi/elx/efct/efct_hw_queues.c:494:2-494:56: struct hw_rq *rq = hw->hw_rq[hw->hw_rq_lookup[rqindex]];
-
drivers/scsi/elx/efct/efct_hw_queues.c:531:2-531:16: struct hw_rq *rq;
-
drivers/scsi/elx/efct/efct_hw_queues.c:607:2-607:42: struct hw_rq *rq = hw->hw_rq[hw_rq_index];
-
drivers/scsi/elx/libefc_sli/sli4.c:371:2-371:33: struct sli4_rqst_rq_create_v1 *rq;
-
drivers/scsi/esas2r/esas2r.h:1408:2-1408:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_disc.c:160:2-160:34: struct esas2r_request *rq = &a->general_req;
-
drivers/scsi/esas2r/esas2r_disc.c:313:2-313:34: struct esas2r_request *rq = &a->general_req;
-
drivers/scsi/esas2r/esas2r_disc.c:1163:2-1163:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_init.c:770:2-770:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_init.c:1246:2-1246:34: struct esas2r_request *rq = &a->general_req;
-
drivers/scsi/esas2r/esas2r_init.c:1287:2-1287:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_int.c:210:2-210:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_int.c:309:2-309:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_int.c:390:2-390:34: struct esas2r_request *rq = &a->general_req;
-
drivers/scsi/esas2r/esas2r_int.c:448:2-448:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_io.c:190:2-190:35: struct esas2r_request *rq = sgc->first_req;
-
drivers/scsi/esas2r/esas2r_io.c:373:2-373:35: struct esas2r_request *rq = sgc->first_req;
-
drivers/scsi/esas2r/esas2r_io.c:527:2-527:35: struct esas2r_request *rq = sgc->first_req;
-
drivers/scsi/esas2r/esas2r_io.c:770:2-770:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_ioctl.c:111:2-111:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_ioctl.c:208:2-208:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_ioctl.c:1277:2-1277:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_ioctl.c:1830:3-1830:26: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_ioctl.c:1952:3-1952:26: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_main.c:145:2-145:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_main.c:822:2-822:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_main.c:909:2-909:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_main.c:1108:2-1108:25: struct esas2r_request *rq;
-
drivers/scsi/esas2r/esas2r_main.c:1493:2-1493:25: struct esas2r_request *rq;
-
drivers/scsi/fnic/fnic_scsi.c:1717:2-1717:46: struct request *const rq = scsi_cmd_to_rq(sc);
-
drivers/scsi/fnic/fnic_scsi.c:2192:2-2192:40: struct request *rq = scsi_cmd_to_rq(sc);
-
drivers/scsi/fnic/fnic_scsi.c:2225:2-2225:40: struct request *rq = scsi_cmd_to_rq(sc);
-
drivers/scsi/hpsa.c:996:2-996:51: struct reply_queue_buffer *rq = &h->reply_queue[q];
-
drivers/scsi/hpsa.c:5987:2-5987:6: int rq;
-
drivers/scsi/hpsa.h:491:2-491:51: struct reply_queue_buffer *rq = &h->reply_queue[q];
-
drivers/scsi/hpsa.h:593:2-593:51: struct reply_queue_buffer *rq = &h->reply_queue[q];
-
drivers/scsi/mpi3mr/mpi3mr_os.c:4001:2-4001:42: struct request *rq = scsi_cmd_to_rq(scmd);
-
drivers/scsi/mpt3sas/mpt3sas_scsih.c:5131:2-5131:42: struct request *rq = scsi_cmd_to_rq(scmd);
-
drivers/scsi/myrb.c:1267:2-1267:42: struct request *rq = scsi_cmd_to_rq(scmd);
-
drivers/scsi/myrs.c:1587:2-1587:42: struct request *rq = scsi_cmd_to_rq(scmd);
-
drivers/scsi/scsi_bsg.c:16:2-16:18: struct request *rq;
-
drivers/scsi/scsi_error.c:2381:2-2381:18: struct request *rq;
-
drivers/scsi/scsi_ioctl.c:413:2-413:18: struct request *rq;
-
drivers/scsi/scsi_ioctl.c:520:2-520:18: struct request *rq;
-
drivers/scsi/scsi_lib.c:123:2-123:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/scsi_lib.c:1017:2-1017:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/scsi_lib.c:1131:2-1131:18: struct request *rq;
-
drivers/scsi/scsi_lib.c:1155:2-1155:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/scsi_logging.c:31:2-31:62: struct request *rq = scsi_cmd_to_rq((struct scsi_cmnd *)scmd);
-
drivers/scsi/sd.c:746:2-746:42: struct request *rq = scsi_cmd_to_rq(scmd);
-
drivers/scsi/sd.c:837:2-837:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/sd.c:873:2-873:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/sd.c:905:2-905:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/sd.c:935:2-935:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/sd.c:1023:2-1023:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/sd.c:1115:2-1115:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/sd.c:1229:2-1229:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/sd.c:1273:2-1273:43: struct request *rq = scsi_cmd_to_rq(SCpnt);
-
drivers/scsi/sd_zbc.c:327:2-327:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/sd_zbc.c:409:2-409:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/sd_zbc.c:474:2-474:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/sd_zbc.c:529:2-529:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/sd_zbc.c:602:2-602:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/scsi/sg.c:1723:2-1723:18: struct request *rq;
-
drivers/scsi/sr.c:298:2-298:43: struct request *rq = scsi_cmd_to_rq(SCpnt);
-
drivers/scsi/sr.c:360:2-360:43: struct request *rq = scsi_cmd_to_rq(SCpnt);
-
drivers/scsi/sr.c:932:2-932:18: struct request *rq;
-
drivers/scsi/virtio_scsi.c:523:2-523:40: struct request *rq = scsi_cmd_to_rq(sc);
-
drivers/tty/ipwireless/hardware.c:1734:2-1734:29: struct ipw_rx_packet *rp, *rq;
-
drivers/ufs/core/ufshcd.c:319:2-319:42: struct utp_upiu_req *rq = hba->lrb[tag].ucd_req_ptr;
-
drivers/ufs/core/ufshcd.c:393:2-393:41: struct request *rq = scsi_cmd_to_rq(cmd);
-
drivers/ufs/core/ufshpb.c:440:2-440:21: struct ufshpb_req *rq;
-
drivers/usb/misc/uss720.c:80:2-80:36: struct uss720_async_request *rq = container_of(kref, struct uss720_async_request, ref_count);
-
drivers/usb/misc/uss720.c:98:2-98:31: struct uss720_async_request *rq;
-
drivers/usb/misc/uss720.c:128:2-128:31: struct uss720_async_request *rq;
-
drivers/usb/misc/uss720.c:178:2-178:31: struct uss720_async_request *rq;
-
drivers/usb/misc/uss720.c:196:2-196:31: struct uss720_async_request *rq;
-
drivers/usb/misc/uss720.c:232:2-232:31: struct uss720_async_request *rq;
-
fs/erofs/decompressor.c:68:2-68:43: struct z_erofs_decompress_req *rq = ctx->rq;
-
fs/erofs/decompressor.c:124:2-124:43: struct z_erofs_decompress_req *rq = ctx->rq;
-
fs/erofs/decompressor.c:206:2-206:43: struct z_erofs_decompress_req *rq = ctx->rq;
-
include/scsi/scsi_cmnd.h:300:2-300:46: struct request *rq = blk_mq_rq_from_pdu(scmd);
-
kernel/sched/core.c:616:2-616:13: struct rq *rq;
-
kernel/sched/core.c:641:2-641:13: struct rq *rq;
-
kernel/sched/core.c:773:2-773:18: struct rq *rq = container_of(timer, struct rq, hrtick_timer);
-
kernel/sched/core.c:801:2-801:18: struct rq *rq = arg;
-
kernel/sched/core.c:1057:2-1057:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:1123:2-1123:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:1167:2-1167:18: struct rq *rq = info;
-
kernel/sched/core.c:1451:2-1451:13: struct rq *rq;
-
kernel/sched/core.c:1703:2-1703:13: struct rq *rq;
-
kernel/sched/core.c:2394:2-2394:18: struct rq *rq = this_rq();
-
kernel/sched/core.c:2493:2-2493:37: struct rq *lowest_rq = NULL, *rq = this_rq();
-
kernel/sched/core.c:2552:2-2552:18: struct rq *rq = task_rq(p);
-
kernel/sched/core.c:2944:2-2944:13: struct rq *rq;
-
kernel/sched/core.c:2969:2-2969:13: struct rq *rq;
-
kernel/sched/core.c:3281:2-3281:13: struct rq *rq;
-
kernel/sched/core.c:3581:2-3581:13: struct rq *rq;
-
kernel/sched/core.c:3708:2-3708:13: struct rq *rq;
-
kernel/sched/core.c:3727:2-3727:18: struct rq *rq = this_rq();
-
kernel/sched/core.c:3759:2-3759:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:3775:2-3775:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:3785:2-3785:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:3864:2-3864:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:4222:2-4222:18: struct rq *rq = NULL;
-
kernel/sched/core.c:4624:2-4624:13: struct rq *rq;
-
kernel/sched/core.c:4986:2-4986:18: struct rq *rq = this_rq();
-
kernel/sched/core.c:5311:2-5311:13: struct rq *rq;
-
kernel/sched/core.c:5403:2-5403:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:6165:2-6165:18: struct rq *rq = cpu_rq(cpu), *core_rq = NULL;
-
kernel/sched/core.c:6208:2-6208:18: struct rq *rq = cpu_rq(cpu), *core_rq = NULL;
-
kernel/sched/core.c:6262:2-6262:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:6345:2-6345:13: struct rq *rq;
-
kernel/sched/core.c:6828:2-6828:13: struct rq *rq;
-
kernel/sched/core.c:6951:2-6951:13: struct rq *rq;
-
kernel/sched/core.c:7074:2-7074:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:7144:2-7144:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:7301:2-7301:13: struct rq *rq;
-
kernel/sched/core.c:8198:2-8198:13: struct rq *rq;
-
kernel/sched/core.c:8553:2-8553:13: struct rq *rq, *p_rq;
-
kernel/sched/core.c:8708:2-8708:13: struct rq *rq;
-
kernel/sched/core.c:8871:2-8871:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:9008:2-9008:13: struct rq *rq;
-
kernel/sched/core.c:9052:2-9052:18: struct rq *rq = this_rq();
-
kernel/sched/core.c:9145:2-9145:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:9166:2-9166:18: struct rq *rq = this_rq();
-
kernel/sched/core.c:9271:2-9271:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:9316:2-9316:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:9383:2-9383:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:9454:2-9454:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/core.c:9609:3-9609:14: struct rq *rq;
-
kernel/sched/core.c:10129:2-10129:13: struct rq *rq;
-
kernel/sched/core.c:10227:2-10227:13: struct rq *rq;
-
kernel/sched/core.c:10549:3-10549:27: struct rq *rq = cfs_rq->rq;
-
kernel/sched/core_sched.c:58:2-58:13: struct rq *rq;
-
kernel/sched/cpufreq_schedutil.c:159:2-159:18: struct rq *rq = cpu_rq(sg_cpu->cpu);
-
kernel/sched/cputime.c:221:2-221:18: struct rq *rq = this_rq();
-
kernel/sched/deadline.c:66:2-66:18: struct rq *rq = task_rq(p);
-
kernel/sched/deadline.c:173:3-173:19: struct rq *rq = cpu_rq(i);
-
kernel/sched/deadline.c:309:2-309:13: struct rq *rq;
-
kernel/sched/deadline.c:393:2-393:35: struct rq *rq = rq_of_dl_rq(dl_rq);
-
kernel/sched/deadline.c:786:2-786:35: struct rq *rq = rq_of_dl_rq(dl_rq);
-
kernel/sched/deadline.c:829:2-829:35: struct rq *rq = rq_of_dl_rq(dl_rq);
-
kernel/sched/deadline.c:1013:2-1013:35: struct rq *rq = rq_of_dl_rq(dl_rq);
-
kernel/sched/deadline.c:1049:2-1049:18: struct rq *rq = task_rq(p);
-
kernel/sched/deadline.c:1110:2-1110:13: struct rq *rq;
-
kernel/sched/deadline.c:1239:2-1239:48: struct rq *rq = rq_of_dl_rq(dl_rq_of_se(dl_se));
-
kernel/sched/deadline.c:1414:2-1414:13: struct rq *rq;
-
kernel/sched/deadline.c:1464:2-1464:35: struct rq *rq = rq_of_dl_rq(dl_rq);
-
kernel/sched/deadline.c:1477:2-1477:35: struct rq *rq = rq_of_dl_rq(dl_rq);
-
kernel/sched/deadline.c:1819:2-1819:13: struct rq *rq;
-
kernel/sched/deadline.c:1868:2-1868:13: struct rq *rq;
-
kernel/sched/deadline.c:2488:2-2488:13: struct rq *rq;
-
kernel/sched/deadline.c:2550:2-2550:13: struct rq *rq;
-
kernel/sched/debug.c:584:2-584:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/debug.c:717:2-717:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/fair.c:349:2-349:30: struct rq *rq = rq_of(cfs_rq);
-
kernel/sched/fair.c:418:3-418:31: struct rq *rq = rq_of(cfs_rq);
-
kernel/sched/fair.c:1613:3-1613:19: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/fair.c:1645:2-1645:18: struct rq *rq = cpu_rq(env->dst_cpu);
-
kernel/sched/fair.c:3011:3-3011:31: struct rq *rq = rq_of(cfs_rq);
-
kernel/sched/fair.c:3294:2-3294:30: struct rq *rq = rq_of(cfs_rq);
-
kernel/sched/fair.c:3333:3-3333:31: struct rq *rq = rq_of(cfs_rq);
-
kernel/sched/fair.c:4856:2-4856:18: struct rq *rq = data;
-
kernel/sched/fair.c:4874:2-4874:18: struct rq *rq = data;
-
kernel/sched/fair.c:4889:2-4889:30: struct rq *rq = rq_of(cfs_rq);
-
kernel/sched/fair.c:4976:2-4976:30: struct rq *rq = rq_of(cfs_rq);
-
kernel/sched/fair.c:5077:3-5077:31: struct rq *rq = rq_of(cfs_rq);
-
kernel/sched/fair.c:6097:3-6097:19: struct rq *rq = cpu_rq(i);
-
kernel/sched/fair.c:8201:2-8201:30: struct rq *rq = rq_of(cfs_rq);
-
kernel/sched/fair.c:8262:2-8262:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/fair.c:8341:2-8341:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/fair.c:8693:3-8693:19: struct rq *rq = cpu_rq(i);
-
kernel/sched/fair.c:8931:2-8931:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/fair.c:8967:3-8967:19: struct rq *rq = cpu_rq(i);
-
kernel/sched/fair.c:9624:2-9624:30: struct rq *busiest = NULL, *rq;
-
kernel/sched/fair.c:10607:2-10607:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/fair.c:10695:2-10695:13: struct rq *rq;
-
kernel/sched/fair.c:11119:2-11119:18: struct rq *rq = task_rq(a);
-
kernel/sched/fair.c:11203:2-11203:18: struct rq *rq = this_rq();
-
kernel/sched/fair.c:11528:2-11528:13: struct rq *rq;
-
kernel/sched/fair.c:11545:2-11545:13: struct rq *rq;
-
kernel/sched/fair.c:11573:2-11573:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/fair.c:11621:3-11621:19: struct rq *rq = cpu_rq(i);
-
kernel/sched/fair.c:11672:3-11672:19: struct rq *rq = cpu_rq(i);
-
kernel/sched/membarrier.c:234:2-234:18: struct rq *rq = this_rq();
-
kernel/sched/membarrier.c:465:3-465:19: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/psi.c:904:2-904:13: struct rq *rq;
-
kernel/sched/psi.c:934:2-934:13: struct rq *rq;
-
kernel/sched/psi.c:994:2-994:13: struct rq *rq;
-
kernel/sched/rt.c:218:2-218:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/rt.c:578:2-578:35: struct rq *rq = rq_of_rt_rq(rt_rq);
-
kernel/sched/rt.c:923:3-923:36: struct rq *rq = rq_of_rt_rq(rt_rq);
-
kernel/sched/rt.c:1094:2-1094:35: struct rq *rq = rq_of_rt_rq(rt_rq);
-
kernel/sched/rt.c:1111:2-1111:35: struct rq *rq = rq_of_rt_rq(rt_rq);
-
kernel/sched/rt.c:1135:2-1135:35: struct rq *rq = rq_of_rt_rq(rt_rq);
-
kernel/sched/rt.c:1151:2-1151:35: struct rq *rq = rq_of_rt_rq(rt_rq);
-
kernel/sched/rt.c:1505:2-1505:35: struct rq *rq = rq_of_rt_se(rt_se);
-
kernel/sched/rt.c:1517:2-1517:35: struct rq *rq = rq_of_rt_se(rt_se);
-
kernel/sched/rt.c:1603:2-1603:13: struct rq *rq;
-
kernel/sched/rt.c:2314:2-2314:13: struct rq *rq;
-
kernel/sched/sched.h:1634:2-1634:13: struct rq *rq;
-
kernel/sched/stats.c:126:3-126:14: struct rq *rq;
-
kernel/sched/stats.h:172:3-172:14: struct rq *rq;
-
kernel/sched/topology.c:709:2-709:18: struct rq *rq = cpu_rq(cpu);
-
kernel/sched/topology.c:2253:2-2253:18: struct rq *rq = NULL;
-
net/atm/common.c:227:2-227:30: struct sk_buff_head queue, *rq;
-
net/key/af_key.c:1992:2-1992:48: struct sadb_x_ipsecrequest *rq = (void*)(pol+1);
-
net/key/af_key.c:2151:3-2151:31: struct sadb_x_ipsecrequest *rq;
-
net/key/af_key.c:2548:2-2548:30: struct sadb_x_ipsecrequest *rq;
-
net/key/af_key.c:3505:2-3505:30: struct sadb_x_ipsecrequest *rq;
-
net/sunrpc/cache.c:830:2-830:24: struct cache_request *rq;