Lines Matching defs:cur_ops

158 	struct lock_torture_ops *cur_ops;
290 cxt.cur_ops->flags = flags;
297 spin_unlock_irqrestore(&torture_spinlock, cxt.cur_ops->flags);
343 cxt.cur_ops->flags = flags;
350 raw_spin_unlock_irqrestore(&torture_raw_spinlock, cxt.cur_ops->flags);
435 cxt.cur_ops->flags = flags;
442 write_unlock_irqrestore(&torture_rwlock, cxt.cur_ops->flags);
451 cxt.cur_ops->flags = flags;
458 read_unlock_irqrestore(&torture_rwlock, cxt.cur_ops->flags);
872 cxt.cur_ops->task_boost(&rand);
873 if (cxt.cur_ops->nested_lock)
874 cxt.cur_ops->nested_lock(tid, lockset_mask);
879 cxt.cur_ops->writelock(tid);
893 cxt.cur_ops->write_delay(&rand);
897 cxt.cur_ops->writeunlock(tid);
899 if (cxt.cur_ops->nested_unlock)
900 cxt.cur_ops->nested_unlock(tid, lockset_mask);
905 cxt.cur_ops->task_boost(NULL); /* reset prio */
927 cxt.cur_ops->readlock(tid);
933 cxt.cur_ops->read_delay(&rand);
935 cxt.cur_ops->readunlock(tid);
989 if (cxt.cur_ops->readlock)
1003 if (cxt.cur_ops->readlock) {
1038 lock_torture_print_module_parms(struct lock_torture_ops *cur_ops,
1111 * However cxt->cur_ops.init() may have been invoked, so beside
1112 * perform the underlying torture-specific cleanups, cur_ops.exit()
1137 lock_torture_print_module_parms(cxt.cur_ops,
1140 lock_torture_print_module_parms(cxt.cur_ops,
1143 lock_torture_print_module_parms(cxt.cur_ops,
1155 if (cxt.cur_ops->exit)
1156 cxt.cur_ops->exit();
1185 cxt.cur_ops = torture_ops[i];
1186 if (strcmp(torture_type, cxt.cur_ops->name) == 0)
1201 (!cxt.cur_ops->readlock || nreaders_stress == 0)) {
1212 if (cxt.cur_ops->init) {
1213 cxt.cur_ops->init();
1249 if (cxt.cur_ops->readlock) {
1286 lock_torture_print_module_parms(cxt.cur_ops, "Start of test");
1327 if (cxt.cur_ops->readlock) {
1363 if (cxt.cur_ops->readlock == NULL || (j >= cxt.nrealreaders_stress))