License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 21:07:57 +07:00
|
|
|
/* SPDX-License-Identifier: GPL-2.0 */
|
2013-03-24 06:11:31 +07:00
|
|
|
#ifndef _LINUX_CLOSURE_H
|
|
|
|
#define _LINUX_CLOSURE_H
|
|
|
|
|
|
|
|
#include <linux/llist.h>
|
|
|
|
#include <linux/sched.h>
|
2017-02-09 00:51:37 +07:00
|
|
|
#include <linux/sched/task_stack.h>
|
2013-03-24 06:11:31 +07:00
|
|
|
#include <linux/workqueue.h>
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Closure is perhaps the most overused and abused term in computer science, but
|
|
|
|
* since I've been unable to come up with anything better you're stuck with it
|
|
|
|
* again.
|
|
|
|
*
|
|
|
|
* What are closures?
|
|
|
|
*
|
|
|
|
* They embed a refcount. The basic idea is they count "things that are in
|
|
|
|
* progress" - in flight bios, some other thread that's doing something else -
|
|
|
|
* anything you might want to wait on.
|
|
|
|
*
|
|
|
|
* The refcount may be manipulated with closure_get() and closure_put().
|
|
|
|
* closure_put() is where many of the interesting things happen, when it causes
|
|
|
|
* the refcount to go to 0.
|
|
|
|
*
|
|
|
|
* Closures can be used to wait on things both synchronously and asynchronously,
|
|
|
|
* and synchronous and asynchronous use can be mixed without restriction. To
|
|
|
|
* wait synchronously, use closure_sync() - you will sleep until your closure's
|
|
|
|
* refcount hits 1.
|
|
|
|
*
|
|
|
|
* To wait asynchronously, use
|
|
|
|
* continue_at(cl, next_function, workqueue);
|
|
|
|
*
|
|
|
|
* passing it, as you might expect, the function to run when nothing is pending
|
|
|
|
* and the workqueue to run that function out of.
|
|
|
|
*
|
2016-07-04 08:23:34 +07:00
|
|
|
* continue_at() also, critically, requires a 'return' immediately following the
|
|
|
|
* location where this macro is referenced, to return to the calling function.
|
2013-03-24 06:11:31 +07:00
|
|
|
* There's good reason for this.
|
|
|
|
*
|
|
|
|
* To use safely closures asynchronously, they must always have a refcount while
|
|
|
|
* they are running owned by the thread that is running them. Otherwise, suppose
|
|
|
|
* you submit some bios and wish to have a function run when they all complete:
|
|
|
|
*
|
2015-07-20 20:29:37 +07:00
|
|
|
* foo_endio(struct bio *bio)
|
2013-03-24 06:11:31 +07:00
|
|
|
* {
|
|
|
|
* closure_put(cl);
|
|
|
|
* }
|
|
|
|
*
|
|
|
|
* closure_init(cl);
|
|
|
|
*
|
|
|
|
* do_stuff();
|
|
|
|
* closure_get(cl);
|
|
|
|
* bio1->bi_endio = foo_endio;
|
|
|
|
* bio_submit(bio1);
|
|
|
|
*
|
|
|
|
* do_more_stuff();
|
|
|
|
* closure_get(cl);
|
|
|
|
* bio2->bi_endio = foo_endio;
|
|
|
|
* bio_submit(bio2);
|
|
|
|
*
|
|
|
|
* continue_at(cl, complete_some_read, system_wq);
|
|
|
|
*
|
|
|
|
* If closure's refcount started at 0, complete_some_read() could run before the
|
|
|
|
* second bio was submitted - which is almost always not what you want! More
|
|
|
|
* importantly, it wouldn't be possible to say whether the original thread or
|
|
|
|
* complete_some_read()'s thread owned the closure - and whatever state it was
|
|
|
|
* associated with!
|
|
|
|
*
|
|
|
|
* So, closure_init() initializes a closure's refcount to 1 - and when a
|
|
|
|
* closure_fn is run, the refcount will be reset to 1 first.
|
|
|
|
*
|
|
|
|
* Then, the rule is - if you got the refcount with closure_get(), release it
|
|
|
|
* with closure_put() (i.e, in a bio->bi_endio function). If you have a refcount
|
|
|
|
* on a closure because you called closure_init() or you were run out of a
|
|
|
|
* closure - _always_ use continue_at(). Doing so consistently will help
|
|
|
|
* eliminate an entire class of particularly pernicious races.
|
|
|
|
*
|
|
|
|
* Lastly, you might have a wait list dedicated to a specific event, and have no
|
|
|
|
* need for specifying the condition - you just want to wait until someone runs
|
|
|
|
* closure_wake_up() on the appropriate wait list. In that case, just use
|
|
|
|
* closure_wait(). It will return either true or false, depending on whether the
|
|
|
|
* closure was already on a wait list or not - a closure can only be on one wait
|
|
|
|
* list at a time.
|
|
|
|
*
|
|
|
|
* Parents:
|
|
|
|
*
|
|
|
|
* closure_init() takes two arguments - it takes the closure to initialize, and
|
|
|
|
* a (possibly null) parent.
|
|
|
|
*
|
|
|
|
* If parent is non null, the new closure will have a refcount for its lifetime;
|
|
|
|
* a closure is considered to be "finished" when its refcount hits 0 and the
|
|
|
|
* function to run is null. Hence
|
|
|
|
*
|
|
|
|
* continue_at(cl, NULL, NULL);
|
|
|
|
*
|
|
|
|
* returns up the (spaghetti) stack of closures, precisely like normal return
|
|
|
|
* returns up the C stack. continue_at() with non null fn is better thought of
|
|
|
|
* as doing a tail call.
|
|
|
|
*
|
|
|
|
* All this implies that a closure should typically be embedded in a particular
|
|
|
|
* struct (which its refcount will normally control the lifetime of), and that
|
|
|
|
* struct can very much be thought of as a stack frame.
|
|
|
|
*/
|
|
|
|
|
|
|
|
struct closure;
|
2018-01-09 03:21:25 +07:00
|
|
|
struct closure_syncer;
|
2013-03-24 06:11:31 +07:00
|
|
|
typedef void (closure_fn) (struct closure *);
|
2018-03-19 07:36:23 +07:00
|
|
|
extern struct dentry *bcache_debug;
|
2013-03-24 06:11:31 +07:00
|
|
|
|
|
|
|
struct closure_waitlist {
|
|
|
|
struct llist_head list;
|
|
|
|
};
|
|
|
|
|
|
|
|
enum closure_state {
|
|
|
|
/*
|
|
|
|
* CLOSURE_WAITING: Set iff the closure is on a waitlist. Must be set by
|
|
|
|
* the thread that owns the closure, and cleared by the thread that's
|
|
|
|
* waking up the closure.
|
|
|
|
*
|
|
|
|
* The rest are for debugging and don't affect behaviour:
|
|
|
|
*
|
|
|
|
* CLOSURE_RUNNING: Set when a closure is running (i.e. by
|
|
|
|
* closure_init() and when closure_put() runs then next function), and
|
|
|
|
* must be cleared before remaining hits 0. Primarily to help guard
|
|
|
|
* against incorrect usage and accidentally transferring references.
|
|
|
|
* continue_at() and closure_return() clear it for you, if you're doing
|
|
|
|
* something unusual you can use closure_set_dead() which also helps
|
|
|
|
* annotate where references are being transferred.
|
|
|
|
*/
|
|
|
|
|
2018-01-10 02:13:23 +07:00
|
|
|
CLOSURE_BITS_START = (1U << 26),
|
|
|
|
CLOSURE_DESTRUCTOR = (1U << 26),
|
|
|
|
CLOSURE_WAITING = (1U << 28),
|
|
|
|
CLOSURE_RUNNING = (1U << 30),
|
2013-03-24 06:11:31 +07:00
|
|
|
};
|
|
|
|
|
|
|
|
#define CLOSURE_GUARD_MASK \
|
2018-01-09 03:21:25 +07:00
|
|
|
((CLOSURE_DESTRUCTOR|CLOSURE_WAITING|CLOSURE_RUNNING) << 1)
|
2013-03-24 06:11:31 +07:00
|
|
|
|
|
|
|
#define CLOSURE_REMAINING_MASK (CLOSURE_BITS_START - 1)
|
|
|
|
#define CLOSURE_REMAINING_INITIALIZER (1|CLOSURE_RUNNING)
|
|
|
|
|
|
|
|
struct closure {
|
|
|
|
union {
|
|
|
|
struct {
|
|
|
|
struct workqueue_struct *wq;
|
2018-01-09 03:21:25 +07:00
|
|
|
struct closure_syncer *s;
|
2013-03-24 06:11:31 +07:00
|
|
|
struct llist_node list;
|
|
|
|
closure_fn *fn;
|
|
|
|
};
|
|
|
|
struct work_struct work;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct closure *parent;
|
|
|
|
|
|
|
|
atomic_t remaining;
|
|
|
|
|
|
|
|
#ifdef CONFIG_BCACHE_CLOSURES_DEBUG
|
|
|
|
#define CLOSURE_MAGIC_DEAD 0xc054dead
|
|
|
|
#define CLOSURE_MAGIC_ALIVE 0xc054a11e
|
|
|
|
|
|
|
|
unsigned magic;
|
|
|
|
struct list_head all;
|
|
|
|
unsigned long ip;
|
|
|
|
unsigned long waiting_on;
|
|
|
|
#endif
|
|
|
|
};
|
|
|
|
|
|
|
|
void closure_sub(struct closure *cl, int v);
|
|
|
|
void closure_put(struct closure *cl);
|
|
|
|
void __closure_wake_up(struct closure_waitlist *list);
|
|
|
|
bool closure_wait(struct closure_waitlist *list, struct closure *cl);
|
2018-01-09 03:21:25 +07:00
|
|
|
void __closure_sync(struct closure *cl);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* closure_sync - sleep until a closure a closure has nothing left to wait on
|
|
|
|
*
|
|
|
|
* Sleeps until the refcount hits 1 - the thread that's running the closure owns
|
|
|
|
* the last refcount.
|
|
|
|
*/
|
|
|
|
static inline void closure_sync(struct closure *cl)
|
|
|
|
{
|
|
|
|
if ((atomic_read(&cl->remaining) & CLOSURE_REMAINING_MASK) != 1)
|
|
|
|
__closure_sync(cl);
|
|
|
|
}
|
2013-03-24 06:11:31 +07:00
|
|
|
|
|
|
|
#ifdef CONFIG_BCACHE_CLOSURES_DEBUG
|
|
|
|
|
2018-03-19 07:36:23 +07:00
|
|
|
int closure_debug_init(void);
|
2013-03-24 06:11:31 +07:00
|
|
|
void closure_debug_create(struct closure *cl);
|
|
|
|
void closure_debug_destroy(struct closure *cl);
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
2018-03-19 07:36:23 +07:00
|
|
|
static inline int closure_debug_init(void) { return 0; }
|
2013-03-24 06:11:31 +07:00
|
|
|
static inline void closure_debug_create(struct closure *cl) {}
|
|
|
|
static inline void closure_debug_destroy(struct closure *cl) {}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
static inline void closure_set_ip(struct closure *cl)
|
|
|
|
{
|
|
|
|
#ifdef CONFIG_BCACHE_CLOSURES_DEBUG
|
|
|
|
cl->ip = _THIS_IP_;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void closure_set_ret_ip(struct closure *cl)
|
|
|
|
{
|
|
|
|
#ifdef CONFIG_BCACHE_CLOSURES_DEBUG
|
|
|
|
cl->ip = _RET_IP_;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2013-12-21 06:55:23 +07:00
|
|
|
static inline void closure_set_waiting(struct closure *cl, unsigned long f)
|
2013-03-24 06:11:31 +07:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_BCACHE_CLOSURES_DEBUG
|
2013-12-21 06:55:23 +07:00
|
|
|
cl->waiting_on = f;
|
2013-03-24 06:11:31 +07:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void closure_set_stopped(struct closure *cl)
|
|
|
|
{
|
|
|
|
atomic_sub(CLOSURE_RUNNING, &cl->remaining);
|
|
|
|
}
|
|
|
|
|
2013-12-21 06:55:23 +07:00
|
|
|
static inline void set_closure_fn(struct closure *cl, closure_fn *fn,
|
|
|
|
struct workqueue_struct *wq)
|
2013-03-24 06:11:31 +07:00
|
|
|
{
|
2013-12-21 06:55:23 +07:00
|
|
|
closure_set_ip(cl);
|
|
|
|
cl->fn = fn;
|
|
|
|
cl->wq = wq;
|
|
|
|
/* between atomic_dec() in closure_put() */
|
2014-03-18 00:06:10 +07:00
|
|
|
smp_mb__before_atomic();
|
2013-03-24 06:11:31 +07:00
|
|
|
}
|
|
|
|
|
2013-12-21 06:55:23 +07:00
|
|
|
static inline void closure_queue(struct closure *cl)
|
2013-03-24 06:11:31 +07:00
|
|
|
{
|
2013-12-21 06:55:23 +07:00
|
|
|
struct workqueue_struct *wq = cl->wq;
|
2017-10-14 06:35:40 +07:00
|
|
|
/**
|
|
|
|
* Changes made to closure, work_struct, or a couple of other structs
|
|
|
|
* may cause work.func not pointing to the right location.
|
|
|
|
*/
|
|
|
|
BUILD_BUG_ON(offsetof(struct closure, fn)
|
|
|
|
!= offsetof(struct work_struct, func));
|
2013-12-21 06:55:23 +07:00
|
|
|
if (wq) {
|
|
|
|
INIT_WORK(&cl->work, cl->work.func);
|
|
|
|
BUG_ON(!queue_work(wq, &cl->work));
|
2013-03-24 06:11:31 +07:00
|
|
|
} else
|
2013-12-21 06:55:23 +07:00
|
|
|
cl->fn(cl);
|
2013-03-24 06:11:31 +07:00
|
|
|
}
|
|
|
|
|
2013-12-21 06:55:23 +07:00
|
|
|
/**
|
|
|
|
* closure_get - increment a closure's refcount
|
2013-03-24 06:11:31 +07:00
|
|
|
*/
|
2013-12-21 06:55:23 +07:00
|
|
|
static inline void closure_get(struct closure *cl)
|
|
|
|
{
|
|
|
|
#ifdef CONFIG_BCACHE_CLOSURES_DEBUG
|
|
|
|
BUG_ON((atomic_inc_return(&cl->remaining) &
|
|
|
|
CLOSURE_REMAINING_MASK) <= 1);
|
|
|
|
#else
|
|
|
|
atomic_inc(&cl->remaining);
|
|
|
|
#endif
|
|
|
|
}
|
2013-03-24 06:11:31 +07:00
|
|
|
|
|
|
|
/**
|
2013-12-21 06:55:23 +07:00
|
|
|
* closure_init - Initialize a closure, setting the refcount to 1
|
2013-03-24 06:11:31 +07:00
|
|
|
* @cl: closure to initialize
|
|
|
|
* @parent: parent of the new closure. cl will take a refcount on it for its
|
|
|
|
* lifetime; may be NULL.
|
|
|
|
*/
|
2013-12-21 06:55:23 +07:00
|
|
|
static inline void closure_init(struct closure *cl, struct closure *parent)
|
2013-03-24 06:11:31 +07:00
|
|
|
{
|
|
|
|
memset(cl, 0, sizeof(struct closure));
|
2013-12-21 06:55:23 +07:00
|
|
|
cl->parent = parent;
|
|
|
|
if (parent)
|
|
|
|
closure_get(parent);
|
2013-03-24 06:11:31 +07:00
|
|
|
|
2013-12-21 06:55:23 +07:00
|
|
|
atomic_set(&cl->remaining, CLOSURE_REMAINING_INITIALIZER);
|
2013-03-24 06:11:31 +07:00
|
|
|
|
2013-12-21 06:55:23 +07:00
|
|
|
closure_debug_create(cl);
|
|
|
|
closure_set_ip(cl);
|
2013-03-24 06:11:31 +07:00
|
|
|
}
|
|
|
|
|
2013-12-21 06:55:23 +07:00
|
|
|
static inline void closure_init_stack(struct closure *cl)
|
2013-03-24 06:11:31 +07:00
|
|
|
{
|
2013-12-21 06:55:23 +07:00
|
|
|
memset(cl, 0, sizeof(struct closure));
|
2018-01-09 03:21:25 +07:00
|
|
|
atomic_set(&cl->remaining, CLOSURE_REMAINING_INITIALIZER);
|
2013-03-24 06:11:31 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2013-12-21 06:55:23 +07:00
|
|
|
* closure_wake_up - wake up all closures on a wait list.
|
2013-03-24 06:11:31 +07:00
|
|
|
*/
|
|
|
|
static inline void closure_wake_up(struct closure_waitlist *list)
|
|
|
|
{
|
|
|
|
smp_mb();
|
|
|
|
__closure_wake_up(list);
|
|
|
|
}
|
|
|
|
|
2013-12-21 06:55:23 +07:00
|
|
|
/**
|
|
|
|
* continue_at - jump to another function with barrier
|
|
|
|
*
|
|
|
|
* After @cl is no longer waiting on anything (i.e. all outstanding refs have
|
|
|
|
* been dropped with closure_put()), it will resume execution at @fn running out
|
|
|
|
* of @wq (or, if @wq is NULL, @fn will be called by closure_put() directly).
|
|
|
|
*
|
|
|
|
* This is because after calling continue_at() you no longer have a ref on @cl,
|
|
|
|
* and whatever @cl owns may be freed out from under you - a running closure fn
|
|
|
|
* has a ref on its own closure which continue_at() drops.
|
2018-01-09 03:21:25 +07:00
|
|
|
*
|
|
|
|
* Note you are expected to immediately return after using this macro.
|
2013-03-24 06:11:31 +07:00
|
|
|
*/
|
|
|
|
#define continue_at(_cl, _fn, _wq) \
|
|
|
|
do { \
|
|
|
|
set_closure_fn(_cl, _fn, _wq); \
|
|
|
|
closure_sub(_cl, CLOSURE_RUNNING + 1); \
|
|
|
|
} while (0)
|
|
|
|
|
2013-12-21 06:55:23 +07:00
|
|
|
/**
|
|
|
|
* closure_return - finish execution of a closure
|
|
|
|
*
|
|
|
|
* This is used to indicate that @cl is finished: when all outstanding refs on
|
|
|
|
* @cl have been dropped @cl's ref on its parent closure (as passed to
|
|
|
|
* closure_init()) will be dropped, if one was specified - thus this can be
|
|
|
|
* thought of as returning to the parent closure.
|
|
|
|
*/
|
2013-03-24 06:11:31 +07:00
|
|
|
#define closure_return(_cl) continue_at((_cl), NULL, NULL)
|
|
|
|
|
2013-12-21 06:55:23 +07:00
|
|
|
/**
|
|
|
|
* continue_at_nobarrier - jump to another function without barrier
|
|
|
|
*
|
|
|
|
* Causes @fn to be executed out of @cl, in @wq context (or called directly if
|
|
|
|
* @wq is NULL).
|
|
|
|
*
|
|
|
|
* The ref the caller of continue_at_nobarrier() had on @cl is now owned by @fn,
|
|
|
|
* thus it's not safe to touch anything protected by @cl after a
|
|
|
|
* continue_at_nobarrier().
|
|
|
|
*/
|
2013-03-24 06:11:31 +07:00
|
|
|
#define continue_at_nobarrier(_cl, _fn, _wq) \
|
|
|
|
do { \
|
|
|
|
set_closure_fn(_cl, _fn, _wq); \
|
2013-10-25 07:07:04 +07:00
|
|
|
closure_queue(_cl); \
|
2013-03-24 06:11:31 +07:00
|
|
|
} while (0)
|
|
|
|
|
2013-12-21 06:55:23 +07:00
|
|
|
/**
|
|
|
|
* closure_return - finish execution of a closure, with destructor
|
|
|
|
*
|
|
|
|
* Works like closure_return(), except @destructor will be called when all
|
|
|
|
* outstanding refs on @cl have been dropped; @destructor may be used to safely
|
|
|
|
* free the memory occupied by @cl, and it is called with the ref on the parent
|
|
|
|
* closure still held - so @destructor could safely return an item to a
|
|
|
|
* freelist protected by @cl's parent.
|
|
|
|
*/
|
2013-03-24 06:11:31 +07:00
|
|
|
#define closure_return_with_destructor(_cl, _destructor) \
|
|
|
|
do { \
|
|
|
|
set_closure_fn(_cl, _destructor, NULL); \
|
|
|
|
closure_sub(_cl, CLOSURE_RUNNING - CLOSURE_DESTRUCTOR + 1); \
|
|
|
|
} while (0)
|
|
|
|
|
2013-12-21 06:55:23 +07:00
|
|
|
/**
|
|
|
|
* closure_call - execute @fn out of a new, uninitialized closure
|
|
|
|
*
|
|
|
|
* Typically used when running out of one closure, and we want to run @fn
|
|
|
|
* asynchronously out of a new closure - @parent will then wait for @cl to
|
|
|
|
* finish.
|
|
|
|
*/
|
2013-03-24 06:11:31 +07:00
|
|
|
static inline void closure_call(struct closure *cl, closure_fn fn,
|
|
|
|
struct workqueue_struct *wq,
|
|
|
|
struct closure *parent)
|
|
|
|
{
|
|
|
|
closure_init(cl, parent);
|
|
|
|
continue_at_nobarrier(cl, fn, wq);
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* _LINUX_CLOSURE_H */
|