mirror of https://gitee.com/openkylin/linux.git
Merge branch 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tj/wq
* 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tj/wq: workqueue: remove in_workqueue_context() workqueue: Clarify that schedule_on_each_cpu is synchronous memory_hotplug: drop spurious calls to flush_scheduled_work() shpchp: update workqueue usage pciehp: update workqueue usage isdn/eicon: don't call flush_scheduled_work() from diva_os_remove_soft_isr() workqueue: add and use WQ_MEM_RECLAIM flag workqueue: fix HIGHPRI handling in keep_working() workqueue: add queue_work and activate_work trace points workqueue: prepare for more tracepoints workqueue: implement flush[_delayed]_work_sync() workqueue: factor out start_flush_work() workqueue: cleanup flush/cancel functions workqueue: implement alloc_ordered_workqueue() Fix up trivial conflict in fs/gfs2/main.c as per Tejun
This commit is contained in:
commit
91b745016c
|
@ -196,11 +196,11 @@ resources, scheduled and executed.
|
||||||
suspend operations. Work items on the wq are drained and no
|
suspend operations. Work items on the wq are drained and no
|
||||||
new work item starts execution until thawed.
|
new work item starts execution until thawed.
|
||||||
|
|
||||||
WQ_RESCUER
|
WQ_MEM_RECLAIM
|
||||||
|
|
||||||
All wq which might be used in the memory reclaim paths _MUST_
|
All wq which might be used in the memory reclaim paths _MUST_
|
||||||
have this flag set. This reserves one worker exclusively for
|
have this flag set. The wq is guaranteed to have at least one
|
||||||
the execution of this wq under memory pressure.
|
execution context regardless of memory pressure.
|
||||||
|
|
||||||
WQ_HIGHPRI
|
WQ_HIGHPRI
|
||||||
|
|
||||||
|
@ -356,11 +356,11 @@ If q1 has WQ_CPU_INTENSIVE set,
|
||||||
|
|
||||||
6. Guidelines
|
6. Guidelines
|
||||||
|
|
||||||
* Do not forget to use WQ_RESCUER if a wq may process work items which
|
* Do not forget to use WQ_MEM_RECLAIM if a wq may process work items
|
||||||
are used during memory reclaim. Each wq with WQ_RESCUER set has one
|
which are used during memory reclaim. Each wq with WQ_MEM_RECLAIM
|
||||||
rescuer thread reserved for it. If there is dependency among
|
set has an execution context reserved for it. If there is
|
||||||
multiple work items used during memory reclaim, they should be
|
dependency among multiple work items used during memory reclaim,
|
||||||
queued to separate wq each with WQ_RESCUER.
|
they should be queued to separate wq each with WQ_MEM_RECLAIM.
|
||||||
|
|
||||||
* Unless strict ordering is required, there is no need to use ST wq.
|
* Unless strict ordering is required, there is no need to use ST wq.
|
||||||
|
|
||||||
|
@ -368,12 +368,13 @@ If q1 has WQ_CPU_INTENSIVE set,
|
||||||
recommended. In most use cases, concurrency level usually stays
|
recommended. In most use cases, concurrency level usually stays
|
||||||
well under the default limit.
|
well under the default limit.
|
||||||
|
|
||||||
* A wq serves as a domain for forward progress guarantee (WQ_RESCUER),
|
* A wq serves as a domain for forward progress guarantee
|
||||||
flush and work item attributes. Work items which are not involved
|
(WQ_MEM_RECLAIM, flush and work item attributes. Work items which
|
||||||
in memory reclaim and don't need to be flushed as a part of a group
|
are not involved in memory reclaim and don't need to be flushed as a
|
||||||
of work items, and don't require any special attribute, can use one
|
part of a group of work items, and don't require any special
|
||||||
of the system wq. There is no difference in execution
|
attribute, can use one of the system wq. There is no difference in
|
||||||
characteristics between using a dedicated wq and a system wq.
|
execution characteristics between using a dedicated wq and a system
|
||||||
|
wq.
|
||||||
|
|
||||||
* Unless work items are expected to consume a huge amount of CPU
|
* Unless work items are expected to consume a huge amount of CPU
|
||||||
cycles, using a bound wq is usually beneficial due to the increased
|
cycles, using a bound wq is usually beneficial due to the increased
|
||||||
|
|
|
@ -3335,7 +3335,7 @@ void ata_sff_port_init(struct ata_port *ap)
|
||||||
|
|
||||||
int __init ata_sff_init(void)
|
int __init ata_sff_init(void)
|
||||||
{
|
{
|
||||||
ata_sff_wq = alloc_workqueue("ata_sff", WQ_RESCUER, WQ_MAX_ACTIVE);
|
ata_sff_wq = alloc_workqueue("ata_sff", WQ_MEM_RECLAIM, WQ_MAX_ACTIVE);
|
||||||
if (!ata_sff_wq)
|
if (!ata_sff_wq)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
|
|
@ -15,7 +15,6 @@
|
||||||
#include <asm/uaccess.h>
|
#include <asm/uaccess.h>
|
||||||
#include <asm/io.h>
|
#include <asm/io.h>
|
||||||
#include <linux/ioport.h>
|
#include <linux/ioport.h>
|
||||||
#include <linux/workqueue.h>
|
|
||||||
#include <linux/pci.h>
|
#include <linux/pci.h>
|
||||||
#include <linux/interrupt.h>
|
#include <linux/interrupt.h>
|
||||||
#include <linux/list.h>
|
#include <linux/list.h>
|
||||||
|
@ -546,7 +545,6 @@ void diva_os_remove_soft_isr(diva_os_soft_isr_t * psoft_isr)
|
||||||
void *mem;
|
void *mem;
|
||||||
|
|
||||||
tasklet_kill(&pdpc->divas_task);
|
tasklet_kill(&pdpc->divas_task);
|
||||||
flush_scheduled_work();
|
|
||||||
mem = psoft_isr->object;
|
mem = psoft_isr->object;
|
||||||
psoft_isr->object = NULL;
|
psoft_isr->object = NULL;
|
||||||
diva_os_free(0, mem);
|
diva_os_free(0, mem);
|
||||||
|
|
|
@ -36,6 +36,7 @@
|
||||||
#include <linux/sched.h> /* signal_pending() */
|
#include <linux/sched.h> /* signal_pending() */
|
||||||
#include <linux/pcieport_if.h>
|
#include <linux/pcieport_if.h>
|
||||||
#include <linux/mutex.h>
|
#include <linux/mutex.h>
|
||||||
|
#include <linux/workqueue.h>
|
||||||
|
|
||||||
#define MY_NAME "pciehp"
|
#define MY_NAME "pciehp"
|
||||||
|
|
||||||
|
@ -44,6 +45,7 @@ extern int pciehp_poll_time;
|
||||||
extern int pciehp_debug;
|
extern int pciehp_debug;
|
||||||
extern int pciehp_force;
|
extern int pciehp_force;
|
||||||
extern struct workqueue_struct *pciehp_wq;
|
extern struct workqueue_struct *pciehp_wq;
|
||||||
|
extern struct workqueue_struct *pciehp_ordered_wq;
|
||||||
|
|
||||||
#define dbg(format, arg...) \
|
#define dbg(format, arg...) \
|
||||||
do { \
|
do { \
|
||||||
|
|
|
@ -43,6 +43,7 @@ int pciehp_poll_mode;
|
||||||
int pciehp_poll_time;
|
int pciehp_poll_time;
|
||||||
int pciehp_force;
|
int pciehp_force;
|
||||||
struct workqueue_struct *pciehp_wq;
|
struct workqueue_struct *pciehp_wq;
|
||||||
|
struct workqueue_struct *pciehp_ordered_wq;
|
||||||
|
|
||||||
#define DRIVER_VERSION "0.4"
|
#define DRIVER_VERSION "0.4"
|
||||||
#define DRIVER_AUTHOR "Dan Zink <dan.zink@compaq.com>, Greg Kroah-Hartman <greg@kroah.com>, Dely Sy <dely.l.sy@intel.com>"
|
#define DRIVER_AUTHOR "Dan Zink <dan.zink@compaq.com>, Greg Kroah-Hartman <greg@kroah.com>, Dely Sy <dely.l.sy@intel.com>"
|
||||||
|
@ -340,18 +341,33 @@ static int __init pcied_init(void)
|
||||||
{
|
{
|
||||||
int retval = 0;
|
int retval = 0;
|
||||||
|
|
||||||
|
pciehp_wq = alloc_workqueue("pciehp", 0, 0);
|
||||||
|
if (!pciehp_wq)
|
||||||
|
return -ENOMEM;
|
||||||
|
|
||||||
|
pciehp_ordered_wq = alloc_ordered_workqueue("pciehp_ordered", 0);
|
||||||
|
if (!pciehp_ordered_wq) {
|
||||||
|
destroy_workqueue(pciehp_wq);
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
|
||||||
pciehp_firmware_init();
|
pciehp_firmware_init();
|
||||||
retval = pcie_port_service_register(&hpdriver_portdrv);
|
retval = pcie_port_service_register(&hpdriver_portdrv);
|
||||||
dbg("pcie_port_service_register = %d\n", retval);
|
dbg("pcie_port_service_register = %d\n", retval);
|
||||||
info(DRIVER_DESC " version: " DRIVER_VERSION "\n");
|
info(DRIVER_DESC " version: " DRIVER_VERSION "\n");
|
||||||
if (retval)
|
if (retval) {
|
||||||
|
destroy_workqueue(pciehp_ordered_wq);
|
||||||
|
destroy_workqueue(pciehp_wq);
|
||||||
dbg("Failure to register service\n");
|
dbg("Failure to register service\n");
|
||||||
|
}
|
||||||
return retval;
|
return retval;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __exit pcied_cleanup(void)
|
static void __exit pcied_cleanup(void)
|
||||||
{
|
{
|
||||||
dbg("unload_pciehpd()\n");
|
dbg("unload_pciehpd()\n");
|
||||||
|
destroy_workqueue(pciehp_ordered_wq);
|
||||||
|
destroy_workqueue(pciehp_wq);
|
||||||
pcie_port_service_unregister(&hpdriver_portdrv);
|
pcie_port_service_unregister(&hpdriver_portdrv);
|
||||||
info(DRIVER_DESC " version: " DRIVER_VERSION " unloaded\n");
|
info(DRIVER_DESC " version: " DRIVER_VERSION " unloaded\n");
|
||||||
}
|
}
|
||||||
|
|
|
@ -32,7 +32,6 @@
|
||||||
#include <linux/types.h>
|
#include <linux/types.h>
|
||||||
#include <linux/slab.h>
|
#include <linux/slab.h>
|
||||||
#include <linux/pci.h>
|
#include <linux/pci.h>
|
||||||
#include <linux/workqueue.h>
|
|
||||||
#include "../pci.h"
|
#include "../pci.h"
|
||||||
#include "pciehp.h"
|
#include "pciehp.h"
|
||||||
|
|
||||||
|
@ -50,7 +49,7 @@ static int queue_interrupt_event(struct slot *p_slot, u32 event_type)
|
||||||
info->p_slot = p_slot;
|
info->p_slot = p_slot;
|
||||||
INIT_WORK(&info->work, interrupt_event_handler);
|
INIT_WORK(&info->work, interrupt_event_handler);
|
||||||
|
|
||||||
schedule_work(&info->work);
|
queue_work(pciehp_wq, &info->work);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
@ -345,7 +344,7 @@ void pciehp_queue_pushbutton_work(struct work_struct *work)
|
||||||
kfree(info);
|
kfree(info);
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
queue_work(pciehp_wq, &info->work);
|
queue_work(pciehp_ordered_wq, &info->work);
|
||||||
out:
|
out:
|
||||||
mutex_unlock(&p_slot->lock);
|
mutex_unlock(&p_slot->lock);
|
||||||
}
|
}
|
||||||
|
@ -378,7 +377,7 @@ static void handle_button_press_event(struct slot *p_slot)
|
||||||
if (ATTN_LED(ctrl))
|
if (ATTN_LED(ctrl))
|
||||||
pciehp_set_attention_status(p_slot, 0);
|
pciehp_set_attention_status(p_slot, 0);
|
||||||
|
|
||||||
schedule_delayed_work(&p_slot->work, 5*HZ);
|
queue_delayed_work(pciehp_wq, &p_slot->work, 5*HZ);
|
||||||
break;
|
break;
|
||||||
case BLINKINGOFF_STATE:
|
case BLINKINGOFF_STATE:
|
||||||
case BLINKINGON_STATE:
|
case BLINKINGON_STATE:
|
||||||
|
@ -440,7 +439,7 @@ static void handle_surprise_event(struct slot *p_slot)
|
||||||
else
|
else
|
||||||
p_slot->state = POWERON_STATE;
|
p_slot->state = POWERON_STATE;
|
||||||
|
|
||||||
queue_work(pciehp_wq, &info->work);
|
queue_work(pciehp_ordered_wq, &info->work);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void interrupt_event_handler(struct work_struct *work)
|
static void interrupt_event_handler(struct work_struct *work)
|
||||||
|
|
|
@ -41,8 +41,6 @@
|
||||||
#include "../pci.h"
|
#include "../pci.h"
|
||||||
#include "pciehp.h"
|
#include "pciehp.h"
|
||||||
|
|
||||||
static atomic_t pciehp_num_controllers = ATOMIC_INIT(0);
|
|
||||||
|
|
||||||
static inline int pciehp_readw(struct controller *ctrl, int reg, u16 *value)
|
static inline int pciehp_readw(struct controller *ctrl, int reg, u16 *value)
|
||||||
{
|
{
|
||||||
struct pci_dev *dev = ctrl->pcie->port;
|
struct pci_dev *dev = ctrl->pcie->port;
|
||||||
|
@ -805,8 +803,8 @@ static void pcie_cleanup_slot(struct controller *ctrl)
|
||||||
{
|
{
|
||||||
struct slot *slot = ctrl->slot;
|
struct slot *slot = ctrl->slot;
|
||||||
cancel_delayed_work(&slot->work);
|
cancel_delayed_work(&slot->work);
|
||||||
flush_scheduled_work();
|
|
||||||
flush_workqueue(pciehp_wq);
|
flush_workqueue(pciehp_wq);
|
||||||
|
flush_workqueue(pciehp_ordered_wq);
|
||||||
kfree(slot);
|
kfree(slot);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -912,16 +910,6 @@ struct controller *pcie_init(struct pcie_device *dev)
|
||||||
/* Disable sotfware notification */
|
/* Disable sotfware notification */
|
||||||
pcie_disable_notification(ctrl);
|
pcie_disable_notification(ctrl);
|
||||||
|
|
||||||
/*
|
|
||||||
* If this is the first controller to be initialized,
|
|
||||||
* initialize the pciehp work queue
|
|
||||||
*/
|
|
||||||
if (atomic_add_return(1, &pciehp_num_controllers) == 1) {
|
|
||||||
pciehp_wq = create_singlethread_workqueue("pciehpd");
|
|
||||||
if (!pciehp_wq)
|
|
||||||
goto abort_ctrl;
|
|
||||||
}
|
|
||||||
|
|
||||||
ctrl_info(ctrl, "HPC vendor_id %x device_id %x ss_vid %x ss_did %x\n",
|
ctrl_info(ctrl, "HPC vendor_id %x device_id %x ss_vid %x ss_did %x\n",
|
||||||
pdev->vendor, pdev->device, pdev->subsystem_vendor,
|
pdev->vendor, pdev->device, pdev->subsystem_vendor,
|
||||||
pdev->subsystem_device);
|
pdev->subsystem_device);
|
||||||
|
@ -941,11 +929,5 @@ void pciehp_release_ctrl(struct controller *ctrl)
|
||||||
{
|
{
|
||||||
pcie_shutdown_notification(ctrl);
|
pcie_shutdown_notification(ctrl);
|
||||||
pcie_cleanup_slot(ctrl);
|
pcie_cleanup_slot(ctrl);
|
||||||
/*
|
|
||||||
* If this is the last controller to be released, destroy the
|
|
||||||
* pciehp work queue
|
|
||||||
*/
|
|
||||||
if (atomic_dec_and_test(&pciehp_num_controllers))
|
|
||||||
destroy_workqueue(pciehp_wq);
|
|
||||||
kfree(ctrl);
|
kfree(ctrl);
|
||||||
}
|
}
|
||||||
|
|
|
@ -35,6 +35,7 @@
|
||||||
#include <linux/delay.h>
|
#include <linux/delay.h>
|
||||||
#include <linux/sched.h> /* signal_pending(), struct timer_list */
|
#include <linux/sched.h> /* signal_pending(), struct timer_list */
|
||||||
#include <linux/mutex.h>
|
#include <linux/mutex.h>
|
||||||
|
#include <linux/workqueue.h>
|
||||||
|
|
||||||
#if !defined(MODULE)
|
#if !defined(MODULE)
|
||||||
#define MY_NAME "shpchp"
|
#define MY_NAME "shpchp"
|
||||||
|
@ -46,6 +47,7 @@ extern int shpchp_poll_mode;
|
||||||
extern int shpchp_poll_time;
|
extern int shpchp_poll_time;
|
||||||
extern int shpchp_debug;
|
extern int shpchp_debug;
|
||||||
extern struct workqueue_struct *shpchp_wq;
|
extern struct workqueue_struct *shpchp_wq;
|
||||||
|
extern struct workqueue_struct *shpchp_ordered_wq;
|
||||||
|
|
||||||
#define dbg(format, arg...) \
|
#define dbg(format, arg...) \
|
||||||
do { \
|
do { \
|
||||||
|
|
|
@ -33,7 +33,6 @@
|
||||||
#include <linux/types.h>
|
#include <linux/types.h>
|
||||||
#include <linux/slab.h>
|
#include <linux/slab.h>
|
||||||
#include <linux/pci.h>
|
#include <linux/pci.h>
|
||||||
#include <linux/workqueue.h>
|
|
||||||
#include "shpchp.h"
|
#include "shpchp.h"
|
||||||
|
|
||||||
/* Global variables */
|
/* Global variables */
|
||||||
|
@ -41,6 +40,7 @@ int shpchp_debug;
|
||||||
int shpchp_poll_mode;
|
int shpchp_poll_mode;
|
||||||
int shpchp_poll_time;
|
int shpchp_poll_time;
|
||||||
struct workqueue_struct *shpchp_wq;
|
struct workqueue_struct *shpchp_wq;
|
||||||
|
struct workqueue_struct *shpchp_ordered_wq;
|
||||||
|
|
||||||
#define DRIVER_VERSION "0.4"
|
#define DRIVER_VERSION "0.4"
|
||||||
#define DRIVER_AUTHOR "Dan Zink <dan.zink@compaq.com>, Greg Kroah-Hartman <greg@kroah.com>, Dely Sy <dely.l.sy@intel.com>"
|
#define DRIVER_AUTHOR "Dan Zink <dan.zink@compaq.com>, Greg Kroah-Hartman <greg@kroah.com>, Dely Sy <dely.l.sy@intel.com>"
|
||||||
|
@ -174,8 +174,8 @@ void cleanup_slots(struct controller *ctrl)
|
||||||
slot = list_entry(tmp, struct slot, slot_list);
|
slot = list_entry(tmp, struct slot, slot_list);
|
||||||
list_del(&slot->slot_list);
|
list_del(&slot->slot_list);
|
||||||
cancel_delayed_work(&slot->work);
|
cancel_delayed_work(&slot->work);
|
||||||
flush_scheduled_work();
|
|
||||||
flush_workqueue(shpchp_wq);
|
flush_workqueue(shpchp_wq);
|
||||||
|
flush_workqueue(shpchp_ordered_wq);
|
||||||
pci_hp_deregister(slot->hotplug_slot);
|
pci_hp_deregister(slot->hotplug_slot);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -360,9 +360,23 @@ static int __init shpcd_init(void)
|
||||||
{
|
{
|
||||||
int retval = 0;
|
int retval = 0;
|
||||||
|
|
||||||
|
shpchp_wq = alloc_ordered_workqueue("shpchp", 0);
|
||||||
|
if (!shpchp_wq)
|
||||||
|
return -ENOMEM;
|
||||||
|
|
||||||
|
shpchp_ordered_wq = alloc_ordered_workqueue("shpchp_ordered", 0);
|
||||||
|
if (!shpchp_ordered_wq) {
|
||||||
|
destroy_workqueue(shpchp_wq);
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
|
||||||
retval = pci_register_driver(&shpc_driver);
|
retval = pci_register_driver(&shpc_driver);
|
||||||
dbg("%s: pci_register_driver = %d\n", __func__, retval);
|
dbg("%s: pci_register_driver = %d\n", __func__, retval);
|
||||||
info(DRIVER_DESC " version: " DRIVER_VERSION "\n");
|
info(DRIVER_DESC " version: " DRIVER_VERSION "\n");
|
||||||
|
if (retval) {
|
||||||
|
destroy_workqueue(shpchp_ordered_wq);
|
||||||
|
destroy_workqueue(shpchp_wq);
|
||||||
|
}
|
||||||
return retval;
|
return retval;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -370,6 +384,8 @@ static void __exit shpcd_cleanup(void)
|
||||||
{
|
{
|
||||||
dbg("unload_shpchpd()\n");
|
dbg("unload_shpchpd()\n");
|
||||||
pci_unregister_driver(&shpc_driver);
|
pci_unregister_driver(&shpc_driver);
|
||||||
|
destroy_workqueue(shpchp_ordered_wq);
|
||||||
|
destroy_workqueue(shpchp_wq);
|
||||||
info(DRIVER_DESC " version: " DRIVER_VERSION " unloaded\n");
|
info(DRIVER_DESC " version: " DRIVER_VERSION " unloaded\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -32,7 +32,6 @@
|
||||||
#include <linux/types.h>
|
#include <linux/types.h>
|
||||||
#include <linux/slab.h>
|
#include <linux/slab.h>
|
||||||
#include <linux/pci.h>
|
#include <linux/pci.h>
|
||||||
#include <linux/workqueue.h>
|
|
||||||
#include "../pci.h"
|
#include "../pci.h"
|
||||||
#include "shpchp.h"
|
#include "shpchp.h"
|
||||||
|
|
||||||
|
@ -52,7 +51,7 @@ static int queue_interrupt_event(struct slot *p_slot, u32 event_type)
|
||||||
info->p_slot = p_slot;
|
info->p_slot = p_slot;
|
||||||
INIT_WORK(&info->work, interrupt_event_handler);
|
INIT_WORK(&info->work, interrupt_event_handler);
|
||||||
|
|
||||||
schedule_work(&info->work);
|
queue_work(shpchp_wq, &info->work);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
@ -457,7 +456,7 @@ void shpchp_queue_pushbutton_work(struct work_struct *work)
|
||||||
kfree(info);
|
kfree(info);
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
queue_work(shpchp_wq, &info->work);
|
queue_work(shpchp_ordered_wq, &info->work);
|
||||||
out:
|
out:
|
||||||
mutex_unlock(&p_slot->lock);
|
mutex_unlock(&p_slot->lock);
|
||||||
}
|
}
|
||||||
|
@ -505,7 +504,7 @@ static void handle_button_press_event(struct slot *p_slot)
|
||||||
p_slot->hpc_ops->green_led_blink(p_slot);
|
p_slot->hpc_ops->green_led_blink(p_slot);
|
||||||
p_slot->hpc_ops->set_attention_status(p_slot, 0);
|
p_slot->hpc_ops->set_attention_status(p_slot, 0);
|
||||||
|
|
||||||
schedule_delayed_work(&p_slot->work, 5*HZ);
|
queue_delayed_work(shpchp_wq, &p_slot->work, 5*HZ);
|
||||||
break;
|
break;
|
||||||
case BLINKINGOFF_STATE:
|
case BLINKINGOFF_STATE:
|
||||||
case BLINKINGON_STATE:
|
case BLINKINGON_STATE:
|
||||||
|
|
|
@ -179,8 +179,6 @@
|
||||||
#define SLOT_EVENT_LATCH 0x2
|
#define SLOT_EVENT_LATCH 0x2
|
||||||
#define SLOT_SERR_INT_MASK 0x3
|
#define SLOT_SERR_INT_MASK 0x3
|
||||||
|
|
||||||
static atomic_t shpchp_num_controllers = ATOMIC_INIT(0);
|
|
||||||
|
|
||||||
static irqreturn_t shpc_isr(int irq, void *dev_id);
|
static irqreturn_t shpc_isr(int irq, void *dev_id);
|
||||||
static void start_int_poll_timer(struct controller *ctrl, int sec);
|
static void start_int_poll_timer(struct controller *ctrl, int sec);
|
||||||
static int hpc_check_cmd_status(struct controller *ctrl);
|
static int hpc_check_cmd_status(struct controller *ctrl);
|
||||||
|
@ -614,13 +612,6 @@ static void hpc_release_ctlr(struct controller *ctrl)
|
||||||
|
|
||||||
iounmap(ctrl->creg);
|
iounmap(ctrl->creg);
|
||||||
release_mem_region(ctrl->mmio_base, ctrl->mmio_size);
|
release_mem_region(ctrl->mmio_base, ctrl->mmio_size);
|
||||||
|
|
||||||
/*
|
|
||||||
* If this is the last controller to be released, destroy the
|
|
||||||
* shpchpd work queue
|
|
||||||
*/
|
|
||||||
if (atomic_dec_and_test(&shpchp_num_controllers))
|
|
||||||
destroy_workqueue(shpchp_wq);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int hpc_power_on_slot(struct slot * slot)
|
static int hpc_power_on_slot(struct slot * slot)
|
||||||
|
@ -1077,9 +1068,8 @@ int shpc_init(struct controller *ctrl, struct pci_dev *pdev)
|
||||||
|
|
||||||
rc = request_irq(ctrl->pci_dev->irq, shpc_isr, IRQF_SHARED,
|
rc = request_irq(ctrl->pci_dev->irq, shpc_isr, IRQF_SHARED,
|
||||||
MY_NAME, (void *)ctrl);
|
MY_NAME, (void *)ctrl);
|
||||||
ctrl_dbg(ctrl, "request_irq %d for hpc%d (returns %d)\n",
|
ctrl_dbg(ctrl, "request_irq %d (returns %d)\n",
|
||||||
ctrl->pci_dev->irq,
|
ctrl->pci_dev->irq, rc);
|
||||||
atomic_read(&shpchp_num_controllers), rc);
|
|
||||||
if (rc) {
|
if (rc) {
|
||||||
ctrl_err(ctrl, "Can't get irq %d for the hotplug "
|
ctrl_err(ctrl, "Can't get irq %d for the hotplug "
|
||||||
"controller\n", ctrl->pci_dev->irq);
|
"controller\n", ctrl->pci_dev->irq);
|
||||||
|
@ -1091,18 +1081,6 @@ int shpc_init(struct controller *ctrl, struct pci_dev *pdev)
|
||||||
shpc_get_max_bus_speed(ctrl);
|
shpc_get_max_bus_speed(ctrl);
|
||||||
shpc_get_cur_bus_speed(ctrl);
|
shpc_get_cur_bus_speed(ctrl);
|
||||||
|
|
||||||
/*
|
|
||||||
* If this is the first controller to be initialized,
|
|
||||||
* initialize the shpchpd work queue
|
|
||||||
*/
|
|
||||||
if (atomic_add_return(1, &shpchp_num_controllers) == 1) {
|
|
||||||
shpchp_wq = create_singlethread_workqueue("shpchpd");
|
|
||||||
if (!shpchp_wq) {
|
|
||||||
rc = -ENOMEM;
|
|
||||||
goto abort_iounmap;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Unmask all event interrupts of all slots
|
* Unmask all event interrupts of all slots
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -144,7 +144,7 @@ static int __init init_gfs2_fs(void)
|
||||||
|
|
||||||
error = -ENOMEM;
|
error = -ENOMEM;
|
||||||
gfs_recovery_wq = alloc_workqueue("gfs_recovery",
|
gfs_recovery_wq = alloc_workqueue("gfs_recovery",
|
||||||
WQ_RESCUER | WQ_FREEZEABLE, 0);
|
WQ_MEM_RECLAIM | WQ_FREEZEABLE, 0);
|
||||||
if (!gfs_recovery_wq)
|
if (!gfs_recovery_wq)
|
||||||
goto fail_wq;
|
goto fail_wq;
|
||||||
|
|
||||||
|
|
|
@ -1921,7 +1921,7 @@ xfs_buf_init(void)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
xfslogd_workqueue = alloc_workqueue("xfslogd",
|
xfslogd_workqueue = alloc_workqueue("xfslogd",
|
||||||
WQ_RESCUER | WQ_HIGHPRI, 1);
|
WQ_MEM_RECLAIM | WQ_HIGHPRI, 1);
|
||||||
if (!xfslogd_workqueue)
|
if (!xfslogd_workqueue)
|
||||||
goto out_free_buf_zone;
|
goto out_free_buf_zone;
|
||||||
|
|
||||||
|
|
|
@ -243,11 +243,12 @@ enum {
|
||||||
WQ_NON_REENTRANT = 1 << 0, /* guarantee non-reentrance */
|
WQ_NON_REENTRANT = 1 << 0, /* guarantee non-reentrance */
|
||||||
WQ_UNBOUND = 1 << 1, /* not bound to any cpu */
|
WQ_UNBOUND = 1 << 1, /* not bound to any cpu */
|
||||||
WQ_FREEZEABLE = 1 << 2, /* freeze during suspend */
|
WQ_FREEZEABLE = 1 << 2, /* freeze during suspend */
|
||||||
WQ_RESCUER = 1 << 3, /* has an rescue worker */
|
WQ_MEM_RECLAIM = 1 << 3, /* may be used for memory reclaim */
|
||||||
WQ_HIGHPRI = 1 << 4, /* high priority */
|
WQ_HIGHPRI = 1 << 4, /* high priority */
|
||||||
WQ_CPU_INTENSIVE = 1 << 5, /* cpu instensive workqueue */
|
WQ_CPU_INTENSIVE = 1 << 5, /* cpu instensive workqueue */
|
||||||
|
|
||||||
WQ_DYING = 1 << 6, /* internal: workqueue is dying */
|
WQ_DYING = 1 << 6, /* internal: workqueue is dying */
|
||||||
|
WQ_RESCUER = 1 << 7, /* internal: workqueue has rescuer */
|
||||||
|
|
||||||
WQ_MAX_ACTIVE = 512, /* I like 512, better ideas? */
|
WQ_MAX_ACTIVE = 512, /* I like 512, better ideas? */
|
||||||
WQ_MAX_UNBOUND_PER_CPU = 4, /* 4 * #cpus for unbound wq */
|
WQ_MAX_UNBOUND_PER_CPU = 4, /* 4 * #cpus for unbound wq */
|
||||||
|
@ -306,12 +307,30 @@ __alloc_workqueue_key(const char *name, unsigned int flags, int max_active,
|
||||||
__alloc_workqueue_key((name), (flags), (max_active), NULL, NULL)
|
__alloc_workqueue_key((name), (flags), (max_active), NULL, NULL)
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
/**
|
||||||
|
* alloc_ordered_workqueue - allocate an ordered workqueue
|
||||||
|
* @name: name of the workqueue
|
||||||
|
* @flags: WQ_* flags (only WQ_FREEZEABLE and WQ_MEM_RECLAIM are meaningful)
|
||||||
|
*
|
||||||
|
* Allocate an ordered workqueue. An ordered workqueue executes at
|
||||||
|
* most one work item at any given time in the queued order. They are
|
||||||
|
* implemented as unbound workqueues with @max_active of one.
|
||||||
|
*
|
||||||
|
* RETURNS:
|
||||||
|
* Pointer to the allocated workqueue on success, %NULL on failure.
|
||||||
|
*/
|
||||||
|
static inline struct workqueue_struct *
|
||||||
|
alloc_ordered_workqueue(const char *name, unsigned int flags)
|
||||||
|
{
|
||||||
|
return alloc_workqueue(name, WQ_UNBOUND | flags, 1);
|
||||||
|
}
|
||||||
|
|
||||||
#define create_workqueue(name) \
|
#define create_workqueue(name) \
|
||||||
alloc_workqueue((name), WQ_RESCUER, 1)
|
alloc_workqueue((name), WQ_MEM_RECLAIM, 1)
|
||||||
#define create_freezeable_workqueue(name) \
|
#define create_freezeable_workqueue(name) \
|
||||||
alloc_workqueue((name), WQ_FREEZEABLE | WQ_UNBOUND | WQ_RESCUER, 1)
|
alloc_workqueue((name), WQ_FREEZEABLE | WQ_UNBOUND | WQ_MEM_RECLAIM, 1)
|
||||||
#define create_singlethread_workqueue(name) \
|
#define create_singlethread_workqueue(name) \
|
||||||
alloc_workqueue((name), WQ_UNBOUND | WQ_RESCUER, 1)
|
alloc_workqueue((name), WQ_UNBOUND | WQ_MEM_RECLAIM, 1)
|
||||||
|
|
||||||
extern void destroy_workqueue(struct workqueue_struct *wq);
|
extern void destroy_workqueue(struct workqueue_struct *wq);
|
||||||
|
|
||||||
|
@ -325,7 +344,6 @@ extern int queue_delayed_work_on(int cpu, struct workqueue_struct *wq,
|
||||||
|
|
||||||
extern void flush_workqueue(struct workqueue_struct *wq);
|
extern void flush_workqueue(struct workqueue_struct *wq);
|
||||||
extern void flush_scheduled_work(void);
|
extern void flush_scheduled_work(void);
|
||||||
extern void flush_delayed_work(struct delayed_work *work);
|
|
||||||
|
|
||||||
extern int schedule_work(struct work_struct *work);
|
extern int schedule_work(struct work_struct *work);
|
||||||
extern int schedule_work_on(int cpu, struct work_struct *work);
|
extern int schedule_work_on(int cpu, struct work_struct *work);
|
||||||
|
@ -337,8 +355,13 @@ extern int keventd_up(void);
|
||||||
|
|
||||||
int execute_in_process_context(work_func_t fn, struct execute_work *);
|
int execute_in_process_context(work_func_t fn, struct execute_work *);
|
||||||
|
|
||||||
extern int flush_work(struct work_struct *work);
|
extern bool flush_work(struct work_struct *work);
|
||||||
extern int cancel_work_sync(struct work_struct *work);
|
extern bool flush_work_sync(struct work_struct *work);
|
||||||
|
extern bool cancel_work_sync(struct work_struct *work);
|
||||||
|
|
||||||
|
extern bool flush_delayed_work(struct delayed_work *dwork);
|
||||||
|
extern bool flush_delayed_work_sync(struct delayed_work *work);
|
||||||
|
extern bool cancel_delayed_work_sync(struct delayed_work *dwork);
|
||||||
|
|
||||||
extern void workqueue_set_max_active(struct workqueue_struct *wq,
|
extern void workqueue_set_max_active(struct workqueue_struct *wq,
|
||||||
int max_active);
|
int max_active);
|
||||||
|
@ -352,9 +375,9 @@ extern unsigned int work_busy(struct work_struct *work);
|
||||||
* it returns 1 and the work doesn't re-arm itself. Run flush_workqueue() or
|
* it returns 1 and the work doesn't re-arm itself. Run flush_workqueue() or
|
||||||
* cancel_work_sync() to wait on it.
|
* cancel_work_sync() to wait on it.
|
||||||
*/
|
*/
|
||||||
static inline int cancel_delayed_work(struct delayed_work *work)
|
static inline bool cancel_delayed_work(struct delayed_work *work)
|
||||||
{
|
{
|
||||||
int ret;
|
bool ret;
|
||||||
|
|
||||||
ret = del_timer_sync(&work->timer);
|
ret = del_timer_sync(&work->timer);
|
||||||
if (ret)
|
if (ret)
|
||||||
|
@ -367,9 +390,9 @@ static inline int cancel_delayed_work(struct delayed_work *work)
|
||||||
* if it returns 0 the timer function may be running and the queueing is in
|
* if it returns 0 the timer function may be running and the queueing is in
|
||||||
* progress.
|
* progress.
|
||||||
*/
|
*/
|
||||||
static inline int __cancel_delayed_work(struct delayed_work *work)
|
static inline bool __cancel_delayed_work(struct delayed_work *work)
|
||||||
{
|
{
|
||||||
int ret;
|
bool ret;
|
||||||
|
|
||||||
ret = del_timer(&work->timer);
|
ret = del_timer(&work->timer);
|
||||||
if (ret)
|
if (ret)
|
||||||
|
@ -377,8 +400,6 @@ static inline int __cancel_delayed_work(struct delayed_work *work)
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
extern int cancel_delayed_work_sync(struct delayed_work *work);
|
|
||||||
|
|
||||||
/* Obsolete. use cancel_delayed_work_sync() */
|
/* Obsolete. use cancel_delayed_work_sync() */
|
||||||
static inline
|
static inline
|
||||||
void cancel_rearming_delayed_workqueue(struct workqueue_struct *wq,
|
void cancel_rearming_delayed_workqueue(struct workqueue_struct *wq,
|
||||||
|
@ -409,8 +430,4 @@ extern bool freeze_workqueues_busy(void);
|
||||||
extern void thaw_workqueues(void);
|
extern void thaw_workqueues(void);
|
||||||
#endif /* CONFIG_FREEZER */
|
#endif /* CONFIG_FREEZER */
|
||||||
|
|
||||||
#ifdef CONFIG_LOCKDEP
|
|
||||||
int in_workqueue_context(struct workqueue_struct *wq);
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -7,6 +7,76 @@
|
||||||
#include <linux/tracepoint.h>
|
#include <linux/tracepoint.h>
|
||||||
#include <linux/workqueue.h>
|
#include <linux/workqueue.h>
|
||||||
|
|
||||||
|
DECLARE_EVENT_CLASS(workqueue_work,
|
||||||
|
|
||||||
|
TP_PROTO(struct work_struct *work),
|
||||||
|
|
||||||
|
TP_ARGS(work),
|
||||||
|
|
||||||
|
TP_STRUCT__entry(
|
||||||
|
__field( void *, work )
|
||||||
|
),
|
||||||
|
|
||||||
|
TP_fast_assign(
|
||||||
|
__entry->work = work;
|
||||||
|
),
|
||||||
|
|
||||||
|
TP_printk("work struct %p", __entry->work)
|
||||||
|
);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* workqueue_queue_work - called when a work gets queued
|
||||||
|
* @req_cpu: the requested cpu
|
||||||
|
* @cwq: pointer to struct cpu_workqueue_struct
|
||||||
|
* @work: pointer to struct work_struct
|
||||||
|
*
|
||||||
|
* This event occurs when a work is queued immediately or once a
|
||||||
|
* delayed work is actually queued on a workqueue (ie: once the delay
|
||||||
|
* has been reached).
|
||||||
|
*/
|
||||||
|
TRACE_EVENT(workqueue_queue_work,
|
||||||
|
|
||||||
|
TP_PROTO(unsigned int req_cpu, struct cpu_workqueue_struct *cwq,
|
||||||
|
struct work_struct *work),
|
||||||
|
|
||||||
|
TP_ARGS(req_cpu, cwq, work),
|
||||||
|
|
||||||
|
TP_STRUCT__entry(
|
||||||
|
__field( void *, work )
|
||||||
|
__field( void *, function)
|
||||||
|
__field( void *, workqueue)
|
||||||
|
__field( unsigned int, req_cpu )
|
||||||
|
__field( unsigned int, cpu )
|
||||||
|
),
|
||||||
|
|
||||||
|
TP_fast_assign(
|
||||||
|
__entry->work = work;
|
||||||
|
__entry->function = work->func;
|
||||||
|
__entry->workqueue = cwq->wq;
|
||||||
|
__entry->req_cpu = req_cpu;
|
||||||
|
__entry->cpu = cwq->gcwq->cpu;
|
||||||
|
),
|
||||||
|
|
||||||
|
TP_printk("work struct=%p function=%pf workqueue=%p req_cpu=%u cpu=%u",
|
||||||
|
__entry->work, __entry->function, __entry->workqueue,
|
||||||
|
__entry->req_cpu, __entry->cpu)
|
||||||
|
);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* workqueue_activate_work - called when a work gets activated
|
||||||
|
* @work: pointer to struct work_struct
|
||||||
|
*
|
||||||
|
* This event occurs when a queued work is put on the active queue,
|
||||||
|
* which happens immediately after queueing unless @max_active limit
|
||||||
|
* is reached.
|
||||||
|
*/
|
||||||
|
DEFINE_EVENT(workqueue_work, workqueue_activate_work,
|
||||||
|
|
||||||
|
TP_PROTO(struct work_struct *work),
|
||||||
|
|
||||||
|
TP_ARGS(work)
|
||||||
|
);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* workqueue_execute_start - called immediately before the workqueue callback
|
* workqueue_execute_start - called immediately before the workqueue callback
|
||||||
* @work: pointer to struct work_struct
|
* @work: pointer to struct work_struct
|
||||||
|
@ -38,24 +108,13 @@ TRACE_EVENT(workqueue_execute_start,
|
||||||
*
|
*
|
||||||
* Allows to track workqueue execution.
|
* Allows to track workqueue execution.
|
||||||
*/
|
*/
|
||||||
TRACE_EVENT(workqueue_execute_end,
|
DEFINE_EVENT(workqueue_work, workqueue_execute_end,
|
||||||
|
|
||||||
TP_PROTO(struct work_struct *work),
|
TP_PROTO(struct work_struct *work),
|
||||||
|
|
||||||
TP_ARGS(work),
|
TP_ARGS(work)
|
||||||
|
|
||||||
TP_STRUCT__entry(
|
|
||||||
__field( void *, work )
|
|
||||||
),
|
|
||||||
|
|
||||||
TP_fast_assign(
|
|
||||||
__entry->work = work;
|
|
||||||
),
|
|
||||||
|
|
||||||
TP_printk("work struct %p", __entry->work)
|
|
||||||
);
|
);
|
||||||
|
|
||||||
|
|
||||||
#endif /* _TRACE_WORKQUEUE_H */
|
#endif /* _TRACE_WORKQUEUE_H */
|
||||||
|
|
||||||
/* This part must be outside protection */
|
/* This part must be outside protection */
|
||||||
|
|
|
@ -42,9 +42,6 @@
|
||||||
#include <linux/lockdep.h>
|
#include <linux/lockdep.h>
|
||||||
#include <linux/idr.h>
|
#include <linux/idr.h>
|
||||||
|
|
||||||
#define CREATE_TRACE_POINTS
|
|
||||||
#include <trace/events/workqueue.h>
|
|
||||||
|
|
||||||
#include "workqueue_sched.h"
|
#include "workqueue_sched.h"
|
||||||
|
|
||||||
enum {
|
enum {
|
||||||
|
@ -257,6 +254,9 @@ EXPORT_SYMBOL_GPL(system_long_wq);
|
||||||
EXPORT_SYMBOL_GPL(system_nrt_wq);
|
EXPORT_SYMBOL_GPL(system_nrt_wq);
|
||||||
EXPORT_SYMBOL_GPL(system_unbound_wq);
|
EXPORT_SYMBOL_GPL(system_unbound_wq);
|
||||||
|
|
||||||
|
#define CREATE_TRACE_POINTS
|
||||||
|
#include <trace/events/workqueue.h>
|
||||||
|
|
||||||
#define for_each_busy_worker(worker, i, pos, gcwq) \
|
#define for_each_busy_worker(worker, i, pos, gcwq) \
|
||||||
for (i = 0; i < BUSY_WORKER_HASH_SIZE; i++) \
|
for (i = 0; i < BUSY_WORKER_HASH_SIZE; i++) \
|
||||||
hlist_for_each_entry(worker, pos, &gcwq->busy_hash[i], hentry)
|
hlist_for_each_entry(worker, pos, &gcwq->busy_hash[i], hentry)
|
||||||
|
@ -310,21 +310,6 @@ static inline int __next_wq_cpu(int cpu, const struct cpumask *mask,
|
||||||
(cpu) < WORK_CPU_NONE; \
|
(cpu) < WORK_CPU_NONE; \
|
||||||
(cpu) = __next_wq_cpu((cpu), cpu_possible_mask, (wq)))
|
(cpu) = __next_wq_cpu((cpu), cpu_possible_mask, (wq)))
|
||||||
|
|
||||||
#ifdef CONFIG_LOCKDEP
|
|
||||||
/**
|
|
||||||
* in_workqueue_context() - in context of specified workqueue?
|
|
||||||
* @wq: the workqueue of interest
|
|
||||||
*
|
|
||||||
* Checks lockdep state to see if the current task is executing from
|
|
||||||
* within a workqueue item. This function exists only if lockdep is
|
|
||||||
* enabled.
|
|
||||||
*/
|
|
||||||
int in_workqueue_context(struct workqueue_struct *wq)
|
|
||||||
{
|
|
||||||
return lock_is_held(&wq->lockdep_map);
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifdef CONFIG_DEBUG_OBJECTS_WORK
|
#ifdef CONFIG_DEBUG_OBJECTS_WORK
|
||||||
|
|
||||||
static struct debug_obj_descr work_debug_descr;
|
static struct debug_obj_descr work_debug_descr;
|
||||||
|
@ -604,7 +589,9 @@ static bool keep_working(struct global_cwq *gcwq)
|
||||||
{
|
{
|
||||||
atomic_t *nr_running = get_gcwq_nr_running(gcwq->cpu);
|
atomic_t *nr_running = get_gcwq_nr_running(gcwq->cpu);
|
||||||
|
|
||||||
return !list_empty(&gcwq->worklist) && atomic_read(nr_running) <= 1;
|
return !list_empty(&gcwq->worklist) &&
|
||||||
|
(atomic_read(nr_running) <= 1 ||
|
||||||
|
gcwq->flags & GCWQ_HIGHPRI_PENDING);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Do we need a new worker? Called from manager. */
|
/* Do we need a new worker? Called from manager. */
|
||||||
|
@ -997,6 +984,7 @@ static void __queue_work(unsigned int cpu, struct workqueue_struct *wq,
|
||||||
|
|
||||||
/* gcwq determined, get cwq and queue */
|
/* gcwq determined, get cwq and queue */
|
||||||
cwq = get_cwq(gcwq->cpu, wq);
|
cwq = get_cwq(gcwq->cpu, wq);
|
||||||
|
trace_workqueue_queue_work(cpu, cwq, work);
|
||||||
|
|
||||||
BUG_ON(!list_empty(&work->entry));
|
BUG_ON(!list_empty(&work->entry));
|
||||||
|
|
||||||
|
@ -1004,6 +992,7 @@ static void __queue_work(unsigned int cpu, struct workqueue_struct *wq,
|
||||||
work_flags = work_color_to_flags(cwq->work_color);
|
work_flags = work_color_to_flags(cwq->work_color);
|
||||||
|
|
||||||
if (likely(cwq->nr_active < cwq->max_active)) {
|
if (likely(cwq->nr_active < cwq->max_active)) {
|
||||||
|
trace_workqueue_activate_work(work);
|
||||||
cwq->nr_active++;
|
cwq->nr_active++;
|
||||||
worklist = gcwq_determine_ins_pos(gcwq, cwq);
|
worklist = gcwq_determine_ins_pos(gcwq, cwq);
|
||||||
} else {
|
} else {
|
||||||
|
@ -1679,6 +1668,7 @@ static void cwq_activate_first_delayed(struct cpu_workqueue_struct *cwq)
|
||||||
struct work_struct, entry);
|
struct work_struct, entry);
|
||||||
struct list_head *pos = gcwq_determine_ins_pos(cwq->gcwq, cwq);
|
struct list_head *pos = gcwq_determine_ins_pos(cwq->gcwq, cwq);
|
||||||
|
|
||||||
|
trace_workqueue_activate_work(work);
|
||||||
move_linked_works(work, pos, NULL);
|
move_linked_works(work, pos, NULL);
|
||||||
__clear_bit(WORK_STRUCT_DELAYED_BIT, work_data_bits(work));
|
__clear_bit(WORK_STRUCT_DELAYED_BIT, work_data_bits(work));
|
||||||
cwq->nr_active++;
|
cwq->nr_active++;
|
||||||
|
@ -2326,27 +2316,17 @@ void flush_workqueue(struct workqueue_struct *wq)
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(flush_workqueue);
|
EXPORT_SYMBOL_GPL(flush_workqueue);
|
||||||
|
|
||||||
/**
|
static bool start_flush_work(struct work_struct *work, struct wq_barrier *barr,
|
||||||
* flush_work - block until a work_struct's callback has terminated
|
bool wait_executing)
|
||||||
* @work: the work which is to be flushed
|
|
||||||
*
|
|
||||||
* Returns false if @work has already terminated.
|
|
||||||
*
|
|
||||||
* It is expected that, prior to calling flush_work(), the caller has
|
|
||||||
* arranged for the work to not be requeued, otherwise it doesn't make
|
|
||||||
* sense to use this function.
|
|
||||||
*/
|
|
||||||
int flush_work(struct work_struct *work)
|
|
||||||
{
|
{
|
||||||
struct worker *worker = NULL;
|
struct worker *worker = NULL;
|
||||||
struct global_cwq *gcwq;
|
struct global_cwq *gcwq;
|
||||||
struct cpu_workqueue_struct *cwq;
|
struct cpu_workqueue_struct *cwq;
|
||||||
struct wq_barrier barr;
|
|
||||||
|
|
||||||
might_sleep();
|
might_sleep();
|
||||||
gcwq = get_work_gcwq(work);
|
gcwq = get_work_gcwq(work);
|
||||||
if (!gcwq)
|
if (!gcwq)
|
||||||
return 0;
|
return false;
|
||||||
|
|
||||||
spin_lock_irq(&gcwq->lock);
|
spin_lock_irq(&gcwq->lock);
|
||||||
if (!list_empty(&work->entry)) {
|
if (!list_empty(&work->entry)) {
|
||||||
|
@ -2359,28 +2339,127 @@ int flush_work(struct work_struct *work)
|
||||||
cwq = get_work_cwq(work);
|
cwq = get_work_cwq(work);
|
||||||
if (unlikely(!cwq || gcwq != cwq->gcwq))
|
if (unlikely(!cwq || gcwq != cwq->gcwq))
|
||||||
goto already_gone;
|
goto already_gone;
|
||||||
} else {
|
} else if (wait_executing) {
|
||||||
worker = find_worker_executing_work(gcwq, work);
|
worker = find_worker_executing_work(gcwq, work);
|
||||||
if (!worker)
|
if (!worker)
|
||||||
goto already_gone;
|
goto already_gone;
|
||||||
cwq = worker->current_cwq;
|
cwq = worker->current_cwq;
|
||||||
}
|
} else
|
||||||
|
goto already_gone;
|
||||||
|
|
||||||
insert_wq_barrier(cwq, &barr, work, worker);
|
insert_wq_barrier(cwq, barr, work, worker);
|
||||||
spin_unlock_irq(&gcwq->lock);
|
spin_unlock_irq(&gcwq->lock);
|
||||||
|
|
||||||
lock_map_acquire(&cwq->wq->lockdep_map);
|
lock_map_acquire(&cwq->wq->lockdep_map);
|
||||||
lock_map_release(&cwq->wq->lockdep_map);
|
lock_map_release(&cwq->wq->lockdep_map);
|
||||||
|
return true;
|
||||||
wait_for_completion(&barr.done);
|
|
||||||
destroy_work_on_stack(&barr.work);
|
|
||||||
return 1;
|
|
||||||
already_gone:
|
already_gone:
|
||||||
spin_unlock_irq(&gcwq->lock);
|
spin_unlock_irq(&gcwq->lock);
|
||||||
return 0;
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* flush_work - wait for a work to finish executing the last queueing instance
|
||||||
|
* @work: the work to flush
|
||||||
|
*
|
||||||
|
* Wait until @work has finished execution. This function considers
|
||||||
|
* only the last queueing instance of @work. If @work has been
|
||||||
|
* enqueued across different CPUs on a non-reentrant workqueue or on
|
||||||
|
* multiple workqueues, @work might still be executing on return on
|
||||||
|
* some of the CPUs from earlier queueing.
|
||||||
|
*
|
||||||
|
* If @work was queued only on a non-reentrant, ordered or unbound
|
||||||
|
* workqueue, @work is guaranteed to be idle on return if it hasn't
|
||||||
|
* been requeued since flush started.
|
||||||
|
*
|
||||||
|
* RETURNS:
|
||||||
|
* %true if flush_work() waited for the work to finish execution,
|
||||||
|
* %false if it was already idle.
|
||||||
|
*/
|
||||||
|
bool flush_work(struct work_struct *work)
|
||||||
|
{
|
||||||
|
struct wq_barrier barr;
|
||||||
|
|
||||||
|
if (start_flush_work(work, &barr, true)) {
|
||||||
|
wait_for_completion(&barr.done);
|
||||||
|
destroy_work_on_stack(&barr.work);
|
||||||
|
return true;
|
||||||
|
} else
|
||||||
|
return false;
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(flush_work);
|
EXPORT_SYMBOL_GPL(flush_work);
|
||||||
|
|
||||||
|
static bool wait_on_cpu_work(struct global_cwq *gcwq, struct work_struct *work)
|
||||||
|
{
|
||||||
|
struct wq_barrier barr;
|
||||||
|
struct worker *worker;
|
||||||
|
|
||||||
|
spin_lock_irq(&gcwq->lock);
|
||||||
|
|
||||||
|
worker = find_worker_executing_work(gcwq, work);
|
||||||
|
if (unlikely(worker))
|
||||||
|
insert_wq_barrier(worker->current_cwq, &barr, work, worker);
|
||||||
|
|
||||||
|
spin_unlock_irq(&gcwq->lock);
|
||||||
|
|
||||||
|
if (unlikely(worker)) {
|
||||||
|
wait_for_completion(&barr.done);
|
||||||
|
destroy_work_on_stack(&barr.work);
|
||||||
|
return true;
|
||||||
|
} else
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
static bool wait_on_work(struct work_struct *work)
|
||||||
|
{
|
||||||
|
bool ret = false;
|
||||||
|
int cpu;
|
||||||
|
|
||||||
|
might_sleep();
|
||||||
|
|
||||||
|
lock_map_acquire(&work->lockdep_map);
|
||||||
|
lock_map_release(&work->lockdep_map);
|
||||||
|
|
||||||
|
for_each_gcwq_cpu(cpu)
|
||||||
|
ret |= wait_on_cpu_work(get_gcwq(cpu), work);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* flush_work_sync - wait until a work has finished execution
|
||||||
|
* @work: the work to flush
|
||||||
|
*
|
||||||
|
* Wait until @work has finished execution. On return, it's
|
||||||
|
* guaranteed that all queueing instances of @work which happened
|
||||||
|
* before this function is called are finished. In other words, if
|
||||||
|
* @work hasn't been requeued since this function was called, @work is
|
||||||
|
* guaranteed to be idle on return.
|
||||||
|
*
|
||||||
|
* RETURNS:
|
||||||
|
* %true if flush_work_sync() waited for the work to finish execution,
|
||||||
|
* %false if it was already idle.
|
||||||
|
*/
|
||||||
|
bool flush_work_sync(struct work_struct *work)
|
||||||
|
{
|
||||||
|
struct wq_barrier barr;
|
||||||
|
bool pending, waited;
|
||||||
|
|
||||||
|
/* we'll wait for executions separately, queue barr only if pending */
|
||||||
|
pending = start_flush_work(work, &barr, false);
|
||||||
|
|
||||||
|
/* wait for executions to finish */
|
||||||
|
waited = wait_on_work(work);
|
||||||
|
|
||||||
|
/* wait for the pending one */
|
||||||
|
if (pending) {
|
||||||
|
wait_for_completion(&barr.done);
|
||||||
|
destroy_work_on_stack(&barr.work);
|
||||||
|
}
|
||||||
|
|
||||||
|
return pending || waited;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(flush_work_sync);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Upon a successful return (>= 0), the caller "owns" WORK_STRUCT_PENDING bit,
|
* Upon a successful return (>= 0), the caller "owns" WORK_STRUCT_PENDING bit,
|
||||||
* so this work can't be re-armed in any way.
|
* so this work can't be re-armed in any way.
|
||||||
|
@ -2423,39 +2502,7 @@ static int try_to_grab_pending(struct work_struct *work)
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void wait_on_cpu_work(struct global_cwq *gcwq, struct work_struct *work)
|
static bool __cancel_work_timer(struct work_struct *work,
|
||||||
{
|
|
||||||
struct wq_barrier barr;
|
|
||||||
struct worker *worker;
|
|
||||||
|
|
||||||
spin_lock_irq(&gcwq->lock);
|
|
||||||
|
|
||||||
worker = find_worker_executing_work(gcwq, work);
|
|
||||||
if (unlikely(worker))
|
|
||||||
insert_wq_barrier(worker->current_cwq, &barr, work, worker);
|
|
||||||
|
|
||||||
spin_unlock_irq(&gcwq->lock);
|
|
||||||
|
|
||||||
if (unlikely(worker)) {
|
|
||||||
wait_for_completion(&barr.done);
|
|
||||||
destroy_work_on_stack(&barr.work);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
static void wait_on_work(struct work_struct *work)
|
|
||||||
{
|
|
||||||
int cpu;
|
|
||||||
|
|
||||||
might_sleep();
|
|
||||||
|
|
||||||
lock_map_acquire(&work->lockdep_map);
|
|
||||||
lock_map_release(&work->lockdep_map);
|
|
||||||
|
|
||||||
for_each_gcwq_cpu(cpu)
|
|
||||||
wait_on_cpu_work(get_gcwq(cpu), work);
|
|
||||||
}
|
|
||||||
|
|
||||||
static int __cancel_work_timer(struct work_struct *work,
|
|
||||||
struct timer_list* timer)
|
struct timer_list* timer)
|
||||||
{
|
{
|
||||||
int ret;
|
int ret;
|
||||||
|
@ -2472,42 +2519,81 @@ static int __cancel_work_timer(struct work_struct *work,
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* cancel_work_sync - block until a work_struct's callback has terminated
|
* cancel_work_sync - cancel a work and wait for it to finish
|
||||||
* @work: the work which is to be flushed
|
* @work: the work to cancel
|
||||||
*
|
*
|
||||||
* Returns true if @work was pending.
|
* Cancel @work and wait for its execution to finish. This function
|
||||||
|
* can be used even if the work re-queues itself or migrates to
|
||||||
|
* another workqueue. On return from this function, @work is
|
||||||
|
* guaranteed to be not pending or executing on any CPU.
|
||||||
*
|
*
|
||||||
* cancel_work_sync() will cancel the work if it is queued. If the work's
|
* cancel_work_sync(&delayed_work->work) must not be used for
|
||||||
* callback appears to be running, cancel_work_sync() will block until it
|
* delayed_work's. Use cancel_delayed_work_sync() instead.
|
||||||
* has completed.
|
|
||||||
*
|
*
|
||||||
* It is possible to use this function if the work re-queues itself. It can
|
* The caller must ensure that the workqueue on which @work was last
|
||||||
* cancel the work even if it migrates to another workqueue, however in that
|
|
||||||
* case it only guarantees that work->func() has completed on the last queued
|
|
||||||
* workqueue.
|
|
||||||
*
|
|
||||||
* cancel_work_sync(&delayed_work->work) should be used only if ->timer is not
|
|
||||||
* pending, otherwise it goes into a busy-wait loop until the timer expires.
|
|
||||||
*
|
|
||||||
* The caller must ensure that workqueue_struct on which this work was last
|
|
||||||
* queued can't be destroyed before this function returns.
|
* queued can't be destroyed before this function returns.
|
||||||
|
*
|
||||||
|
* RETURNS:
|
||||||
|
* %true if @work was pending, %false otherwise.
|
||||||
*/
|
*/
|
||||||
int cancel_work_sync(struct work_struct *work)
|
bool cancel_work_sync(struct work_struct *work)
|
||||||
{
|
{
|
||||||
return __cancel_work_timer(work, NULL);
|
return __cancel_work_timer(work, NULL);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(cancel_work_sync);
|
EXPORT_SYMBOL_GPL(cancel_work_sync);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* cancel_delayed_work_sync - reliably kill off a delayed work.
|
* flush_delayed_work - wait for a dwork to finish executing the last queueing
|
||||||
* @dwork: the delayed work struct
|
* @dwork: the delayed work to flush
|
||||||
*
|
*
|
||||||
* Returns true if @dwork was pending.
|
* Delayed timer is cancelled and the pending work is queued for
|
||||||
|
* immediate execution. Like flush_work(), this function only
|
||||||
|
* considers the last queueing instance of @dwork.
|
||||||
*
|
*
|
||||||
* It is possible to use this function if @dwork rearms itself via queue_work()
|
* RETURNS:
|
||||||
* or queue_delayed_work(). See also the comment for cancel_work_sync().
|
* %true if flush_work() waited for the work to finish execution,
|
||||||
|
* %false if it was already idle.
|
||||||
*/
|
*/
|
||||||
int cancel_delayed_work_sync(struct delayed_work *dwork)
|
bool flush_delayed_work(struct delayed_work *dwork)
|
||||||
|
{
|
||||||
|
if (del_timer_sync(&dwork->timer))
|
||||||
|
__queue_work(raw_smp_processor_id(),
|
||||||
|
get_work_cwq(&dwork->work)->wq, &dwork->work);
|
||||||
|
return flush_work(&dwork->work);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(flush_delayed_work);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* flush_delayed_work_sync - wait for a dwork to finish
|
||||||
|
* @dwork: the delayed work to flush
|
||||||
|
*
|
||||||
|
* Delayed timer is cancelled and the pending work is queued for
|
||||||
|
* execution immediately. Other than timer handling, its behavior
|
||||||
|
* is identical to flush_work_sync().
|
||||||
|
*
|
||||||
|
* RETURNS:
|
||||||
|
* %true if flush_work_sync() waited for the work to finish execution,
|
||||||
|
* %false if it was already idle.
|
||||||
|
*/
|
||||||
|
bool flush_delayed_work_sync(struct delayed_work *dwork)
|
||||||
|
{
|
||||||
|
if (del_timer_sync(&dwork->timer))
|
||||||
|
__queue_work(raw_smp_processor_id(),
|
||||||
|
get_work_cwq(&dwork->work)->wq, &dwork->work);
|
||||||
|
return flush_work_sync(&dwork->work);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(flush_delayed_work_sync);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* cancel_delayed_work_sync - cancel a delayed work and wait for it to finish
|
||||||
|
* @dwork: the delayed work cancel
|
||||||
|
*
|
||||||
|
* This is cancel_work_sync() for delayed works.
|
||||||
|
*
|
||||||
|
* RETURNS:
|
||||||
|
* %true if @dwork was pending, %false otherwise.
|
||||||
|
*/
|
||||||
|
bool cancel_delayed_work_sync(struct delayed_work *dwork)
|
||||||
{
|
{
|
||||||
return __cancel_work_timer(&dwork->work, &dwork->timer);
|
return __cancel_work_timer(&dwork->work, &dwork->timer);
|
||||||
}
|
}
|
||||||
|
@ -2558,23 +2644,6 @@ int schedule_delayed_work(struct delayed_work *dwork,
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(schedule_delayed_work);
|
EXPORT_SYMBOL(schedule_delayed_work);
|
||||||
|
|
||||||
/**
|
|
||||||
* flush_delayed_work - block until a dwork_struct's callback has terminated
|
|
||||||
* @dwork: the delayed work which is to be flushed
|
|
||||||
*
|
|
||||||
* Any timeout is cancelled, and any pending work is run immediately.
|
|
||||||
*/
|
|
||||||
void flush_delayed_work(struct delayed_work *dwork)
|
|
||||||
{
|
|
||||||
if (del_timer_sync(&dwork->timer)) {
|
|
||||||
__queue_work(get_cpu(), get_work_cwq(&dwork->work)->wq,
|
|
||||||
&dwork->work);
|
|
||||||
put_cpu();
|
|
||||||
}
|
|
||||||
flush_work(&dwork->work);
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL(flush_delayed_work);
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* schedule_delayed_work_on - queue work in global workqueue on CPU after delay
|
* schedule_delayed_work_on - queue work in global workqueue on CPU after delay
|
||||||
* @cpu: cpu to use
|
* @cpu: cpu to use
|
||||||
|
@ -2592,13 +2661,15 @@ int schedule_delayed_work_on(int cpu,
|
||||||
EXPORT_SYMBOL(schedule_delayed_work_on);
|
EXPORT_SYMBOL(schedule_delayed_work_on);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* schedule_on_each_cpu - call a function on each online CPU from keventd
|
* schedule_on_each_cpu - execute a function synchronously on each online CPU
|
||||||
* @func: the function to call
|
* @func: the function to call
|
||||||
*
|
*
|
||||||
* Returns zero on success.
|
* schedule_on_each_cpu() executes @func on each online CPU using the
|
||||||
* Returns -ve errno on failure.
|
* system workqueue and blocks until all CPUs have completed.
|
||||||
*
|
|
||||||
* schedule_on_each_cpu() is very slow.
|
* schedule_on_each_cpu() is very slow.
|
||||||
|
*
|
||||||
|
* RETURNS:
|
||||||
|
* 0 on success, -errno on failure.
|
||||||
*/
|
*/
|
||||||
int schedule_on_each_cpu(work_func_t func)
|
int schedule_on_each_cpu(work_func_t func)
|
||||||
{
|
{
|
||||||
|
@ -2763,6 +2834,13 @@ struct workqueue_struct *__alloc_workqueue_key(const char *name,
|
||||||
struct workqueue_struct *wq;
|
struct workqueue_struct *wq;
|
||||||
unsigned int cpu;
|
unsigned int cpu;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Workqueues which may be used during memory reclaim should
|
||||||
|
* have a rescuer to guarantee forward progress.
|
||||||
|
*/
|
||||||
|
if (flags & WQ_MEM_RECLAIM)
|
||||||
|
flags |= WQ_RESCUER;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Unbound workqueues aren't concurrency managed and should be
|
* Unbound workqueues aren't concurrency managed and should be
|
||||||
* dispatched to workers immediately.
|
* dispatched to workers immediately.
|
||||||
|
|
|
@ -840,7 +840,6 @@ static int offline_pages(unsigned long start_pfn,
|
||||||
ret = 0;
|
ret = 0;
|
||||||
if (drain) {
|
if (drain) {
|
||||||
lru_add_drain_all();
|
lru_add_drain_all();
|
||||||
flush_scheduled_work();
|
|
||||||
cond_resched();
|
cond_resched();
|
||||||
drain_all_pages();
|
drain_all_pages();
|
||||||
}
|
}
|
||||||
|
@ -862,7 +861,6 @@ static int offline_pages(unsigned long start_pfn,
|
||||||
}
|
}
|
||||||
/* drain all zone's lru pagevec, this is asyncronous... */
|
/* drain all zone's lru pagevec, this is asyncronous... */
|
||||||
lru_add_drain_all();
|
lru_add_drain_all();
|
||||||
flush_scheduled_work();
|
|
||||||
yield();
|
yield();
|
||||||
/* drain pcp pages , this is synchrouns. */
|
/* drain pcp pages , this is synchrouns. */
|
||||||
drain_all_pages();
|
drain_all_pages();
|
||||||
|
|
Loading…
Reference in New Issue