mirror of https://gitee.com/openkylin/linux.git
Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/net
This commit is contained in:
commit
c63d352f05
11
Makefile
11
Makefile
|
@ -1,7 +1,7 @@
|
||||||
VERSION = 4
|
VERSION = 4
|
||||||
PATCHLEVEL = 9
|
PATCHLEVEL = 9
|
||||||
SUBLEVEL = 0
|
SUBLEVEL = 0
|
||||||
EXTRAVERSION = -rc7
|
EXTRAVERSION = -rc8
|
||||||
NAME = Psychotic Stoned Sheep
|
NAME = Psychotic Stoned Sheep
|
||||||
|
|
||||||
# *DOCUMENTATION*
|
# *DOCUMENTATION*
|
||||||
|
@ -607,6 +607,13 @@ else
|
||||||
include/config/auto.conf: ;
|
include/config/auto.conf: ;
|
||||||
endif # $(dot-config)
|
endif # $(dot-config)
|
||||||
|
|
||||||
|
# For the kernel to actually contain only the needed exported symbols,
|
||||||
|
# we have to build modules as well to determine what those symbols are.
|
||||||
|
# (this can be evaluated only once include/config/auto.conf has been included)
|
||||||
|
ifdef CONFIG_TRIM_UNUSED_KSYMS
|
||||||
|
KBUILD_MODULES := 1
|
||||||
|
endif
|
||||||
|
|
||||||
# The all: target is the default when no target is given on the
|
# The all: target is the default when no target is given on the
|
||||||
# command line.
|
# command line.
|
||||||
# This allow a user to issue only 'make' to build a kernel including modules
|
# This allow a user to issue only 'make' to build a kernel including modules
|
||||||
|
@ -944,7 +951,7 @@ ifdef CONFIG_GDB_SCRIPTS
|
||||||
endif
|
endif
|
||||||
ifdef CONFIG_TRIM_UNUSED_KSYMS
|
ifdef CONFIG_TRIM_UNUSED_KSYMS
|
||||||
$(Q)$(CONFIG_SHELL) $(srctree)/scripts/adjust_autoksyms.sh \
|
$(Q)$(CONFIG_SHELL) $(srctree)/scripts/adjust_autoksyms.sh \
|
||||||
"$(MAKE) KBUILD_MODULES=1 -f $(srctree)/Makefile vmlinux_prereq"
|
"$(MAKE) -f $(srctree)/Makefile vmlinux"
|
||||||
endif
|
endif
|
||||||
|
|
||||||
# standalone target for easier testing
|
# standalone target for easier testing
|
||||||
|
|
|
@ -100,7 +100,8 @@ src-wlib-y := string.S crt0.S crtsavres.S stdio.c decompress.c main.c \
|
||||||
ns16550.c serial.c simple_alloc.c div64.S util.S \
|
ns16550.c serial.c simple_alloc.c div64.S util.S \
|
||||||
elf_util.c $(zlib-y) devtree.c stdlib.c \
|
elf_util.c $(zlib-y) devtree.c stdlib.c \
|
||||||
oflib.c ofconsole.c cuboot.c mpsc.c cpm-serial.c \
|
oflib.c ofconsole.c cuboot.c mpsc.c cpm-serial.c \
|
||||||
uartlite.c mpc52xx-psc.c opal.c opal-calls.S
|
uartlite.c mpc52xx-psc.c opal.c
|
||||||
|
src-wlib-$(CONFIG_PPC64_BOOT_WRAPPER) += opal-calls.S
|
||||||
src-wlib-$(CONFIG_40x) += 4xx.c planetcore.c
|
src-wlib-$(CONFIG_40x) += 4xx.c planetcore.c
|
||||||
src-wlib-$(CONFIG_44x) += 4xx.c ebony.c bamboo.c
|
src-wlib-$(CONFIG_44x) += 4xx.c ebony.c bamboo.c
|
||||||
src-wlib-$(CONFIG_8xx) += mpc8xx.c planetcore.c fsl-soc.c
|
src-wlib-$(CONFIG_8xx) += mpc8xx.c planetcore.c fsl-soc.c
|
||||||
|
|
|
@ -13,7 +13,7 @@
|
||||||
#include <libfdt.h>
|
#include <libfdt.h>
|
||||||
#include "../include/asm/opal-api.h"
|
#include "../include/asm/opal-api.h"
|
||||||
|
|
||||||
#ifdef __powerpc64__
|
#ifdef CONFIG_PPC64_BOOT_WRAPPER
|
||||||
|
|
||||||
/* Global OPAL struct used by opal-call.S */
|
/* Global OPAL struct used by opal-call.S */
|
||||||
struct opal {
|
struct opal {
|
||||||
|
|
|
@ -671,8 +671,10 @@ static int eeh_reset_device(struct eeh_pe *pe, struct pci_bus *bus,
|
||||||
|
|
||||||
/* Clear frozen state */
|
/* Clear frozen state */
|
||||||
rc = eeh_clear_pe_frozen_state(pe, false);
|
rc = eeh_clear_pe_frozen_state(pe, false);
|
||||||
if (rc)
|
if (rc) {
|
||||||
|
pci_unlock_rescan_remove();
|
||||||
return rc;
|
return rc;
|
||||||
|
}
|
||||||
|
|
||||||
/* Give the system 5 seconds to finish running the user-space
|
/* Give the system 5 seconds to finish running the user-space
|
||||||
* hotplug shutdown scripts, e.g. ifdown for ethernet. Yes,
|
* hotplug shutdown scripts, e.g. ifdown for ethernet. Yes,
|
||||||
|
|
|
@ -94,8 +94,17 @@ SECTIONS
|
||||||
* detected, and will result in a crash at boot due to offsets being
|
* detected, and will result in a crash at boot due to offsets being
|
||||||
* wrong.
|
* wrong.
|
||||||
*/
|
*/
|
||||||
|
#ifdef CONFIG_PPC64
|
||||||
|
/*
|
||||||
|
* BLOCK(0) overrides the default output section alignment because
|
||||||
|
* this needs to start right after .head.text in order for fixed
|
||||||
|
* section placement to work.
|
||||||
|
*/
|
||||||
|
.text BLOCK(0) : AT(ADDR(.text) - LOAD_OFFSET) {
|
||||||
|
#else
|
||||||
.text : AT(ADDR(.text) - LOAD_OFFSET) {
|
.text : AT(ADDR(.text) - LOAD_OFFSET) {
|
||||||
ALIGN_FUNCTION();
|
ALIGN_FUNCTION();
|
||||||
|
#endif
|
||||||
/* careful! __ftr_alt_* sections need to be close to .text */
|
/* careful! __ftr_alt_* sections need to be close to .text */
|
||||||
*(.text .fixup __ftr_alt_* .ref.text)
|
*(.text .fixup __ftr_alt_* .ref.text)
|
||||||
SCHED_TEXT
|
SCHED_TEXT
|
||||||
|
|
|
@ -55,7 +55,7 @@ int __hash_page_4K(unsigned long ea, unsigned long access, unsigned long vsid,
|
||||||
*/
|
*/
|
||||||
rflags = htab_convert_pte_flags(new_pte);
|
rflags = htab_convert_pte_flags(new_pte);
|
||||||
|
|
||||||
if (!cpu_has_feature(CPU_FTR_NOEXECUTE) &&
|
if (cpu_has_feature(CPU_FTR_NOEXECUTE) &&
|
||||||
!cpu_has_feature(CPU_FTR_COHERENT_ICACHE))
|
!cpu_has_feature(CPU_FTR_COHERENT_ICACHE))
|
||||||
rflags = hash_page_do_lazy_icache(rflags, __pte(old_pte), trap);
|
rflags = hash_page_do_lazy_icache(rflags, __pte(old_pte), trap);
|
||||||
|
|
||||||
|
|
|
@ -87,7 +87,7 @@ int __hash_page_4K(unsigned long ea, unsigned long access, unsigned long vsid,
|
||||||
subpg_pte = new_pte & ~subpg_prot;
|
subpg_pte = new_pte & ~subpg_prot;
|
||||||
rflags = htab_convert_pte_flags(subpg_pte);
|
rflags = htab_convert_pte_flags(subpg_pte);
|
||||||
|
|
||||||
if (!cpu_has_feature(CPU_FTR_NOEXECUTE) &&
|
if (cpu_has_feature(CPU_FTR_NOEXECUTE) &&
|
||||||
!cpu_has_feature(CPU_FTR_COHERENT_ICACHE)) {
|
!cpu_has_feature(CPU_FTR_COHERENT_ICACHE)) {
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -258,7 +258,7 @@ int __hash_page_64K(unsigned long ea, unsigned long access,
|
||||||
|
|
||||||
rflags = htab_convert_pte_flags(new_pte);
|
rflags = htab_convert_pte_flags(new_pte);
|
||||||
|
|
||||||
if (!cpu_has_feature(CPU_FTR_NOEXECUTE) &&
|
if (cpu_has_feature(CPU_FTR_NOEXECUTE) &&
|
||||||
!cpu_has_feature(CPU_FTR_COHERENT_ICACHE))
|
!cpu_has_feature(CPU_FTR_COHERENT_ICACHE))
|
||||||
rflags = hash_page_do_lazy_icache(rflags, __pte(old_pte), trap);
|
rflags = hash_page_do_lazy_icache(rflags, __pte(old_pte), trap);
|
||||||
|
|
||||||
|
|
|
@ -40,6 +40,7 @@ obj-$(CONFIG_CRYPTO_ECDH) += ecdh_generic.o
|
||||||
|
|
||||||
$(obj)/rsapubkey-asn1.o: $(obj)/rsapubkey-asn1.c $(obj)/rsapubkey-asn1.h
|
$(obj)/rsapubkey-asn1.o: $(obj)/rsapubkey-asn1.c $(obj)/rsapubkey-asn1.h
|
||||||
$(obj)/rsaprivkey-asn1.o: $(obj)/rsaprivkey-asn1.c $(obj)/rsaprivkey-asn1.h
|
$(obj)/rsaprivkey-asn1.o: $(obj)/rsaprivkey-asn1.c $(obj)/rsaprivkey-asn1.h
|
||||||
|
$(obj)/rsa_helper.o: $(obj)/rsapubkey-asn1.h $(obj)/rsaprivkey-asn1.h
|
||||||
clean-files += rsapubkey-asn1.c rsapubkey-asn1.h
|
clean-files += rsapubkey-asn1.c rsapubkey-asn1.h
|
||||||
clean-files += rsaprivkey-asn1.c rsaprivkey-asn1.h
|
clean-files += rsaprivkey-asn1.c rsaprivkey-asn1.h
|
||||||
|
|
||||||
|
|
|
@ -262,6 +262,7 @@ static int drbg_kcapi_sym_ctr(struct drbg_state *drbg,
|
||||||
u8 *inbuf, u32 inbuflen,
|
u8 *inbuf, u32 inbuflen,
|
||||||
u8 *outbuf, u32 outlen);
|
u8 *outbuf, u32 outlen);
|
||||||
#define DRBG_CTR_NULL_LEN 128
|
#define DRBG_CTR_NULL_LEN 128
|
||||||
|
#define DRBG_OUTSCRATCHLEN DRBG_CTR_NULL_LEN
|
||||||
|
|
||||||
/* BCC function for CTR DRBG as defined in 10.4.3 */
|
/* BCC function for CTR DRBG as defined in 10.4.3 */
|
||||||
static int drbg_ctr_bcc(struct drbg_state *drbg,
|
static int drbg_ctr_bcc(struct drbg_state *drbg,
|
||||||
|
@ -1644,6 +1645,9 @@ static int drbg_fini_sym_kernel(struct drbg_state *drbg)
|
||||||
kfree(drbg->ctr_null_value_buf);
|
kfree(drbg->ctr_null_value_buf);
|
||||||
drbg->ctr_null_value = NULL;
|
drbg->ctr_null_value = NULL;
|
||||||
|
|
||||||
|
kfree(drbg->outscratchpadbuf);
|
||||||
|
drbg->outscratchpadbuf = NULL;
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1708,6 +1712,15 @@ static int drbg_init_sym_kernel(struct drbg_state *drbg)
|
||||||
drbg->ctr_null_value = (u8 *)PTR_ALIGN(drbg->ctr_null_value_buf,
|
drbg->ctr_null_value = (u8 *)PTR_ALIGN(drbg->ctr_null_value_buf,
|
||||||
alignmask + 1);
|
alignmask + 1);
|
||||||
|
|
||||||
|
drbg->outscratchpadbuf = kmalloc(DRBG_OUTSCRATCHLEN + alignmask,
|
||||||
|
GFP_KERNEL);
|
||||||
|
if (!drbg->outscratchpadbuf) {
|
||||||
|
drbg_fini_sym_kernel(drbg);
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
drbg->outscratchpad = (u8 *)PTR_ALIGN(drbg->outscratchpadbuf,
|
||||||
|
alignmask + 1);
|
||||||
|
|
||||||
return alignmask;
|
return alignmask;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1737,15 +1750,16 @@ static int drbg_kcapi_sym_ctr(struct drbg_state *drbg,
|
||||||
u8 *outbuf, u32 outlen)
|
u8 *outbuf, u32 outlen)
|
||||||
{
|
{
|
||||||
struct scatterlist sg_in;
|
struct scatterlist sg_in;
|
||||||
|
int ret;
|
||||||
|
|
||||||
sg_init_one(&sg_in, inbuf, inlen);
|
sg_init_one(&sg_in, inbuf, inlen);
|
||||||
|
|
||||||
while (outlen) {
|
while (outlen) {
|
||||||
u32 cryptlen = min_t(u32, inlen, outlen);
|
u32 cryptlen = min3(inlen, outlen, (u32)DRBG_OUTSCRATCHLEN);
|
||||||
struct scatterlist sg_out;
|
struct scatterlist sg_out;
|
||||||
int ret;
|
|
||||||
|
|
||||||
sg_init_one(&sg_out, outbuf, cryptlen);
|
/* Output buffer may not be valid for SGL, use scratchpad */
|
||||||
|
sg_init_one(&sg_out, drbg->outscratchpad, cryptlen);
|
||||||
skcipher_request_set_crypt(drbg->ctr_req, &sg_in, &sg_out,
|
skcipher_request_set_crypt(drbg->ctr_req, &sg_in, &sg_out,
|
||||||
cryptlen, drbg->V);
|
cryptlen, drbg->V);
|
||||||
ret = crypto_skcipher_encrypt(drbg->ctr_req);
|
ret = crypto_skcipher_encrypt(drbg->ctr_req);
|
||||||
|
@ -1761,14 +1775,19 @@ static int drbg_kcapi_sym_ctr(struct drbg_state *drbg,
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
default:
|
default:
|
||||||
return ret;
|
goto out;
|
||||||
}
|
}
|
||||||
init_completion(&drbg->ctr_completion);
|
init_completion(&drbg->ctr_completion);
|
||||||
|
|
||||||
|
memcpy(outbuf, drbg->outscratchpad, cryptlen);
|
||||||
|
|
||||||
outlen -= cryptlen;
|
outlen -= cryptlen;
|
||||||
}
|
}
|
||||||
|
ret = 0;
|
||||||
|
|
||||||
return 0;
|
out:
|
||||||
|
memzero_explicit(drbg->outscratchpad, DRBG_OUTSCRATCHLEN);
|
||||||
|
return ret;
|
||||||
}
|
}
|
||||||
#endif /* CONFIG_CRYPTO_DRBG_CTR */
|
#endif /* CONFIG_CRYPTO_DRBG_CTR */
|
||||||
|
|
||||||
|
|
|
@ -1727,7 +1727,7 @@ static int eni_do_init(struct atm_dev *dev)
|
||||||
printk("\n");
|
printk("\n");
|
||||||
printk(KERN_ERR DEV_LABEL "(itf %d): can't set up page "
|
printk(KERN_ERR DEV_LABEL "(itf %d): can't set up page "
|
||||||
"mapping\n",dev->number);
|
"mapping\n",dev->number);
|
||||||
return error;
|
return -ENOMEM;
|
||||||
}
|
}
|
||||||
eni_dev->ioaddr = base;
|
eni_dev->ioaddr = base;
|
||||||
eni_dev->base_diff = real_base - (unsigned long) base;
|
eni_dev->base_diff = real_base - (unsigned long) base;
|
||||||
|
|
|
@ -2143,6 +2143,7 @@ static int lanai_dev_open(struct atm_dev *atmdev)
|
||||||
lanai->base = (bus_addr_t) ioremap(raw_base, LANAI_MAPPING_SIZE);
|
lanai->base = (bus_addr_t) ioremap(raw_base, LANAI_MAPPING_SIZE);
|
||||||
if (lanai->base == NULL) {
|
if (lanai->base == NULL) {
|
||||||
printk(KERN_ERR DEV_LABEL ": couldn't remap I/O space\n");
|
printk(KERN_ERR DEV_LABEL ": couldn't remap I/O space\n");
|
||||||
|
result = -ENOMEM;
|
||||||
goto error_pci;
|
goto error_pci;
|
||||||
}
|
}
|
||||||
/* 3.3: Reset lanai and PHY */
|
/* 3.3: Reset lanai and PHY */
|
||||||
|
|
|
@ -422,7 +422,7 @@ static inline void get_aes_decrypt_key(unsigned char *dec_key,
|
||||||
{
|
{
|
||||||
u32 temp;
|
u32 temp;
|
||||||
u32 w_ring[MAX_NK];
|
u32 w_ring[MAX_NK];
|
||||||
int i, j, k = 0;
|
int i, j, k;
|
||||||
u8 nr, nk;
|
u8 nr, nk;
|
||||||
|
|
||||||
switch (keylength) {
|
switch (keylength) {
|
||||||
|
@ -460,6 +460,7 @@ static inline void get_aes_decrypt_key(unsigned char *dec_key,
|
||||||
temp = w_ring[i % nk];
|
temp = w_ring[i % nk];
|
||||||
i++;
|
i++;
|
||||||
}
|
}
|
||||||
|
i--;
|
||||||
for (k = 0, j = i % nk; k < nk; k++) {
|
for (k = 0, j = i % nk; k < nk; k++) {
|
||||||
*((u32 *)dec_key + k) = htonl(w_ring[j]);
|
*((u32 *)dec_key + k) = htonl(w_ring[j]);
|
||||||
j--;
|
j--;
|
||||||
|
|
|
@ -485,7 +485,6 @@ static int amdgpu_atpx_power_state(enum vga_switcheroo_client_id id,
|
||||||
*/
|
*/
|
||||||
static bool amdgpu_atpx_pci_probe_handle(struct pci_dev *pdev)
|
static bool amdgpu_atpx_pci_probe_handle(struct pci_dev *pdev)
|
||||||
{
|
{
|
||||||
struct pci_dev *parent_pdev = pci_upstream_bridge(pdev);
|
|
||||||
acpi_handle dhandle, atpx_handle;
|
acpi_handle dhandle, atpx_handle;
|
||||||
acpi_status status;
|
acpi_status status;
|
||||||
|
|
||||||
|
@ -500,7 +499,6 @@ static bool amdgpu_atpx_pci_probe_handle(struct pci_dev *pdev)
|
||||||
}
|
}
|
||||||
amdgpu_atpx_priv.dhandle = dhandle;
|
amdgpu_atpx_priv.dhandle = dhandle;
|
||||||
amdgpu_atpx_priv.atpx.handle = atpx_handle;
|
amdgpu_atpx_priv.atpx.handle = atpx_handle;
|
||||||
amdgpu_atpx_priv.bridge_pm_usable = parent_pdev && parent_pdev->bridge_d3;
|
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -562,17 +560,25 @@ static bool amdgpu_atpx_detect(void)
|
||||||
struct pci_dev *pdev = NULL;
|
struct pci_dev *pdev = NULL;
|
||||||
bool has_atpx = false;
|
bool has_atpx = false;
|
||||||
int vga_count = 0;
|
int vga_count = 0;
|
||||||
|
bool d3_supported = false;
|
||||||
|
struct pci_dev *parent_pdev;
|
||||||
|
|
||||||
while ((pdev = pci_get_class(PCI_CLASS_DISPLAY_VGA << 8, pdev)) != NULL) {
|
while ((pdev = pci_get_class(PCI_CLASS_DISPLAY_VGA << 8, pdev)) != NULL) {
|
||||||
vga_count++;
|
vga_count++;
|
||||||
|
|
||||||
has_atpx |= (amdgpu_atpx_pci_probe_handle(pdev) == true);
|
has_atpx |= (amdgpu_atpx_pci_probe_handle(pdev) == true);
|
||||||
|
|
||||||
|
parent_pdev = pci_upstream_bridge(pdev);
|
||||||
|
d3_supported |= parent_pdev && parent_pdev->bridge_d3;
|
||||||
}
|
}
|
||||||
|
|
||||||
while ((pdev = pci_get_class(PCI_CLASS_DISPLAY_OTHER << 8, pdev)) != NULL) {
|
while ((pdev = pci_get_class(PCI_CLASS_DISPLAY_OTHER << 8, pdev)) != NULL) {
|
||||||
vga_count++;
|
vga_count++;
|
||||||
|
|
||||||
has_atpx |= (amdgpu_atpx_pci_probe_handle(pdev) == true);
|
has_atpx |= (amdgpu_atpx_pci_probe_handle(pdev) == true);
|
||||||
|
|
||||||
|
parent_pdev = pci_upstream_bridge(pdev);
|
||||||
|
d3_supported |= parent_pdev && parent_pdev->bridge_d3;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (has_atpx && vga_count == 2) {
|
if (has_atpx && vga_count == 2) {
|
||||||
|
@ -580,6 +586,7 @@ static bool amdgpu_atpx_detect(void)
|
||||||
printk(KERN_INFO "vga_switcheroo: detected switching method %s handle\n",
|
printk(KERN_INFO "vga_switcheroo: detected switching method %s handle\n",
|
||||||
acpi_method_name);
|
acpi_method_name);
|
||||||
amdgpu_atpx_priv.atpx_detected = true;
|
amdgpu_atpx_priv.atpx_detected = true;
|
||||||
|
amdgpu_atpx_priv.bridge_pm_usable = d3_supported;
|
||||||
amdgpu_atpx_init();
|
amdgpu_atpx_init();
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
|
@ -2214,6 +2214,7 @@ uint32_t polaris10_get_mac_definition(uint32_t value)
|
||||||
int polaris10_process_firmware_header(struct pp_hwmgr *hwmgr)
|
int polaris10_process_firmware_header(struct pp_hwmgr *hwmgr)
|
||||||
{
|
{
|
||||||
struct polaris10_smumgr *smu_data = (struct polaris10_smumgr *)(hwmgr->smumgr->backend);
|
struct polaris10_smumgr *smu_data = (struct polaris10_smumgr *)(hwmgr->smumgr->backend);
|
||||||
|
struct smu7_hwmgr *data = (struct smu7_hwmgr *)(hwmgr->backend);
|
||||||
uint32_t tmp;
|
uint32_t tmp;
|
||||||
int result;
|
int result;
|
||||||
bool error = false;
|
bool error = false;
|
||||||
|
@ -2233,8 +2234,10 @@ int polaris10_process_firmware_header(struct pp_hwmgr *hwmgr)
|
||||||
offsetof(SMU74_Firmware_Header, SoftRegisters),
|
offsetof(SMU74_Firmware_Header, SoftRegisters),
|
||||||
&tmp, SMC_RAM_END);
|
&tmp, SMC_RAM_END);
|
||||||
|
|
||||||
if (!result)
|
if (!result) {
|
||||||
|
data->soft_regs_start = tmp;
|
||||||
smu_data->smu7_data.soft_regs_start = tmp;
|
smu_data->smu7_data.soft_regs_start = tmp;
|
||||||
|
}
|
||||||
|
|
||||||
error |= (0 != result);
|
error |= (0 != result);
|
||||||
|
|
||||||
|
|
|
@ -375,7 +375,6 @@ static int hdlcd_drm_bind(struct device *dev)
|
||||||
|
|
||||||
err_fbdev:
|
err_fbdev:
|
||||||
drm_kms_helper_poll_fini(drm);
|
drm_kms_helper_poll_fini(drm);
|
||||||
drm_mode_config_cleanup(drm);
|
|
||||||
drm_vblank_cleanup(drm);
|
drm_vblank_cleanup(drm);
|
||||||
err_vblank:
|
err_vblank:
|
||||||
pm_runtime_disable(drm->dev);
|
pm_runtime_disable(drm->dev);
|
||||||
|
@ -387,6 +386,7 @@ static int hdlcd_drm_bind(struct device *dev)
|
||||||
drm_irq_uninstall(drm);
|
drm_irq_uninstall(drm);
|
||||||
of_reserved_mem_device_release(drm->dev);
|
of_reserved_mem_device_release(drm->dev);
|
||||||
err_free:
|
err_free:
|
||||||
|
drm_mode_config_cleanup(drm);
|
||||||
dev_set_drvdata(dev, NULL);
|
dev_set_drvdata(dev, NULL);
|
||||||
drm_dev_unref(drm);
|
drm_dev_unref(drm);
|
||||||
|
|
||||||
|
|
|
@ -254,10 +254,12 @@ static int drm_getcap(struct drm_device *dev, void *data, struct drm_file *file_
|
||||||
req->value = dev->mode_config.async_page_flip;
|
req->value = dev->mode_config.async_page_flip;
|
||||||
break;
|
break;
|
||||||
case DRM_CAP_PAGE_FLIP_TARGET:
|
case DRM_CAP_PAGE_FLIP_TARGET:
|
||||||
req->value = 1;
|
if (drm_core_check_feature(dev, DRIVER_MODESET)) {
|
||||||
drm_for_each_crtc(crtc, dev) {
|
req->value = 1;
|
||||||
if (!crtc->funcs->page_flip_target)
|
drm_for_each_crtc(crtc, dev) {
|
||||||
req->value = 0;
|
if (!crtc->funcs->page_flip_target)
|
||||||
|
req->value = 0;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
case DRM_CAP_CURSOR_WIDTH:
|
case DRM_CAP_CURSOR_WIDTH:
|
||||||
|
|
|
@ -2268,7 +2268,7 @@ i915_gem_object_get_pages_gtt(struct drm_i915_gem_object *obj)
|
||||||
page = shmem_read_mapping_page(mapping, i);
|
page = shmem_read_mapping_page(mapping, i);
|
||||||
if (IS_ERR(page)) {
|
if (IS_ERR(page)) {
|
||||||
ret = PTR_ERR(page);
|
ret = PTR_ERR(page);
|
||||||
goto err_pages;
|
goto err_sg;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#ifdef CONFIG_SWIOTLB
|
#ifdef CONFIG_SWIOTLB
|
||||||
|
@ -2311,8 +2311,9 @@ i915_gem_object_get_pages_gtt(struct drm_i915_gem_object *obj)
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
err_pages:
|
err_sg:
|
||||||
sg_mark_end(sg);
|
sg_mark_end(sg);
|
||||||
|
err_pages:
|
||||||
for_each_sgt_page(page, sgt_iter, st)
|
for_each_sgt_page(page, sgt_iter, st)
|
||||||
put_page(page);
|
put_page(page);
|
||||||
sg_free_table(st);
|
sg_free_table(st);
|
||||||
|
|
|
@ -12260,7 +12260,7 @@ static int intel_crtc_page_flip(struct drm_crtc *crtc,
|
||||||
intel_crtc->reset_count = i915_reset_count(&dev_priv->gpu_error);
|
intel_crtc->reset_count = i915_reset_count(&dev_priv->gpu_error);
|
||||||
if (i915_reset_in_progress_or_wedged(&dev_priv->gpu_error)) {
|
if (i915_reset_in_progress_or_wedged(&dev_priv->gpu_error)) {
|
||||||
ret = -EIO;
|
ret = -EIO;
|
||||||
goto cleanup;
|
goto unlock;
|
||||||
}
|
}
|
||||||
|
|
||||||
atomic_inc(&intel_crtc->unpin_work_count);
|
atomic_inc(&intel_crtc->unpin_work_count);
|
||||||
|
@ -12352,6 +12352,7 @@ static int intel_crtc_page_flip(struct drm_crtc *crtc,
|
||||||
intel_unpin_fb_obj(fb, crtc->primary->state->rotation);
|
intel_unpin_fb_obj(fb, crtc->primary->state->rotation);
|
||||||
cleanup_pending:
|
cleanup_pending:
|
||||||
atomic_dec(&intel_crtc->unpin_work_count);
|
atomic_dec(&intel_crtc->unpin_work_count);
|
||||||
|
unlock:
|
||||||
mutex_unlock(&dev->struct_mutex);
|
mutex_unlock(&dev->struct_mutex);
|
||||||
cleanup:
|
cleanup:
|
||||||
crtc->primary->fb = old_fb;
|
crtc->primary->fb = old_fb;
|
||||||
|
|
|
@ -479,7 +479,6 @@ static int radeon_atpx_power_state(enum vga_switcheroo_client_id id,
|
||||||
*/
|
*/
|
||||||
static bool radeon_atpx_pci_probe_handle(struct pci_dev *pdev)
|
static bool radeon_atpx_pci_probe_handle(struct pci_dev *pdev)
|
||||||
{
|
{
|
||||||
struct pci_dev *parent_pdev = pci_upstream_bridge(pdev);
|
|
||||||
acpi_handle dhandle, atpx_handle;
|
acpi_handle dhandle, atpx_handle;
|
||||||
acpi_status status;
|
acpi_status status;
|
||||||
|
|
||||||
|
@ -493,7 +492,6 @@ static bool radeon_atpx_pci_probe_handle(struct pci_dev *pdev)
|
||||||
|
|
||||||
radeon_atpx_priv.dhandle = dhandle;
|
radeon_atpx_priv.dhandle = dhandle;
|
||||||
radeon_atpx_priv.atpx.handle = atpx_handle;
|
radeon_atpx_priv.atpx.handle = atpx_handle;
|
||||||
radeon_atpx_priv.bridge_pm_usable = parent_pdev && parent_pdev->bridge_d3;
|
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -555,11 +553,16 @@ static bool radeon_atpx_detect(void)
|
||||||
struct pci_dev *pdev = NULL;
|
struct pci_dev *pdev = NULL;
|
||||||
bool has_atpx = false;
|
bool has_atpx = false;
|
||||||
int vga_count = 0;
|
int vga_count = 0;
|
||||||
|
bool d3_supported = false;
|
||||||
|
struct pci_dev *parent_pdev;
|
||||||
|
|
||||||
while ((pdev = pci_get_class(PCI_CLASS_DISPLAY_VGA << 8, pdev)) != NULL) {
|
while ((pdev = pci_get_class(PCI_CLASS_DISPLAY_VGA << 8, pdev)) != NULL) {
|
||||||
vga_count++;
|
vga_count++;
|
||||||
|
|
||||||
has_atpx |= (radeon_atpx_pci_probe_handle(pdev) == true);
|
has_atpx |= (radeon_atpx_pci_probe_handle(pdev) == true);
|
||||||
|
|
||||||
|
parent_pdev = pci_upstream_bridge(pdev);
|
||||||
|
d3_supported |= parent_pdev && parent_pdev->bridge_d3;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* some newer PX laptops mark the dGPU as a non-VGA display device */
|
/* some newer PX laptops mark the dGPU as a non-VGA display device */
|
||||||
|
@ -567,6 +570,9 @@ static bool radeon_atpx_detect(void)
|
||||||
vga_count++;
|
vga_count++;
|
||||||
|
|
||||||
has_atpx |= (radeon_atpx_pci_probe_handle(pdev) == true);
|
has_atpx |= (radeon_atpx_pci_probe_handle(pdev) == true);
|
||||||
|
|
||||||
|
parent_pdev = pci_upstream_bridge(pdev);
|
||||||
|
d3_supported |= parent_pdev && parent_pdev->bridge_d3;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (has_atpx && vga_count == 2) {
|
if (has_atpx && vga_count == 2) {
|
||||||
|
@ -574,6 +580,7 @@ static bool radeon_atpx_detect(void)
|
||||||
printk(KERN_INFO "vga_switcheroo: detected switching method %s handle\n",
|
printk(KERN_INFO "vga_switcheroo: detected switching method %s handle\n",
|
||||||
acpi_method_name);
|
acpi_method_name);
|
||||||
radeon_atpx_priv.atpx_detected = true;
|
radeon_atpx_priv.atpx_detected = true;
|
||||||
|
radeon_atpx_priv.bridge_pm_usable = d3_supported;
|
||||||
radeon_atpx_init();
|
radeon_atpx_init();
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
|
@ -1499,6 +1499,7 @@ hfc4s8s_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
|
||||||
printk(KERN_INFO
|
printk(KERN_INFO
|
||||||
"HFC-4S/8S: failed to request address space at 0x%04x\n",
|
"HFC-4S/8S: failed to request address space at 0x%04x\n",
|
||||||
hw->iobase);
|
hw->iobase);
|
||||||
|
err = -EBUSY;
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1872,8 +1872,16 @@ static void bnx2x_get_ringparam(struct net_device *dev,
|
||||||
|
|
||||||
ering->rx_max_pending = MAX_RX_AVAIL;
|
ering->rx_max_pending = MAX_RX_AVAIL;
|
||||||
|
|
||||||
|
/* If size isn't already set, we give an estimation of the number
|
||||||
|
* of buffers we'll have. We're neglecting some possible conditions
|
||||||
|
* [we couldn't know for certain at this point if number of queues
|
||||||
|
* might shrink] but the number would be correct for the likely
|
||||||
|
* scenario.
|
||||||
|
*/
|
||||||
if (bp->rx_ring_size)
|
if (bp->rx_ring_size)
|
||||||
ering->rx_pending = bp->rx_ring_size;
|
ering->rx_pending = bp->rx_ring_size;
|
||||||
|
else if (BNX2X_NUM_RX_QUEUES(bp))
|
||||||
|
ering->rx_pending = MAX_RX_AVAIL / BNX2X_NUM_RX_QUEUES(bp);
|
||||||
else
|
else
|
||||||
ering->rx_pending = MAX_RX_AVAIL;
|
ering->rx_pending = MAX_RX_AVAIL;
|
||||||
|
|
||||||
|
|
|
@ -10138,7 +10138,7 @@ static void __bnx2x_add_udp_port(struct bnx2x *bp, u16 port,
|
||||||
{
|
{
|
||||||
struct bnx2x_udp_tunnel *udp_port = &bp->udp_tunnel_ports[type];
|
struct bnx2x_udp_tunnel *udp_port = &bp->udp_tunnel_ports[type];
|
||||||
|
|
||||||
if (!netif_running(bp->dev) || !IS_PF(bp))
|
if (!netif_running(bp->dev) || !IS_PF(bp) || CHIP_IS_E1x(bp))
|
||||||
return;
|
return;
|
||||||
|
|
||||||
if (udp_port->count && udp_port->dst_port == port) {
|
if (udp_port->count && udp_port->dst_port == port) {
|
||||||
|
@ -10163,7 +10163,7 @@ static void __bnx2x_del_udp_port(struct bnx2x *bp, u16 port,
|
||||||
{
|
{
|
||||||
struct bnx2x_udp_tunnel *udp_port = &bp->udp_tunnel_ports[type];
|
struct bnx2x_udp_tunnel *udp_port = &bp->udp_tunnel_ports[type];
|
||||||
|
|
||||||
if (!IS_PF(bp))
|
if (!IS_PF(bp) || CHIP_IS_E1x(bp))
|
||||||
return;
|
return;
|
||||||
|
|
||||||
if (!udp_port->count || udp_port->dst_port != port) {
|
if (!udp_port->count || udp_port->dst_port != port) {
|
||||||
|
@ -13508,6 +13508,7 @@ static int bnx2x_init_firmware(struct bnx2x *bp)
|
||||||
|
|
||||||
/* Initialize the pointers to the init arrays */
|
/* Initialize the pointers to the init arrays */
|
||||||
/* Blob */
|
/* Blob */
|
||||||
|
rc = -ENOMEM;
|
||||||
BNX2X_ALLOC_AND_SET(init_data, request_firmware_exit, be32_to_cpu_n);
|
BNX2X_ALLOC_AND_SET(init_data, request_firmware_exit, be32_to_cpu_n);
|
||||||
|
|
||||||
/* Opcodes */
|
/* Opcodes */
|
||||||
|
|
|
@ -4116,7 +4116,7 @@ static int bnxt_hwrm_stat_ctx_alloc(struct bnxt *bp)
|
||||||
bp->grp_info[i].fw_stats_ctx = cpr->hw_stats_ctx_id;
|
bp->grp_info[i].fw_stats_ctx = cpr->hw_stats_ctx_id;
|
||||||
}
|
}
|
||||||
mutex_unlock(&bp->hwrm_cmd_lock);
|
mutex_unlock(&bp->hwrm_cmd_lock);
|
||||||
return 0;
|
return rc;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int bnxt_hwrm_func_qcfg(struct bnxt *bp)
|
static int bnxt_hwrm_func_qcfg(struct bnxt *bp)
|
||||||
|
|
|
@ -468,6 +468,9 @@ static void ep93xx_free_buffers(struct ep93xx_priv *ep)
|
||||||
struct device *dev = ep->dev->dev.parent;
|
struct device *dev = ep->dev->dev.parent;
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
|
if (!ep->descs)
|
||||||
|
return;
|
||||||
|
|
||||||
for (i = 0; i < RX_QUEUE_ENTRIES; i++) {
|
for (i = 0; i < RX_QUEUE_ENTRIES; i++) {
|
||||||
dma_addr_t d;
|
dma_addr_t d;
|
||||||
|
|
||||||
|
@ -490,6 +493,7 @@ static void ep93xx_free_buffers(struct ep93xx_priv *ep)
|
||||||
|
|
||||||
dma_free_coherent(dev, sizeof(struct ep93xx_descs), ep->descs,
|
dma_free_coherent(dev, sizeof(struct ep93xx_descs), ep->descs,
|
||||||
ep->descs_dma_addr);
|
ep->descs_dma_addr);
|
||||||
|
ep->descs = NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int ep93xx_alloc_buffers(struct ep93xx_priv *ep)
|
static int ep93xx_alloc_buffers(struct ep93xx_priv *ep)
|
||||||
|
|
|
@ -90,7 +90,8 @@ static struct be_cmd_priv_map cmd_priv_map[] = {
|
||||||
{
|
{
|
||||||
OPCODE_COMMON_SET_HSW_CONFIG,
|
OPCODE_COMMON_SET_HSW_CONFIG,
|
||||||
CMD_SUBSYSTEM_COMMON,
|
CMD_SUBSYSTEM_COMMON,
|
||||||
BE_PRIV_DEVCFG | BE_PRIV_VHADM
|
BE_PRIV_DEVCFG | BE_PRIV_VHADM |
|
||||||
|
BE_PRIV_DEVSEC
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
OPCODE_COMMON_GET_EXT_FAT_CAPABILITIES,
|
OPCODE_COMMON_GET_EXT_FAT_CAPABILITIES,
|
||||||
|
|
|
@ -2310,6 +2310,8 @@ static const struct fec_stat {
|
||||||
{ "IEEE_rx_octets_ok", IEEE_R_OCTETS_OK },
|
{ "IEEE_rx_octets_ok", IEEE_R_OCTETS_OK },
|
||||||
};
|
};
|
||||||
|
|
||||||
|
#define FEC_STATS_SIZE (ARRAY_SIZE(fec_stats) * sizeof(u64))
|
||||||
|
|
||||||
static void fec_enet_update_ethtool_stats(struct net_device *dev)
|
static void fec_enet_update_ethtool_stats(struct net_device *dev)
|
||||||
{
|
{
|
||||||
struct fec_enet_private *fep = netdev_priv(dev);
|
struct fec_enet_private *fep = netdev_priv(dev);
|
||||||
|
@ -2327,7 +2329,7 @@ static void fec_enet_get_ethtool_stats(struct net_device *dev,
|
||||||
if (netif_running(dev))
|
if (netif_running(dev))
|
||||||
fec_enet_update_ethtool_stats(dev);
|
fec_enet_update_ethtool_stats(dev);
|
||||||
|
|
||||||
memcpy(data, fep->ethtool_stats, ARRAY_SIZE(fec_stats) * sizeof(u64));
|
memcpy(data, fep->ethtool_stats, FEC_STATS_SIZE);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void fec_enet_get_strings(struct net_device *netdev,
|
static void fec_enet_get_strings(struct net_device *netdev,
|
||||||
|
@ -2352,6 +2354,12 @@ static int fec_enet_get_sset_count(struct net_device *dev, int sset)
|
||||||
return -EOPNOTSUPP;
|
return -EOPNOTSUPP;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#else /* !defined(CONFIG_M5272) */
|
||||||
|
#define FEC_STATS_SIZE 0
|
||||||
|
static inline void fec_enet_update_ethtool_stats(struct net_device *dev)
|
||||||
|
{
|
||||||
|
}
|
||||||
#endif /* !defined(CONFIG_M5272) */
|
#endif /* !defined(CONFIG_M5272) */
|
||||||
|
|
||||||
/* ITR clock source is enet system clock (clk_ahb).
|
/* ITR clock source is enet system clock (clk_ahb).
|
||||||
|
@ -3279,8 +3287,7 @@ fec_probe(struct platform_device *pdev)
|
||||||
|
|
||||||
/* Init network device */
|
/* Init network device */
|
||||||
ndev = alloc_etherdev_mqs(sizeof(struct fec_enet_private) +
|
ndev = alloc_etherdev_mqs(sizeof(struct fec_enet_private) +
|
||||||
ARRAY_SIZE(fec_stats) * sizeof(u64),
|
FEC_STATS_SIZE, num_tx_qs, num_rx_qs);
|
||||||
num_tx_qs, num_rx_qs);
|
|
||||||
if (!ndev)
|
if (!ndev)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
|
|
@ -260,11 +260,6 @@ static void dump_buf(void *buf, int size, int data_only, int offset)
|
||||||
pr_debug("\n");
|
pr_debug("\n");
|
||||||
}
|
}
|
||||||
|
|
||||||
enum {
|
|
||||||
MLX5_DRIVER_STATUS_ABORTED = 0xfe,
|
|
||||||
MLX5_DRIVER_SYND = 0xbadd00de,
|
|
||||||
};
|
|
||||||
|
|
||||||
static int mlx5_internal_err_ret_value(struct mlx5_core_dev *dev, u16 op,
|
static int mlx5_internal_err_ret_value(struct mlx5_core_dev *dev, u16 op,
|
||||||
u32 *synd, u8 *status)
|
u32 *synd, u8 *status)
|
||||||
{
|
{
|
||||||
|
|
|
@ -262,7 +262,7 @@ struct mlx5e_tstamp {
|
||||||
};
|
};
|
||||||
|
|
||||||
enum {
|
enum {
|
||||||
MLX5E_RQ_STATE_FLUSH,
|
MLX5E_RQ_STATE_ENABLED,
|
||||||
MLX5E_RQ_STATE_UMR_WQE_IN_PROGRESS,
|
MLX5E_RQ_STATE_UMR_WQE_IN_PROGRESS,
|
||||||
MLX5E_RQ_STATE_AM,
|
MLX5E_RQ_STATE_AM,
|
||||||
};
|
};
|
||||||
|
@ -415,7 +415,7 @@ struct mlx5e_sq_dma {
|
||||||
};
|
};
|
||||||
|
|
||||||
enum {
|
enum {
|
||||||
MLX5E_SQ_STATE_FLUSH,
|
MLX5E_SQ_STATE_ENABLED,
|
||||||
MLX5E_SQ_STATE_BF_ENABLE,
|
MLX5E_SQ_STATE_BF_ENABLE,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -831,6 +831,7 @@ static int mlx5e_open_rq(struct mlx5e_channel *c,
|
||||||
if (err)
|
if (err)
|
||||||
goto err_destroy_rq;
|
goto err_destroy_rq;
|
||||||
|
|
||||||
|
set_bit(MLX5E_RQ_STATE_ENABLED, &rq->state);
|
||||||
err = mlx5e_modify_rq_state(rq, MLX5_RQC_STATE_RST, MLX5_RQC_STATE_RDY);
|
err = mlx5e_modify_rq_state(rq, MLX5_RQC_STATE_RST, MLX5_RQC_STATE_RDY);
|
||||||
if (err)
|
if (err)
|
||||||
goto err_disable_rq;
|
goto err_disable_rq;
|
||||||
|
@ -845,6 +846,7 @@ static int mlx5e_open_rq(struct mlx5e_channel *c,
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
err_disable_rq:
|
err_disable_rq:
|
||||||
|
clear_bit(MLX5E_RQ_STATE_ENABLED, &rq->state);
|
||||||
mlx5e_disable_rq(rq);
|
mlx5e_disable_rq(rq);
|
||||||
err_destroy_rq:
|
err_destroy_rq:
|
||||||
mlx5e_destroy_rq(rq);
|
mlx5e_destroy_rq(rq);
|
||||||
|
@ -854,7 +856,7 @@ static int mlx5e_open_rq(struct mlx5e_channel *c,
|
||||||
|
|
||||||
static void mlx5e_close_rq(struct mlx5e_rq *rq)
|
static void mlx5e_close_rq(struct mlx5e_rq *rq)
|
||||||
{
|
{
|
||||||
set_bit(MLX5E_RQ_STATE_FLUSH, &rq->state);
|
clear_bit(MLX5E_RQ_STATE_ENABLED, &rq->state);
|
||||||
napi_synchronize(&rq->channel->napi); /* prevent mlx5e_post_rx_wqes */
|
napi_synchronize(&rq->channel->napi); /* prevent mlx5e_post_rx_wqes */
|
||||||
cancel_work_sync(&rq->am.work);
|
cancel_work_sync(&rq->am.work);
|
||||||
|
|
||||||
|
@ -1078,7 +1080,6 @@ static int mlx5e_enable_sq(struct mlx5e_sq *sq, struct mlx5e_sq_param *param)
|
||||||
MLX5_SET(sqc, sqc, min_wqe_inline_mode, sq->min_inline_mode);
|
MLX5_SET(sqc, sqc, min_wqe_inline_mode, sq->min_inline_mode);
|
||||||
MLX5_SET(sqc, sqc, state, MLX5_SQC_STATE_RST);
|
MLX5_SET(sqc, sqc, state, MLX5_SQC_STATE_RST);
|
||||||
MLX5_SET(sqc, sqc, tis_lst_sz, param->type == MLX5E_SQ_ICO ? 0 : 1);
|
MLX5_SET(sqc, sqc, tis_lst_sz, param->type == MLX5E_SQ_ICO ? 0 : 1);
|
||||||
MLX5_SET(sqc, sqc, flush_in_error_en, 1);
|
|
||||||
|
|
||||||
MLX5_SET(wq, wq, wq_type, MLX5_WQ_TYPE_CYCLIC);
|
MLX5_SET(wq, wq, wq_type, MLX5_WQ_TYPE_CYCLIC);
|
||||||
MLX5_SET(wq, wq, uar_page, sq->uar.index);
|
MLX5_SET(wq, wq, uar_page, sq->uar.index);
|
||||||
|
@ -1155,6 +1156,7 @@ static int mlx5e_open_sq(struct mlx5e_channel *c,
|
||||||
if (err)
|
if (err)
|
||||||
goto err_destroy_sq;
|
goto err_destroy_sq;
|
||||||
|
|
||||||
|
set_bit(MLX5E_SQ_STATE_ENABLED, &sq->state);
|
||||||
err = mlx5e_modify_sq(sq, MLX5_SQC_STATE_RST, MLX5_SQC_STATE_RDY,
|
err = mlx5e_modify_sq(sq, MLX5_SQC_STATE_RST, MLX5_SQC_STATE_RDY,
|
||||||
false, 0);
|
false, 0);
|
||||||
if (err)
|
if (err)
|
||||||
|
@ -1168,6 +1170,7 @@ static int mlx5e_open_sq(struct mlx5e_channel *c,
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
err_disable_sq:
|
err_disable_sq:
|
||||||
|
clear_bit(MLX5E_SQ_STATE_ENABLED, &sq->state);
|
||||||
mlx5e_disable_sq(sq);
|
mlx5e_disable_sq(sq);
|
||||||
err_destroy_sq:
|
err_destroy_sq:
|
||||||
mlx5e_destroy_sq(sq);
|
mlx5e_destroy_sq(sq);
|
||||||
|
@ -1184,7 +1187,7 @@ static inline void netif_tx_disable_queue(struct netdev_queue *txq)
|
||||||
|
|
||||||
static void mlx5e_close_sq(struct mlx5e_sq *sq)
|
static void mlx5e_close_sq(struct mlx5e_sq *sq)
|
||||||
{
|
{
|
||||||
set_bit(MLX5E_SQ_STATE_FLUSH, &sq->state);
|
clear_bit(MLX5E_SQ_STATE_ENABLED, &sq->state);
|
||||||
/* prevent netif_tx_wake_queue */
|
/* prevent netif_tx_wake_queue */
|
||||||
napi_synchronize(&sq->channel->napi);
|
napi_synchronize(&sq->channel->napi);
|
||||||
|
|
||||||
|
@ -3163,7 +3166,7 @@ static void mlx5e_tx_timeout(struct net_device *dev)
|
||||||
if (!netif_xmit_stopped(netdev_get_tx_queue(dev, i)))
|
if (!netif_xmit_stopped(netdev_get_tx_queue(dev, i)))
|
||||||
continue;
|
continue;
|
||||||
sched_work = true;
|
sched_work = true;
|
||||||
set_bit(MLX5E_SQ_STATE_FLUSH, &sq->state);
|
clear_bit(MLX5E_SQ_STATE_ENABLED, &sq->state);
|
||||||
netdev_err(dev, "TX timeout on queue: %d, SQ: 0x%x, CQ: 0x%x, SQ Cons: 0x%x SQ Prod: 0x%x\n",
|
netdev_err(dev, "TX timeout on queue: %d, SQ: 0x%x, CQ: 0x%x, SQ Cons: 0x%x SQ Prod: 0x%x\n",
|
||||||
i, sq->sqn, sq->cq.mcq.cqn, sq->cc, sq->pc);
|
i, sq->sqn, sq->cq.mcq.cqn, sq->cc, sq->pc);
|
||||||
}
|
}
|
||||||
|
@ -3227,13 +3230,13 @@ static int mlx5e_xdp_set(struct net_device *netdev, struct bpf_prog *prog)
|
||||||
for (i = 0; i < priv->params.num_channels; i++) {
|
for (i = 0; i < priv->params.num_channels; i++) {
|
||||||
struct mlx5e_channel *c = priv->channel[i];
|
struct mlx5e_channel *c = priv->channel[i];
|
||||||
|
|
||||||
set_bit(MLX5E_RQ_STATE_FLUSH, &c->rq.state);
|
clear_bit(MLX5E_RQ_STATE_ENABLED, &c->rq.state);
|
||||||
napi_synchronize(&c->napi);
|
napi_synchronize(&c->napi);
|
||||||
/* prevent mlx5e_poll_rx_cq from accessing rq->xdp_prog */
|
/* prevent mlx5e_poll_rx_cq from accessing rq->xdp_prog */
|
||||||
|
|
||||||
old_prog = xchg(&c->rq.xdp_prog, prog);
|
old_prog = xchg(&c->rq.xdp_prog, prog);
|
||||||
|
|
||||||
clear_bit(MLX5E_RQ_STATE_FLUSH, &c->rq.state);
|
set_bit(MLX5E_RQ_STATE_ENABLED, &c->rq.state);
|
||||||
/* napi_schedule in case we have missed anything */
|
/* napi_schedule in case we have missed anything */
|
||||||
set_bit(MLX5E_CHANNEL_NAPI_SCHED, &c->flags);
|
set_bit(MLX5E_CHANNEL_NAPI_SCHED, &c->flags);
|
||||||
napi_schedule(&c->napi);
|
napi_schedule(&c->napi);
|
||||||
|
|
|
@ -340,7 +340,7 @@ static inline void mlx5e_post_umr_wqe(struct mlx5e_rq *rq, u16 ix)
|
||||||
while ((pi = (sq->pc & wq->sz_m1)) > sq->edge) {
|
while ((pi = (sq->pc & wq->sz_m1)) > sq->edge) {
|
||||||
sq->db.ico_wqe[pi].opcode = MLX5_OPCODE_NOP;
|
sq->db.ico_wqe[pi].opcode = MLX5_OPCODE_NOP;
|
||||||
sq->db.ico_wqe[pi].num_wqebbs = 1;
|
sq->db.ico_wqe[pi].num_wqebbs = 1;
|
||||||
mlx5e_send_nop(sq, true);
|
mlx5e_send_nop(sq, false);
|
||||||
}
|
}
|
||||||
|
|
||||||
wqe = mlx5_wq_cyc_get_wqe(wq, pi);
|
wqe = mlx5_wq_cyc_get_wqe(wq, pi);
|
||||||
|
@ -412,7 +412,7 @@ void mlx5e_post_rx_mpwqe(struct mlx5e_rq *rq)
|
||||||
|
|
||||||
clear_bit(MLX5E_RQ_STATE_UMR_WQE_IN_PROGRESS, &rq->state);
|
clear_bit(MLX5E_RQ_STATE_UMR_WQE_IN_PROGRESS, &rq->state);
|
||||||
|
|
||||||
if (unlikely(test_bit(MLX5E_RQ_STATE_FLUSH, &rq->state))) {
|
if (unlikely(!test_bit(MLX5E_RQ_STATE_ENABLED, &rq->state))) {
|
||||||
mlx5e_free_rx_mpwqe(rq, &rq->mpwqe.info[wq->head]);
|
mlx5e_free_rx_mpwqe(rq, &rq->mpwqe.info[wq->head]);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
@ -445,7 +445,7 @@ void mlx5e_dealloc_rx_mpwqe(struct mlx5e_rq *rq, u16 ix)
|
||||||
}
|
}
|
||||||
|
|
||||||
#define RQ_CANNOT_POST(rq) \
|
#define RQ_CANNOT_POST(rq) \
|
||||||
(test_bit(MLX5E_RQ_STATE_FLUSH, &rq->state) || \
|
(!test_bit(MLX5E_RQ_STATE_ENABLED, &rq->state) || \
|
||||||
test_bit(MLX5E_RQ_STATE_UMR_WQE_IN_PROGRESS, &rq->state))
|
test_bit(MLX5E_RQ_STATE_UMR_WQE_IN_PROGRESS, &rq->state))
|
||||||
|
|
||||||
bool mlx5e_post_rx_wqes(struct mlx5e_rq *rq)
|
bool mlx5e_post_rx_wqes(struct mlx5e_rq *rq)
|
||||||
|
@ -928,7 +928,7 @@ int mlx5e_poll_rx_cq(struct mlx5e_cq *cq, int budget)
|
||||||
struct mlx5e_sq *xdp_sq = &rq->channel->xdp_sq;
|
struct mlx5e_sq *xdp_sq = &rq->channel->xdp_sq;
|
||||||
int work_done = 0;
|
int work_done = 0;
|
||||||
|
|
||||||
if (unlikely(test_bit(MLX5E_RQ_STATE_FLUSH, &rq->state)))
|
if (unlikely(!test_bit(MLX5E_RQ_STATE_ENABLED, &rq->state)))
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
if (cq->decmprs_left)
|
if (cq->decmprs_left)
|
||||||
|
|
|
@ -409,7 +409,7 @@ bool mlx5e_poll_tx_cq(struct mlx5e_cq *cq, int napi_budget)
|
||||||
|
|
||||||
sq = container_of(cq, struct mlx5e_sq, cq);
|
sq = container_of(cq, struct mlx5e_sq, cq);
|
||||||
|
|
||||||
if (unlikely(test_bit(MLX5E_SQ_STATE_FLUSH, &sq->state)))
|
if (unlikely(!test_bit(MLX5E_SQ_STATE_ENABLED, &sq->state)))
|
||||||
return false;
|
return false;
|
||||||
|
|
||||||
npkts = 0;
|
npkts = 0;
|
||||||
|
|
|
@ -56,7 +56,7 @@ static void mlx5e_poll_ico_cq(struct mlx5e_cq *cq)
|
||||||
struct mlx5_cqe64 *cqe;
|
struct mlx5_cqe64 *cqe;
|
||||||
u16 sqcc;
|
u16 sqcc;
|
||||||
|
|
||||||
if (unlikely(test_bit(MLX5E_SQ_STATE_FLUSH, &sq->state)))
|
if (unlikely(!test_bit(MLX5E_SQ_STATE_ENABLED, &sq->state)))
|
||||||
return;
|
return;
|
||||||
|
|
||||||
cqe = mlx5e_get_cqe(cq);
|
cqe = mlx5e_get_cqe(cq);
|
||||||
|
@ -113,7 +113,7 @@ static inline bool mlx5e_poll_xdp_tx_cq(struct mlx5e_cq *cq)
|
||||||
|
|
||||||
sq = container_of(cq, struct mlx5e_sq, cq);
|
sq = container_of(cq, struct mlx5e_sq, cq);
|
||||||
|
|
||||||
if (unlikely(test_bit(MLX5E_SQ_STATE_FLUSH, &sq->state)))
|
if (unlikely(!test_bit(MLX5E_SQ_STATE_ENABLED, &sq->state)))
|
||||||
return false;
|
return false;
|
||||||
|
|
||||||
/* sq->cc must be updated only after mlx5_cqwq_update_db_record(),
|
/* sq->cc must be updated only after mlx5_cqwq_update_db_record(),
|
||||||
|
|
|
@ -62,13 +62,13 @@ MODULE_DESCRIPTION("Mellanox Connect-IB, ConnectX-4 core driver");
|
||||||
MODULE_LICENSE("Dual BSD/GPL");
|
MODULE_LICENSE("Dual BSD/GPL");
|
||||||
MODULE_VERSION(DRIVER_VERSION);
|
MODULE_VERSION(DRIVER_VERSION);
|
||||||
|
|
||||||
int mlx5_core_debug_mask;
|
unsigned int mlx5_core_debug_mask;
|
||||||
module_param_named(debug_mask, mlx5_core_debug_mask, int, 0644);
|
module_param_named(debug_mask, mlx5_core_debug_mask, uint, 0644);
|
||||||
MODULE_PARM_DESC(debug_mask, "debug mask: 1 = dump cmd data, 2 = dump cmd exec time, 3 = both. Default=0");
|
MODULE_PARM_DESC(debug_mask, "debug mask: 1 = dump cmd data, 2 = dump cmd exec time, 3 = both. Default=0");
|
||||||
|
|
||||||
#define MLX5_DEFAULT_PROF 2
|
#define MLX5_DEFAULT_PROF 2
|
||||||
static int prof_sel = MLX5_DEFAULT_PROF;
|
static unsigned int prof_sel = MLX5_DEFAULT_PROF;
|
||||||
module_param_named(prof_sel, prof_sel, int, 0444);
|
module_param_named(prof_sel, prof_sel, uint, 0444);
|
||||||
MODULE_PARM_DESC(prof_sel, "profile selector. Valid range 0 - 2");
|
MODULE_PARM_DESC(prof_sel, "profile selector. Valid range 0 - 2");
|
||||||
|
|
||||||
enum {
|
enum {
|
||||||
|
@ -767,13 +767,15 @@ static int mlx5_core_set_issi(struct mlx5_core_dev *dev)
|
||||||
u8 status;
|
u8 status;
|
||||||
|
|
||||||
mlx5_cmd_mbox_status(query_out, &status, &syndrome);
|
mlx5_cmd_mbox_status(query_out, &status, &syndrome);
|
||||||
if (status == MLX5_CMD_STAT_BAD_OP_ERR) {
|
if (!status || syndrome == MLX5_DRIVER_SYND) {
|
||||||
pr_debug("Only ISSI 0 is supported\n");
|
mlx5_core_err(dev, "Failed to query ISSI err(%d) status(%d) synd(%d)\n",
|
||||||
return 0;
|
err, status, syndrome);
|
||||||
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
pr_err("failed to query ISSI err(%d)\n", err);
|
mlx5_core_warn(dev, "Query ISSI is not supported by FW, ISSI is 0\n");
|
||||||
return err;
|
dev->issi = 0;
|
||||||
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
sup_issi = MLX5_GET(query_issi_out, query_out, supported_issi_dw0);
|
sup_issi = MLX5_GET(query_issi_out, query_out, supported_issi_dw0);
|
||||||
|
@ -787,7 +789,8 @@ static int mlx5_core_set_issi(struct mlx5_core_dev *dev)
|
||||||
err = mlx5_cmd_exec(dev, set_in, sizeof(set_in),
|
err = mlx5_cmd_exec(dev, set_in, sizeof(set_in),
|
||||||
set_out, sizeof(set_out));
|
set_out, sizeof(set_out));
|
||||||
if (err) {
|
if (err) {
|
||||||
pr_err("failed to set ISSI=1 err(%d)\n", err);
|
mlx5_core_err(dev, "Failed to set ISSI to 1 err(%d)\n",
|
||||||
|
err);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1266,13 +1269,6 @@ static int init_one(struct pci_dev *pdev,
|
||||||
|
|
||||||
dev->pdev = pdev;
|
dev->pdev = pdev;
|
||||||
dev->event = mlx5_core_event;
|
dev->event = mlx5_core_event;
|
||||||
|
|
||||||
if (prof_sel < 0 || prof_sel >= ARRAY_SIZE(profile)) {
|
|
||||||
mlx5_core_warn(dev,
|
|
||||||
"selected profile out of range, selecting default (%d)\n",
|
|
||||||
MLX5_DEFAULT_PROF);
|
|
||||||
prof_sel = MLX5_DEFAULT_PROF;
|
|
||||||
}
|
|
||||||
dev->profile = &profile[prof_sel];
|
dev->profile = &profile[prof_sel];
|
||||||
|
|
||||||
INIT_LIST_HEAD(&priv->ctx_list);
|
INIT_LIST_HEAD(&priv->ctx_list);
|
||||||
|
@ -1490,10 +1486,22 @@ static struct pci_driver mlx5_core_driver = {
|
||||||
.sriov_configure = mlx5_core_sriov_configure,
|
.sriov_configure = mlx5_core_sriov_configure,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
static void mlx5_core_verify_params(void)
|
||||||
|
{
|
||||||
|
if (prof_sel >= ARRAY_SIZE(profile)) {
|
||||||
|
pr_warn("mlx5_core: WARNING: Invalid module parameter prof_sel %d, valid range 0-%zu, changing back to default(%d)\n",
|
||||||
|
prof_sel,
|
||||||
|
ARRAY_SIZE(profile) - 1,
|
||||||
|
MLX5_DEFAULT_PROF);
|
||||||
|
prof_sel = MLX5_DEFAULT_PROF;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
static int __init init(void)
|
static int __init init(void)
|
||||||
{
|
{
|
||||||
int err;
|
int err;
|
||||||
|
|
||||||
|
mlx5_core_verify_params();
|
||||||
mlx5_register_debugfs();
|
mlx5_register_debugfs();
|
||||||
|
|
||||||
err = pci_register_driver(&mlx5_core_driver);
|
err = pci_register_driver(&mlx5_core_driver);
|
||||||
|
|
|
@ -44,11 +44,11 @@
|
||||||
|
|
||||||
#define MLX5_TOTAL_VPORTS(mdev) (1 + pci_sriov_get_totalvfs(mdev->pdev))
|
#define MLX5_TOTAL_VPORTS(mdev) (1 + pci_sriov_get_totalvfs(mdev->pdev))
|
||||||
|
|
||||||
extern int mlx5_core_debug_mask;
|
extern uint mlx5_core_debug_mask;
|
||||||
|
|
||||||
#define mlx5_core_dbg(__dev, format, ...) \
|
#define mlx5_core_dbg(__dev, format, ...) \
|
||||||
dev_dbg(&(__dev)->pdev->dev, "%s:%s:%d:(pid %d): " format, \
|
dev_dbg(&(__dev)->pdev->dev, "%s:%d:(pid %d): " format, \
|
||||||
(__dev)->priv.name, __func__, __LINE__, current->pid, \
|
__func__, __LINE__, current->pid, \
|
||||||
##__VA_ARGS__)
|
##__VA_ARGS__)
|
||||||
|
|
||||||
#define mlx5_core_dbg_mask(__dev, mask, format, ...) \
|
#define mlx5_core_dbg_mask(__dev, mask, format, ...) \
|
||||||
|
@ -63,8 +63,8 @@ do { \
|
||||||
##__VA_ARGS__)
|
##__VA_ARGS__)
|
||||||
|
|
||||||
#define mlx5_core_warn(__dev, format, ...) \
|
#define mlx5_core_warn(__dev, format, ...) \
|
||||||
dev_warn(&(__dev)->pdev->dev, "%s:%s:%d:(pid %d): " format, \
|
dev_warn(&(__dev)->pdev->dev, "%s:%d:(pid %d): " format, \
|
||||||
(__dev)->priv.name, __func__, __LINE__, current->pid, \
|
__func__, __LINE__, current->pid, \
|
||||||
##__VA_ARGS__)
|
##__VA_ARGS__)
|
||||||
|
|
||||||
#define mlx5_core_info(__dev, format, ...) \
|
#define mlx5_core_info(__dev, format, ...) \
|
||||||
|
@ -75,6 +75,11 @@ enum {
|
||||||
MLX5_CMD_TIME, /* print command execution time */
|
MLX5_CMD_TIME, /* print command execution time */
|
||||||
};
|
};
|
||||||
|
|
||||||
|
enum {
|
||||||
|
MLX5_DRIVER_STATUS_ABORTED = 0xfe,
|
||||||
|
MLX5_DRIVER_SYND = 0xbadd00de,
|
||||||
|
};
|
||||||
|
|
||||||
int mlx5_query_hca_caps(struct mlx5_core_dev *dev);
|
int mlx5_query_hca_caps(struct mlx5_core_dev *dev);
|
||||||
int mlx5_query_board_id(struct mlx5_core_dev *dev);
|
int mlx5_query_board_id(struct mlx5_core_dev *dev);
|
||||||
int mlx5_cmd_init_hca(struct mlx5_core_dev *dev);
|
int mlx5_cmd_init_hca(struct mlx5_core_dev *dev);
|
||||||
|
|
|
@ -2229,6 +2229,7 @@ static int qed_ll2_start_xmit(struct qed_dev *cdev, struct sk_buff *skb)
|
||||||
mapping))) {
|
mapping))) {
|
||||||
DP_NOTICE(cdev,
|
DP_NOTICE(cdev,
|
||||||
"Unable to map frag - dropping packet\n");
|
"Unable to map frag - dropping packet\n");
|
||||||
|
rc = -ENOMEM;
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
|
|
|
@ -43,9 +43,11 @@ static void dwmac1000_dma_axi(void __iomem *ioaddr, struct stmmac_axi *axi)
|
||||||
if (axi->axi_xit_frm)
|
if (axi->axi_xit_frm)
|
||||||
value |= DMA_AXI_LPI_XIT_FRM;
|
value |= DMA_AXI_LPI_XIT_FRM;
|
||||||
|
|
||||||
|
value &= ~DMA_AXI_WR_OSR_LMT;
|
||||||
value |= (axi->axi_wr_osr_lmt & DMA_AXI_WR_OSR_LMT_MASK) <<
|
value |= (axi->axi_wr_osr_lmt & DMA_AXI_WR_OSR_LMT_MASK) <<
|
||||||
DMA_AXI_WR_OSR_LMT_SHIFT;
|
DMA_AXI_WR_OSR_LMT_SHIFT;
|
||||||
|
|
||||||
|
value &= ~DMA_AXI_RD_OSR_LMT;
|
||||||
value |= (axi->axi_rd_osr_lmt & DMA_AXI_RD_OSR_LMT_MASK) <<
|
value |= (axi->axi_rd_osr_lmt & DMA_AXI_RD_OSR_LMT_MASK) <<
|
||||||
DMA_AXI_RD_OSR_LMT_SHIFT;
|
DMA_AXI_RD_OSR_LMT_SHIFT;
|
||||||
|
|
||||||
|
|
|
@ -30,9 +30,11 @@ static void dwmac4_dma_axi(void __iomem *ioaddr, struct stmmac_axi *axi)
|
||||||
if (axi->axi_xit_frm)
|
if (axi->axi_xit_frm)
|
||||||
value |= DMA_AXI_LPI_XIT_FRM;
|
value |= DMA_AXI_LPI_XIT_FRM;
|
||||||
|
|
||||||
|
value &= ~DMA_AXI_WR_OSR_LMT;
|
||||||
value |= (axi->axi_wr_osr_lmt & DMA_AXI_OSR_MAX) <<
|
value |= (axi->axi_wr_osr_lmt & DMA_AXI_OSR_MAX) <<
|
||||||
DMA_AXI_WR_OSR_LMT_SHIFT;
|
DMA_AXI_WR_OSR_LMT_SHIFT;
|
||||||
|
|
||||||
|
value &= ~DMA_AXI_RD_OSR_LMT;
|
||||||
value |= (axi->axi_rd_osr_lmt & DMA_AXI_OSR_MAX) <<
|
value |= (axi->axi_rd_osr_lmt & DMA_AXI_OSR_MAX) <<
|
||||||
DMA_AXI_RD_OSR_LMT_SHIFT;
|
DMA_AXI_RD_OSR_LMT_SHIFT;
|
||||||
|
|
||||||
|
|
|
@ -126,8 +126,10 @@ static struct stmmac_axi *stmmac_axi_setup(struct platform_device *pdev)
|
||||||
axi->axi_mb = of_property_read_bool(np, "snps,axi_mb");
|
axi->axi_mb = of_property_read_bool(np, "snps,axi_mb");
|
||||||
axi->axi_rb = of_property_read_bool(np, "snps,axi_rb");
|
axi->axi_rb = of_property_read_bool(np, "snps,axi_rb");
|
||||||
|
|
||||||
of_property_read_u32(np, "snps,wr_osr_lmt", &axi->axi_wr_osr_lmt);
|
if (of_property_read_u32(np, "snps,wr_osr_lmt", &axi->axi_wr_osr_lmt))
|
||||||
of_property_read_u32(np, "snps,rd_osr_lmt", &axi->axi_rd_osr_lmt);
|
axi->axi_wr_osr_lmt = 1;
|
||||||
|
if (of_property_read_u32(np, "snps,rd_osr_lmt", &axi->axi_rd_osr_lmt))
|
||||||
|
axi->axi_rd_osr_lmt = 1;
|
||||||
of_property_read_u32_array(np, "snps,blen", axi->axi_blen, AXI_BLEN);
|
of_property_read_u32_array(np, "snps,blen", axi->axi_blen, AXI_BLEN);
|
||||||
of_node_put(np);
|
of_node_put(np);
|
||||||
|
|
||||||
|
|
|
@ -1723,6 +1723,7 @@ static int irda_usb_probe(struct usb_interface *intf,
|
||||||
/* Don't change this buffer size and allocation without doing
|
/* Don't change this buffer size and allocation without doing
|
||||||
* some heavy and complete testing. Don't ask why :-(
|
* some heavy and complete testing. Don't ask why :-(
|
||||||
* Jean II */
|
* Jean II */
|
||||||
|
ret = -ENOMEM;
|
||||||
self->speed_buff = kzalloc(IRDA_USB_SPEED_MTU, GFP_KERNEL);
|
self->speed_buff = kzalloc(IRDA_USB_SPEED_MTU, GFP_KERNEL);
|
||||||
if (!self->speed_buff)
|
if (!self->speed_buff)
|
||||||
goto err_out_3;
|
goto err_out_3;
|
||||||
|
|
|
@ -3494,6 +3494,7 @@ static int lan78xx_probe(struct usb_interface *intf,
|
||||||
if (buf) {
|
if (buf) {
|
||||||
dev->urb_intr = usb_alloc_urb(0, GFP_KERNEL);
|
dev->urb_intr = usb_alloc_urb(0, GFP_KERNEL);
|
||||||
if (!dev->urb_intr) {
|
if (!dev->urb_intr) {
|
||||||
|
ret = -ENOMEM;
|
||||||
kfree(buf);
|
kfree(buf);
|
||||||
goto out3;
|
goto out3;
|
||||||
} else {
|
} else {
|
||||||
|
|
|
@ -969,12 +969,17 @@ static int virtnet_set_mac_address(struct net_device *dev, void *p)
|
||||||
struct virtnet_info *vi = netdev_priv(dev);
|
struct virtnet_info *vi = netdev_priv(dev);
|
||||||
struct virtio_device *vdev = vi->vdev;
|
struct virtio_device *vdev = vi->vdev;
|
||||||
int ret;
|
int ret;
|
||||||
struct sockaddr *addr = p;
|
struct sockaddr *addr;
|
||||||
struct scatterlist sg;
|
struct scatterlist sg;
|
||||||
|
|
||||||
ret = eth_prepare_mac_addr_change(dev, p);
|
addr = kmalloc(sizeof(*addr), GFP_KERNEL);
|
||||||
|
if (!addr)
|
||||||
|
return -ENOMEM;
|
||||||
|
memcpy(addr, p, sizeof(*addr));
|
||||||
|
|
||||||
|
ret = eth_prepare_mac_addr_change(dev, addr);
|
||||||
if (ret)
|
if (ret)
|
||||||
return ret;
|
goto out;
|
||||||
|
|
||||||
if (virtio_has_feature(vdev, VIRTIO_NET_F_CTRL_MAC_ADDR)) {
|
if (virtio_has_feature(vdev, VIRTIO_NET_F_CTRL_MAC_ADDR)) {
|
||||||
sg_init_one(&sg, addr->sa_data, dev->addr_len);
|
sg_init_one(&sg, addr->sa_data, dev->addr_len);
|
||||||
|
@ -982,7 +987,8 @@ static int virtnet_set_mac_address(struct net_device *dev, void *p)
|
||||||
VIRTIO_NET_CTRL_MAC_ADDR_SET, &sg)) {
|
VIRTIO_NET_CTRL_MAC_ADDR_SET, &sg)) {
|
||||||
dev_warn(&vdev->dev,
|
dev_warn(&vdev->dev,
|
||||||
"Failed to set mac address by vq command.\n");
|
"Failed to set mac address by vq command.\n");
|
||||||
return -EINVAL;
|
ret = -EINVAL;
|
||||||
|
goto out;
|
||||||
}
|
}
|
||||||
} else if (virtio_has_feature(vdev, VIRTIO_NET_F_MAC) &&
|
} else if (virtio_has_feature(vdev, VIRTIO_NET_F_MAC) &&
|
||||||
!virtio_has_feature(vdev, VIRTIO_F_VERSION_1)) {
|
!virtio_has_feature(vdev, VIRTIO_F_VERSION_1)) {
|
||||||
|
@ -996,8 +1002,11 @@ static int virtnet_set_mac_address(struct net_device *dev, void *p)
|
||||||
}
|
}
|
||||||
|
|
||||||
eth_commit_mac_addr_change(dev, p);
|
eth_commit_mac_addr_change(dev, p);
|
||||||
|
ret = 0;
|
||||||
|
|
||||||
return 0;
|
out:
|
||||||
|
kfree(addr);
|
||||||
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct rtnl_link_stats64 *virtnet_stats(struct net_device *dev,
|
static struct rtnl_link_stats64 *virtnet_stats(struct net_device *dev,
|
||||||
|
|
|
@ -124,6 +124,8 @@ struct drbg_state {
|
||||||
struct skcipher_request *ctr_req; /* CTR mode request handle */
|
struct skcipher_request *ctr_req; /* CTR mode request handle */
|
||||||
__u8 *ctr_null_value_buf; /* CTR mode unaligned buffer */
|
__u8 *ctr_null_value_buf; /* CTR mode unaligned buffer */
|
||||||
__u8 *ctr_null_value; /* CTR mode aligned zero buf */
|
__u8 *ctr_null_value; /* CTR mode aligned zero buf */
|
||||||
|
__u8 *outscratchpadbuf; /* CTR mode output scratchpad */
|
||||||
|
__u8 *outscratchpad; /* CTR mode aligned outbuf */
|
||||||
struct completion ctr_completion; /* CTR mode async handler */
|
struct completion ctr_completion; /* CTR mode async handler */
|
||||||
int ctr_async_err; /* CTR mode async error */
|
int ctr_async_err; /* CTR mode async error */
|
||||||
|
|
||||||
|
|
|
@ -31,7 +31,7 @@
|
||||||
#include <linux/hdlc/ioctl.h>
|
#include <linux/hdlc/ioctl.h>
|
||||||
|
|
||||||
/* For glibc compatibility. An empty enum does not compile. */
|
/* For glibc compatibility. An empty enum does not compile. */
|
||||||
#if __UAPI_DEF_IF_NET_DEVICE_FLAGS_LOWER_UP_DORMANT_ECHO != 0 && \
|
#if __UAPI_DEF_IF_NET_DEVICE_FLAGS_LOWER_UP_DORMANT_ECHO != 0 || \
|
||||||
__UAPI_DEF_IF_NET_DEVICE_FLAGS != 0
|
__UAPI_DEF_IF_NET_DEVICE_FLAGS != 0
|
||||||
/**
|
/**
|
||||||
* enum net_device_flags - &struct net_device flags
|
* enum net_device_flags - &struct net_device flags
|
||||||
|
@ -99,7 +99,7 @@ enum net_device_flags {
|
||||||
IFF_ECHO = 1<<18, /* volatile */
|
IFF_ECHO = 1<<18, /* volatile */
|
||||||
#endif /* __UAPI_DEF_IF_NET_DEVICE_FLAGS_LOWER_UP_DORMANT_ECHO */
|
#endif /* __UAPI_DEF_IF_NET_DEVICE_FLAGS_LOWER_UP_DORMANT_ECHO */
|
||||||
};
|
};
|
||||||
#endif /* __UAPI_DEF_IF_NET_DEVICE_FLAGS_LOWER_UP_DORMANT_ECHO != 0 && __UAPI_DEF_IF_NET_DEVICE_FLAGS != 0 */
|
#endif /* __UAPI_DEF_IF_NET_DEVICE_FLAGS_LOWER_UP_DORMANT_ECHO != 0 || __UAPI_DEF_IF_NET_DEVICE_FLAGS != 0 */
|
||||||
|
|
||||||
/* for compatibility with glibc net/if.h */
|
/* for compatibility with glibc net/if.h */
|
||||||
#if __UAPI_DEF_IF_NET_DEVICE_FLAGS
|
#if __UAPI_DEF_IF_NET_DEVICE_FLAGS
|
||||||
|
|
|
@ -5,6 +5,7 @@ header-y += nf_conntrack_ftp.h
|
||||||
header-y += nf_conntrack_sctp.h
|
header-y += nf_conntrack_sctp.h
|
||||||
header-y += nf_conntrack_tcp.h
|
header-y += nf_conntrack_tcp.h
|
||||||
header-y += nf_conntrack_tuple_common.h
|
header-y += nf_conntrack_tuple_common.h
|
||||||
|
header-y += nf_log.h
|
||||||
header-y += nf_tables.h
|
header-y += nf_tables.h
|
||||||
header-y += nf_tables_compat.h
|
header-y += nf_tables_compat.h
|
||||||
header-y += nf_nat.h
|
header-y += nf_nat.h
|
||||||
|
|
|
@ -12,3 +12,4 @@ header-y += tc_bpf.h
|
||||||
header-y += tc_connmark.h
|
header-y += tc_connmark.h
|
||||||
header-y += tc_ife.h
|
header-y += tc_ife.h
|
||||||
header-y += tc_tunnel_key.h
|
header-y += tc_tunnel_key.h
|
||||||
|
header-y += tc_skbmod.h
|
||||||
|
|
15
mm/shmem.c
15
mm/shmem.c
|
@ -1848,6 +1848,18 @@ alloc_nohuge: page = shmem_alloc_and_acct_page(gfp, info, sbinfo,
|
||||||
return error;
|
return error;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* This is like autoremove_wake_function, but it removes the wait queue
|
||||||
|
* entry unconditionally - even if something else had already woken the
|
||||||
|
* target.
|
||||||
|
*/
|
||||||
|
static int synchronous_wake_function(wait_queue_t *wait, unsigned mode, int sync, void *key)
|
||||||
|
{
|
||||||
|
int ret = default_wake_function(wait, mode, sync, key);
|
||||||
|
list_del_init(&wait->task_list);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
static int shmem_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
|
static int shmem_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
|
||||||
{
|
{
|
||||||
struct inode *inode = file_inode(vma->vm_file);
|
struct inode *inode = file_inode(vma->vm_file);
|
||||||
|
@ -1883,7 +1895,7 @@ static int shmem_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
|
||||||
vmf->pgoff >= shmem_falloc->start &&
|
vmf->pgoff >= shmem_falloc->start &&
|
||||||
vmf->pgoff < shmem_falloc->next) {
|
vmf->pgoff < shmem_falloc->next) {
|
||||||
wait_queue_head_t *shmem_falloc_waitq;
|
wait_queue_head_t *shmem_falloc_waitq;
|
||||||
DEFINE_WAIT(shmem_fault_wait);
|
DEFINE_WAIT_FUNC(shmem_fault_wait, synchronous_wake_function);
|
||||||
|
|
||||||
ret = VM_FAULT_NOPAGE;
|
ret = VM_FAULT_NOPAGE;
|
||||||
if ((vmf->flags & FAULT_FLAG_ALLOW_RETRY) &&
|
if ((vmf->flags & FAULT_FLAG_ALLOW_RETRY) &&
|
||||||
|
@ -2665,6 +2677,7 @@ static long shmem_fallocate(struct file *file, int mode, loff_t offset,
|
||||||
spin_lock(&inode->i_lock);
|
spin_lock(&inode->i_lock);
|
||||||
inode->i_private = NULL;
|
inode->i_private = NULL;
|
||||||
wake_up_all(&shmem_falloc_waitq);
|
wake_up_all(&shmem_falloc_waitq);
|
||||||
|
WARN_ON_ONCE(!list_empty(&shmem_falloc_waitq.task_list));
|
||||||
spin_unlock(&inode->i_lock);
|
spin_unlock(&inode->i_lock);
|
||||||
error = 0;
|
error = 0;
|
||||||
goto out;
|
goto out;
|
||||||
|
|
|
@ -2354,6 +2354,8 @@ static void shrink_node_memcg(struct pglist_data *pgdat, struct mem_cgroup *memc
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
cond_resched();
|
||||||
|
|
||||||
if (nr_reclaimed < nr_to_reclaim || scan_adjusted)
|
if (nr_reclaimed < nr_to_reclaim || scan_adjusted)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
|
|
|
@ -348,7 +348,7 @@ static unsigned long count_shadow_nodes(struct shrinker *shrinker,
|
||||||
shadow_nodes = list_lru_shrink_count(&workingset_shadow_nodes, sc);
|
shadow_nodes = list_lru_shrink_count(&workingset_shadow_nodes, sc);
|
||||||
local_irq_enable();
|
local_irq_enable();
|
||||||
|
|
||||||
if (memcg_kmem_enabled()) {
|
if (sc->memcg) {
|
||||||
pages = mem_cgroup_node_nr_lru_pages(sc->memcg, sc->nid,
|
pages = mem_cgroup_node_nr_lru_pages(sc->memcg, sc->nid,
|
||||||
LRU_ALL_FILE);
|
LRU_ALL_FILE);
|
||||||
} else {
|
} else {
|
||||||
|
|
|
@ -3287,7 +3287,7 @@ static bool batadv_send_my_tt_response(struct batadv_priv *bat_priv,
|
||||||
&tvlv_tt_data,
|
&tvlv_tt_data,
|
||||||
&tt_change,
|
&tt_change,
|
||||||
&tt_len);
|
&tt_len);
|
||||||
if (!tt_len)
|
if (!tt_len || !tvlv_len)
|
||||||
goto unlock;
|
goto unlock;
|
||||||
|
|
||||||
/* Copy the last orig_node's OGM buffer */
|
/* Copy the last orig_node's OGM buffer */
|
||||||
|
@ -3305,7 +3305,7 @@ static bool batadv_send_my_tt_response(struct batadv_priv *bat_priv,
|
||||||
&tvlv_tt_data,
|
&tvlv_tt_data,
|
||||||
&tt_change,
|
&tt_change,
|
||||||
&tt_len);
|
&tt_len);
|
||||||
if (!tt_len)
|
if (!tt_len || !tvlv_len)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
/* fill the rest of the tvlv with the real TT entries */
|
/* fill the rest of the tvlv with the real TT entries */
|
||||||
|
|
|
@ -938,6 +938,7 @@ int br_sysfs_addbr(struct net_device *dev)
|
||||||
if (!br->ifobj) {
|
if (!br->ifobj) {
|
||||||
pr_info("%s: can't add kobject (directory) %s/%s\n",
|
pr_info("%s: can't add kobject (directory) %s/%s\n",
|
||||||
__func__, dev->name, SYSFS_BRIDGE_PORT_SUBDIR);
|
__func__, dev->name, SYSFS_BRIDGE_PORT_SUBDIR);
|
||||||
|
err = -ENOMEM;
|
||||||
goto out3;
|
goto out3;
|
||||||
}
|
}
|
||||||
return 0;
|
return 0;
|
||||||
|
|
|
@ -1107,10 +1107,7 @@ static struct net_proto_family caif_family_ops = {
|
||||||
|
|
||||||
static int __init caif_sktinit_module(void)
|
static int __init caif_sktinit_module(void)
|
||||||
{
|
{
|
||||||
int err = sock_register(&caif_family_ops);
|
return sock_register(&caif_family_ops);
|
||||||
if (!err)
|
|
||||||
return err;
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __exit caif_sktexit_module(void)
|
static void __exit caif_sktexit_module(void)
|
||||||
|
|
|
@ -1353,6 +1353,7 @@ static int dcbnl_cee_fill(struct sk_buff *skb, struct net_device *netdev)
|
||||||
dcb_unlock:
|
dcb_unlock:
|
||||||
spin_unlock_bh(&dcb_lock);
|
spin_unlock_bh(&dcb_lock);
|
||||||
nla_put_failure:
|
nla_put_failure:
|
||||||
|
err = -EMSGSIZE;
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -808,6 +808,13 @@ static unsigned char update_suffix(struct key_vector *tn)
|
||||||
{
|
{
|
||||||
unsigned char slen = tn->pos;
|
unsigned char slen = tn->pos;
|
||||||
unsigned long stride, i;
|
unsigned long stride, i;
|
||||||
|
unsigned char slen_max;
|
||||||
|
|
||||||
|
/* only vector 0 can have a suffix length greater than or equal to
|
||||||
|
* tn->pos + tn->bits, the second highest node will have a suffix
|
||||||
|
* length at most of tn->pos + tn->bits - 1
|
||||||
|
*/
|
||||||
|
slen_max = min_t(unsigned char, tn->pos + tn->bits - 1, tn->slen);
|
||||||
|
|
||||||
/* search though the list of children looking for nodes that might
|
/* search though the list of children looking for nodes that might
|
||||||
* have a suffix greater than the one we currently have. This is
|
* have a suffix greater than the one we currently have. This is
|
||||||
|
@ -825,12 +832,8 @@ static unsigned char update_suffix(struct key_vector *tn)
|
||||||
slen = n->slen;
|
slen = n->slen;
|
||||||
i &= ~(stride - 1);
|
i &= ~(stride - 1);
|
||||||
|
|
||||||
/* if slen covers all but the last bit we can stop here
|
/* stop searching if we have hit the maximum possible value */
|
||||||
* there will be nothing longer than that since only node
|
if (slen >= slen_max)
|
||||||
* 0 and 1 << (bits - 1) could have that as their suffix
|
|
||||||
* length.
|
|
||||||
*/
|
|
||||||
if ((slen + 1) >= (tn->pos + tn->bits))
|
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1002,39 +1005,27 @@ static struct key_vector *resize(struct trie *t, struct key_vector *tn)
|
||||||
return collapse(t, tn);
|
return collapse(t, tn);
|
||||||
|
|
||||||
/* update parent in case halve failed */
|
/* update parent in case halve failed */
|
||||||
tp = node_parent(tn);
|
return node_parent(tn);
|
||||||
|
|
||||||
/* Return if at least one deflate was run */
|
|
||||||
if (max_work != MAX_WORK)
|
|
||||||
return tp;
|
|
||||||
|
|
||||||
/* push the suffix length to the parent node */
|
|
||||||
if (tn->slen > tn->pos) {
|
|
||||||
unsigned char slen = update_suffix(tn);
|
|
||||||
|
|
||||||
if (slen > tp->slen)
|
|
||||||
tp->slen = slen;
|
|
||||||
}
|
|
||||||
|
|
||||||
return tp;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void leaf_pull_suffix(struct key_vector *tp, struct key_vector *l)
|
static void node_pull_suffix(struct key_vector *tn, unsigned char slen)
|
||||||
{
|
{
|
||||||
while ((tp->slen > tp->pos) && (tp->slen > l->slen)) {
|
unsigned char node_slen = tn->slen;
|
||||||
if (update_suffix(tp) > l->slen)
|
|
||||||
|
while ((node_slen > tn->pos) && (node_slen > slen)) {
|
||||||
|
slen = update_suffix(tn);
|
||||||
|
if (node_slen == slen)
|
||||||
break;
|
break;
|
||||||
tp = node_parent(tp);
|
|
||||||
|
tn = node_parent(tn);
|
||||||
|
node_slen = tn->slen;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void leaf_push_suffix(struct key_vector *tn, struct key_vector *l)
|
static void node_push_suffix(struct key_vector *tn, unsigned char slen)
|
||||||
{
|
{
|
||||||
/* if this is a new leaf then tn will be NULL and we can sort
|
while (tn->slen < slen) {
|
||||||
* out parent suffix lengths as a part of trie_rebalance
|
tn->slen = slen;
|
||||||
*/
|
|
||||||
while (tn->slen < l->slen) {
|
|
||||||
tn->slen = l->slen;
|
|
||||||
tn = node_parent(tn);
|
tn = node_parent(tn);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1155,6 +1146,7 @@ static int fib_insert_node(struct trie *t, struct key_vector *tp,
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Case 3: n is NULL, and will just insert a new leaf */
|
/* Case 3: n is NULL, and will just insert a new leaf */
|
||||||
|
node_push_suffix(tp, new->fa_slen);
|
||||||
NODE_INIT_PARENT(l, tp);
|
NODE_INIT_PARENT(l, tp);
|
||||||
put_child_root(tp, key, l);
|
put_child_root(tp, key, l);
|
||||||
trie_rebalance(t, tp);
|
trie_rebalance(t, tp);
|
||||||
|
@ -1196,7 +1188,7 @@ static int fib_insert_alias(struct trie *t, struct key_vector *tp,
|
||||||
/* if we added to the tail node then we need to update slen */
|
/* if we added to the tail node then we need to update slen */
|
||||||
if (l->slen < new->fa_slen) {
|
if (l->slen < new->fa_slen) {
|
||||||
l->slen = new->fa_slen;
|
l->slen = new->fa_slen;
|
||||||
leaf_push_suffix(tp, l);
|
node_push_suffix(tp, new->fa_slen);
|
||||||
}
|
}
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -1588,6 +1580,8 @@ static void fib_remove_alias(struct trie *t, struct key_vector *tp,
|
||||||
* out parent suffix lengths as a part of trie_rebalance
|
* out parent suffix lengths as a part of trie_rebalance
|
||||||
*/
|
*/
|
||||||
if (hlist_empty(&l->leaf)) {
|
if (hlist_empty(&l->leaf)) {
|
||||||
|
if (tp->slen == l->slen)
|
||||||
|
node_pull_suffix(tp, tp->pos);
|
||||||
put_child_root(tp, l->key, NULL);
|
put_child_root(tp, l->key, NULL);
|
||||||
node_free(l);
|
node_free(l);
|
||||||
trie_rebalance(t, tp);
|
trie_rebalance(t, tp);
|
||||||
|
@ -1600,7 +1594,7 @@ static void fib_remove_alias(struct trie *t, struct key_vector *tp,
|
||||||
|
|
||||||
/* update the trie with the latest suffix length */
|
/* update the trie with the latest suffix length */
|
||||||
l->slen = fa->fa_slen;
|
l->slen = fa->fa_slen;
|
||||||
leaf_pull_suffix(tp, l);
|
node_pull_suffix(tp, fa->fa_slen);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Caller must hold RTNL. */
|
/* Caller must hold RTNL. */
|
||||||
|
@ -1872,6 +1866,10 @@ void fib_table_flush_external(struct fib_table *tb)
|
||||||
if (IS_TRIE(pn))
|
if (IS_TRIE(pn))
|
||||||
break;
|
break;
|
||||||
|
|
||||||
|
/* update the suffix to address pulled leaves */
|
||||||
|
if (pn->slen > pn->pos)
|
||||||
|
update_suffix(pn);
|
||||||
|
|
||||||
/* resize completed node */
|
/* resize completed node */
|
||||||
pn = resize(t, pn);
|
pn = resize(t, pn);
|
||||||
cindex = get_index(pkey, pn);
|
cindex = get_index(pkey, pn);
|
||||||
|
@ -1938,6 +1936,10 @@ int fib_table_flush(struct net *net, struct fib_table *tb)
|
||||||
if (IS_TRIE(pn))
|
if (IS_TRIE(pn))
|
||||||
break;
|
break;
|
||||||
|
|
||||||
|
/* update the suffix to address pulled leaves */
|
||||||
|
if (pn->slen > pn->pos)
|
||||||
|
update_suffix(pn);
|
||||||
|
|
||||||
/* resize completed node */
|
/* resize completed node */
|
||||||
pn = resize(t, pn);
|
pn = resize(t, pn);
|
||||||
cindex = get_index(pkey, pn);
|
cindex = get_index(pkey, pn);
|
||||||
|
|
|
@ -657,6 +657,10 @@ int ping_common_sendmsg(int family, struct msghdr *msg, size_t len,
|
||||||
if (len > 0xFFFF)
|
if (len > 0xFFFF)
|
||||||
return -EMSGSIZE;
|
return -EMSGSIZE;
|
||||||
|
|
||||||
|
/* Must have at least a full ICMP header. */
|
||||||
|
if (len < icmph_len)
|
||||||
|
return -EINVAL;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Check the flags.
|
* Check the flags.
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -129,6 +129,23 @@ int sysctl_tcp_invalid_ratelimit __read_mostly = HZ/2;
|
||||||
#define REXMIT_LOST 1 /* retransmit packets marked lost */
|
#define REXMIT_LOST 1 /* retransmit packets marked lost */
|
||||||
#define REXMIT_NEW 2 /* FRTO-style transmit of unsent/new packets */
|
#define REXMIT_NEW 2 /* FRTO-style transmit of unsent/new packets */
|
||||||
|
|
||||||
|
static void tcp_gro_dev_warn(struct sock *sk, const struct sk_buff *skb)
|
||||||
|
{
|
||||||
|
static bool __once __read_mostly;
|
||||||
|
|
||||||
|
if (!__once) {
|
||||||
|
struct net_device *dev;
|
||||||
|
|
||||||
|
__once = true;
|
||||||
|
|
||||||
|
rcu_read_lock();
|
||||||
|
dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif);
|
||||||
|
pr_warn("%s: Driver has suspect GRO implementation, TCP performance may be compromised.\n",
|
||||||
|
dev ? dev->name : "Unknown driver");
|
||||||
|
rcu_read_unlock();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/* Adapt the MSS value used to make delayed ack decision to the
|
/* Adapt the MSS value used to make delayed ack decision to the
|
||||||
* real world.
|
* real world.
|
||||||
*/
|
*/
|
||||||
|
@ -145,7 +162,10 @@ static void tcp_measure_rcv_mss(struct sock *sk, const struct sk_buff *skb)
|
||||||
*/
|
*/
|
||||||
len = skb_shinfo(skb)->gso_size ? : skb->len;
|
len = skb_shinfo(skb)->gso_size ? : skb->len;
|
||||||
if (len >= icsk->icsk_ack.rcv_mss) {
|
if (len >= icsk->icsk_ack.rcv_mss) {
|
||||||
icsk->icsk_ack.rcv_mss = len;
|
icsk->icsk_ack.rcv_mss = min_t(unsigned int, len,
|
||||||
|
tcp_sk(sk)->advmss);
|
||||||
|
if (unlikely(icsk->icsk_ack.rcv_mss != len))
|
||||||
|
tcp_gro_dev_warn(sk, skb);
|
||||||
} else {
|
} else {
|
||||||
/* Otherwise, we make more careful check taking into account,
|
/* Otherwise, we make more careful check taking into account,
|
||||||
* that SACKs block is variable.
|
* that SACKs block is variable.
|
||||||
|
|
|
@ -1252,7 +1252,7 @@ static int rtm_to_route_config(struct sk_buff *skb, struct nlmsghdr *nlh,
|
||||||
if (!nla)
|
if (!nla)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
switch(index) {
|
switch (index) {
|
||||||
case RTA_OIF:
|
case RTA_OIF:
|
||||||
cfg->rc_ifindex = nla_get_u32(nla);
|
cfg->rc_ifindex = nla_get_u32(nla);
|
||||||
break;
|
break;
|
||||||
|
|
|
@ -322,11 +322,13 @@ static void netlink_skb_set_owner_r(struct sk_buff *skb, struct sock *sk)
|
||||||
sk_mem_charge(sk, skb->truesize);
|
sk_mem_charge(sk, skb->truesize);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __netlink_sock_destruct(struct sock *sk)
|
static void netlink_sock_destruct(struct sock *sk)
|
||||||
{
|
{
|
||||||
struct netlink_sock *nlk = nlk_sk(sk);
|
struct netlink_sock *nlk = nlk_sk(sk);
|
||||||
|
|
||||||
if (nlk->cb_running) {
|
if (nlk->cb_running) {
|
||||||
|
if (nlk->cb.done)
|
||||||
|
nlk->cb.done(&nlk->cb);
|
||||||
module_put(nlk->cb.module);
|
module_put(nlk->cb.module);
|
||||||
kfree_skb(nlk->cb.skb);
|
kfree_skb(nlk->cb.skb);
|
||||||
}
|
}
|
||||||
|
@ -348,21 +350,7 @@ static void netlink_sock_destruct_work(struct work_struct *work)
|
||||||
struct netlink_sock *nlk = container_of(work, struct netlink_sock,
|
struct netlink_sock *nlk = container_of(work, struct netlink_sock,
|
||||||
work);
|
work);
|
||||||
|
|
||||||
nlk->cb.done(&nlk->cb);
|
sk_free(&nlk->sk);
|
||||||
__netlink_sock_destruct(&nlk->sk);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void netlink_sock_destruct(struct sock *sk)
|
|
||||||
{
|
|
||||||
struct netlink_sock *nlk = nlk_sk(sk);
|
|
||||||
|
|
||||||
if (nlk->cb_running && nlk->cb.done) {
|
|
||||||
INIT_WORK(&nlk->work, netlink_sock_destruct_work);
|
|
||||||
schedule_work(&nlk->work);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
__netlink_sock_destruct(sk);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* This lock without WQ_FLAG_EXCLUSIVE is good on UP and it is _very_ bad on
|
/* This lock without WQ_FLAG_EXCLUSIVE is good on UP and it is _very_ bad on
|
||||||
|
@ -667,8 +655,18 @@ static int netlink_create(struct net *net, struct socket *sock, int protocol,
|
||||||
static void deferred_put_nlk_sk(struct rcu_head *head)
|
static void deferred_put_nlk_sk(struct rcu_head *head)
|
||||||
{
|
{
|
||||||
struct netlink_sock *nlk = container_of(head, struct netlink_sock, rcu);
|
struct netlink_sock *nlk = container_of(head, struct netlink_sock, rcu);
|
||||||
|
struct sock *sk = &nlk->sk;
|
||||||
|
|
||||||
sock_put(&nlk->sk);
|
if (!atomic_dec_and_test(&sk->sk_refcnt))
|
||||||
|
return;
|
||||||
|
|
||||||
|
if (nlk->cb_running && nlk->cb.done) {
|
||||||
|
INIT_WORK(&nlk->work, netlink_sock_destruct_work);
|
||||||
|
schedule_work(&nlk->work);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
sk_free(sk);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int netlink_release(struct socket *sock)
|
static int netlink_release(struct socket *sock)
|
||||||
|
|
|
@ -1702,7 +1702,7 @@ interrupts are disabled.
|
||||||
static void xmit_descs(struct snd_dbri *dbri)
|
static void xmit_descs(struct snd_dbri *dbri)
|
||||||
{
|
{
|
||||||
struct dbri_streaminfo *info;
|
struct dbri_streaminfo *info;
|
||||||
u32 dvma_addr = (u32)dbri->dma_dvma;
|
u32 dvma_addr;
|
||||||
s32 *cmd;
|
s32 *cmd;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
int first_td;
|
int first_td;
|
||||||
|
@ -1710,6 +1710,7 @@ static void xmit_descs(struct snd_dbri *dbri)
|
||||||
if (dbri == NULL)
|
if (dbri == NULL)
|
||||||
return; /* Disabled */
|
return; /* Disabled */
|
||||||
|
|
||||||
|
dvma_addr = (u32)dbri->dma_dvma;
|
||||||
info = &dbri->stream_info[DBRI_REC];
|
info = &dbri->stream_info[DBRI_REC];
|
||||||
spin_lock_irqsave(&dbri->lock, flags);
|
spin_lock_irqsave(&dbri->lock, flags);
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue