public inbox for gentoo-commits@lists.gentoo.org
 help / color / mirror / Atom feed
From: "Mike Pagano" <mpagano@gentoo.org>
To: gentoo-commits@lists.gentoo.org
Subject: [gentoo-commits] proj/linux-patches:6.1 commit in: /
Date: Thu, 13 Apr 2023 16:09:32 +0000 (UTC)	[thread overview]
Message-ID: <1681402159.ca7e265c42f5d0ef307cd4e76877fd77b86dae6c.mpagano@gentoo> (raw)

commit:     ca7e265c42f5d0ef307cd4e76877fd77b86dae6c
Author:     Mike Pagano <mpagano <AT> gentoo <DOT> org>
AuthorDate: Thu Apr 13 16:09:19 2023 +0000
Commit:     Mike Pagano <mpagano <AT> gentoo <DOT> org>
CommitDate: Thu Apr 13 16:09:19 2023 +0000
URL:        https://gitweb.gentoo.org/proj/linux-patches.git/commit/?id=ca7e265c

Linux patch 6.1.24

Signed-off-by: Mike Pagano <mpagano <AT> gentoo.org>

 0000_README             |     4 +
 1023_linux-6.1.24.patch | 14535 ++++++++++++++++++++++++++++++++++++++++++++++
 2 files changed, 14539 insertions(+)

diff --git a/0000_README b/0000_README
index 82189ada..ef975a1f 100644
--- a/0000_README
+++ b/0000_README
@@ -135,6 +135,10 @@ Patch:  1022_linux-6.1.23.patch
 From:   https://www.kernel.org
 Desc:   Linux 6.1.23
 
+Patch:  1023_linux-6.1.24.patch
+From:   https://www.kernel.org
+Desc:   Linux 6.1.24
+
 Patch:  1500_XATTR_USER_PREFIX.patch
 From:   https://bugs.gentoo.org/show_bug.cgi?id=470644
 Desc:   Support for namespace user.pax.* on tmpfs.

diff --git a/1023_linux-6.1.24.patch b/1023_linux-6.1.24.patch
new file mode 100644
index 00000000..546adf85
--- /dev/null
+++ b/1023_linux-6.1.24.patch
@@ -0,0 +1,14535 @@
+diff --git a/Documentation/devicetree/bindings/serial/renesas,scif.yaml b/Documentation/devicetree/bindings/serial/renesas,scif.yaml
+index f930e7f1349fc..52c0948738637 100644
+--- a/Documentation/devicetree/bindings/serial/renesas,scif.yaml
++++ b/Documentation/devicetree/bindings/serial/renesas,scif.yaml
+@@ -91,7 +91,7 @@ properties:
+           - description: Error interrupt
+           - description: Receive buffer full interrupt
+           - description: Transmit buffer empty interrupt
+-          - description: Transmit End interrupt
++          - description: Break interrupt
+       - items:
+           - description: Error interrupt
+           - description: Receive buffer full interrupt
+@@ -106,7 +106,7 @@ properties:
+           - const: eri
+           - const: rxi
+           - const: txi
+-          - const: tei
++          - const: bri
+       - items:
+           - const: eri
+           - const: rxi
+diff --git a/Documentation/mm/zsmalloc.rst b/Documentation/mm/zsmalloc.rst
+index 6e79893d61326..d98607aa9c1fd 100644
+--- a/Documentation/mm/zsmalloc.rst
++++ b/Documentation/mm/zsmalloc.rst
+@@ -68,6 +68,8 @@ pages_used
+ 	the number of pages allocated for the class
+ pages_per_zspage
+ 	the number of 0-order pages to make a zspage
++freeable
++	the approximate number of pages class compaction can free
+ 
+ We assign a zspage to ZS_ALMOST_EMPTY fullness group when n <= N / f, where
+ 
+diff --git a/Makefile b/Makefile
+index a162f6cdf77c6..4f9939dec2c26 100644
+--- a/Makefile
++++ b/Makefile
+@@ -1,7 +1,7 @@
+ # SPDX-License-Identifier: GPL-2.0
+ VERSION = 6
+ PATCHLEVEL = 1
+-SUBLEVEL = 23
++SUBLEVEL = 24
+ EXTRAVERSION =
+ NAME = Hurr durr I'ma ninja sloth
+ 
+diff --git a/arch/arm64/kernel/compat_alignment.c b/arch/arm64/kernel/compat_alignment.c
+index 5edec2f49ec98..deff21bfa6800 100644
+--- a/arch/arm64/kernel/compat_alignment.c
++++ b/arch/arm64/kernel/compat_alignment.c
+@@ -314,36 +314,32 @@ int do_compat_alignment_fixup(unsigned long addr, struct pt_regs *regs)
+ 	int (*handler)(unsigned long addr, u32 instr, struct pt_regs *regs);
+ 	unsigned int type;
+ 	u32 instr = 0;
+-	u16 tinstr = 0;
+ 	int isize = 4;
+ 	int thumb2_32b = 0;
+-	int fault;
+ 
+ 	instrptr = instruction_pointer(regs);
+ 
+ 	if (compat_thumb_mode(regs)) {
+ 		__le16 __user *ptr = (__le16 __user *)(instrptr & ~1);
++		u16 tinstr, tinst2;
+ 
+-		fault = alignment_get_thumb(regs, ptr, &tinstr);
+-		if (!fault) {
+-			if (IS_T32(tinstr)) {
+-				/* Thumb-2 32-bit */
+-				u16 tinst2;
+-				fault = alignment_get_thumb(regs, ptr + 1, &tinst2);
+-				instr = ((u32)tinstr << 16) | tinst2;
+-				thumb2_32b = 1;
+-			} else {
+-				isize = 2;
+-				instr = thumb2arm(tinstr);
+-			}
++		if (alignment_get_thumb(regs, ptr, &tinstr))
++			return 1;
++
++		if (IS_T32(tinstr)) { /* Thumb-2 32-bit */
++			if (alignment_get_thumb(regs, ptr + 1, &tinst2))
++				return 1;
++			instr = ((u32)tinstr << 16) | tinst2;
++			thumb2_32b = 1;
++		} else {
++			isize = 2;
++			instr = thumb2arm(tinstr);
+ 		}
+ 	} else {
+-		fault = alignment_get_arm(regs, (__le32 __user *)instrptr, &instr);
++		if (alignment_get_arm(regs, (__le32 __user *)instrptr, &instr))
++			return 1;
+ 	}
+ 
+-	if (fault)
+-		return 1;
+-
+ 	switch (CODING_BITS(instr)) {
+ 	case 0x00000000:	/* 3.13.4 load/store instruction extensions */
+ 		if (LDSTHD_I_BIT(instr))
+diff --git a/arch/arm64/kvm/pmu-emul.c b/arch/arm64/kvm/pmu-emul.c
+index 0003c7d37533a..67770375c5eed 100644
+--- a/arch/arm64/kvm/pmu-emul.c
++++ b/arch/arm64/kvm/pmu-emul.c
+@@ -15,16 +15,14 @@
+ #include <kvm/arm_pmu.h>
+ #include <kvm/arm_vgic.h>
+ 
++#define PERF_ATTR_CFG1_COUNTER_64BIT	BIT(0)
++
+ DEFINE_STATIC_KEY_FALSE(kvm_arm_pmu_available);
+ 
+ static LIST_HEAD(arm_pmus);
+ static DEFINE_MUTEX(arm_pmus_lock);
+ 
+ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx);
+-static void kvm_pmu_update_pmc_chained(struct kvm_vcpu *vcpu, u64 select_idx);
+-static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc);
+-
+-#define PERF_ATTR_CFG1_KVM_PMU_CHAINED 0x1
+ 
+ static u32 kvm_pmu_event_mask(struct kvm *kvm)
+ {
+@@ -52,11 +50,22 @@ static u32 kvm_pmu_event_mask(struct kvm *kvm)
+  * @select_idx: The counter index
+  */
+ static bool kvm_pmu_idx_is_64bit(struct kvm_vcpu *vcpu, u64 select_idx)
++{
++	return (select_idx == ARMV8_PMU_CYCLE_IDX);
++}
++
++static bool kvm_pmu_idx_has_64bit_overflow(struct kvm_vcpu *vcpu, u64 select_idx)
+ {
+ 	return (select_idx == ARMV8_PMU_CYCLE_IDX &&
+ 		__vcpu_sys_reg(vcpu, PMCR_EL0) & ARMV8_PMU_PMCR_LC);
+ }
+ 
++static bool kvm_pmu_counter_can_chain(struct kvm_vcpu *vcpu, u64 idx)
++{
++	return (!(idx & 1) && (idx + 1) < ARMV8_PMU_CYCLE_IDX &&
++		!kvm_pmu_idx_has_64bit_overflow(vcpu, idx));
++}
++
+ static struct kvm_vcpu *kvm_pmc_to_vcpu(struct kvm_pmc *pmc)
+ {
+ 	struct kvm_pmu *pmu;
+@@ -69,91 +78,22 @@ static struct kvm_vcpu *kvm_pmc_to_vcpu(struct kvm_pmc *pmc)
+ }
+ 
+ /**
+- * kvm_pmu_pmc_is_chained - determine if the pmc is chained
+- * @pmc: The PMU counter pointer
+- */
+-static bool kvm_pmu_pmc_is_chained(struct kvm_pmc *pmc)
+-{
+-	struct kvm_vcpu *vcpu = kvm_pmc_to_vcpu(pmc);
+-
+-	return test_bit(pmc->idx >> 1, vcpu->arch.pmu.chained);
+-}
+-
+-/**
+- * kvm_pmu_idx_is_high_counter - determine if select_idx is a high/low counter
+- * @select_idx: The counter index
+- */
+-static bool kvm_pmu_idx_is_high_counter(u64 select_idx)
+-{
+-	return select_idx & 0x1;
+-}
+-
+-/**
+- * kvm_pmu_get_canonical_pmc - obtain the canonical pmc
+- * @pmc: The PMU counter pointer
+- *
+- * When a pair of PMCs are chained together we use the low counter (canonical)
+- * to hold the underlying perf event.
+- */
+-static struct kvm_pmc *kvm_pmu_get_canonical_pmc(struct kvm_pmc *pmc)
+-{
+-	if (kvm_pmu_pmc_is_chained(pmc) &&
+-	    kvm_pmu_idx_is_high_counter(pmc->idx))
+-		return pmc - 1;
+-
+-	return pmc;
+-}
+-static struct kvm_pmc *kvm_pmu_get_alternate_pmc(struct kvm_pmc *pmc)
+-{
+-	if (kvm_pmu_idx_is_high_counter(pmc->idx))
+-		return pmc - 1;
+-	else
+-		return pmc + 1;
+-}
+-
+-/**
+- * kvm_pmu_idx_has_chain_evtype - determine if the event type is chain
++ * kvm_pmu_get_counter_value - get PMU counter value
+  * @vcpu: The vcpu pointer
+  * @select_idx: The counter index
+  */
+-static bool kvm_pmu_idx_has_chain_evtype(struct kvm_vcpu *vcpu, u64 select_idx)
+-{
+-	u64 eventsel, reg;
+-
+-	select_idx |= 0x1;
+-
+-	if (select_idx == ARMV8_PMU_CYCLE_IDX)
+-		return false;
+-
+-	reg = PMEVTYPER0_EL0 + select_idx;
+-	eventsel = __vcpu_sys_reg(vcpu, reg) & kvm_pmu_event_mask(vcpu->kvm);
+-
+-	return eventsel == ARMV8_PMUV3_PERFCTR_CHAIN;
+-}
+-
+-/**
+- * kvm_pmu_get_pair_counter_value - get PMU counter value
+- * @vcpu: The vcpu pointer
+- * @pmc: The PMU counter pointer
+- */
+-static u64 kvm_pmu_get_pair_counter_value(struct kvm_vcpu *vcpu,
+-					  struct kvm_pmc *pmc)
++u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx)
+ {
+-	u64 counter, counter_high, reg, enabled, running;
+-
+-	if (kvm_pmu_pmc_is_chained(pmc)) {
+-		pmc = kvm_pmu_get_canonical_pmc(pmc);
+-		reg = PMEVCNTR0_EL0 + pmc->idx;
++	u64 counter, reg, enabled, running;
++	struct kvm_pmu *pmu = &vcpu->arch.pmu;
++	struct kvm_pmc *pmc = &pmu->pmc[select_idx];
+ 
+-		counter = __vcpu_sys_reg(vcpu, reg);
+-		counter_high = __vcpu_sys_reg(vcpu, reg + 1);
++	if (!kvm_vcpu_has_pmu(vcpu))
++		return 0;
+ 
+-		counter = lower_32_bits(counter) | (counter_high << 32);
+-	} else {
+-		reg = (pmc->idx == ARMV8_PMU_CYCLE_IDX)
+-		      ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + pmc->idx;
+-		counter = __vcpu_sys_reg(vcpu, reg);
+-	}
++	reg = (pmc->idx == ARMV8_PMU_CYCLE_IDX)
++		? PMCCNTR_EL0 : PMEVCNTR0_EL0 + pmc->idx;
++	counter = __vcpu_sys_reg(vcpu, reg);
+ 
+ 	/*
+ 	 * The real counter value is equal to the value of counter register plus
+@@ -163,29 +103,7 @@ static u64 kvm_pmu_get_pair_counter_value(struct kvm_vcpu *vcpu,
+ 		counter += perf_event_read_value(pmc->perf_event, &enabled,
+ 						 &running);
+ 
+-	return counter;
+-}
+-
+-/**
+- * kvm_pmu_get_counter_value - get PMU counter value
+- * @vcpu: The vcpu pointer
+- * @select_idx: The counter index
+- */
+-u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx)
+-{
+-	u64 counter;
+-	struct kvm_pmu *pmu = &vcpu->arch.pmu;
+-	struct kvm_pmc *pmc = &pmu->pmc[select_idx];
+-
+-	if (!kvm_vcpu_has_pmu(vcpu))
+-		return 0;
+-
+-	counter = kvm_pmu_get_pair_counter_value(vcpu, pmc);
+-
+-	if (kvm_pmu_pmc_is_chained(pmc) &&
+-	    kvm_pmu_idx_is_high_counter(select_idx))
+-		counter = upper_32_bits(counter);
+-	else if (select_idx != ARMV8_PMU_CYCLE_IDX)
++	if (!kvm_pmu_idx_is_64bit(vcpu, select_idx))
+ 		counter = lower_32_bits(counter);
+ 
+ 	return counter;
+@@ -218,7 +136,6 @@ void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val)
+  */
+ static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc)
+ {
+-	pmc = kvm_pmu_get_canonical_pmc(pmc);
+ 	if (pmc->perf_event) {
+ 		perf_event_disable(pmc->perf_event);
+ 		perf_event_release_kernel(pmc->perf_event);
+@@ -236,11 +153,10 @@ static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc)
+ {
+ 	u64 counter, reg, val;
+ 
+-	pmc = kvm_pmu_get_canonical_pmc(pmc);
+ 	if (!pmc->perf_event)
+ 		return;
+ 
+-	counter = kvm_pmu_get_pair_counter_value(vcpu, pmc);
++	counter = kvm_pmu_get_counter_value(vcpu, pmc->idx);
+ 
+ 	if (pmc->idx == ARMV8_PMU_CYCLE_IDX) {
+ 		reg = PMCCNTR_EL0;
+@@ -252,9 +168,6 @@ static void kvm_pmu_stop_counter(struct kvm_vcpu *vcpu, struct kvm_pmc *pmc)
+ 
+ 	__vcpu_sys_reg(vcpu, reg) = val;
+ 
+-	if (kvm_pmu_pmc_is_chained(pmc))
+-		__vcpu_sys_reg(vcpu, reg + 1) = upper_32_bits(counter);
+-
+ 	kvm_pmu_release_perf_event(pmc);
+ }
+ 
+@@ -285,8 +198,6 @@ void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu)
+ 
+ 	for_each_set_bit(i, &mask, 32)
+ 		kvm_pmu_stop_counter(vcpu, &pmu->pmc[i]);
+-
+-	bitmap_zero(vcpu->arch.pmu.chained, ARMV8_PMU_MAX_COUNTER_PAIRS);
+ }
+ 
+ /**
+@@ -340,12 +251,9 @@ void kvm_pmu_enable_counter_mask(struct kvm_vcpu *vcpu, u64 val)
+ 
+ 		pmc = &pmu->pmc[i];
+ 
+-		/* A change in the enable state may affect the chain state */
+-		kvm_pmu_update_pmc_chained(vcpu, i);
+-		kvm_pmu_create_perf_event(vcpu, i);
+-
+-		/* At this point, pmc must be the canonical */
+-		if (pmc->perf_event) {
++		if (!pmc->perf_event) {
++			kvm_pmu_create_perf_event(vcpu, i);
++		} else {
+ 			perf_event_enable(pmc->perf_event);
+ 			if (pmc->perf_event->state != PERF_EVENT_STATE_ACTIVE)
+ 				kvm_debug("fail to enable perf event\n");
+@@ -375,11 +283,6 @@ void kvm_pmu_disable_counter_mask(struct kvm_vcpu *vcpu, u64 val)
+ 
+ 		pmc = &pmu->pmc[i];
+ 
+-		/* A change in the enable state may affect the chain state */
+-		kvm_pmu_update_pmc_chained(vcpu, i);
+-		kvm_pmu_create_perf_event(vcpu, i);
+-
+-		/* At this point, pmc must be the canonical */
+ 		if (pmc->perf_event)
+ 			perf_event_disable(pmc->perf_event);
+ 	}
+@@ -484,6 +387,65 @@ static void kvm_pmu_perf_overflow_notify_vcpu(struct irq_work *work)
+ 	kvm_vcpu_kick(vcpu);
+ }
+ 
++/*
++ * Perform an increment on any of the counters described in @mask,
++ * generating the overflow if required, and propagate it as a chained
++ * event if possible.
++ */
++static void kvm_pmu_counter_increment(struct kvm_vcpu *vcpu,
++				      unsigned long mask, u32 event)
++{
++	int i;
++
++	if (!(__vcpu_sys_reg(vcpu, PMCR_EL0) & ARMV8_PMU_PMCR_E))
++		return;
++
++	/* Weed out disabled counters */
++	mask &= __vcpu_sys_reg(vcpu, PMCNTENSET_EL0);
++
++	for_each_set_bit(i, &mask, ARMV8_PMU_CYCLE_IDX) {
++		u64 type, reg;
++
++		/* Filter on event type */
++		type = __vcpu_sys_reg(vcpu, PMEVTYPER0_EL0 + i);
++		type &= kvm_pmu_event_mask(vcpu->kvm);
++		if (type != event)
++			continue;
++
++		/* Increment this counter */
++		reg = __vcpu_sys_reg(vcpu, PMEVCNTR0_EL0 + i) + 1;
++		reg = lower_32_bits(reg);
++		__vcpu_sys_reg(vcpu, PMEVCNTR0_EL0 + i) = reg;
++
++		if (reg) /* No overflow? move on */
++			continue;
++
++		/* Mark overflow */
++		__vcpu_sys_reg(vcpu, PMOVSSET_EL0) |= BIT(i);
++
++		if (kvm_pmu_counter_can_chain(vcpu, i))
++			kvm_pmu_counter_increment(vcpu, BIT(i + 1),
++						  ARMV8_PMUV3_PERFCTR_CHAIN);
++	}
++}
++
++/* Compute the sample period for a given counter value */
++static u64 compute_period(struct kvm_vcpu *vcpu, u64 select_idx, u64 counter)
++{
++	u64 val;
++
++	if (kvm_pmu_idx_is_64bit(vcpu, select_idx)) {
++		if (!kvm_pmu_idx_has_64bit_overflow(vcpu, select_idx))
++			val = -(counter & GENMASK(31, 0));
++		else
++			val = (-counter) & GENMASK(63, 0);
++	} else {
++		val = (-counter) & GENMASK(31, 0);
++	}
++
++	return val;
++}
++
+ /**
+  * When the perf event overflows, set the overflow status and inform the vcpu.
+  */
+@@ -503,10 +465,7 @@ static void kvm_pmu_perf_overflow(struct perf_event *perf_event,
+ 	 * Reset the sample period to the architectural limit,
+ 	 * i.e. the point where the counter overflows.
+ 	 */
+-	period = -(local64_read(&perf_event->count));
+-
+-	if (!kvm_pmu_idx_is_64bit(vcpu, pmc->idx))
+-		period &= GENMASK(31, 0);
++	period = compute_period(vcpu, idx, local64_read(&perf_event->count));
+ 
+ 	local64_set(&perf_event->hw.period_left, 0);
+ 	perf_event->attr.sample_period = period;
+@@ -514,6 +473,10 @@ static void kvm_pmu_perf_overflow(struct perf_event *perf_event,
+ 
+ 	__vcpu_sys_reg(vcpu, PMOVSSET_EL0) |= BIT(idx);
+ 
++	if (kvm_pmu_counter_can_chain(vcpu, idx))
++		kvm_pmu_counter_increment(vcpu, BIT(idx + 1),
++					  ARMV8_PMUV3_PERFCTR_CHAIN);
++
+ 	if (kvm_pmu_overflow_status(vcpu)) {
+ 		kvm_make_request(KVM_REQ_IRQ_PENDING, vcpu);
+ 
+@@ -533,50 +496,7 @@ static void kvm_pmu_perf_overflow(struct perf_event *perf_event,
+  */
+ void kvm_pmu_software_increment(struct kvm_vcpu *vcpu, u64 val)
+ {
+-	struct kvm_pmu *pmu = &vcpu->arch.pmu;
+-	int i;
+-
+-	if (!kvm_vcpu_has_pmu(vcpu))
+-		return;
+-
+-	if (!(__vcpu_sys_reg(vcpu, PMCR_EL0) & ARMV8_PMU_PMCR_E))
+-		return;
+-
+-	/* Weed out disabled counters */
+-	val &= __vcpu_sys_reg(vcpu, PMCNTENSET_EL0);
+-
+-	for (i = 0; i < ARMV8_PMU_CYCLE_IDX; i++) {
+-		u64 type, reg;
+-
+-		if (!(val & BIT(i)))
+-			continue;
+-
+-		/* PMSWINC only applies to ... SW_INC! */
+-		type = __vcpu_sys_reg(vcpu, PMEVTYPER0_EL0 + i);
+-		type &= kvm_pmu_event_mask(vcpu->kvm);
+-		if (type != ARMV8_PMUV3_PERFCTR_SW_INCR)
+-			continue;
+-
+-		/* increment this even SW_INC counter */
+-		reg = __vcpu_sys_reg(vcpu, PMEVCNTR0_EL0 + i) + 1;
+-		reg = lower_32_bits(reg);
+-		__vcpu_sys_reg(vcpu, PMEVCNTR0_EL0 + i) = reg;
+-
+-		if (reg) /* no overflow on the low part */
+-			continue;
+-
+-		if (kvm_pmu_pmc_is_chained(&pmu->pmc[i])) {
+-			/* increment the high counter */
+-			reg = __vcpu_sys_reg(vcpu, PMEVCNTR0_EL0 + i + 1) + 1;
+-			reg = lower_32_bits(reg);
+-			__vcpu_sys_reg(vcpu, PMEVCNTR0_EL0 + i + 1) = reg;
+-			if (!reg) /* mark overflow on the high counter */
+-				__vcpu_sys_reg(vcpu, PMOVSSET_EL0) |= BIT(i + 1);
+-		} else {
+-			/* mark overflow on low counter */
+-			__vcpu_sys_reg(vcpu, PMOVSSET_EL0) |= BIT(i);
+-		}
+-	}
++	kvm_pmu_counter_increment(vcpu, val, ARMV8_PMUV3_PERFCTR_SW_INCR);
+ }
+ 
+ /**
+@@ -591,6 +511,9 @@ void kvm_pmu_handle_pmcr(struct kvm_vcpu *vcpu, u64 val)
+ 	if (!kvm_vcpu_has_pmu(vcpu))
+ 		return;
+ 
++	/* The reset bits don't indicate any state, and shouldn't be saved. */
++	__vcpu_sys_reg(vcpu, PMCR_EL0) = val & ~(ARMV8_PMU_PMCR_C | ARMV8_PMU_PMCR_P);
++
+ 	if (val & ARMV8_PMU_PMCR_E) {
+ 		kvm_pmu_enable_counter_mask(vcpu,
+ 		       __vcpu_sys_reg(vcpu, PMCNTENSET_EL0));
+@@ -625,18 +548,11 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx)
+ {
+ 	struct arm_pmu *arm_pmu = vcpu->kvm->arch.arm_pmu;
+ 	struct kvm_pmu *pmu = &vcpu->arch.pmu;
+-	struct kvm_pmc *pmc;
++	struct kvm_pmc *pmc = &pmu->pmc[select_idx];
+ 	struct perf_event *event;
+ 	struct perf_event_attr attr;
+ 	u64 eventsel, counter, reg, data;
+ 
+-	/*
+-	 * For chained counters the event type and filtering attributes are
+-	 * obtained from the low/even counter. We also use this counter to
+-	 * determine if the event is enabled/disabled.
+-	 */
+-	pmc = kvm_pmu_get_canonical_pmc(&pmu->pmc[select_idx]);
+-
+ 	reg = (pmc->idx == ARMV8_PMU_CYCLE_IDX)
+ 	      ? PMCCFILTR_EL0 : PMEVTYPER0_EL0 + pmc->idx;
+ 	data = __vcpu_sys_reg(vcpu, reg);
+@@ -647,8 +563,12 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx)
+ 	else
+ 		eventsel = data & kvm_pmu_event_mask(vcpu->kvm);
+ 
+-	/* Software increment event doesn't need to be backed by a perf event */
+-	if (eventsel == ARMV8_PMUV3_PERFCTR_SW_INCR)
++	/*
++	 * Neither SW increment nor chained events need to be backed
++	 * by a perf event.
++	 */
++	if (eventsel == ARMV8_PMUV3_PERFCTR_SW_INCR ||
++	    eventsel == ARMV8_PMUV3_PERFCTR_CHAIN)
+ 		return;
+ 
+ 	/*
+@@ -670,30 +590,20 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx)
+ 	attr.exclude_host = 1; /* Don't count host events */
+ 	attr.config = eventsel;
+ 
+-	counter = kvm_pmu_get_pair_counter_value(vcpu, pmc);
++	counter = kvm_pmu_get_counter_value(vcpu, select_idx);
+ 
+-	if (kvm_pmu_pmc_is_chained(pmc)) {
+-		/**
+-		 * The initial sample period (overflow count) of an event. For
+-		 * chained counters we only support overflow interrupts on the
+-		 * high counter.
+-		 */
+-		attr.sample_period = (-counter) & GENMASK(63, 0);
+-		attr.config1 |= PERF_ATTR_CFG1_KVM_PMU_CHAINED;
++	/*
++	 * If counting with a 64bit counter, advertise it to the perf
++	 * code, carefully dealing with the initial sample period
++	 * which also depends on the overflow.
++	 */
++	if (kvm_pmu_idx_is_64bit(vcpu, select_idx))
++		attr.config1 |= PERF_ATTR_CFG1_COUNTER_64BIT;
+ 
+-		event = perf_event_create_kernel_counter(&attr, -1, current,
+-							 kvm_pmu_perf_overflow,
+-							 pmc + 1);
+-	} else {
+-		/* The initial sample period (overflow count) of an event. */
+-		if (kvm_pmu_idx_is_64bit(vcpu, pmc->idx))
+-			attr.sample_period = (-counter) & GENMASK(63, 0);
+-		else
+-			attr.sample_period = (-counter) & GENMASK(31, 0);
++	attr.sample_period = compute_period(vcpu, select_idx, counter);
+ 
+-		event = perf_event_create_kernel_counter(&attr, -1, current,
++	event = perf_event_create_kernel_counter(&attr, -1, current,
+ 						 kvm_pmu_perf_overflow, pmc);
+-	}
+ 
+ 	if (IS_ERR(event)) {
+ 		pr_err_once("kvm: pmu event creation failed %ld\n",
+@@ -704,41 +614,6 @@ static void kvm_pmu_create_perf_event(struct kvm_vcpu *vcpu, u64 select_idx)
+ 	pmc->perf_event = event;
+ }
+ 
+-/**
+- * kvm_pmu_update_pmc_chained - update chained bitmap
+- * @vcpu: The vcpu pointer
+- * @select_idx: The number of selected counter
+- *
+- * Update the chained bitmap based on the event type written in the
+- * typer register and the enable state of the odd register.
+- */
+-static void kvm_pmu_update_pmc_chained(struct kvm_vcpu *vcpu, u64 select_idx)
+-{
+-	struct kvm_pmu *pmu = &vcpu->arch.pmu;
+-	struct kvm_pmc *pmc = &pmu->pmc[select_idx], *canonical_pmc;
+-	bool new_state, old_state;
+-
+-	old_state = kvm_pmu_pmc_is_chained(pmc);
+-	new_state = kvm_pmu_idx_has_chain_evtype(vcpu, pmc->idx) &&
+-		    kvm_pmu_counter_is_enabled(vcpu, pmc->idx | 0x1);
+-
+-	if (old_state == new_state)
+-		return;
+-
+-	canonical_pmc = kvm_pmu_get_canonical_pmc(pmc);
+-	kvm_pmu_stop_counter(vcpu, canonical_pmc);
+-	if (new_state) {
+-		/*
+-		 * During promotion from !chained to chained we must ensure
+-		 * the adjacent counter is stopped and its event destroyed
+-		 */
+-		kvm_pmu_stop_counter(vcpu, kvm_pmu_get_alternate_pmc(pmc));
+-		set_bit(pmc->idx >> 1, vcpu->arch.pmu.chained);
+-		return;
+-	}
+-	clear_bit(pmc->idx >> 1, vcpu->arch.pmu.chained);
+-}
+-
+ /**
+  * kvm_pmu_set_counter_event_type - set selected counter to monitor some event
+  * @vcpu: The vcpu pointer
+@@ -766,7 +641,6 @@ void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu, u64 data,
+ 
+ 	__vcpu_sys_reg(vcpu, reg) = data & mask;
+ 
+-	kvm_pmu_update_pmc_chained(vcpu, select_idx);
+ 	kvm_pmu_create_perf_event(vcpu, select_idx);
+ }
+ 
+diff --git a/arch/arm64/kvm/sys_regs.c b/arch/arm64/kvm/sys_regs.c
+index bfe4f17232b3e..1f80e17a64608 100644
+--- a/arch/arm64/kvm/sys_regs.c
++++ b/arch/arm64/kvm/sys_regs.c
+@@ -697,13 +697,15 @@ static bool access_pmcr(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
+ 		return false;
+ 
+ 	if (p->is_write) {
+-		/* Only update writeable bits of PMCR */
++		/*
++		 * Only update writeable bits of PMCR (continuing into
++		 * kvm_pmu_handle_pmcr() as well)
++		 */
+ 		val = __vcpu_sys_reg(vcpu, PMCR_EL0);
+ 		val &= ~ARMV8_PMU_PMCR_MASK;
+ 		val |= p->regval & ARMV8_PMU_PMCR_MASK;
+ 		if (!kvm_supports_32bit_el0())
+ 			val |= ARMV8_PMU_PMCR_LC;
+-		__vcpu_sys_reg(vcpu, PMCR_EL0) = val;
+ 		kvm_pmu_handle_pmcr(vcpu, val);
+ 		kvm_vcpu_pmu_restore_guest(vcpu);
+ 	} else {
+diff --git a/arch/s390/kvm/intercept.c b/arch/s390/kvm/intercept.c
+index 88112065d9411..ee7478a601442 100644
+--- a/arch/s390/kvm/intercept.c
++++ b/arch/s390/kvm/intercept.c
+@@ -271,10 +271,18 @@ static int handle_prog(struct kvm_vcpu *vcpu)
+  * handle_external_interrupt - used for external interruption interceptions
+  * @vcpu: virtual cpu
+  *
+- * This interception only occurs if the CPUSTAT_EXT_INT bit was set, or if
+- * the new PSW does not have external interrupts disabled. In the first case,
+- * we've got to deliver the interrupt manually, and in the second case, we
+- * drop to userspace to handle the situation there.
++ * This interception occurs if:
++ * - the CPUSTAT_EXT_INT bit was already set when the external interrupt
++ *   occurred. In this case, the interrupt needs to be injected manually to
++ *   preserve interrupt priority.
++ * - the external new PSW has external interrupts enabled, which will cause an
++ *   interruption loop. We drop to userspace in this case.
++ *
++ * The latter case can be detected by inspecting the external mask bit in the
++ * external new psw.
++ *
++ * Under PV, only the latter case can occur, since interrupt priorities are
++ * handled in the ultravisor.
+  */
+ static int handle_external_interrupt(struct kvm_vcpu *vcpu)
+ {
+@@ -285,10 +293,18 @@ static int handle_external_interrupt(struct kvm_vcpu *vcpu)
+ 
+ 	vcpu->stat.exit_external_interrupt++;
+ 
+-	rc = read_guest_lc(vcpu, __LC_EXT_NEW_PSW, &newpsw, sizeof(psw_t));
+-	if (rc)
+-		return rc;
+-	/* We can not handle clock comparator or timer interrupt with bad PSW */
++	if (kvm_s390_pv_cpu_is_protected(vcpu)) {
++		newpsw = vcpu->arch.sie_block->gpsw;
++	} else {
++		rc = read_guest_lc(vcpu, __LC_EXT_NEW_PSW, &newpsw, sizeof(psw_t));
++		if (rc)
++			return rc;
++	}
++
++	/*
++	 * Clock comparator or timer interrupt with external interrupt enabled
++	 * will cause interrupt loop. Drop to userspace.
++	 */
+ 	if ((eic == EXT_IRQ_CLK_COMP || eic == EXT_IRQ_CPU_TIMER) &&
+ 	    (newpsw.mask & PSW_MASK_EXT))
+ 		return -EOPNOTSUPP;
+diff --git a/arch/x86/kernel/acpi/boot.c b/arch/x86/kernel/acpi/boot.c
+index 518bda50068cb..0f762070a5e10 100644
+--- a/arch/x86/kernel/acpi/boot.c
++++ b/arch/x86/kernel/acpi/boot.c
+@@ -146,7 +146,11 @@ static int __init acpi_parse_madt(struct acpi_table_header *table)
+ 
+ 		pr_debug("Local APIC address 0x%08x\n", madt->address);
+ 	}
+-	if (madt->header.revision >= 5)
++
++	/* ACPI 6.3 and newer support the online capable bit. */
++	if (acpi_gbl_FADT.header.revision > 6 ||
++	    (acpi_gbl_FADT.header.revision == 6 &&
++	     acpi_gbl_FADT.minor_revision >= 3))
+ 		acpi_support_online_capable = true;
+ 
+ 	default_acpi_madt_oem_check(madt->header.oem_id,
+@@ -193,7 +197,8 @@ static bool __init acpi_is_processor_usable(u32 lapic_flags)
+ 	if (lapic_flags & ACPI_MADT_ENABLED)
+ 		return true;
+ 
+-	if (acpi_support_online_capable && (lapic_flags & ACPI_MADT_ONLINE_CAPABLE))
++	if (!acpi_support_online_capable ||
++	    (lapic_flags & ACPI_MADT_ONLINE_CAPABLE))
+ 		return true;
+ 
+ 	return false;
+diff --git a/arch/x86/kvm/vmx/nested.c b/arch/x86/kvm/vmx/nested.c
+index 1d00f7824da1e..8e56ec6e72e9d 100644
+--- a/arch/x86/kvm/vmx/nested.c
++++ b/arch/x86/kvm/vmx/nested.c
+@@ -3845,7 +3845,12 @@ static void nested_vmx_inject_exception_vmexit(struct kvm_vcpu *vcpu)
+ 		exit_qual = 0;
+ 	}
+ 
+-	if (ex->has_error_code) {
++	/*
++	 * Unlike AMD's Paged Real Mode, which reports an error code on #PF
++	 * VM-Exits even if the CPU is in Real Mode, Intel VMX never sets the
++	 * "has error code" flags on VM-Exit if the CPU is in Real Mode.
++	 */
++	if (ex->has_error_code && is_protmode(vcpu)) {
+ 		/*
+ 		 * Intel CPUs do not generate error codes with bits 31:16 set,
+ 		 * and more importantly VMX disallows setting bits 31:16 in the
+diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
+index ab09d292bdede..3463ef7f30196 100644
+--- a/arch/x86/kvm/x86.c
++++ b/arch/x86/kvm/x86.c
+@@ -9853,13 +9853,20 @@ int kvm_check_nested_events(struct kvm_vcpu *vcpu)
+ 
+ static void kvm_inject_exception(struct kvm_vcpu *vcpu)
+ {
++	/*
++	 * Suppress the error code if the vCPU is in Real Mode, as Real Mode
++	 * exceptions don't report error codes.  The presence of an error code
++	 * is carried with the exception and only stripped when the exception
++	 * is injected as intercepted #PF VM-Exits for AMD's Paged Real Mode do
++	 * report an error code despite the CPU being in Real Mode.
++	 */
++	vcpu->arch.exception.has_error_code &= is_protmode(vcpu);
++
+ 	trace_kvm_inj_exception(vcpu->arch.exception.vector,
+ 				vcpu->arch.exception.has_error_code,
+ 				vcpu->arch.exception.error_code,
+ 				vcpu->arch.exception.injected);
+ 
+-	if (vcpu->arch.exception.error_code && !is_protmode(vcpu))
+-		vcpu->arch.exception.error_code = false;
+ 	static_call(kvm_x86_inject_exception)(vcpu);
+ }
+ 
+diff --git a/block/blk-mq.c b/block/blk-mq.c
+index aa67a52c5a069..1b04a1c48ee50 100644
+--- a/block/blk-mq.c
++++ b/block/blk-mq.c
+@@ -1327,8 +1327,6 @@ bool blk_rq_is_poll(struct request *rq)
+ 		return false;
+ 	if (rq->mq_hctx->type != HCTX_TYPE_POLL)
+ 		return false;
+-	if (WARN_ON_ONCE(!rq->bio))
+-		return false;
+ 	return true;
+ }
+ EXPORT_SYMBOL_GPL(blk_rq_is_poll);
+@@ -1336,7 +1334,7 @@ EXPORT_SYMBOL_GPL(blk_rq_is_poll);
+ static void blk_rq_poll_completion(struct request *rq, struct completion *wait)
+ {
+ 	do {
+-		bio_poll(rq->bio, NULL, 0);
++		blk_mq_poll(rq->q, blk_rq_to_qc(rq), NULL, 0);
+ 		cond_resched();
+ 	} while (!completion_done(wait));
+ }
+diff --git a/block/blk-throttle.c b/block/blk-throttle.c
+index 847721dc2b2b8..f1bc600c4ded6 100644
+--- a/block/blk-throttle.c
++++ b/block/blk-throttle.c
+@@ -1066,7 +1066,6 @@ static void tg_dispatch_one_bio(struct throtl_grp *tg, bool rw)
+ 	sq->nr_queued[rw]--;
+ 
+ 	throtl_charge_bio(tg, bio);
+-	bio_set_flag(bio, BIO_BPS_THROTTLED);
+ 
+ 	/*
+ 	 * If our parent is another tg, we just need to transfer @bio to
+@@ -1079,6 +1078,7 @@ static void tg_dispatch_one_bio(struct throtl_grp *tg, bool rw)
+ 		throtl_add_bio_tg(bio, &tg->qnode_on_parent[rw], parent_tg);
+ 		start_parent_slice_with_credit(tg, parent_tg, rw);
+ 	} else {
++		bio_set_flag(bio, BIO_BPS_THROTTLED);
+ 		throtl_qnode_add_bio(bio, &tg->qnode_on_parent[rw],
+ 				     &parent_sq->queued[rw]);
+ 		BUG_ON(tg->td->nr_queued[rw] <= 0);
+diff --git a/block/genhd.c b/block/genhd.c
+index 0b6928e948f31..62a61388e752d 100644
+--- a/block/genhd.c
++++ b/block/genhd.c
+@@ -368,7 +368,6 @@ int disk_scan_partitions(struct gendisk *disk, fmode_t mode)
+ 	if (disk->open_partitions)
+ 		return -EBUSY;
+ 
+-	set_bit(GD_NEED_PART_SCAN, &disk->state);
+ 	/*
+ 	 * If the device is opened exclusively by current thread already, it's
+ 	 * safe to scan partitons, otherwise, use bd_prepare_to_claim() to
+@@ -381,12 +380,19 @@ int disk_scan_partitions(struct gendisk *disk, fmode_t mode)
+ 			return ret;
+ 	}
+ 
++	set_bit(GD_NEED_PART_SCAN, &disk->state);
+ 	bdev = blkdev_get_by_dev(disk_devt(disk), mode & ~FMODE_EXCL, NULL);
+ 	if (IS_ERR(bdev))
+ 		ret =  PTR_ERR(bdev);
+ 	else
+ 		blkdev_put(bdev, mode & ~FMODE_EXCL);
+ 
++	/*
++	 * If blkdev_get_by_dev() failed early, GD_NEED_PART_SCAN is still set,
++	 * and this will cause that re-assemble partitioned raid device will
++	 * creat partition for underlying disk.
++	 */
++	clear_bit(GD_NEED_PART_SCAN, &disk->state);
+ 	if (!(mode & FMODE_EXCL))
+ 		bd_abort_claiming(disk->part0, disk_scan_partitions);
+ 	return ret;
+diff --git a/drivers/acpi/acpi_video.c b/drivers/acpi/acpi_video.c
+index 0c79f463fbfd4..ed318485eb192 100644
+--- a/drivers/acpi/acpi_video.c
++++ b/drivers/acpi/acpi_video.c
+@@ -1984,6 +1984,7 @@ static int instance;
+ static int acpi_video_bus_add(struct acpi_device *device)
+ {
+ 	struct acpi_video_bus *video;
++	bool auto_detect;
+ 	int error;
+ 	acpi_status status;
+ 
+@@ -2045,10 +2046,20 @@ static int acpi_video_bus_add(struct acpi_device *device)
+ 	mutex_unlock(&video_list_lock);
+ 
+ 	/*
+-	 * The userspace visible backlight_device gets registered separately
+-	 * from acpi_video_register_backlight().
++	 * If backlight-type auto-detection is used then a native backlight may
++	 * show up later and this may change the result from video to native.
++	 * Therefor normally the userspace visible /sys/class/backlight device
++	 * gets registered separately by the GPU driver calling
++	 * acpi_video_register_backlight() when an internal panel is detected.
++	 * Register the backlight now when not using auto-detection, so that
++	 * when the kernel cmdline or DMI-quirks are used the backlight will
++	 * get registered even if acpi_video_register_backlight() is not called.
+ 	 */
+ 	acpi_video_run_bcl_for_osi(video);
++	if (__acpi_video_get_backlight_type(false, &auto_detect) == acpi_backlight_video &&
++	    !auto_detect)
++		acpi_video_bus_register_backlight(video);
++
+ 	acpi_video_bus_add_notify_handler(video);
+ 
+ 	return 0;
+diff --git a/drivers/acpi/video_detect.c b/drivers/acpi/video_detect.c
+index f06b3d3556710..42b5af5490a11 100644
+--- a/drivers/acpi/video_detect.c
++++ b/drivers/acpi/video_detect.c
+@@ -274,6 +274,43 @@ static const struct dmi_system_id video_detect_dmi_table[] = {
+ 		},
+ 	},
+ 
++	/*
++	 * Models which need acpi_video backlight control where the GPU drivers
++	 * do not call acpi_video_register_backlight() because no internal panel
++	 * is detected. Typically these are all-in-ones (monitors with builtin
++	 * PC) where the panel connection shows up as regular DP instead of eDP.
++	 */
++	{
++	 .callback = video_detect_force_video,
++	 /* Apple iMac14,1 */
++	 .matches = {
++		DMI_MATCH(DMI_SYS_VENDOR, "Apple Inc."),
++		DMI_MATCH(DMI_PRODUCT_NAME, "iMac14,1"),
++		},
++	},
++	{
++	 .callback = video_detect_force_video,
++	 /* Apple iMac14,2 */
++	 .matches = {
++		DMI_MATCH(DMI_SYS_VENDOR, "Apple Inc."),
++		DMI_MATCH(DMI_PRODUCT_NAME, "iMac14,2"),
++		},
++	},
++
++	/*
++	 * Older models with nvidia GPU which need acpi_video backlight
++	 * control and where the old nvidia binary driver series does not
++	 * call acpi_video_register_backlight().
++	 */
++	{
++	 .callback = video_detect_force_video,
++	 /* ThinkPad W530 */
++	 .matches = {
++		DMI_MATCH(DMI_SYS_VENDOR, "LENOVO"),
++		DMI_MATCH(DMI_PRODUCT_VERSION, "ThinkPad W530"),
++		},
++	},
++
+ 	/*
+ 	 * These models have a working acpi_video backlight control, and using
+ 	 * native backlight causes a regression where backlight does not work
+@@ -772,7 +809,7 @@ static bool prefer_native_over_acpi_video(void)
+  * Determine which type of backlight interface to use on this system,
+  * First check cmdline, then dmi quirks, then do autodetect.
+  */
+-static enum acpi_backlight_type __acpi_video_get_backlight_type(bool native)
++enum acpi_backlight_type __acpi_video_get_backlight_type(bool native, bool *auto_detect)
+ {
+ 	static DEFINE_MUTEX(init_mutex);
+ 	static bool nvidia_wmi_ec_present;
+@@ -797,6 +834,9 @@ static enum acpi_backlight_type __acpi_video_get_backlight_type(bool native)
+ 		native_available = true;
+ 	mutex_unlock(&init_mutex);
+ 
++	if (auto_detect)
++		*auto_detect = false;
++
+ 	/*
+ 	 * The below heuristics / detection steps are in order of descending
+ 	 * presedence. The commandline takes presedence over anything else.
+@@ -808,6 +848,9 @@ static enum acpi_backlight_type __acpi_video_get_backlight_type(bool native)
+ 	if (acpi_backlight_dmi != acpi_backlight_undef)
+ 		return acpi_backlight_dmi;
+ 
++	if (auto_detect)
++		*auto_detect = true;
++
+ 	/* Special cases such as nvidia_wmi_ec and apple gmux. */
+ 	if (nvidia_wmi_ec_present)
+ 		return acpi_backlight_nvidia_wmi_ec;
+@@ -827,15 +870,4 @@ static enum acpi_backlight_type __acpi_video_get_backlight_type(bool native)
+ 	/* No ACPI video/native (old hw), use vendor specific fw methods. */
+ 	return acpi_backlight_vendor;
+ }
+-
+-enum acpi_backlight_type acpi_video_get_backlight_type(void)
+-{
+-	return __acpi_video_get_backlight_type(false);
+-}
+-EXPORT_SYMBOL(acpi_video_get_backlight_type);
+-
+-bool acpi_video_backlight_use_native(void)
+-{
+-	return __acpi_video_get_backlight_type(true) == acpi_backlight_native;
+-}
+-EXPORT_SYMBOL(acpi_video_backlight_use_native);
++EXPORT_SYMBOL(__acpi_video_get_backlight_type);
+diff --git a/drivers/block/ublk_drv.c b/drivers/block/ublk_drv.c
+index 22a790d512842..2ed994a313a91 100644
+--- a/drivers/block/ublk_drv.c
++++ b/drivers/block/ublk_drv.c
+@@ -233,7 +233,7 @@ static int ublk_validate_params(const struct ublk_device *ub)
+ 	if (ub->params.types & UBLK_PARAM_TYPE_BASIC) {
+ 		const struct ublk_param_basic *p = &ub->params.basic;
+ 
+-		if (p->logical_bs_shift > PAGE_SHIFT)
++		if (p->logical_bs_shift > PAGE_SHIFT || p->logical_bs_shift < 9)
+ 			return -EINVAL;
+ 
+ 		if (p->logical_bs_shift > p->physical_bs_shift)
+@@ -1202,9 +1202,10 @@ static void ublk_handle_need_get_data(struct ublk_device *ub, int q_id,
+ 	ublk_queue_cmd(ubq, req);
+ }
+ 
+-static int ublk_ch_uring_cmd(struct io_uring_cmd *cmd, unsigned int issue_flags)
++static int __ublk_ch_uring_cmd(struct io_uring_cmd *cmd,
++			       unsigned int issue_flags,
++			       struct ublksrv_io_cmd *ub_cmd)
+ {
+-	struct ublksrv_io_cmd *ub_cmd = (struct ublksrv_io_cmd *)cmd->cmd;
+ 	struct ublk_device *ub = cmd->file->private_data;
+ 	struct ublk_queue *ubq;
+ 	struct ublk_io *io;
+@@ -1306,6 +1307,23 @@ static int ublk_ch_uring_cmd(struct io_uring_cmd *cmd, unsigned int issue_flags)
+ 	return -EIOCBQUEUED;
+ }
+ 
++static int ublk_ch_uring_cmd(struct io_uring_cmd *cmd, unsigned int issue_flags)
++{
++	struct ublksrv_io_cmd *ub_src = (struct ublksrv_io_cmd *) cmd->cmd;
++	struct ublksrv_io_cmd ub_cmd;
++
++	/*
++	 * Not necessary for async retry, but let's keep it simple and always
++	 * copy the values to avoid any potential reuse.
++	 */
++	ub_cmd.q_id = READ_ONCE(ub_src->q_id);
++	ub_cmd.tag = READ_ONCE(ub_src->tag);
++	ub_cmd.result = READ_ONCE(ub_src->result);
++	ub_cmd.addr = READ_ONCE(ub_src->addr);
++
++	return __ublk_ch_uring_cmd(cmd, issue_flags, &ub_cmd);
++}
++
+ static const struct file_operations ublk_ch_fops = {
+ 	.owner = THIS_MODULE,
+ 	.open = ublk_ch_open,
+@@ -1886,6 +1904,8 @@ static int ublk_ctrl_set_params(struct io_uring_cmd *cmd)
+ 		/* clear all we don't support yet */
+ 		ub->params.types &= UBLK_PARAM_TYPE_ALL;
+ 		ret = ublk_validate_params(ub);
++		if (ret)
++			ub->params.types = 0;
+ 	}
+ 	mutex_unlock(&ub->mutex);
+ 	ublk_put_device(ub);
+diff --git a/drivers/counter/104-quad-8.c b/drivers/counter/104-quad-8.c
+index deed4afadb298..d9cb937665cfc 100644
+--- a/drivers/counter/104-quad-8.c
++++ b/drivers/counter/104-quad-8.c
+@@ -97,10 +97,6 @@ struct quad8 {
+ 	struct quad8_reg __iomem *reg;
+ };
+ 
+-/* Borrow Toggle flip-flop */
+-#define QUAD8_FLAG_BT BIT(0)
+-/* Carry Toggle flip-flop */
+-#define QUAD8_FLAG_CT BIT(1)
+ /* Error flag */
+ #define QUAD8_FLAG_E BIT(4)
+ /* Up/Down flag */
+@@ -133,6 +129,9 @@ struct quad8 {
+ #define QUAD8_CMR_QUADRATURE_X2 0x10
+ #define QUAD8_CMR_QUADRATURE_X4 0x18
+ 
++/* Each Counter is 24 bits wide */
++#define LS7267_CNTR_MAX GENMASK(23, 0)
++
+ static int quad8_signal_read(struct counter_device *counter,
+ 			     struct counter_signal *signal,
+ 			     enum counter_signal_level *level)
+@@ -156,18 +155,10 @@ static int quad8_count_read(struct counter_device *counter,
+ {
+ 	struct quad8 *const priv = counter_priv(counter);
+ 	struct channel_reg __iomem *const chan = priv->reg->channel + count->id;
+-	unsigned int flags;
+-	unsigned int borrow;
+-	unsigned int carry;
+ 	unsigned long irqflags;
+ 	int i;
+ 
+-	flags = ioread8(&chan->control);
+-	borrow = flags & QUAD8_FLAG_BT;
+-	carry = !!(flags & QUAD8_FLAG_CT);
+-
+-	/* Borrow XOR Carry effectively doubles count range */
+-	*val = (unsigned long)(borrow ^ carry) << 24;
++	*val = 0;
+ 
+ 	spin_lock_irqsave(&priv->lock, irqflags);
+ 
+@@ -191,8 +182,7 @@ static int quad8_count_write(struct counter_device *counter,
+ 	unsigned long irqflags;
+ 	int i;
+ 
+-	/* Only 24-bit values are supported */
+-	if (val > 0xFFFFFF)
++	if (val > LS7267_CNTR_MAX)
+ 		return -ERANGE;
+ 
+ 	spin_lock_irqsave(&priv->lock, irqflags);
+@@ -378,7 +368,7 @@ static int quad8_action_read(struct counter_device *counter,
+ 
+ 	/* Handle Index signals */
+ 	if (synapse->signal->id >= 16) {
+-		if (priv->preset_enable[count->id])
++		if (!priv->preset_enable[count->id])
+ 			*action = COUNTER_SYNAPSE_ACTION_RISING_EDGE;
+ 		else
+ 			*action = COUNTER_SYNAPSE_ACTION_NONE;
+@@ -806,8 +796,7 @@ static int quad8_count_preset_write(struct counter_device *counter,
+ 	struct quad8 *const priv = counter_priv(counter);
+ 	unsigned long irqflags;
+ 
+-	/* Only 24-bit values are supported */
+-	if (preset > 0xFFFFFF)
++	if (preset > LS7267_CNTR_MAX)
+ 		return -ERANGE;
+ 
+ 	spin_lock_irqsave(&priv->lock, irqflags);
+@@ -834,8 +823,7 @@ static int quad8_count_ceiling_read(struct counter_device *counter,
+ 		*ceiling = priv->preset[count->id];
+ 		break;
+ 	default:
+-		/* By default 0x1FFFFFF (25 bits unsigned) is maximum count */
+-		*ceiling = 0x1FFFFFF;
++		*ceiling = LS7267_CNTR_MAX;
+ 		break;
+ 	}
+ 
+@@ -850,8 +838,7 @@ static int quad8_count_ceiling_write(struct counter_device *counter,
+ 	struct quad8 *const priv = counter_priv(counter);
+ 	unsigned long irqflags;
+ 
+-	/* Only 24-bit values are supported */
+-	if (ceiling > 0xFFFFFF)
++	if (ceiling > LS7267_CNTR_MAX)
+ 		return -ERANGE;
+ 
+ 	spin_lock_irqsave(&priv->lock, irqflags);
+diff --git a/drivers/cxl/core/pci.c b/drivers/cxl/core/pci.c
+index 9240df53ed875..a8456d5441fc7 100644
+--- a/drivers/cxl/core/pci.c
++++ b/drivers/cxl/core/pci.c
+@@ -483,7 +483,7 @@ static struct pci_doe_mb *find_cdat_doe(struct device *uport)
+ 	return NULL;
+ }
+ 
+-#define CDAT_DOE_REQ(entry_handle)					\
++#define CDAT_DOE_REQ(entry_handle) cpu_to_le32				\
+ 	(FIELD_PREP(CXL_DOE_TABLE_ACCESS_REQ_CODE,			\
+ 		    CXL_DOE_TABLE_ACCESS_REQ_CODE_READ) |		\
+ 	 FIELD_PREP(CXL_DOE_TABLE_ACCESS_TABLE_TYPE,			\
+@@ -496,8 +496,8 @@ static void cxl_doe_task_complete(struct pci_doe_task *task)
+ }
+ 
+ struct cdat_doe_task {
+-	u32 request_pl;
+-	u32 response_pl[32];
++	__le32 request_pl;
++	__le32 response_pl[32];
+ 	struct completion c;
+ 	struct pci_doe_task task;
+ };
+@@ -531,10 +531,10 @@ static int cxl_cdat_get_length(struct device *dev,
+ 		return rc;
+ 	}
+ 	wait_for_completion(&t.c);
+-	if (t.task.rv < sizeof(u32))
++	if (t.task.rv < 2 * sizeof(__le32))
+ 		return -EIO;
+ 
+-	*length = t.response_pl[1];
++	*length = le32_to_cpu(t.response_pl[1]);
+ 	dev_dbg(dev, "CDAT length %zu\n", *length);
+ 
+ 	return 0;
+@@ -545,13 +545,13 @@ static int cxl_cdat_read_table(struct device *dev,
+ 			       struct cxl_cdat *cdat)
+ {
+ 	size_t length = cdat->length;
+-	u32 *data = cdat->table;
++	__le32 *data = cdat->table;
+ 	int entry_handle = 0;
+ 
+ 	do {
+ 		DECLARE_CDAT_DOE_TASK(CDAT_DOE_REQ(entry_handle), t);
++		struct cdat_entry_header *entry;
+ 		size_t entry_dw;
+-		u32 *entry;
+ 		int rc;
+ 
+ 		rc = pci_doe_submit_task(cdat_doe, &t.task);
+@@ -560,26 +560,34 @@ static int cxl_cdat_read_table(struct device *dev,
+ 			return rc;
+ 		}
+ 		wait_for_completion(&t.c);
+-		/* 1 DW header + 1 DW data min */
+-		if (t.task.rv < (2 * sizeof(u32)))
++
++		/* 1 DW Table Access Response Header + CDAT entry */
++		entry = (struct cdat_entry_header *)(t.response_pl + 1);
++		if ((entry_handle == 0 &&
++		     t.task.rv != sizeof(__le32) + sizeof(struct cdat_header)) ||
++		    (entry_handle > 0 &&
++		     (t.task.rv < sizeof(__le32) + sizeof(*entry) ||
++		      t.task.rv != sizeof(__le32) + le16_to_cpu(entry->length))))
+ 			return -EIO;
+ 
+ 		/* Get the CXL table access header entry handle */
+ 		entry_handle = FIELD_GET(CXL_DOE_TABLE_ACCESS_ENTRY_HANDLE,
+-					 t.response_pl[0]);
+-		entry = t.response_pl + 1;
+-		entry_dw = t.task.rv / sizeof(u32);
++					 le32_to_cpu(t.response_pl[0]));
++		entry_dw = t.task.rv / sizeof(__le32);
+ 		/* Skip Header */
+ 		entry_dw -= 1;
+-		entry_dw = min(length / sizeof(u32), entry_dw);
++		entry_dw = min(length / sizeof(__le32), entry_dw);
+ 		/* Prevent length < 1 DW from causing a buffer overflow */
+ 		if (entry_dw) {
+-			memcpy(data, entry, entry_dw * sizeof(u32));
+-			length -= entry_dw * sizeof(u32);
++			memcpy(data, entry, entry_dw * sizeof(__le32));
++			length -= entry_dw * sizeof(__le32);
+ 			data += entry_dw;
+ 		}
+ 	} while (entry_handle != CXL_DOE_TABLE_ACCESS_LAST_ENTRY);
+ 
++	/* Length in CDAT header may exceed concatenation of CDAT entries */
++	cdat->length -= length;
++
+ 	return 0;
+ }
+ 
+diff --git a/drivers/cxl/cxlpci.h b/drivers/cxl/cxlpci.h
+index eec597dbe763a..79e5603dfc82d 100644
+--- a/drivers/cxl/cxlpci.h
++++ b/drivers/cxl/cxlpci.h
+@@ -71,6 +71,20 @@ static inline resource_size_t cxl_regmap_to_base(struct pci_dev *pdev,
+ 	return pci_resource_start(pdev, map->barno) + map->block_offset;
+ }
+ 
++struct cdat_header {
++	__le32 length;
++	u8 revision;
++	u8 checksum;
++	u8 reserved[6];
++	__le32 sequence;
++} __packed;
++
++struct cdat_entry_header {
++	u8 type;
++	u8 reserved;
++	__le16 length;
++} __packed;
++
+ int devm_cxl_port_enumerate_dports(struct cxl_port *port);
+ struct cxl_dev_state;
+ int cxl_hdm_decode_init(struct cxl_dev_state *cxlds, struct cxl_hdm *cxlhdm);
+diff --git a/drivers/gpio/Kconfig b/drivers/gpio/Kconfig
+index a01af11806164..e3af86f06c630 100644
+--- a/drivers/gpio/Kconfig
++++ b/drivers/gpio/Kconfig
+@@ -100,7 +100,7 @@ config GPIO_GENERIC
+ 	tristate
+ 
+ config GPIO_REGMAP
+-	depends on REGMAP
++	select REGMAP
+ 	tristate
+ 
+ # put drivers in the right section, in alphabetical order
+diff --git a/drivers/gpio/gpio-davinci.c b/drivers/gpio/gpio-davinci.c
+index 59c4c48d8296b..69f3d864f69d3 100644
+--- a/drivers/gpio/gpio-davinci.c
++++ b/drivers/gpio/gpio-davinci.c
+@@ -328,7 +328,7 @@ static struct irq_chip gpio_irqchip = {
+ 	.irq_enable	= gpio_irq_enable,
+ 	.irq_disable	= gpio_irq_disable,
+ 	.irq_set_type	= gpio_irq_type,
+-	.flags		= IRQCHIP_SET_TYPE_MASKED,
++	.flags		= IRQCHIP_SET_TYPE_MASKED | IRQCHIP_SKIP_SET_WAKE,
+ };
+ 
+ static void gpio_irq_handler(struct irq_desc *desc)
+@@ -645,9 +645,6 @@ static void davinci_gpio_save_context(struct davinci_gpio_controller *chips,
+ 		context->set_falling = readl_relaxed(&g->set_falling);
+ 	}
+ 
+-	/* Clear Bank interrupt enable bit */
+-	writel_relaxed(0, base + BINTEN);
+-
+ 	/* Clear all interrupt status registers */
+ 	writel_relaxed(GENMASK(31, 0), &g->intstat);
+ }
+diff --git a/drivers/gpio/gpio-mvebu.c b/drivers/gpio/gpio-mvebu.c
+index 1bb317b8dccea..91a4232ee58c2 100644
+--- a/drivers/gpio/gpio-mvebu.c
++++ b/drivers/gpio/gpio-mvebu.c
+@@ -657,9 +657,10 @@ static void mvebu_pwm_free(struct pwm_chip *chip, struct pwm_device *pwm)
+ 	spin_unlock_irqrestore(&mvpwm->lock, flags);
+ }
+ 
+-static void mvebu_pwm_get_state(struct pwm_chip *chip,
+-				struct pwm_device *pwm,
+-				struct pwm_state *state) {
++static int mvebu_pwm_get_state(struct pwm_chip *chip,
++			       struct pwm_device *pwm,
++			       struct pwm_state *state)
++{
+ 
+ 	struct mvebu_pwm *mvpwm = to_mvebu_pwm(chip);
+ 	struct mvebu_gpio_chip *mvchip = mvpwm->mvchip;
+@@ -693,6 +694,8 @@ static void mvebu_pwm_get_state(struct pwm_chip *chip,
+ 		state->enabled = false;
+ 
+ 	spin_unlock_irqrestore(&mvpwm->lock, flags);
++
++	return 0;
+ }
+ 
+ static int mvebu_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
+index 9ef83f3ab3a7e..9df5dcedaf3e2 100644
+--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
+@@ -3034,6 +3034,24 @@ static int amdgpu_device_ip_suspend_phase2(struct amdgpu_device *adev)
+ 		     adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_MES))
+ 			continue;
+ 
++		/* SDMA 5.x+ is part of GFX power domain so it's covered by GFXOFF */
++		if (adev->in_s0ix &&
++		    (adev->ip_versions[SDMA0_HWIP][0] >= IP_VERSION(5, 0, 0)) &&
++		    (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_SDMA))
++			continue;
++
++		/* Once swPSP provides the IMU, RLC FW binaries to TOS during cold-boot.
++		 * These are in TMR, hence are expected to be reused by PSP-TOS to reload
++		 * from this location and RLC Autoload automatically also gets loaded
++		 * from here based on PMFW -> PSP message during re-init sequence.
++		 * Therefore, the psp suspend & resume should be skipped to avoid destroy
++		 * the TMR and reload FWs again for IMU enabled APU ASICs.
++		 */
++		if (amdgpu_in_reset(adev) &&
++		    (adev->flags & AMD_IS_APU) && adev->gfx.imu.funcs &&
++		    adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_PSP)
++			continue;
++
+ 		/* XXX handle errors */
+ 		r = adev->ip_blocks[i].version->funcs->suspend(adev);
+ 		/* XXX handle errors */
+diff --git a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
+index 050e7a52c8f62..6c5ea99223bab 100644
+--- a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
++++ b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
+@@ -2175,6 +2175,8 @@ static int detect_mst_link_for_all_connectors(struct drm_device *dev)
+ 				DRM_ERROR("DM_MST: Failed to start MST\n");
+ 				aconnector->dc_link->type =
+ 					dc_connection_single;
++				ret = dm_helpers_dp_mst_stop_top_mgr(aconnector->dc_link->ctx,
++								     aconnector->dc_link);
+ 				break;
+ 			}
+ 		}
+diff --git a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_helpers.c b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_helpers.c
+index 57454967617f8..6bd7e45370141 100644
+--- a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_helpers.c
++++ b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_helpers.c
+@@ -206,7 +206,7 @@ bool dm_helpers_dp_mst_write_payload_allocation_table(
+ 	if (enable)
+ 		drm_dp_add_payload_part1(mst_mgr, mst_state, payload);
+ 	else
+-		drm_dp_remove_payload(mst_mgr, mst_state, payload);
++		drm_dp_remove_payload(mst_mgr, mst_state, payload, payload);
+ 
+ 	/* mst_mgr->->payloads are VC payload notify MST branch using DPCD or
+ 	 * AUX message. The sequence is slot 1-63 allocated sequence for each
+diff --git a/drivers/gpu/drm/bridge/lontium-lt9611.c b/drivers/gpu/drm/bridge/lontium-lt9611.c
+index 3b77238ca4aff..ae8c6d9d4095f 100644
+--- a/drivers/gpu/drm/bridge/lontium-lt9611.c
++++ b/drivers/gpu/drm/bridge/lontium-lt9611.c
+@@ -258,6 +258,7 @@ static int lt9611_pll_setup(struct lt9611 *lt9611, const struct drm_display_mode
+ 		{ 0x8126, 0x55 },
+ 		{ 0x8127, 0x66 },
+ 		{ 0x8128, 0x88 },
++		{ 0x812a, 0x20 },
+ 	};
+ 
+ 	regmap_multi_reg_write(lt9611->regmap, reg_cfg, ARRAY_SIZE(reg_cfg));
+diff --git a/drivers/gpu/drm/bridge/ti-sn65dsi86.c b/drivers/gpu/drm/bridge/ti-sn65dsi86.c
+index eb24322df721a..aeca9c066bf29 100644
+--- a/drivers/gpu/drm/bridge/ti-sn65dsi86.c
++++ b/drivers/gpu/drm/bridge/ti-sn65dsi86.c
+@@ -1500,8 +1500,8 @@ out:
+ 	return ret;
+ }
+ 
+-static void ti_sn_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-				struct pwm_state *state)
++static int ti_sn_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++			       struct pwm_state *state)
+ {
+ 	struct ti_sn65dsi86 *pdata = pwm_chip_to_ti_sn_bridge(chip);
+ 	unsigned int pwm_en_inv;
+@@ -1512,19 +1512,19 @@ static void ti_sn_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 
+ 	ret = regmap_read(pdata->regmap, SN_PWM_EN_INV_REG, &pwm_en_inv);
+ 	if (ret)
+-		return;
++		return 0;
+ 
+ 	ret = ti_sn65dsi86_read_u16(pdata, SN_BACKLIGHT_SCALE_REG, &scale);
+ 	if (ret)
+-		return;
++		return 0;
+ 
+ 	ret = ti_sn65dsi86_read_u16(pdata, SN_BACKLIGHT_REG, &backlight);
+ 	if (ret)
+-		return;
++		return 0;
+ 
+ 	ret = regmap_read(pdata->regmap, SN_PWM_PRE_DIV_REG, &pre_div);
+ 	if (ret)
+-		return;
++		return 0;
+ 
+ 	state->enabled = FIELD_GET(SN_PWM_EN_MASK, pwm_en_inv);
+ 	if (FIELD_GET(SN_PWM_INV_MASK, pwm_en_inv))
+@@ -1539,6 +1539,8 @@ static void ti_sn_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 
+ 	if (state->duty_cycle > state->period)
+ 		state->duty_cycle = state->period;
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops ti_sn_pwm_ops = {
+diff --git a/drivers/gpu/drm/display/drm_dp_mst_topology.c b/drivers/gpu/drm/display/drm_dp_mst_topology.c
+index e77c674b37ca2..38dab76ae69ea 100644
+--- a/drivers/gpu/drm/display/drm_dp_mst_topology.c
++++ b/drivers/gpu/drm/display/drm_dp_mst_topology.c
+@@ -3342,7 +3342,8 @@ EXPORT_SYMBOL(drm_dp_add_payload_part1);
+  * drm_dp_remove_payload() - Remove an MST payload
+  * @mgr: Manager to use.
+  * @mst_state: The MST atomic state
+- * @payload: The payload to write
++ * @old_payload: The payload with its old state
++ * @new_payload: The payload to write
+  *
+  * Removes a payload from an MST topology if it was successfully assigned a start slot. Also updates
+  * the starting time slots of all other payloads which would have been shifted towards the start of
+@@ -3350,36 +3351,37 @@ EXPORT_SYMBOL(drm_dp_add_payload_part1);
+  */
+ void drm_dp_remove_payload(struct drm_dp_mst_topology_mgr *mgr,
+ 			   struct drm_dp_mst_topology_state *mst_state,
+-			   struct drm_dp_mst_atomic_payload *payload)
++			   const struct drm_dp_mst_atomic_payload *old_payload,
++			   struct drm_dp_mst_atomic_payload *new_payload)
+ {
+ 	struct drm_dp_mst_atomic_payload *pos;
+ 	bool send_remove = false;
+ 
+ 	/* We failed to make the payload, so nothing to do */
+-	if (payload->vc_start_slot == -1)
++	if (new_payload->vc_start_slot == -1)
+ 		return;
+ 
+ 	mutex_lock(&mgr->lock);
+-	send_remove = drm_dp_mst_port_downstream_of_branch(payload->port, mgr->mst_primary);
++	send_remove = drm_dp_mst_port_downstream_of_branch(new_payload->port, mgr->mst_primary);
+ 	mutex_unlock(&mgr->lock);
+ 
+ 	if (send_remove)
+-		drm_dp_destroy_payload_step1(mgr, mst_state, payload);
++		drm_dp_destroy_payload_step1(mgr, mst_state, new_payload);
+ 	else
+ 		drm_dbg_kms(mgr->dev, "Payload for VCPI %d not in topology, not sending remove\n",
+-			    payload->vcpi);
++			    new_payload->vcpi);
+ 
+ 	list_for_each_entry(pos, &mst_state->payloads, next) {
+-		if (pos != payload && pos->vc_start_slot > payload->vc_start_slot)
+-			pos->vc_start_slot -= payload->time_slots;
++		if (pos != new_payload && pos->vc_start_slot > new_payload->vc_start_slot)
++			pos->vc_start_slot -= old_payload->time_slots;
+ 	}
+-	payload->vc_start_slot = -1;
++	new_payload->vc_start_slot = -1;
+ 
+ 	mgr->payload_count--;
+-	mgr->next_start_slot -= payload->time_slots;
++	mgr->next_start_slot -= old_payload->time_slots;
+ 
+-	if (payload->delete)
+-		drm_dp_mst_put_port_malloc(payload->port);
++	if (new_payload->delete)
++		drm_dp_mst_put_port_malloc(new_payload->port);
+ }
+ EXPORT_SYMBOL(drm_dp_remove_payload);
+ 
+diff --git a/drivers/gpu/drm/i915/display/intel_color.c b/drivers/gpu/drm/i915/display/intel_color.c
+index c85757f55112b..78211e583fd7e 100644
+--- a/drivers/gpu/drm/i915/display/intel_color.c
++++ b/drivers/gpu/drm/i915/display/intel_color.c
+@@ -2098,6 +2098,25 @@ static void glk_read_luts(struct intel_crtc_state *crtc_state)
+ 	}
+ }
+ 
++static void icl_color_commit_arm(const struct intel_crtc_state *crtc_state)
++{
++	struct intel_crtc *crtc = to_intel_crtc(crtc_state->uapi.crtc);
++	struct drm_i915_private *i915 = to_i915(crtc->base.dev);
++	enum pipe pipe = crtc->pipe;
++
++	/*
++	 * We don't (yet) allow userspace to control the pipe background color,
++	 * so force it to black.
++	 */
++	intel_de_write(i915, SKL_BOTTOM_COLOR(pipe), 0);
++
++	intel_de_write(i915, GAMMA_MODE(crtc->pipe),
++		       crtc_state->gamma_mode);
++
++	intel_de_write_fw(i915, PIPE_CSC_MODE(crtc->pipe),
++			  crtc_state->csc_mode);
++}
++
+ static struct drm_property_blob *
+ icl_read_lut_multi_segment(struct intel_crtc *crtc)
+ {
+@@ -2183,7 +2202,7 @@ static const struct intel_color_funcs i9xx_color_funcs = {
+ static const struct intel_color_funcs icl_color_funcs = {
+ 	.color_check = icl_color_check,
+ 	.color_commit_noarm = icl_color_commit_noarm,
+-	.color_commit_arm = skl_color_commit_arm,
++	.color_commit_arm = icl_color_commit_arm,
+ 	.load_luts = icl_load_luts,
+ 	.read_luts = icl_read_luts,
+ };
+diff --git a/drivers/gpu/drm/i915/display/intel_dp_mst.c b/drivers/gpu/drm/i915/display/intel_dp_mst.c
+index 27c2098dd7070..9a6822256ddf6 100644
+--- a/drivers/gpu/drm/i915/display/intel_dp_mst.c
++++ b/drivers/gpu/drm/i915/display/intel_dp_mst.c
+@@ -364,8 +364,14 @@ static void intel_mst_disable_dp(struct intel_atomic_state *state,
+ 	struct intel_dp *intel_dp = &dig_port->dp;
+ 	struct intel_connector *connector =
+ 		to_intel_connector(old_conn_state->connector);
+-	struct drm_dp_mst_topology_state *mst_state =
+-		drm_atomic_get_mst_topology_state(&state->base, &intel_dp->mst_mgr);
++	struct drm_dp_mst_topology_state *old_mst_state =
++		drm_atomic_get_old_mst_topology_state(&state->base, &intel_dp->mst_mgr);
++	struct drm_dp_mst_topology_state *new_mst_state =
++		drm_atomic_get_new_mst_topology_state(&state->base, &intel_dp->mst_mgr);
++	const struct drm_dp_mst_atomic_payload *old_payload =
++		drm_atomic_get_mst_payload_state(old_mst_state, connector->port);
++	struct drm_dp_mst_atomic_payload *new_payload =
++		drm_atomic_get_mst_payload_state(new_mst_state, connector->port);
+ 	struct drm_i915_private *i915 = to_i915(connector->base.dev);
+ 
+ 	drm_dbg_kms(&i915->drm, "active links %d\n",
+@@ -373,8 +379,8 @@ static void intel_mst_disable_dp(struct intel_atomic_state *state,
+ 
+ 	intel_hdcp_disable(intel_mst->connector);
+ 
+-	drm_dp_remove_payload(&intel_dp->mst_mgr, mst_state,
+-			      drm_atomic_get_mst_payload_state(mst_state, connector->port));
++	drm_dp_remove_payload(&intel_dp->mst_mgr, new_mst_state,
++			      old_payload, new_payload);
+ 
+ 	intel_audio_codec_disable(encoder, old_crtc_state, old_conn_state);
+ }
+diff --git a/drivers/gpu/drm/i915/gt/intel_execlists_submission.c b/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
+index bfd1ffc71a489..fc4a846289855 100644
+--- a/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
++++ b/drivers/gpu/drm/i915/gt/intel_execlists_submission.c
+@@ -2017,6 +2017,8 @@ process_csb(struct intel_engine_cs *engine, struct i915_request **inactive)
+ 	 * inspecting the queue to see if we need to resumbit.
+ 	 */
+ 	if (*prev != *execlists->active) { /* elide lite-restores */
++		struct intel_context *prev_ce = NULL, *active_ce = NULL;
++
+ 		/*
+ 		 * Note the inherent discrepancy between the HW runtime,
+ 		 * recorded as part of the context switch, and the CPU
+@@ -2028,9 +2030,15 @@ process_csb(struct intel_engine_cs *engine, struct i915_request **inactive)
+ 		 * and correct overselves later when updating from HW.
+ 		 */
+ 		if (*prev)
+-			lrc_runtime_stop((*prev)->context);
++			prev_ce = (*prev)->context;
+ 		if (*execlists->active)
+-			lrc_runtime_start((*execlists->active)->context);
++			active_ce = (*execlists->active)->context;
++		if (prev_ce != active_ce) {
++			if (prev_ce)
++				lrc_runtime_stop(prev_ce);
++			if (active_ce)
++				lrc_runtime_start(active_ce);
++		}
+ 		new_timeslice(execlists);
+ 	}
+ 
+diff --git a/drivers/gpu/drm/i915/i915_perf.c b/drivers/gpu/drm/i915/i915_perf.c
+index 0defbb43ceea8..3ce49c118b83f 100644
+--- a/drivers/gpu/drm/i915/i915_perf.c
++++ b/drivers/gpu/drm/i915/i915_perf.c
+@@ -4270,13 +4270,13 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data,
+ 		err = oa_config->id;
+ 		goto sysfs_err;
+ 	}
+-
+-	mutex_unlock(&perf->metrics_lock);
++	id = oa_config->id;
+ 
+ 	drm_dbg(&perf->i915->drm,
+ 		"Added config %s id=%i\n", oa_config->uuid, oa_config->id);
++	mutex_unlock(&perf->metrics_lock);
+ 
+-	return oa_config->id;
++	return id;
+ 
+ sysfs_err:
+ 	mutex_unlock(&perf->metrics_lock);
+diff --git a/drivers/gpu/drm/i915/i915_reg.h b/drivers/gpu/drm/i915/i915_reg.h
+index 64eacd11b8bff..2372bfa04aa87 100644
+--- a/drivers/gpu/drm/i915/i915_reg.h
++++ b/drivers/gpu/drm/i915/i915_reg.h
+@@ -3747,9 +3747,10 @@
+ 
+ /* Skylake+ pipe bottom (background) color */
+ #define _SKL_BOTTOM_COLOR_A		0x70034
++#define _SKL_BOTTOM_COLOR_B		0x71034
+ #define   SKL_BOTTOM_COLOR_GAMMA_ENABLE		REG_BIT(31)
+ #define   SKL_BOTTOM_COLOR_CSC_ENABLE		REG_BIT(30)
+-#define SKL_BOTTOM_COLOR(pipe)		_MMIO_PIPE2(pipe, _SKL_BOTTOM_COLOR_A)
++#define SKL_BOTTOM_COLOR(pipe)		_MMIO_PIPE(pipe, _SKL_BOTTOM_COLOR_A, _SKL_BOTTOM_COLOR_B)
+ 
+ #define _ICL_PIPE_A_STATUS			0x70058
+ #define ICL_PIPESTATUS(pipe)			_MMIO_PIPE2(pipe, _ICL_PIPE_A_STATUS)
+diff --git a/drivers/gpu/drm/nouveau/dispnv50/disp.c b/drivers/gpu/drm/nouveau/dispnv50/disp.c
+index 33c97d510999b..006cb76adaa93 100644
+--- a/drivers/gpu/drm/nouveau/dispnv50/disp.c
++++ b/drivers/gpu/drm/nouveau/dispnv50/disp.c
+@@ -410,6 +410,35 @@ nv50_outp_atomic_check_view(struct drm_encoder *encoder,
+ 	return 0;
+ }
+ 
++static void
++nv50_outp_atomic_fix_depth(struct drm_encoder *encoder, struct drm_crtc_state *crtc_state)
++{
++	struct nv50_head_atom *asyh = nv50_head_atom(crtc_state);
++	struct nouveau_encoder *nv_encoder = nouveau_encoder(encoder);
++	struct drm_display_mode *mode = &asyh->state.adjusted_mode;
++	unsigned int max_rate, mode_rate;
++
++	switch (nv_encoder->dcb->type) {
++	case DCB_OUTPUT_DP:
++		max_rate = nv_encoder->dp.link_nr * nv_encoder->dp.link_bw;
++
++		/* we don't support more than 10 anyway */
++		asyh->or.bpc = min_t(u8, asyh->or.bpc, 10);
++
++		/* reduce the bpc until it works out */
++		while (asyh->or.bpc > 6) {
++			mode_rate = DIV_ROUND_UP(mode->clock * asyh->or.bpc * 3, 8);
++			if (mode_rate <= max_rate)
++				break;
++
++			asyh->or.bpc -= 2;
++		}
++		break;
++	default:
++		break;
++	}
++}
++
+ static int
+ nv50_outp_atomic_check(struct drm_encoder *encoder,
+ 		       struct drm_crtc_state *crtc_state,
+@@ -428,6 +457,9 @@ nv50_outp_atomic_check(struct drm_encoder *encoder,
+ 	if (crtc_state->mode_changed || crtc_state->connectors_changed)
+ 		asyh->or.bpc = connector->display_info.bpc;
+ 
++	/* We might have to reduce the bpc */
++	nv50_outp_atomic_fix_depth(encoder, crtc_state);
++
+ 	return 0;
+ }
+ 
+@@ -996,7 +1028,7 @@ nv50_msto_prepare(struct drm_atomic_state *state,
+ 
+ 	// TODO: Figure out if we want to do a better job of handling VCPI allocation failures here?
+ 	if (msto->disabled) {
+-		drm_dp_remove_payload(mgr, mst_state, payload);
++		drm_dp_remove_payload(mgr, mst_state, payload, payload);
+ 	} else {
+ 		if (msto->enabled)
+ 			drm_dp_add_payload_part1(mgr, mst_state, payload);
+diff --git a/drivers/gpu/drm/nouveau/nouveau_dp.c b/drivers/gpu/drm/nouveau/nouveau_dp.c
+index 20db8ea1a0baf..53185746fb3d1 100644
+--- a/drivers/gpu/drm/nouveau/nouveau_dp.c
++++ b/drivers/gpu/drm/nouveau/nouveau_dp.c
+@@ -245,8 +245,6 @@ void nouveau_dp_irq(struct nouveau_drm *drm,
+ }
+ 
+ /* TODO:
+- * - Use the minimum possible BPC here, once we add support for the max bpc
+- *   property.
+  * - Validate against the DP caps advertised by the GPU (we don't check these
+  *   yet)
+  */
+@@ -258,7 +256,11 @@ nv50_dp_mode_valid(struct drm_connector *connector,
+ {
+ 	const unsigned int min_clock = 25000;
+ 	unsigned int max_rate, mode_rate, ds_max_dotclock, clock = mode->clock;
+-	const u8 bpp = connector->display_info.bpc * 3;
++	/* Check with the minmum bpc always, so we can advertise better modes.
++	 * In particlar not doing this causes modes to be dropped on HDR
++	 * displays as we might check with a bpc of 16 even.
++	 */
++	const u8 bpp = 6 * 3;
+ 
+ 	if (mode->flags & DRM_MODE_FLAG_INTERLACE && !outp->caps.dp_interlace)
+ 		return MODE_NO_INTERLACE;
+diff --git a/drivers/gpu/drm/panfrost/panfrost_mmu.c b/drivers/gpu/drm/panfrost/panfrost_mmu.c
+index 666a5e53fe193..e961fa27702ce 100644
+--- a/drivers/gpu/drm/panfrost/panfrost_mmu.c
++++ b/drivers/gpu/drm/panfrost/panfrost_mmu.c
+@@ -504,6 +504,7 @@ static int panfrost_mmu_map_fault_addr(struct panfrost_device *pfdev, int as,
+ 		if (IS_ERR(pages[i])) {
+ 			mutex_unlock(&bo->base.pages_lock);
+ 			ret = PTR_ERR(pages[i]);
++			pages[i] = NULL;
+ 			goto err_pages;
+ 		}
+ 	}
+diff --git a/drivers/hv/connection.c b/drivers/hv/connection.c
+index 9dc27e5d367a2..da51b50787dff 100644
+--- a/drivers/hv/connection.c
++++ b/drivers/hv/connection.c
+@@ -409,6 +409,10 @@ void vmbus_disconnect(void)
+  */
+ struct vmbus_channel *relid2channel(u32 relid)
+ {
++	if (vmbus_connection.channels == NULL) {
++		pr_warn_once("relid2channel: relid=%d: No channels mapped!\n", relid);
++		return NULL;
++	}
+ 	if (WARN_ON(relid >= MAX_CHANNEL_RELIDS))
+ 		return NULL;
+ 	return READ_ONCE(vmbus_connection.channels[relid]);
+diff --git a/drivers/hwtracing/coresight/coresight-etm4x-core.c b/drivers/hwtracing/coresight/coresight-etm4x-core.c
+index c7a65d1524fcb..1cf7478da6ee8 100644
+--- a/drivers/hwtracing/coresight/coresight-etm4x-core.c
++++ b/drivers/hwtracing/coresight/coresight-etm4x-core.c
+@@ -451,7 +451,7 @@ static int etm4_enable_hw(struct etmv4_drvdata *drvdata)
+ 		if (etm4x_sspcicrn_present(drvdata, i))
+ 			etm4x_relaxed_write32(csa, config->ss_pe_cmp[i], TRCSSPCICRn(i));
+ 	}
+-	for (i = 0; i < drvdata->nr_addr_cmp; i++) {
++	for (i = 0; i < drvdata->nr_addr_cmp * 2; i++) {
+ 		etm4x_relaxed_write64(csa, config->addr_val[i], TRCACVRn(i));
+ 		etm4x_relaxed_write64(csa, config->addr_acc[i], TRCACATRn(i));
+ 	}
+@@ -1010,25 +1010,21 @@ static bool etm4_init_iomem_access(struct etmv4_drvdata *drvdata,
+ 				   struct csdev_access *csa)
+ {
+ 	u32 devarch = readl_relaxed(drvdata->base + TRCDEVARCH);
+-	u32 idr1 = readl_relaxed(drvdata->base + TRCIDR1);
+ 
+ 	/*
+ 	 * All ETMs must implement TRCDEVARCH to indicate that
+-	 * the component is an ETMv4. To support any broken
+-	 * implementations we fall back to TRCIDR1 check, which
+-	 * is not really reliable.
++	 * the component is an ETMv4. Even though TRCIDR1 also
++	 * contains the information, it is part of the "Trace"
++	 * register and must be accessed with the OSLK cleared,
++	 * with MMIO. But we cannot touch the OSLK until we are
++	 * sure this is an ETM. So rely only on the TRCDEVARCH.
+ 	 */
+-	if ((devarch & ETM_DEVARCH_ID_MASK) == ETM_DEVARCH_ETMv4x_ARCH) {
+-		drvdata->arch = etm_devarch_to_arch(devarch);
+-	} else {
+-		pr_warn("CPU%d: ETM4x incompatible TRCDEVARCH: %x, falling back to TRCIDR1\n",
+-			smp_processor_id(), devarch);
+-
+-		if (ETM_TRCIDR1_ARCH_MAJOR(idr1) != ETM_TRCIDR1_ARCH_ETMv4)
+-			return false;
+-		drvdata->arch = etm_trcidr_to_arch(idr1);
++	if ((devarch & ETM_DEVARCH_ID_MASK) != ETM_DEVARCH_ETMv4x_ARCH) {
++		pr_warn_once("TRCDEVARCH doesn't match ETMv4 architecture\n");
++		return false;
+ 	}
+ 
++	drvdata->arch = etm_devarch_to_arch(devarch);
+ 	*csa = CSDEV_ACCESS_IOMEM(drvdata->base);
+ 	return true;
+ }
+diff --git a/drivers/hwtracing/coresight/coresight-etm4x.h b/drivers/hwtracing/coresight/coresight-etm4x.h
+index 4b21bb79f1682..0174fbf1a9637 100644
+--- a/drivers/hwtracing/coresight/coresight-etm4x.h
++++ b/drivers/hwtracing/coresight/coresight-etm4x.h
+@@ -753,14 +753,12 @@
+  * TRCDEVARCH	- CoreSight architected register
+  *                - Bits[15:12] - Major version
+  *                - Bits[19:16] - Minor version
+- * TRCIDR1	- ETM architected register
+- *                - Bits[11:8] - Major version
+- *                - Bits[7:4]  - Minor version
+- * We must rely on TRCDEVARCH for the version information,
+- * however we don't want to break the support for potential
+- * old implementations which might not implement it. Thus
+- * we fall back to TRCIDR1 if TRCDEVARCH is not implemented
+- * for memory mapped components.
++ *
++ * We must rely only on TRCDEVARCH for the version information. Even though,
++ * TRCIDR1 also provides the architecture version, it is a "Trace" register
++ * and as such must be accessed only with Trace power domain ON. This may
++ * not be available at probe time.
++ *
+  * Now to make certain decisions easier based on the version
+  * we use an internal representation of the version in the
+  * driver, as follows :
+@@ -786,12 +784,6 @@ static inline u8 etm_devarch_to_arch(u32 devarch)
+ 				ETM_DEVARCH_REVISION(devarch));
+ }
+ 
+-static inline u8 etm_trcidr_to_arch(u32 trcidr1)
+-{
+-	return ETM_ARCH_VERSION(ETM_TRCIDR1_ARCH_MAJOR(trcidr1),
+-				ETM_TRCIDR1_ARCH_MINOR(trcidr1));
+-}
+-
+ enum etm_impdef_type {
+ 	ETM4_IMPDEF_HISI_CORE_COMMIT,
+ 	ETM4_IMPDEF_FEATURE_MAX,
+diff --git a/drivers/iio/adc/ad7791.c b/drivers/iio/adc/ad7791.c
+index fee8d129a5f08..86effe8501b44 100644
+--- a/drivers/iio/adc/ad7791.c
++++ b/drivers/iio/adc/ad7791.c
+@@ -253,7 +253,7 @@ static const struct ad_sigma_delta_info ad7791_sigma_delta_info = {
+ 	.has_registers = true,
+ 	.addr_shift = 4,
+ 	.read_mask = BIT(3),
+-	.irq_flags = IRQF_TRIGGER_LOW,
++	.irq_flags = IRQF_TRIGGER_FALLING,
+ };
+ 
+ static int ad7791_read_raw(struct iio_dev *indio_dev,
+diff --git a/drivers/iio/adc/ltc2497.c b/drivers/iio/adc/ltc2497.c
+index 556f10dfb5023..e1e14f5d237d2 100644
+--- a/drivers/iio/adc/ltc2497.c
++++ b/drivers/iio/adc/ltc2497.c
+@@ -28,7 +28,6 @@ struct ltc2497_driverdata {
+ 	struct ltc2497core_driverdata common_ddata;
+ 	struct i2c_client *client;
+ 	u32 recv_size;
+-	u32 sub_lsb;
+ 	/*
+ 	 * DMA (thus cache coherency maintenance) may require the
+ 	 * transfer buffers to live in their own cache lines.
+@@ -65,10 +64,10 @@ static int ltc2497_result_and_measure(struct ltc2497core_driverdata *ddata,
+ 		 * equivalent to a sign extension.
+ 		 */
+ 		if (st->recv_size == 3) {
+-			*val = (get_unaligned_be24(st->data.d8) >> st->sub_lsb)
++			*val = (get_unaligned_be24(st->data.d8) >> 6)
+ 				- BIT(ddata->chip_info->resolution + 1);
+ 		} else {
+-			*val = (be32_to_cpu(st->data.d32) >> st->sub_lsb)
++			*val = (be32_to_cpu(st->data.d32) >> 6)
+ 				- BIT(ddata->chip_info->resolution + 1);
+ 		}
+ 
+@@ -122,7 +121,6 @@ static int ltc2497_probe(struct i2c_client *client,
+ 	st->common_ddata.chip_info = chip_info;
+ 
+ 	resolution = chip_info->resolution;
+-	st->sub_lsb = 31 - (resolution + 1);
+ 	st->recv_size = BITS_TO_BYTES(resolution) + 1;
+ 
+ 	return ltc2497core_probe(dev, indio_dev);
+diff --git a/drivers/iio/adc/qcom-spmi-adc5.c b/drivers/iio/adc/qcom-spmi-adc5.c
+index 821fee60a7651..d1b86570768a9 100644
+--- a/drivers/iio/adc/qcom-spmi-adc5.c
++++ b/drivers/iio/adc/qcom-spmi-adc5.c
+@@ -626,12 +626,20 @@ static int adc5_get_fw_channel_data(struct adc5_chip *adc,
+ 				    struct fwnode_handle *fwnode,
+ 				    const struct adc5_data *data)
+ {
+-	const char *name = fwnode_get_name(fwnode), *channel_name;
++	const char *channel_name;
++	char *name;
+ 	u32 chan, value, varr[2];
+ 	u32 sid = 0;
+ 	int ret;
+ 	struct device *dev = adc->dev;
+ 
++	name = devm_kasprintf(dev, GFP_KERNEL, "%pfwP", fwnode);
++	if (!name)
++		return -ENOMEM;
++
++	/* Cut the address part */
++	name[strchrnul(name, '@') - name] = '\0';
++
+ 	ret = fwnode_property_read_u32(fwnode, "reg", &chan);
+ 	if (ret) {
+ 		dev_err(dev, "invalid channel number %s\n", name);
+diff --git a/drivers/iio/adc/ti-ads7950.c b/drivers/iio/adc/ti-ads7950.c
+index 2cc9a9bd9db60..263fc3a1b87e1 100644
+--- a/drivers/iio/adc/ti-ads7950.c
++++ b/drivers/iio/adc/ti-ads7950.c
+@@ -634,6 +634,7 @@ static int ti_ads7950_probe(struct spi_device *spi)
+ 	st->chip.label = dev_name(&st->spi->dev);
+ 	st->chip.parent = &st->spi->dev;
+ 	st->chip.owner = THIS_MODULE;
++	st->chip.can_sleep = true;
+ 	st->chip.base = -1;
+ 	st->chip.ngpio = TI_ADS7950_NUM_GPIOS;
+ 	st->chip.get_direction = ti_ads7950_get_direction;
+diff --git a/drivers/iio/dac/cio-dac.c b/drivers/iio/dac/cio-dac.c
+index 791dd999cf291..18a64f72fc188 100644
+--- a/drivers/iio/dac/cio-dac.c
++++ b/drivers/iio/dac/cio-dac.c
+@@ -66,8 +66,8 @@ static int cio_dac_write_raw(struct iio_dev *indio_dev,
+ 	if (mask != IIO_CHAN_INFO_RAW)
+ 		return -EINVAL;
+ 
+-	/* DAC can only accept up to a 16-bit value */
+-	if ((unsigned int)val > 65535)
++	/* DAC can only accept up to a 12-bit value */
++	if ((unsigned int)val > 4095)
+ 		return -EINVAL;
+ 
+ 	priv->chan_out_states[chan->channel] = val;
+diff --git a/drivers/iio/imu/Kconfig b/drivers/iio/imu/Kconfig
+index f1d7d4b5e2224..c2f97629e9cdb 100644
+--- a/drivers/iio/imu/Kconfig
++++ b/drivers/iio/imu/Kconfig
+@@ -47,6 +47,7 @@ config ADIS16480
+ 	depends on SPI
+ 	select IIO_ADIS_LIB
+ 	select IIO_ADIS_LIB_BUFFER if IIO_BUFFER
++	select CRC32
+ 	help
+ 	  Say yes here to build support for Analog Devices ADIS16375, ADIS16480,
+ 	  ADIS16485, ADIS16488 inertial sensors.
+diff --git a/drivers/iio/industrialio-buffer.c b/drivers/iio/industrialio-buffer.c
+index 228598b82a2f3..7e7ee307a3f7d 100644
+--- a/drivers/iio/industrialio-buffer.c
++++ b/drivers/iio/industrialio-buffer.c
+@@ -203,24 +203,27 @@ static ssize_t iio_buffer_write(struct file *filp, const char __user *buf,
+ 				break;
+ 			}
+ 
++			if (filp->f_flags & O_NONBLOCK) {
++				if (!written)
++					ret = -EAGAIN;
++				break;
++			}
++
+ 			wait_woken(&wait, TASK_INTERRUPTIBLE,
+ 					MAX_SCHEDULE_TIMEOUT);
+ 			continue;
+ 		}
+ 
+ 		ret = rb->access->write(rb, n - written, buf + written);
+-		if (ret == 0 && (filp->f_flags & O_NONBLOCK))
+-			ret = -EAGAIN;
++		if (ret < 0)
++			break;
+ 
+-		if (ret > 0) {
+-			written += ret;
+-			if (written != n && !(filp->f_flags & O_NONBLOCK))
+-				continue;
+-		}
+-	} while (ret == 0);
++		written += ret;
++
++	} while (written != n);
+ 	remove_wait_queue(&rb->pollq, &wait);
+ 
+-	return ret < 0 ? ret : n;
++	return ret < 0 ? ret : written;
+ }
+ 
+ /**
+diff --git a/drivers/iio/light/cm32181.c b/drivers/iio/light/cm32181.c
+index b1674a5bfa368..d4a34a3bf00d9 100644
+--- a/drivers/iio/light/cm32181.c
++++ b/drivers/iio/light/cm32181.c
+@@ -429,6 +429,14 @@ static const struct iio_info cm32181_info = {
+ 	.attrs			= &cm32181_attribute_group,
+ };
+ 
++static void cm32181_unregister_dummy_client(void *data)
++{
++	struct i2c_client *client = data;
++
++	/* Unregister the dummy client */
++	i2c_unregister_device(client);
++}
++
+ static int cm32181_probe(struct i2c_client *client)
+ {
+ 	struct device *dev = &client->dev;
+@@ -460,6 +468,10 @@ static int cm32181_probe(struct i2c_client *client)
+ 		client = i2c_acpi_new_device(dev, 1, &board_info);
+ 		if (IS_ERR(client))
+ 			return PTR_ERR(client);
++
++		ret = devm_add_action_or_reset(dev, cm32181_unregister_dummy_client, client);
++		if (ret)
++			return ret;
+ 	}
+ 
+ 	cm32181 = iio_priv(indio_dev);
+diff --git a/drivers/leds/rgb/leds-qcom-lpg.c b/drivers/leds/rgb/leds-qcom-lpg.c
+index c1a56259226fb..f1c2419334e6f 100644
+--- a/drivers/leds/rgb/leds-qcom-lpg.c
++++ b/drivers/leds/rgb/leds-qcom-lpg.c
+@@ -972,8 +972,8 @@ out_unlock:
+ 	return ret;
+ }
+ 
+-static void lpg_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-			      struct pwm_state *state)
++static int lpg_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++			     struct pwm_state *state)
+ {
+ 	struct lpg *lpg = container_of(chip, struct lpg, pwm);
+ 	struct lpg_channel *chan = &lpg->channels[pwm->hwpwm];
+@@ -986,20 +986,20 @@ static void lpg_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 
+ 	ret = regmap_read(lpg->map, chan->base + LPG_SIZE_CLK_REG, &val);
+ 	if (ret)
+-		return;
++		return 0;
+ 
+ 	refclk = lpg_clk_rates[val & PWM_CLK_SELECT_MASK];
+ 	if (refclk) {
+ 		ret = regmap_read(lpg->map, chan->base + LPG_PREDIV_CLK_REG, &val);
+ 		if (ret)
+-			return;
++			return 0;
+ 
+ 		pre_div = lpg_pre_divs[FIELD_GET(PWM_FREQ_PRE_DIV_MASK, val)];
+ 		m = FIELD_GET(PWM_FREQ_EXP_MASK, val);
+ 
+ 		ret = regmap_bulk_read(lpg->map, chan->base + PWM_VALUE_REG, &pwm_value, sizeof(pwm_value));
+ 		if (ret)
+-			return;
++			return 0;
+ 
+ 		state->period = DIV_ROUND_UP_ULL((u64)NSEC_PER_SEC * LPG_RESOLUTION * pre_div * (1 << m), refclk);
+ 		state->duty_cycle = DIV_ROUND_UP_ULL((u64)NSEC_PER_SEC * pwm_value * pre_div * (1 << m), refclk);
+@@ -1010,13 +1010,15 @@ static void lpg_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 
+ 	ret = regmap_read(lpg->map, chan->base + PWM_ENABLE_CONTROL_REG, &val);
+ 	if (ret)
+-		return;
++		return 0;
+ 
+ 	state->enabled = FIELD_GET(LPG_ENABLE_CONTROL_OUTPUT, val);
+ 	state->polarity = PWM_POLARITY_NORMAL;
+ 
+ 	if (state->duty_cycle > state->period)
+ 		state->duty_cycle = state->period;
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops lpg_pwm_ops = {
+diff --git a/drivers/md/dm-bio-prison-v1.c b/drivers/md/dm-bio-prison-v1.c
+index 1f8f98efd97a0..138067abe14b7 100644
+--- a/drivers/md/dm-bio-prison-v1.c
++++ b/drivers/md/dm-bio-prison-v1.c
+@@ -285,14 +285,14 @@ EXPORT_SYMBOL_GPL(dm_cell_promote_or_release);
+ 
+ struct dm_deferred_entry {
+ 	struct dm_deferred_set *ds;
+-	unsigned count;
++	unsigned int count;
+ 	struct list_head work_items;
+ };
+ 
+ struct dm_deferred_set {
+ 	spinlock_t lock;
+-	unsigned current_entry;
+-	unsigned sweeper;
++	unsigned int current_entry;
++	unsigned int sweeper;
+ 	struct dm_deferred_entry entries[DEFERRED_SET_SIZE];
+ };
+ 
+@@ -338,7 +338,7 @@ struct dm_deferred_entry *dm_deferred_entry_inc(struct dm_deferred_set *ds)
+ }
+ EXPORT_SYMBOL_GPL(dm_deferred_entry_inc);
+ 
+-static unsigned ds_next(unsigned index)
++static unsigned int ds_next(unsigned int index)
+ {
+ 	return (index + 1) % DEFERRED_SET_SIZE;
+ }
+@@ -373,7 +373,7 @@ EXPORT_SYMBOL_GPL(dm_deferred_entry_dec);
+ int dm_deferred_set_add_work(struct dm_deferred_set *ds, struct list_head *work)
+ {
+ 	int r = 1;
+-	unsigned next_entry;
++	unsigned int next_entry;
+ 
+ 	spin_lock_irq(&ds->lock);
+ 	if ((ds->sweeper == ds->current_entry) &&
+diff --git a/drivers/md/dm-bio-prison-v2.c b/drivers/md/dm-bio-prison-v2.c
+index 9dec3b61cf70a..0cc0d13c40e51 100644
+--- a/drivers/md/dm-bio-prison-v2.c
++++ b/drivers/md/dm-bio-prison-v2.c
+@@ -148,7 +148,7 @@ static bool __find_or_insert(struct dm_bio_prison_v2 *prison,
+ 
+ static bool __get(struct dm_bio_prison_v2 *prison,
+ 		  struct dm_cell_key_v2 *key,
+-		  unsigned lock_level,
++		  unsigned int lock_level,
+ 		  struct bio *inmate,
+ 		  struct dm_bio_prison_cell_v2 *cell_prealloc,
+ 		  struct dm_bio_prison_cell_v2 **cell)
+@@ -171,7 +171,7 @@ static bool __get(struct dm_bio_prison_v2 *prison,
+ 
+ bool dm_cell_get_v2(struct dm_bio_prison_v2 *prison,
+ 		    struct dm_cell_key_v2 *key,
+-		    unsigned lock_level,
++		    unsigned int lock_level,
+ 		    struct bio *inmate,
+ 		    struct dm_bio_prison_cell_v2 *cell_prealloc,
+ 		    struct dm_bio_prison_cell_v2 **cell_result)
+@@ -224,7 +224,7 @@ EXPORT_SYMBOL_GPL(dm_cell_put_v2);
+ 
+ static int __lock(struct dm_bio_prison_v2 *prison,
+ 		  struct dm_cell_key_v2 *key,
+-		  unsigned lock_level,
++		  unsigned int lock_level,
+ 		  struct dm_bio_prison_cell_v2 *cell_prealloc,
+ 		  struct dm_bio_prison_cell_v2 **cell_result)
+ {
+@@ -255,7 +255,7 @@ static int __lock(struct dm_bio_prison_v2 *prison,
+ 
+ int dm_cell_lock_v2(struct dm_bio_prison_v2 *prison,
+ 		    struct dm_cell_key_v2 *key,
+-		    unsigned lock_level,
++		    unsigned int lock_level,
+ 		    struct dm_bio_prison_cell_v2 *cell_prealloc,
+ 		    struct dm_bio_prison_cell_v2 **cell_result)
+ {
+@@ -291,7 +291,7 @@ EXPORT_SYMBOL_GPL(dm_cell_quiesce_v2);
+ 
+ static int __promote(struct dm_bio_prison_v2 *prison,
+ 		     struct dm_bio_prison_cell_v2 *cell,
+-		     unsigned new_lock_level)
++		     unsigned int new_lock_level)
+ {
+ 	if (!cell->exclusive_lock)
+ 		return -EINVAL;
+@@ -302,7 +302,7 @@ static int __promote(struct dm_bio_prison_v2 *prison,
+ 
+ int dm_cell_lock_promote_v2(struct dm_bio_prison_v2 *prison,
+ 			    struct dm_bio_prison_cell_v2 *cell,
+-			    unsigned new_lock_level)
++			    unsigned int new_lock_level)
+ {
+ 	int r;
+ 
+diff --git a/drivers/md/dm-bio-prison-v2.h b/drivers/md/dm-bio-prison-v2.h
+index 6e04234268db3..5a7d996bbbd80 100644
+--- a/drivers/md/dm-bio-prison-v2.h
++++ b/drivers/md/dm-bio-prison-v2.h
+@@ -44,8 +44,8 @@ struct dm_cell_key_v2 {
+ struct dm_bio_prison_cell_v2 {
+ 	// FIXME: pack these
+ 	bool exclusive_lock;
+-	unsigned exclusive_level;
+-	unsigned shared_count;
++	unsigned int exclusive_level;
++	unsigned int shared_count;
+ 	struct work_struct *quiesce_continuation;
+ 
+ 	struct rb_node node;
+@@ -86,7 +86,7 @@ void dm_bio_prison_free_cell_v2(struct dm_bio_prison_v2 *prison,
+  */
+ bool dm_cell_get_v2(struct dm_bio_prison_v2 *prison,
+ 		    struct dm_cell_key_v2 *key,
+-		    unsigned lock_level,
++		    unsigned int lock_level,
+ 		    struct bio *inmate,
+ 		    struct dm_bio_prison_cell_v2 *cell_prealloc,
+ 		    struct dm_bio_prison_cell_v2 **cell_result);
+@@ -114,7 +114,7 @@ bool dm_cell_put_v2(struct dm_bio_prison_v2 *prison,
+  */
+ int dm_cell_lock_v2(struct dm_bio_prison_v2 *prison,
+ 		    struct dm_cell_key_v2 *key,
+-		    unsigned lock_level,
++		    unsigned int lock_level,
+ 		    struct dm_bio_prison_cell_v2 *cell_prealloc,
+ 		    struct dm_bio_prison_cell_v2 **cell_result);
+ 
+@@ -132,7 +132,7 @@ void dm_cell_quiesce_v2(struct dm_bio_prison_v2 *prison,
+  */
+ int dm_cell_lock_promote_v2(struct dm_bio_prison_v2 *prison,
+ 			    struct dm_bio_prison_cell_v2 *cell,
+-			    unsigned new_lock_level);
++			    unsigned int new_lock_level);
+ 
+ /*
+  * Adds any held bios to the bio list.
+diff --git a/drivers/md/dm-bufio.c b/drivers/md/dm-bufio.c
+index 19caaf684ee34..382c5cc471952 100644
+--- a/drivers/md/dm-bufio.c
++++ b/drivers/md/dm-bufio.c
+@@ -89,7 +89,7 @@ struct dm_bufio_client {
+ 	unsigned long n_buffers[LIST_SIZE];
+ 
+ 	struct block_device *bdev;
+-	unsigned block_size;
++	unsigned int block_size;
+ 	s8 sectors_per_block_bits;
+ 	void (*alloc_callback)(struct dm_buffer *);
+ 	void (*write_callback)(struct dm_buffer *);
+@@ -98,9 +98,9 @@ struct dm_bufio_client {
+ 	struct dm_io_client *dm_io;
+ 
+ 	struct list_head reserved_buffers;
+-	unsigned need_reserved_buffers;
++	unsigned int need_reserved_buffers;
+ 
+-	unsigned minimum_buffers;
++	unsigned int minimum_buffers;
+ 
+ 	struct rb_root buffer_tree;
+ 	wait_queue_head_t free_buffer_wait;
+@@ -145,14 +145,14 @@ struct dm_buffer {
+ 	unsigned char list_mode;		/* LIST_* */
+ 	blk_status_t read_error;
+ 	blk_status_t write_error;
+-	unsigned accessed;
+-	unsigned hold_count;
++	unsigned int accessed;
++	unsigned int hold_count;
+ 	unsigned long state;
+ 	unsigned long last_accessed;
+-	unsigned dirty_start;
+-	unsigned dirty_end;
+-	unsigned write_start;
+-	unsigned write_end;
++	unsigned int dirty_start;
++	unsigned int dirty_end;
++	unsigned int write_start;
++	unsigned int write_end;
+ 	struct dm_bufio_client *c;
+ 	struct list_head write_list;
+ 	void (*end_io)(struct dm_buffer *, blk_status_t);
+@@ -220,7 +220,7 @@ static unsigned long global_num = 0;
+ /*
+  * Buffers are freed after this timeout
+  */
+-static unsigned dm_bufio_max_age = DM_BUFIO_DEFAULT_AGE_SECS;
++static unsigned int dm_bufio_max_age = DM_BUFIO_DEFAULT_AGE_SECS;
+ static unsigned long dm_bufio_retain_bytes = DM_BUFIO_DEFAULT_RETAIN_BYTES;
+ 
+ static unsigned long dm_bufio_peak_allocated;
+@@ -438,7 +438,7 @@ static void *alloc_buffer_data(struct dm_bufio_client *c, gfp_t gfp_mask,
+ 	 * as if GFP_NOIO was specified.
+ 	 */
+ 	if (gfp_mask & __GFP_NORETRY) {
+-		unsigned noio_flag = memalloc_noio_save();
++		unsigned int noio_flag = memalloc_noio_save();
+ 		void *ptr = __vmalloc(c->block_size, gfp_mask);
+ 
+ 		memalloc_noio_restore(noio_flag);
+@@ -591,7 +591,7 @@ static void dmio_complete(unsigned long error, void *context)
+ }
+ 
+ static void use_dmio(struct dm_buffer *b, enum req_op op, sector_t sector,
+-		     unsigned n_sectors, unsigned offset)
++		     unsigned int n_sectors, unsigned int offset)
+ {
+ 	int r;
+ 	struct dm_io_request io_req = {
+@@ -629,11 +629,11 @@ static void bio_complete(struct bio *bio)
+ }
+ 
+ static void use_bio(struct dm_buffer *b, enum req_op op, sector_t sector,
+-		    unsigned n_sectors, unsigned offset)
++		    unsigned int n_sectors, unsigned int offset)
+ {
+ 	struct bio *bio;
+ 	char *ptr;
+-	unsigned vec_size, len;
++	unsigned int vec_size, len;
+ 
+ 	vec_size = b->c->block_size >> PAGE_SHIFT;
+ 	if (unlikely(b->c->sectors_per_block_bits < PAGE_SHIFT - SECTOR_SHIFT))
+@@ -654,7 +654,7 @@ dmio:
+ 	len = n_sectors << SECTOR_SHIFT;
+ 
+ 	do {
+-		unsigned this_step = min((unsigned)(PAGE_SIZE - offset_in_page(ptr)), len);
++		unsigned int this_step = min((unsigned int)(PAGE_SIZE - offset_in_page(ptr)), len);
+ 		if (!bio_add_page(bio, virt_to_page(ptr), this_step,
+ 				  offset_in_page(ptr))) {
+ 			bio_put(bio);
+@@ -684,9 +684,9 @@ static inline sector_t block_to_sector(struct dm_bufio_client *c, sector_t block
+ static void submit_io(struct dm_buffer *b, enum req_op op,
+ 		      void (*end_io)(struct dm_buffer *, blk_status_t))
+ {
+-	unsigned n_sectors;
++	unsigned int n_sectors;
+ 	sector_t sector;
+-	unsigned offset, end;
++	unsigned int offset, end;
+ 
+ 	b->end_io = end_io;
+ 
+@@ -1156,7 +1156,7 @@ void *dm_bufio_new(struct dm_bufio_client *c, sector_t block,
+ EXPORT_SYMBOL_GPL(dm_bufio_new);
+ 
+ void dm_bufio_prefetch(struct dm_bufio_client *c,
+-		       sector_t block, unsigned n_blocks)
++		       sector_t block, unsigned int n_blocks)
+ {
+ 	struct blk_plug plug;
+ 
+@@ -1232,7 +1232,7 @@ void dm_bufio_release(struct dm_buffer *b)
+ EXPORT_SYMBOL_GPL(dm_bufio_release);
+ 
+ void dm_bufio_mark_partial_buffer_dirty(struct dm_buffer *b,
+-					unsigned start, unsigned end)
++					unsigned int start, unsigned int end)
+ {
+ 	struct dm_bufio_client *c = b->c;
+ 
+@@ -1529,13 +1529,13 @@ void dm_bufio_forget_buffers(struct dm_bufio_client *c, sector_t block, sector_t
+ }
+ EXPORT_SYMBOL_GPL(dm_bufio_forget_buffers);
+ 
+-void dm_bufio_set_minimum_buffers(struct dm_bufio_client *c, unsigned n)
++void dm_bufio_set_minimum_buffers(struct dm_bufio_client *c, unsigned int n)
+ {
+ 	c->minimum_buffers = n;
+ }
+ EXPORT_SYMBOL_GPL(dm_bufio_set_minimum_buffers);
+ 
+-unsigned dm_bufio_get_block_size(struct dm_bufio_client *c)
++unsigned int dm_bufio_get_block_size(struct dm_bufio_client *c)
+ {
+ 	return c->block_size;
+ }
+@@ -1734,15 +1734,15 @@ static unsigned long dm_bufio_shrink_count(struct shrinker *shrink, struct shrin
+ /*
+  * Create the buffering interface
+  */
+-struct dm_bufio_client *dm_bufio_client_create(struct block_device *bdev, unsigned block_size,
+-					       unsigned reserved_buffers, unsigned aux_size,
++struct dm_bufio_client *dm_bufio_client_create(struct block_device *bdev, unsigned int block_size,
++					       unsigned int reserved_buffers, unsigned int aux_size,
+ 					       void (*alloc_callback)(struct dm_buffer *),
+ 					       void (*write_callback)(struct dm_buffer *),
+ 					       unsigned int flags)
+ {
+ 	int r;
+ 	struct dm_bufio_client *c;
+-	unsigned i;
++	unsigned int i;
+ 	char slab_name[27];
+ 
+ 	if (!block_size || block_size & ((1 << SECTOR_SHIFT) - 1)) {
+@@ -1796,7 +1796,7 @@ struct dm_bufio_client *dm_bufio_client_create(struct block_device *bdev, unsign
+ 
+ 	if (block_size <= KMALLOC_MAX_SIZE &&
+ 	    (block_size < PAGE_SIZE || !is_power_of_2(block_size))) {
+-		unsigned align = min(1U << __ffs(block_size), (unsigned)PAGE_SIZE);
++		unsigned int align = min(1U << __ffs(block_size), (unsigned int)PAGE_SIZE);
+ 		snprintf(slab_name, sizeof slab_name, "dm_bufio_cache-%u", block_size);
+ 		c->slab_cache = kmem_cache_create(slab_name, block_size, align,
+ 						  SLAB_RECLAIM_ACCOUNT, NULL);
+@@ -1872,7 +1872,7 @@ EXPORT_SYMBOL_GPL(dm_bufio_client_create);
+  */
+ void dm_bufio_client_destroy(struct dm_bufio_client *c)
+ {
+-	unsigned i;
++	unsigned int i;
+ 
+ 	drop_buffers(c);
+ 
+@@ -1920,9 +1920,9 @@ void dm_bufio_set_sector_offset(struct dm_bufio_client *c, sector_t start)
+ }
+ EXPORT_SYMBOL_GPL(dm_bufio_set_sector_offset);
+ 
+-static unsigned get_max_age_hz(void)
++static unsigned int get_max_age_hz(void)
+ {
+-	unsigned max_age = READ_ONCE(dm_bufio_max_age);
++	unsigned int max_age = READ_ONCE(dm_bufio_max_age);
+ 
+ 	if (max_age > UINT_MAX / HZ)
+ 		max_age = UINT_MAX / HZ;
+@@ -1973,7 +1973,7 @@ static void do_global_cleanup(struct work_struct *w)
+ 	struct dm_bufio_client *locked_client = NULL;
+ 	struct dm_bufio_client *current_client;
+ 	struct dm_buffer *b;
+-	unsigned spinlock_hold_count;
++	unsigned int spinlock_hold_count;
+ 	unsigned long threshold = dm_bufio_cache_size -
+ 		dm_bufio_cache_size / DM_BUFIO_LOW_WATERMARK_RATIO;
+ 	unsigned long loops = global_num * 2;
+diff --git a/drivers/md/dm-cache-background-tracker.c b/drivers/md/dm-cache-background-tracker.c
+index 7887f99b82bd5..c606e6bfc3f8b 100644
+--- a/drivers/md/dm-cache-background-tracker.c
++++ b/drivers/md/dm-cache-background-tracker.c
+@@ -17,7 +17,7 @@ struct bt_work {
+ };
+ 
+ struct background_tracker {
+-	unsigned max_work;
++	unsigned int max_work;
+ 	atomic_t pending_promotes;
+ 	atomic_t pending_writebacks;
+ 	atomic_t pending_demotes;
+@@ -29,7 +29,7 @@ struct background_tracker {
+ 	struct kmem_cache *work_cache;
+ };
+ 
+-struct background_tracker *btracker_create(unsigned max_work)
++struct background_tracker *btracker_create(unsigned int max_work)
+ {
+ 	struct background_tracker *b = kmalloc(sizeof(*b), GFP_KERNEL);
+ 
+@@ -155,13 +155,13 @@ static void update_stats(struct background_tracker *b, struct policy_work *w, in
+ 	}
+ }
+ 
+-unsigned btracker_nr_writebacks_queued(struct background_tracker *b)
++unsigned int btracker_nr_writebacks_queued(struct background_tracker *b)
+ {
+ 	return atomic_read(&b->pending_writebacks);
+ }
+ EXPORT_SYMBOL_GPL(btracker_nr_writebacks_queued);
+ 
+-unsigned btracker_nr_demotions_queued(struct background_tracker *b)
++unsigned int btracker_nr_demotions_queued(struct background_tracker *b)
+ {
+ 	return atomic_read(&b->pending_demotes);
+ }
+diff --git a/drivers/md/dm-cache-background-tracker.h b/drivers/md/dm-cache-background-tracker.h
+index 27ab90dbc2752..14d3d53dc77a3 100644
+--- a/drivers/md/dm-cache-background-tracker.h
++++ b/drivers/md/dm-cache-background-tracker.h
+@@ -12,19 +12,44 @@
+ 
+ /*----------------------------------------------------------------*/
+ 
++/*
++ * The cache policy decides what background work should be performed,
++ * such as promotions, demotions and writebacks. The core cache target
++ * is in charge of performing the work, and does so when it sees fit.
++ *
++ * The background_tracker acts as a go between. Keeping track of future
++ * work that the policy has decided upon, and handing (issuing) it to
++ * the core target when requested.
++ *
++ * There is no locking in this, so calls will probably need to be
++ * protected with a spinlock.
++ */
++
+ struct background_work;
+ struct background_tracker;
+ 
+ /*
+- * FIXME: discuss lack of locking in all methods.
++ * Create a new tracker, it will not be able to queue more than
++ * 'max_work' entries.
++ */
++struct background_tracker *btracker_create(unsigned int max_work);
++
++/*
++ * Destroy the tracker. No issued, but not complete, work should
++ * exist when this is called. It is fine to have queued but unissued
++ * work.
+  */
+-struct background_tracker *btracker_create(unsigned max_work);
+ void btracker_destroy(struct background_tracker *b);
+ 
+-unsigned btracker_nr_writebacks_queued(struct background_tracker *b);
+-unsigned btracker_nr_demotions_queued(struct background_tracker *b);
++unsigned int btracker_nr_writebacks_queued(struct background_tracker *b);
++unsigned int btracker_nr_demotions_queued(struct background_tracker *b);
+ 
+ /*
++ * Queue some work within the tracker. 'work' should point to the work
++ * to queue, this will be copied (ownership doesn't pass).  If pwork
++ * is not NULL then it will be set to point to the tracker's internal
++ * copy of the work.
++ *
+  * returns -EINVAL iff the work is already queued.  -ENOMEM if the work
+  * couldn't be queued for another reason.
+  */
+@@ -33,11 +58,20 @@ int btracker_queue(struct background_tracker *b,
+ 		   struct policy_work **pwork);
+ 
+ /*
++ * Hands out the next piece of work to be performed.
+  * Returns -ENODATA if there's no work.
+  */
+ int btracker_issue(struct background_tracker *b, struct policy_work **work);
+-void btracker_complete(struct background_tracker *b,
+-		       struct policy_work *op);
++
++/*
++ * Informs the tracker that the work has been completed and it may forget
++ * about it.
++ */
++void btracker_complete(struct background_tracker *b, struct policy_work *op);
++
++/*
++ * Predicate to see if an origin block is already scheduled for promotion.
++ */
+ bool btracker_promotion_already_present(struct background_tracker *b,
+ 					dm_oblock_t oblock);
+ 
+diff --git a/drivers/md/dm-cache-metadata.c b/drivers/md/dm-cache-metadata.c
+index 83a5975bcc729..f5b4c996dc05f 100644
+--- a/drivers/md/dm-cache-metadata.c
++++ b/drivers/md/dm-cache-metadata.c
+@@ -104,7 +104,7 @@ struct dm_cache_metadata {
+ 	refcount_t ref_count;
+ 	struct list_head list;
+ 
+-	unsigned version;
++	unsigned int version;
+ 	struct block_device *bdev;
+ 	struct dm_block_manager *bm;
+ 	struct dm_space_map *metadata_sm;
+@@ -129,7 +129,7 @@ struct dm_cache_metadata {
+ 	bool clean_when_opened:1;
+ 
+ 	char policy_name[CACHE_POLICY_NAME_SIZE];
+-	unsigned policy_version[CACHE_POLICY_VERSION_SIZE];
++	unsigned int policy_version[CACHE_POLICY_VERSION_SIZE];
+ 	size_t policy_hint_size;
+ 	struct dm_cache_statistics stats;
+ 
+@@ -260,10 +260,10 @@ static int superblock_lock(struct dm_cache_metadata *cmd,
+ static int __superblock_all_zeroes(struct dm_block_manager *bm, bool *result)
+ {
+ 	int r;
+-	unsigned i;
++	unsigned int i;
+ 	struct dm_block *b;
+ 	__le64 *data_le, zero = cpu_to_le64(0);
+-	unsigned sb_block_size = dm_bm_block_size(bm) / sizeof(__le64);
++	unsigned int sb_block_size = dm_bm_block_size(bm) / sizeof(__le64);
+ 
+ 	/*
+ 	 * We can't use a validator here - it may be all zeroes.
+@@ -727,7 +727,7 @@ static int __commit_transaction(struct dm_cache_metadata *cmd,
+  */
+ #define FLAGS_MASK ((1 << 16) - 1)
+ 
+-static __le64 pack_value(dm_oblock_t block, unsigned flags)
++static __le64 pack_value(dm_oblock_t block, unsigned int flags)
+ {
+ 	uint64_t value = from_oblock(block);
+ 	value <<= 16;
+@@ -735,7 +735,7 @@ static __le64 pack_value(dm_oblock_t block, unsigned flags)
+ 	return cpu_to_le64(value);
+ }
+ 
+-static void unpack_value(__le64 value_le, dm_oblock_t *block, unsigned *flags)
++static void unpack_value(__le64 value_le, dm_oblock_t *block, unsigned int *flags)
+ {
+ 	uint64_t value = le64_to_cpu(value_le);
+ 	uint64_t b = value >> 16;
+@@ -749,7 +749,7 @@ static struct dm_cache_metadata *metadata_open(struct block_device *bdev,
+ 					       sector_t data_block_size,
+ 					       bool may_format_device,
+ 					       size_t policy_hint_size,
+-					       unsigned metadata_version)
++					       unsigned int metadata_version)
+ {
+ 	int r;
+ 	struct dm_cache_metadata *cmd;
+@@ -810,7 +810,7 @@ static struct dm_cache_metadata *lookup_or_open(struct block_device *bdev,
+ 						sector_t data_block_size,
+ 						bool may_format_device,
+ 						size_t policy_hint_size,
+-						unsigned metadata_version)
++						unsigned int metadata_version)
+ {
+ 	struct dm_cache_metadata *cmd, *cmd2;
+ 
+@@ -855,7 +855,7 @@ struct dm_cache_metadata *dm_cache_metadata_open(struct block_device *bdev,
+ 						 sector_t data_block_size,
+ 						 bool may_format_device,
+ 						 size_t policy_hint_size,
+-						 unsigned metadata_version)
++						 unsigned int metadata_version)
+ {
+ 	struct dm_cache_metadata *cmd = lookup_or_open(bdev, data_block_size, may_format_device,
+ 						       policy_hint_size, metadata_version);
+@@ -890,7 +890,7 @@ static int block_clean_combined_dirty(struct dm_cache_metadata *cmd, dm_cblock_t
+ 	int r;
+ 	__le64 value;
+ 	dm_oblock_t ob;
+-	unsigned flags;
++	unsigned int flags;
+ 
+ 	r = dm_array_get_value(&cmd->info, cmd->root, from_cblock(b), &value);
+ 	if (r)
+@@ -1288,7 +1288,7 @@ static bool policy_unchanged(struct dm_cache_metadata *cmd,
+ 			     struct dm_cache_policy *policy)
+ {
+ 	const char *policy_name = dm_cache_policy_get_name(policy);
+-	const unsigned *policy_version = dm_cache_policy_get_version(policy);
++	const unsigned int *policy_version = dm_cache_policy_get_version(policy);
+ 	size_t policy_hint_size = dm_cache_policy_get_hint_size(policy);
+ 
+ 	/*
+@@ -1339,7 +1339,7 @@ static int __load_mapping_v1(struct dm_cache_metadata *cmd,
+ 	__le32 *hint_value_le;
+ 
+ 	dm_oblock_t oblock;
+-	unsigned flags;
++	unsigned int flags;
+ 	bool dirty = true;
+ 
+ 	dm_array_cursor_get_value(mapping_cursor, (void **) &mapping_value_le);
+@@ -1381,7 +1381,7 @@ static int __load_mapping_v2(struct dm_cache_metadata *cmd,
+ 	__le32 *hint_value_le;
+ 
+ 	dm_oblock_t oblock;
+-	unsigned flags;
++	unsigned int flags;
+ 	bool dirty = true;
+ 
+ 	dm_array_cursor_get_value(mapping_cursor, (void **) &mapping_value_le);
+@@ -1513,7 +1513,7 @@ static int __dump_mapping(void *context, uint64_t cblock, void *leaf)
+ {
+ 	__le64 value;
+ 	dm_oblock_t oblock;
+-	unsigned flags;
++	unsigned int flags;
+ 
+ 	memcpy(&value, leaf, sizeof(value));
+ 	unpack_value(value, &oblock, &flags);
+@@ -1547,7 +1547,7 @@ int dm_cache_changed_this_transaction(struct dm_cache_metadata *cmd)
+ static int __dirty(struct dm_cache_metadata *cmd, dm_cblock_t cblock, bool dirty)
+ {
+ 	int r;
+-	unsigned flags;
++	unsigned int flags;
+ 	dm_oblock_t oblock;
+ 	__le64 value;
+ 
+@@ -1574,10 +1574,10 @@ static int __dirty(struct dm_cache_metadata *cmd, dm_cblock_t cblock, bool dirty
+ 
+ }
+ 
+-static int __set_dirty_bits_v1(struct dm_cache_metadata *cmd, unsigned nr_bits, unsigned long *bits)
++static int __set_dirty_bits_v1(struct dm_cache_metadata *cmd, unsigned int nr_bits, unsigned long *bits)
+ {
+ 	int r;
+-	unsigned i;
++	unsigned int i;
+ 	for (i = 0; i < nr_bits; i++) {
+ 		r = __dirty(cmd, to_cblock(i), test_bit(i, bits));
+ 		if (r)
+@@ -1594,7 +1594,7 @@ static int is_dirty_callback(uint32_t index, bool *value, void *context)
+ 	return 0;
+ }
+ 
+-static int __set_dirty_bits_v2(struct dm_cache_metadata *cmd, unsigned nr_bits, unsigned long *bits)
++static int __set_dirty_bits_v2(struct dm_cache_metadata *cmd, unsigned int nr_bits, unsigned long *bits)
+ {
+ 	int r = 0;
+ 
+@@ -1613,7 +1613,7 @@ static int __set_dirty_bits_v2(struct dm_cache_metadata *cmd, unsigned nr_bits,
+ }
+ 
+ int dm_cache_set_dirty_bits(struct dm_cache_metadata *cmd,
+-			    unsigned nr_bits,
++			    unsigned int nr_bits,
+ 			    unsigned long *bits)
+ {
+ 	int r;
+@@ -1712,7 +1712,7 @@ static int write_hints(struct dm_cache_metadata *cmd, struct dm_cache_policy *po
+ 	int r;
+ 	size_t hint_size;
+ 	const char *policy_name = dm_cache_policy_get_name(policy);
+-	const unsigned *policy_version = dm_cache_policy_get_version(policy);
++	const unsigned int *policy_version = dm_cache_policy_get_version(policy);
+ 
+ 	if (!policy_name[0] ||
+ 	    (strlen(policy_name) > sizeof(cmd->policy_name) - 1))
+diff --git a/drivers/md/dm-cache-metadata.h b/drivers/md/dm-cache-metadata.h
+index 0905f2c1615e1..b40322bc44cf7 100644
+--- a/drivers/md/dm-cache-metadata.h
++++ b/drivers/md/dm-cache-metadata.h
+@@ -60,7 +60,7 @@ struct dm_cache_metadata *dm_cache_metadata_open(struct block_device *bdev,
+ 						 sector_t data_block_size,
+ 						 bool may_format_device,
+ 						 size_t policy_hint_size,
+-						 unsigned metadata_version);
++						 unsigned int metadata_version);
+ 
+ void dm_cache_metadata_close(struct dm_cache_metadata *cmd);
+ 
+@@ -96,7 +96,7 @@ int dm_cache_load_mappings(struct dm_cache_metadata *cmd,
+ 			   void *context);
+ 
+ int dm_cache_set_dirty_bits(struct dm_cache_metadata *cmd,
+-			    unsigned nr_bits, unsigned long *bits);
++			    unsigned int nr_bits, unsigned long *bits);
+ 
+ struct dm_cache_statistics {
+ 	uint32_t read_hits;
+diff --git a/drivers/md/dm-cache-policy-internal.h b/drivers/md/dm-cache-policy-internal.h
+index 56f0a23f698c0..8e49baa78dc19 100644
+--- a/drivers/md/dm-cache-policy-internal.h
++++ b/drivers/md/dm-cache-policy-internal.h
+@@ -85,7 +85,7 @@ static inline void policy_tick(struct dm_cache_policy *p, bool can_block)
+ }
+ 
+ static inline int policy_emit_config_values(struct dm_cache_policy *p, char *result,
+-					    unsigned maxlen, ssize_t *sz_ptr)
++					    unsigned int maxlen, ssize_t *sz_ptr)
+ {
+ 	ssize_t sz = *sz_ptr;
+ 	if (p->emit_config_values)
+@@ -112,18 +112,18 @@ static inline void policy_allow_migrations(struct dm_cache_policy *p, bool allow
+ /*
+  * Some utility functions commonly used by policies and the core target.
+  */
+-static inline size_t bitset_size_in_bytes(unsigned nr_entries)
++static inline size_t bitset_size_in_bytes(unsigned int nr_entries)
+ {
+ 	return sizeof(unsigned long) * dm_div_up(nr_entries, BITS_PER_LONG);
+ }
+ 
+-static inline unsigned long *alloc_bitset(unsigned nr_entries)
++static inline unsigned long *alloc_bitset(unsigned int nr_entries)
+ {
+ 	size_t s = bitset_size_in_bytes(nr_entries);
+ 	return vzalloc(s);
+ }
+ 
+-static inline void clear_bitset(void *bitset, unsigned nr_entries)
++static inline void clear_bitset(void *bitset, unsigned int nr_entries)
+ {
+ 	size_t s = bitset_size_in_bytes(nr_entries);
+ 	memset(bitset, 0, s);
+@@ -154,7 +154,7 @@ void dm_cache_policy_destroy(struct dm_cache_policy *p);
+  */
+ const char *dm_cache_policy_get_name(struct dm_cache_policy *p);
+ 
+-const unsigned *dm_cache_policy_get_version(struct dm_cache_policy *p);
++const unsigned int *dm_cache_policy_get_version(struct dm_cache_policy *p);
+ 
+ size_t dm_cache_policy_get_hint_size(struct dm_cache_policy *p);
+ 
+diff --git a/drivers/md/dm-cache-policy-smq.c b/drivers/md/dm-cache-policy-smq.c
+index a3d281fc14c3a..54343812223e8 100644
+--- a/drivers/md/dm-cache-policy-smq.c
++++ b/drivers/md/dm-cache-policy-smq.c
+@@ -23,12 +23,12 @@
+ /*
+  * Safe division functions that return zero on divide by zero.
+  */
+-static unsigned safe_div(unsigned n, unsigned d)
++static unsigned int safe_div(unsigned int n, unsigned int d)
+ {
+ 	return d ? n / d : 0u;
+ }
+ 
+-static unsigned safe_mod(unsigned n, unsigned d)
++static unsigned int safe_mod(unsigned int n, unsigned int d)
+ {
+ 	return d ? n % d : 0u;
+ }
+@@ -36,10 +36,10 @@ static unsigned safe_mod(unsigned n, unsigned d)
+ /*----------------------------------------------------------------*/
+ 
+ struct entry {
+-	unsigned hash_next:28;
+-	unsigned prev:28;
+-	unsigned next:28;
+-	unsigned level:6;
++	unsigned int hash_next:28;
++	unsigned int prev:28;
++	unsigned int next:28;
++	unsigned int level:6;
+ 	bool dirty:1;
+ 	bool allocated:1;
+ 	bool sentinel:1;
+@@ -62,7 +62,7 @@ struct entry_space {
+ 	struct entry *end;
+ };
+ 
+-static int space_init(struct entry_space *es, unsigned nr_entries)
++static int space_init(struct entry_space *es, unsigned int nr_entries)
+ {
+ 	if (!nr_entries) {
+ 		es->begin = es->end = NULL;
+@@ -82,7 +82,7 @@ static void space_exit(struct entry_space *es)
+ 	vfree(es->begin);
+ }
+ 
+-static struct entry *__get_entry(struct entry_space *es, unsigned block)
++static struct entry *__get_entry(struct entry_space *es, unsigned int block)
+ {
+ 	struct entry *e;
+ 
+@@ -92,13 +92,13 @@ static struct entry *__get_entry(struct entry_space *es, unsigned block)
+ 	return e;
+ }
+ 
+-static unsigned to_index(struct entry_space *es, struct entry *e)
++static unsigned int to_index(struct entry_space *es, struct entry *e)
+ {
+ 	BUG_ON(e < es->begin || e >= es->end);
+ 	return e - es->begin;
+ }
+ 
+-static struct entry *to_entry(struct entry_space *es, unsigned block)
++static struct entry *to_entry(struct entry_space *es, unsigned int block)
+ {
+ 	if (block == INDEXER_NULL)
+ 		return NULL;
+@@ -109,8 +109,8 @@ static struct entry *to_entry(struct entry_space *es, unsigned block)
+ /*----------------------------------------------------------------*/
+ 
+ struct ilist {
+-	unsigned nr_elts;	/* excluding sentinel entries */
+-	unsigned head, tail;
++	unsigned int nr_elts;	/* excluding sentinel entries */
++	unsigned int head, tail;
+ };
+ 
+ static void l_init(struct ilist *l)
+@@ -252,23 +252,23 @@ static struct entry *l_pop_tail(struct entry_space *es, struct ilist *l)
+ struct queue {
+ 	struct entry_space *es;
+ 
+-	unsigned nr_elts;
+-	unsigned nr_levels;
++	unsigned int nr_elts;
++	unsigned int nr_levels;
+ 	struct ilist qs[MAX_LEVELS];
+ 
+ 	/*
+ 	 * We maintain a count of the number of entries we would like in each
+ 	 * level.
+ 	 */
+-	unsigned last_target_nr_elts;
+-	unsigned nr_top_levels;
+-	unsigned nr_in_top_levels;
+-	unsigned target_count[MAX_LEVELS];
++	unsigned int last_target_nr_elts;
++	unsigned int nr_top_levels;
++	unsigned int nr_in_top_levels;
++	unsigned int target_count[MAX_LEVELS];
+ };
+ 
+-static void q_init(struct queue *q, struct entry_space *es, unsigned nr_levels)
++static void q_init(struct queue *q, struct entry_space *es, unsigned int nr_levels)
+ {
+-	unsigned i;
++	unsigned int i;
+ 
+ 	q->es = es;
+ 	q->nr_elts = 0;
+@@ -284,7 +284,7 @@ static void q_init(struct queue *q, struct entry_space *es, unsigned nr_levels)
+ 	q->nr_in_top_levels = 0u;
+ }
+ 
+-static unsigned q_size(struct queue *q)
++static unsigned int q_size(struct queue *q)
+ {
+ 	return q->nr_elts;
+ }
+@@ -332,9 +332,9 @@ static void q_del(struct queue *q, struct entry *e)
+ /*
+  * Return the oldest entry of the lowest populated level.
+  */
+-static struct entry *q_peek(struct queue *q, unsigned max_level, bool can_cross_sentinel)
++static struct entry *q_peek(struct queue *q, unsigned int max_level, bool can_cross_sentinel)
+ {
+-	unsigned level;
++	unsigned int level;
+ 	struct entry *e;
+ 
+ 	max_level = min(max_level, q->nr_levels);
+@@ -369,7 +369,7 @@ static struct entry *q_pop(struct queue *q)
+  * used by redistribute, so we know this is true.  It also doesn't adjust
+  * the q->nr_elts count.
+  */
+-static struct entry *__redist_pop_from(struct queue *q, unsigned level)
++static struct entry *__redist_pop_from(struct queue *q, unsigned int level)
+ {
+ 	struct entry *e;
+ 
+@@ -383,9 +383,10 @@ static struct entry *__redist_pop_from(struct queue *q, unsigned level)
+ 	return NULL;
+ }
+ 
+-static void q_set_targets_subrange_(struct queue *q, unsigned nr_elts, unsigned lbegin, unsigned lend)
++static void q_set_targets_subrange_(struct queue *q, unsigned int nr_elts,
++				    unsigned int lbegin, unsigned int lend)
+ {
+-	unsigned level, nr_levels, entries_per_level, remainder;
++	unsigned int level, nr_levels, entries_per_level, remainder;
+ 
+ 	BUG_ON(lbegin > lend);
+ 	BUG_ON(lend > q->nr_levels);
+@@ -426,7 +427,7 @@ static void q_set_targets(struct queue *q)
+ 
+ static void q_redistribute(struct queue *q)
+ {
+-	unsigned target, level;
++	unsigned int target, level;
+ 	struct ilist *l, *l_above;
+ 	struct entry *e;
+ 
+@@ -467,12 +468,12 @@ static void q_redistribute(struct queue *q)
+ 	}
+ }
+ 
+-static void q_requeue(struct queue *q, struct entry *e, unsigned extra_levels,
++static void q_requeue(struct queue *q, struct entry *e, unsigned int extra_levels,
+ 		      struct entry *s1, struct entry *s2)
+ {
+ 	struct entry *de;
+-	unsigned sentinels_passed = 0;
+-	unsigned new_level = min(q->nr_levels - 1u, e->level + extra_levels);
++	unsigned int sentinels_passed = 0;
++	unsigned int new_level = min(q->nr_levels - 1u, e->level + extra_levels);
+ 
+ 	/* try and find an entry to swap with */
+ 	if (extra_levels && (e->level < q->nr_levels - 1u)) {
+@@ -512,9 +513,9 @@ static void q_requeue(struct queue *q, struct entry *e, unsigned extra_levels,
+ #define EIGHTH (1u << (FP_SHIFT - 3u))
+ 
+ struct stats {
+-	unsigned hit_threshold;
+-	unsigned hits;
+-	unsigned misses;
++	unsigned int hit_threshold;
++	unsigned int hits;
++	unsigned int misses;
+ };
+ 
+ enum performance {
+@@ -523,7 +524,7 @@ enum performance {
+ 	Q_WELL
+ };
+ 
+-static void stats_init(struct stats *s, unsigned nr_levels)
++static void stats_init(struct stats *s, unsigned int nr_levels)
+ {
+ 	s->hit_threshold = (nr_levels * 3u) / 4u;
+ 	s->hits = 0u;
+@@ -535,7 +536,7 @@ static void stats_reset(struct stats *s)
+ 	s->hits = s->misses = 0u;
+ }
+ 
+-static void stats_level_accessed(struct stats *s, unsigned level)
++static void stats_level_accessed(struct stats *s, unsigned int level)
+ {
+ 	if (level >= s->hit_threshold)
+ 		s->hits++;
+@@ -556,7 +557,7 @@ static void stats_miss(struct stats *s)
+  */
+ static enum performance stats_assess(struct stats *s)
+ {
+-	unsigned confidence = safe_div(s->hits << FP_SHIFT, s->hits + s->misses);
++	unsigned int confidence = safe_div(s->hits << FP_SHIFT, s->hits + s->misses);
+ 
+ 	if (confidence < SIXTEENTH)
+ 		return Q_POOR;
+@@ -573,16 +574,16 @@ static enum performance stats_assess(struct stats *s)
+ struct smq_hash_table {
+ 	struct entry_space *es;
+ 	unsigned long long hash_bits;
+-	unsigned *buckets;
++	unsigned int *buckets;
+ };
+ 
+ /*
+  * All cache entries are stored in a chained hash table.  To save space we
+  * use indexing again, and only store indexes to the next entry.
+  */
+-static int h_init(struct smq_hash_table *ht, struct entry_space *es, unsigned nr_entries)
++static int h_init(struct smq_hash_table *ht, struct entry_space *es, unsigned int nr_entries)
+ {
+-	unsigned i, nr_buckets;
++	unsigned int i, nr_buckets;
+ 
+ 	ht->es = es;
+ 	nr_buckets = roundup_pow_of_two(max(nr_entries / 4u, 16u));
+@@ -603,7 +604,7 @@ static void h_exit(struct smq_hash_table *ht)
+ 	vfree(ht->buckets);
+ }
+ 
+-static struct entry *h_head(struct smq_hash_table *ht, unsigned bucket)
++static struct entry *h_head(struct smq_hash_table *ht, unsigned int bucket)
+ {
+ 	return to_entry(ht->es, ht->buckets[bucket]);
+ }
+@@ -613,7 +614,7 @@ static struct entry *h_next(struct smq_hash_table *ht, struct entry *e)
+ 	return to_entry(ht->es, e->hash_next);
+ }
+ 
+-static void __h_insert(struct smq_hash_table *ht, unsigned bucket, struct entry *e)
++static void __h_insert(struct smq_hash_table *ht, unsigned int bucket, struct entry *e)
+ {
+ 	e->hash_next = ht->buckets[bucket];
+ 	ht->buckets[bucket] = to_index(ht->es, e);
+@@ -621,11 +622,11 @@ static void __h_insert(struct smq_hash_table *ht, unsigned bucket, struct entry
+ 
+ static void h_insert(struct smq_hash_table *ht, struct entry *e)
+ {
+-	unsigned h = hash_64(from_oblock(e->oblock), ht->hash_bits);
++	unsigned int h = hash_64(from_oblock(e->oblock), ht->hash_bits);
+ 	__h_insert(ht, h, e);
+ }
+ 
+-static struct entry *__h_lookup(struct smq_hash_table *ht, unsigned h, dm_oblock_t oblock,
++static struct entry *__h_lookup(struct smq_hash_table *ht, unsigned int h, dm_oblock_t oblock,
+ 				struct entry **prev)
+ {
+ 	struct entry *e;
+@@ -641,7 +642,7 @@ static struct entry *__h_lookup(struct smq_hash_table *ht, unsigned h, dm_oblock
+ 	return NULL;
+ }
+ 
+-static void __h_unlink(struct smq_hash_table *ht, unsigned h,
++static void __h_unlink(struct smq_hash_table *ht, unsigned int h,
+ 		       struct entry *e, struct entry *prev)
+ {
+ 	if (prev)
+@@ -656,7 +657,7 @@ static void __h_unlink(struct smq_hash_table *ht, unsigned h,
+ static struct entry *h_lookup(struct smq_hash_table *ht, dm_oblock_t oblock)
+ {
+ 	struct entry *e, *prev;
+-	unsigned h = hash_64(from_oblock(oblock), ht->hash_bits);
++	unsigned int h = hash_64(from_oblock(oblock), ht->hash_bits);
+ 
+ 	e = __h_lookup(ht, h, oblock, &prev);
+ 	if (e && prev) {
+@@ -673,7 +674,7 @@ static struct entry *h_lookup(struct smq_hash_table *ht, dm_oblock_t oblock)
+ 
+ static void h_remove(struct smq_hash_table *ht, struct entry *e)
+ {
+-	unsigned h = hash_64(from_oblock(e->oblock), ht->hash_bits);
++	unsigned int h = hash_64(from_oblock(e->oblock), ht->hash_bits);
+ 	struct entry *prev;
+ 
+ 	/*
+@@ -689,16 +690,16 @@ static void h_remove(struct smq_hash_table *ht, struct entry *e)
+ 
+ struct entry_alloc {
+ 	struct entry_space *es;
+-	unsigned begin;
++	unsigned int begin;
+ 
+-	unsigned nr_allocated;
++	unsigned int nr_allocated;
+ 	struct ilist free;
+ };
+ 
+ static void init_allocator(struct entry_alloc *ea, struct entry_space *es,
+-			   unsigned begin, unsigned end)
++			   unsigned int begin, unsigned int end)
+ {
+-	unsigned i;
++	unsigned int i;
+ 
+ 	ea->es = es;
+ 	ea->nr_allocated = 0u;
+@@ -742,7 +743,7 @@ static struct entry *alloc_entry(struct entry_alloc *ea)
+ /*
+  * This assumes the cblock hasn't already been allocated.
+  */
+-static struct entry *alloc_particular_entry(struct entry_alloc *ea, unsigned i)
++static struct entry *alloc_particular_entry(struct entry_alloc *ea, unsigned int i)
+ {
+ 	struct entry *e = __get_entry(ea->es, ea->begin + i);
+ 
+@@ -770,12 +771,12 @@ static bool allocator_empty(struct entry_alloc *ea)
+ 	return l_empty(&ea->free);
+ }
+ 
+-static unsigned get_index(struct entry_alloc *ea, struct entry *e)
++static unsigned int get_index(struct entry_alloc *ea, struct entry *e)
+ {
+ 	return to_index(ea->es, e) - ea->begin;
+ }
+ 
+-static struct entry *get_entry(struct entry_alloc *ea, unsigned index)
++static struct entry *get_entry(struct entry_alloc *ea, unsigned int index)
+ {
+ 	return __get_entry(ea->es, ea->begin + index);
+ }
+@@ -800,9 +801,9 @@ struct smq_policy {
+ 	sector_t cache_block_size;
+ 
+ 	sector_t hotspot_block_size;
+-	unsigned nr_hotspot_blocks;
+-	unsigned cache_blocks_per_hotspot_block;
+-	unsigned hotspot_level_jump;
++	unsigned int nr_hotspot_blocks;
++	unsigned int cache_blocks_per_hotspot_block;
++	unsigned int hotspot_level_jump;
+ 
+ 	struct entry_space es;
+ 	struct entry_alloc writeback_sentinel_alloc;
+@@ -831,7 +832,7 @@ struct smq_policy {
+ 	 * Keeps track of time, incremented by the core.  We use this to
+ 	 * avoid attributing multiple hits within the same tick.
+ 	 */
+-	unsigned tick;
++	unsigned int tick;
+ 
+ 	/*
+ 	 * The hash tables allows us to quickly find an entry by origin
+@@ -846,8 +847,8 @@ struct smq_policy {
+ 	bool current_demote_sentinels;
+ 	unsigned long next_demote_period;
+ 
+-	unsigned write_promote_level;
+-	unsigned read_promote_level;
++	unsigned int write_promote_level;
++	unsigned int read_promote_level;
+ 
+ 	unsigned long next_hotspot_period;
+ 	unsigned long next_cache_period;
+@@ -859,24 +860,24 @@ struct smq_policy {
+ 
+ /*----------------------------------------------------------------*/
+ 
+-static struct entry *get_sentinel(struct entry_alloc *ea, unsigned level, bool which)
++static struct entry *get_sentinel(struct entry_alloc *ea, unsigned int level, bool which)
+ {
+ 	return get_entry(ea, which ? level : NR_CACHE_LEVELS + level);
+ }
+ 
+-static struct entry *writeback_sentinel(struct smq_policy *mq, unsigned level)
++static struct entry *writeback_sentinel(struct smq_policy *mq, unsigned int level)
+ {
+ 	return get_sentinel(&mq->writeback_sentinel_alloc, level, mq->current_writeback_sentinels);
+ }
+ 
+-static struct entry *demote_sentinel(struct smq_policy *mq, unsigned level)
++static struct entry *demote_sentinel(struct smq_policy *mq, unsigned int level)
+ {
+ 	return get_sentinel(&mq->demote_sentinel_alloc, level, mq->current_demote_sentinels);
+ }
+ 
+ static void __update_writeback_sentinels(struct smq_policy *mq)
+ {
+-	unsigned level;
++	unsigned int level;
+ 	struct queue *q = &mq->dirty;
+ 	struct entry *sentinel;
+ 
+@@ -889,7 +890,7 @@ static void __update_writeback_sentinels(struct smq_policy *mq)
+ 
+ static void __update_demote_sentinels(struct smq_policy *mq)
+ {
+-	unsigned level;
++	unsigned int level;
+ 	struct queue *q = &mq->clean;
+ 	struct entry *sentinel;
+ 
+@@ -917,7 +918,7 @@ static void update_sentinels(struct smq_policy *mq)
+ 
+ static void __sentinels_init(struct smq_policy *mq)
+ {
+-	unsigned level;
++	unsigned int level;
+ 	struct entry *sentinel;
+ 
+ 	for (level = 0; level < NR_CACHE_LEVELS; level++) {
+@@ -1008,7 +1009,7 @@ static void requeue(struct smq_policy *mq, struct entry *e)
+ 	}
+ }
+ 
+-static unsigned default_promote_level(struct smq_policy *mq)
++static unsigned int default_promote_level(struct smq_policy *mq)
+ {
+ 	/*
+ 	 * The promote level depends on the current performance of the
+@@ -1030,9 +1031,9 @@ static unsigned default_promote_level(struct smq_policy *mq)
+ 		1, 1, 1, 2, 4, 6, 7, 8, 7, 6, 4, 4, 3, 3, 2, 2, 1
+ 	};
+ 
+-	unsigned hits = mq->cache_stats.hits;
+-	unsigned misses = mq->cache_stats.misses;
+-	unsigned index = safe_div(hits << 4u, hits + misses);
++	unsigned int hits = mq->cache_stats.hits;
++	unsigned int misses = mq->cache_stats.misses;
++	unsigned int index = safe_div(hits << 4u, hits + misses);
+ 	return table[index];
+ }
+ 
+@@ -1042,7 +1043,7 @@ static void update_promote_levels(struct smq_policy *mq)
+ 	 * If there are unused cache entries then we want to be really
+ 	 * eager to promote.
+ 	 */
+-	unsigned threshold_level = allocator_empty(&mq->cache_alloc) ?
++	unsigned int threshold_level = allocator_empty(&mq->cache_alloc) ?
+ 		default_promote_level(mq) : (NR_HOTSPOT_LEVELS / 2u);
+ 
+ 	threshold_level = max(threshold_level, NR_HOTSPOT_LEVELS);
+@@ -1124,7 +1125,7 @@ static void end_cache_period(struct smq_policy *mq)
+ #define CLEAN_TARGET 25u
+ #define FREE_TARGET 25u
+ 
+-static unsigned percent_to_target(struct smq_policy *mq, unsigned p)
++static unsigned int percent_to_target(struct smq_policy *mq, unsigned int p)
+ {
+ 	return from_cblock(mq->cache_size) * p / 100u;
+ }
+@@ -1150,7 +1151,7 @@ static bool clean_target_met(struct smq_policy *mq, bool idle)
+ 
+ static bool free_target_met(struct smq_policy *mq)
+ {
+-	unsigned nr_free;
++	unsigned int nr_free;
+ 
+ 	nr_free = from_cblock(mq->cache_size) - mq->cache_alloc.nr_allocated;
+ 	return (nr_free + btracker_nr_demotions_queued(mq->bg_work)) >=
+@@ -1300,7 +1301,7 @@ static dm_oblock_t to_hblock(struct smq_policy *mq, dm_oblock_t b)
+ 
+ static struct entry *update_hotspot_queue(struct smq_policy *mq, dm_oblock_t b)
+ {
+-	unsigned hi;
++	unsigned int hi;
+ 	dm_oblock_t hb = to_hblock(mq, b);
+ 	struct entry *e = h_lookup(&mq->hotspot_table, hb);
+ 
+@@ -1549,7 +1550,7 @@ static void smq_clear_dirty(struct dm_cache_policy *p, dm_cblock_t cblock)
+ 	spin_unlock_irqrestore(&mq->lock, flags);
+ }
+ 
+-static unsigned random_level(dm_cblock_t cblock)
++static unsigned int random_level(dm_cblock_t cblock)
+ {
+ 	return hash_32(from_cblock(cblock), 9) & (NR_CACHE_LEVELS - 1);
+ }
+@@ -1660,7 +1661,7 @@ static int mq_set_config_value(struct dm_cache_policy *p,
+ }
+ 
+ static int mq_emit_config_values(struct dm_cache_policy *p, char *result,
+-				 unsigned maxlen, ssize_t *sz_ptr)
++				 unsigned int maxlen, ssize_t *sz_ptr)
+ {
+ 	ssize_t sz = *sz_ptr;
+ 
+@@ -1699,16 +1700,16 @@ static void init_policy_functions(struct smq_policy *mq, bool mimic_mq)
+ 
+ static bool too_many_hotspot_blocks(sector_t origin_size,
+ 				    sector_t hotspot_block_size,
+-				    unsigned nr_hotspot_blocks)
++				    unsigned int nr_hotspot_blocks)
+ {
+ 	return (hotspot_block_size * nr_hotspot_blocks) > origin_size;
+ }
+ 
+ static void calc_hotspot_params(sector_t origin_size,
+ 				sector_t cache_block_size,
+-				unsigned nr_cache_blocks,
++				unsigned int nr_cache_blocks,
+ 				sector_t *hotspot_block_size,
+-				unsigned *nr_hotspot_blocks)
++				unsigned int *nr_hotspot_blocks)
+ {
+ 	*hotspot_block_size = cache_block_size * 16u;
+ 	*nr_hotspot_blocks = max(nr_cache_blocks / 4u, 1024u);
+@@ -1724,9 +1725,9 @@ static struct dm_cache_policy *__smq_create(dm_cblock_t cache_size,
+ 					    bool mimic_mq,
+ 					    bool migrations_allowed)
+ {
+-	unsigned i;
+-	unsigned nr_sentinels_per_queue = 2u * NR_CACHE_LEVELS;
+-	unsigned total_sentinels = 2u * nr_sentinels_per_queue;
++	unsigned int i;
++	unsigned int nr_sentinels_per_queue = 2u * NR_CACHE_LEVELS;
++	unsigned int total_sentinels = 2u * nr_sentinels_per_queue;
+ 	struct smq_policy *mq = kzalloc(sizeof(*mq), GFP_KERNEL);
+ 
+ 	if (!mq)
+diff --git a/drivers/md/dm-cache-policy.c b/drivers/md/dm-cache-policy.c
+index c1a3cee99b445..2e58bbcf3e3bd 100644
+--- a/drivers/md/dm-cache-policy.c
++++ b/drivers/md/dm-cache-policy.c
+@@ -154,7 +154,7 @@ const char *dm_cache_policy_get_name(struct dm_cache_policy *p)
+ }
+ EXPORT_SYMBOL_GPL(dm_cache_policy_get_name);
+ 
+-const unsigned *dm_cache_policy_get_version(struct dm_cache_policy *p)
++const unsigned int *dm_cache_policy_get_version(struct dm_cache_policy *p)
+ {
+ 	struct dm_cache_policy_type *t = p->private;
+ 
+diff --git a/drivers/md/dm-cache-policy.h b/drivers/md/dm-cache-policy.h
+index 06eb31af626f1..6ba3e9c91af53 100644
+--- a/drivers/md/dm-cache-policy.h
++++ b/drivers/md/dm-cache-policy.h
+@@ -128,7 +128,7 @@ struct dm_cache_policy {
+ 	 * Configuration.
+ 	 */
+ 	int (*emit_config_values)(struct dm_cache_policy *p, char *result,
+-				  unsigned maxlen, ssize_t *sz_ptr);
++				  unsigned int maxlen, ssize_t *sz_ptr);
+ 	int (*set_config_value)(struct dm_cache_policy *p,
+ 				const char *key, const char *value);
+ 
+@@ -157,7 +157,7 @@ struct dm_cache_policy_type {
+ 	 * what gets passed on the target line to select your policy.
+ 	 */
+ 	char name[CACHE_POLICY_NAME_SIZE];
+-	unsigned version[CACHE_POLICY_VERSION_SIZE];
++	unsigned int version[CACHE_POLICY_VERSION_SIZE];
+ 
+ 	/*
+ 	 * For use by an alias dm_cache_policy_type to point to the
+diff --git a/drivers/md/dm-cache-target.c b/drivers/md/dm-cache-target.c
+index 17fde3e5a1f7b..8f7426b71e025 100644
+--- a/drivers/md/dm-cache-target.c
++++ b/drivers/md/dm-cache-target.c
+@@ -275,7 +275,7 @@ enum cache_io_mode {
+ struct cache_features {
+ 	enum cache_metadata_mode mode;
+ 	enum cache_io_mode io_mode;
+-	unsigned metadata_version;
++	unsigned int metadata_version;
+ 	bool discard_passdown:1;
+ };
+ 
+@@ -362,7 +362,7 @@ struct cache {
+ 	 * Rather than reconstructing the table line for the status we just
+ 	 * save it and regurgitate.
+ 	 */
+-	unsigned nr_ctr_args;
++	unsigned int nr_ctr_args;
+ 	const char **ctr_args;
+ 
+ 	struct dm_kcopyd_client *copier;
+@@ -378,7 +378,7 @@ struct cache {
+ 	unsigned long *dirty_bitset;
+ 	atomic_t nr_dirty;
+ 
+-	unsigned policy_nr_args;
++	unsigned int policy_nr_args;
+ 	struct dm_cache_policy *policy;
+ 
+ 	/*
+@@ -409,7 +409,7 @@ struct cache {
+ 
+ struct per_bio_data {
+ 	bool tick:1;
+-	unsigned req_nr:2;
++	unsigned int req_nr:2;
+ 	struct dm_bio_prison_cell_v2 *cell;
+ 	struct dm_hook_info hook_info;
+ 	sector_t len;
+@@ -517,7 +517,7 @@ static void build_key(dm_oblock_t begin, dm_oblock_t end, struct dm_cell_key_v2
+ #define WRITE_LOCK_LEVEL 0
+ #define READ_WRITE_LOCK_LEVEL 1
+ 
+-static unsigned lock_level(struct bio *bio)
++static unsigned int lock_level(struct bio *bio)
+ {
+ 	return bio_data_dir(bio) == WRITE ?
+ 		WRITE_LOCK_LEVEL :
+@@ -1884,7 +1884,7 @@ static void check_migrations(struct work_struct *ws)
+  */
+ static void destroy(struct cache *cache)
+ {
+-	unsigned i;
++	unsigned int i;
+ 
+ 	mempool_exit(&cache->migration_pool);
+ 
+@@ -2124,7 +2124,7 @@ static int parse_features(struct cache_args *ca, struct dm_arg_set *as,
+ 	};
+ 
+ 	int r, mode_ctr = 0;
+-	unsigned argc;
++	unsigned int argc;
+ 	const char *arg;
+ 	struct cache_features *cf = &ca->features;
+ 
+@@ -2544,7 +2544,7 @@ bad:
+ 
+ static int copy_ctr_args(struct cache *cache, int argc, const char **argv)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	const char **copy;
+ 
+ 	copy = kcalloc(argc, sizeof(*copy), GFP_KERNEL);
+@@ -2566,7 +2566,7 @@ static int copy_ctr_args(struct cache *cache, int argc, const char **argv)
+ 	return 0;
+ }
+ 
+-static int cache_ctr(struct dm_target *ti, unsigned argc, char **argv)
++static int cache_ctr(struct dm_target *ti, unsigned int argc, char **argv)
+ {
+ 	int r = -EINVAL;
+ 	struct cache_args *ca;
+@@ -2669,7 +2669,7 @@ static int write_dirty_bitset(struct cache *cache)
+ 
+ static int write_discard_bitset(struct cache *cache)
+ {
+-	unsigned i, r;
++	unsigned int i, r;
+ 
+ 	if (get_cache_mode(cache) >= CM_READ_ONLY)
+ 		return -EINVAL;
+@@ -2983,11 +2983,11 @@ static void cache_resume(struct dm_target *ti)
+ }
+ 
+ static void emit_flags(struct cache *cache, char *result,
+-		       unsigned maxlen, ssize_t *sz_ptr)
++		       unsigned int maxlen, ssize_t *sz_ptr)
+ {
+ 	ssize_t sz = *sz_ptr;
+ 	struct cache_features *cf = &cache->features;
+-	unsigned count = (cf->metadata_version == 2) + !cf->discard_passdown + 1;
++	unsigned int count = (cf->metadata_version == 2) + !cf->discard_passdown + 1;
+ 
+ 	DMEMIT("%u ", count);
+ 
+@@ -3027,10 +3027,10 @@ static void emit_flags(struct cache *cache, char *result,
+  * <policy name> <#policy args> <policy args>* <cache metadata mode> <needs_check>
+  */
+ static void cache_status(struct dm_target *ti, status_type_t type,
+-			 unsigned status_flags, char *result, unsigned maxlen)
++			 unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	int r = 0;
+-	unsigned i;
++	unsigned int i;
+ 	ssize_t sz = 0;
+ 	dm_block_t nr_free_blocks_metadata = 0;
+ 	dm_block_t nr_blocks_metadata = 0;
+@@ -3067,18 +3067,18 @@ static void cache_status(struct dm_target *ti, status_type_t type,
+ 		residency = policy_residency(cache->policy);
+ 
+ 		DMEMIT("%u %llu/%llu %llu %llu/%llu %u %u %u %u %u %u %lu ",
+-		       (unsigned)DM_CACHE_METADATA_BLOCK_SIZE,
++		       (unsigned int)DM_CACHE_METADATA_BLOCK_SIZE,
+ 		       (unsigned long long)(nr_blocks_metadata - nr_free_blocks_metadata),
+ 		       (unsigned long long)nr_blocks_metadata,
+ 		       (unsigned long long)cache->sectors_per_block,
+ 		       (unsigned long long) from_cblock(residency),
+ 		       (unsigned long long) from_cblock(cache->cache_size),
+-		       (unsigned) atomic_read(&cache->stats.read_hit),
+-		       (unsigned) atomic_read(&cache->stats.read_miss),
+-		       (unsigned) atomic_read(&cache->stats.write_hit),
+-		       (unsigned) atomic_read(&cache->stats.write_miss),
+-		       (unsigned) atomic_read(&cache->stats.demotion),
+-		       (unsigned) atomic_read(&cache->stats.promotion),
++		       (unsigned int) atomic_read(&cache->stats.read_hit),
++		       (unsigned int) atomic_read(&cache->stats.read_miss),
++		       (unsigned int) atomic_read(&cache->stats.write_hit),
++		       (unsigned int) atomic_read(&cache->stats.write_miss),
++		       (unsigned int) atomic_read(&cache->stats.demotion),
++		       (unsigned int) atomic_read(&cache->stats.promotion),
+ 		       (unsigned long) atomic_read(&cache->nr_dirty));
+ 
+ 		emit_flags(cache, result, maxlen, &sz);
+@@ -3257,11 +3257,11 @@ static int request_invalidation(struct cache *cache, struct cblock_range *range)
+ 	return r;
+ }
+ 
+-static int process_invalidate_cblocks_message(struct cache *cache, unsigned count,
++static int process_invalidate_cblocks_message(struct cache *cache, unsigned int count,
+ 					      const char **cblock_ranges)
+ {
+ 	int r = 0;
+-	unsigned i;
++	unsigned int i;
+ 	struct cblock_range range;
+ 
+ 	if (!passthrough_mode(cache)) {
+@@ -3298,8 +3298,8 @@ static int process_invalidate_cblocks_message(struct cache *cache, unsigned coun
+  *
+  * The key migration_threshold is supported by the cache target core.
+  */
+-static int cache_message(struct dm_target *ti, unsigned argc, char **argv,
+-			 char *result, unsigned maxlen)
++static int cache_message(struct dm_target *ti, unsigned int argc, char **argv,
++			 char *result, unsigned int maxlen)
+ {
+ 	struct cache *cache = ti->private;
+ 
+diff --git a/drivers/md/dm-core.h b/drivers/md/dm-core.h
+index 6c6bd24774f25..28c641352de9b 100644
+--- a/drivers/md/dm-core.h
++++ b/drivers/md/dm-core.h
+@@ -119,7 +119,7 @@ struct mapped_device {
+ 	struct dm_stats stats;
+ 
+ 	/* the number of internal suspends */
+-	unsigned internal_suspend_count;
++	unsigned int internal_suspend_count;
+ 
+ 	int swap_bios;
+ 	struct semaphore swap_bios_semaphore;
+@@ -326,9 +326,9 @@ static inline struct completion *dm_get_completion_from_kobject(struct kobject *
+ 	return &container_of(kobj, struct dm_kobject_holder, kobj)->completion;
+ }
+ 
+-unsigned __dm_get_module_param(unsigned *module_param, unsigned def, unsigned max);
++unsigned int __dm_get_module_param(unsigned int *module_param, unsigned int def, unsigned int max);
+ 
+-static inline bool dm_message_test_buffer_overflow(char *result, unsigned maxlen)
++static inline bool dm_message_test_buffer_overflow(char *result, unsigned int maxlen)
+ {
+ 	return !maxlen || strlen(result) + 1 >= maxlen;
+ }
+diff --git a/drivers/md/dm-crypt.c b/drivers/md/dm-crypt.c
+index dc2d0d61ade93..ee269b1d09fac 100644
+--- a/drivers/md/dm-crypt.c
++++ b/drivers/md/dm-crypt.c
+@@ -173,14 +173,14 @@ struct crypt_config {
+ 	} iv_gen_private;
+ 	u64 iv_offset;
+ 	unsigned int iv_size;
+-	unsigned short int sector_size;
++	unsigned short sector_size;
+ 	unsigned char sector_shift;
+ 
+ 	union {
+ 		struct crypto_skcipher **tfms;
+ 		struct crypto_aead **tfms_aead;
+ 	} cipher_tfm;
+-	unsigned tfms_count;
++	unsigned int tfms_count;
+ 	unsigned long cipher_flags;
+ 
+ 	/*
+@@ -214,7 +214,7 @@ struct crypt_config {
+ 	 * pool for per bio private data, crypto requests,
+ 	 * encryption requeusts/buffer pages and integrity tags
+ 	 */
+-	unsigned tag_pool_max_sectors;
++	unsigned int tag_pool_max_sectors;
+ 	mempool_t tag_pool;
+ 	mempool_t req_pool;
+ 	mempool_t page_pool;
+@@ -231,7 +231,7 @@ struct crypt_config {
+ #define POOL_ENTRY_SIZE	512
+ 
+ static DEFINE_SPINLOCK(dm_crypt_clients_lock);
+-static unsigned dm_crypt_clients_n = 0;
++static unsigned int dm_crypt_clients_n = 0;
+ static volatile unsigned long dm_crypt_pages_per_client;
+ #define DM_CRYPT_MEMORY_PERCENT			2
+ #define DM_CRYPT_MIN_PAGES_PER_CLIENT		(BIO_MAX_VECS * 16)
+@@ -356,7 +356,7 @@ static int crypt_iv_essiv_gen(struct crypt_config *cc, u8 *iv,
+ static int crypt_iv_benbi_ctr(struct crypt_config *cc, struct dm_target *ti,
+ 			      const char *opts)
+ {
+-	unsigned bs;
++	unsigned int bs;
+ 	int log;
+ 
+ 	if (crypt_integrity_aead(cc))
+@@ -1466,7 +1466,7 @@ static void kcryptd_async_done(struct crypto_async_request *async_req,
+ static int crypt_alloc_req_skcipher(struct crypt_config *cc,
+ 				     struct convert_context *ctx)
+ {
+-	unsigned key_index = ctx->cc_sector & (cc->tfms_count - 1);
++	unsigned int key_index = ctx->cc_sector & (cc->tfms_count - 1);
+ 
+ 	if (!ctx->r.req) {
+ 		ctx->r.req = mempool_alloc(&cc->req_pool, in_interrupt() ? GFP_ATOMIC : GFP_NOIO);
+@@ -1660,13 +1660,13 @@ static void crypt_free_buffer_pages(struct crypt_config *cc, struct bio *clone);
+  * non-blocking allocations without a mutex first but on failure we fallback
+  * to blocking allocations with a mutex.
+  */
+-static struct bio *crypt_alloc_buffer(struct dm_crypt_io *io, unsigned size)
++static struct bio *crypt_alloc_buffer(struct dm_crypt_io *io, unsigned int size)
+ {
+ 	struct crypt_config *cc = io->cc;
+ 	struct bio *clone;
+ 	unsigned int nr_iovecs = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
+ 	gfp_t gfp_mask = GFP_NOWAIT | __GFP_HIGHMEM;
+-	unsigned i, len, remaining_size;
++	unsigned int i, len, remaining_size;
+ 	struct page *page;
+ 
+ retry:
+@@ -1806,7 +1806,7 @@ static void crypt_endio(struct bio *clone)
+ {
+ 	struct dm_crypt_io *io = clone->bi_private;
+ 	struct crypt_config *cc = io->cc;
+-	unsigned rw = bio_data_dir(clone);
++	unsigned int rw = bio_data_dir(clone);
+ 	blk_status_t error;
+ 
+ 	/*
+@@ -2261,7 +2261,7 @@ static void crypt_free_tfms_aead(struct crypt_config *cc)
+ 
+ static void crypt_free_tfms_skcipher(struct crypt_config *cc)
+ {
+-	unsigned i;
++	unsigned int i;
+ 
+ 	if (!cc->cipher_tfm.tfms)
+ 		return;
+@@ -2286,7 +2286,7 @@ static void crypt_free_tfms(struct crypt_config *cc)
+ 
+ static int crypt_alloc_tfms_skcipher(struct crypt_config *cc, char *ciphermode)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	int err;
+ 
+ 	cc->cipher_tfm.tfms = kcalloc(cc->tfms_count,
+@@ -2344,12 +2344,12 @@ static int crypt_alloc_tfms(struct crypt_config *cc, char *ciphermode)
+ 		return crypt_alloc_tfms_skcipher(cc, ciphermode);
+ }
+ 
+-static unsigned crypt_subkey_size(struct crypt_config *cc)
++static unsigned int crypt_subkey_size(struct crypt_config *cc)
+ {
+ 	return (cc->key_size - cc->key_extra_size) >> ilog2(cc->tfms_count);
+ }
+ 
+-static unsigned crypt_authenckey_size(struct crypt_config *cc)
++static unsigned int crypt_authenckey_size(struct crypt_config *cc)
+ {
+ 	return crypt_subkey_size(cc) + RTA_SPACE(sizeof(struct crypto_authenc_key_param));
+ }
+@@ -2360,7 +2360,7 @@ static unsigned crypt_authenckey_size(struct crypt_config *cc)
+  * This funcion converts cc->key to this special format.
+  */
+ static void crypt_copy_authenckey(char *p, const void *key,
+-				  unsigned enckeylen, unsigned authkeylen)
++				  unsigned int enckeylen, unsigned int authkeylen)
+ {
+ 	struct crypto_authenc_key_param *param;
+ 	struct rtattr *rta;
+@@ -2378,7 +2378,7 @@ static void crypt_copy_authenckey(char *p, const void *key,
+ 
+ static int crypt_setkey(struct crypt_config *cc)
+ {
+-	unsigned subkey_size;
++	unsigned int subkey_size;
+ 	int err = 0, i, r;
+ 
+ 	/* Ignore extra keys (which are used for IV etc) */
+@@ -3417,7 +3417,7 @@ static int crypt_map(struct dm_target *ti, struct bio *bio)
+ 	crypt_io_init(io, cc, bio, dm_target_offset(ti, bio->bi_iter.bi_sector));
+ 
+ 	if (cc->on_disk_tag_size) {
+-		unsigned tag_len = cc->on_disk_tag_size * (bio_sectors(bio) >> cc->sector_shift);
++		unsigned int tag_len = cc->on_disk_tag_size * (bio_sectors(bio) >> cc->sector_shift);
+ 
+ 		if (unlikely(tag_len > KMALLOC_MAX_SIZE) ||
+ 		    unlikely(!(io->integrity_metadata = kmalloc(tag_len,
+@@ -3445,14 +3445,14 @@ static int crypt_map(struct dm_target *ti, struct bio *bio)
+ 
+ static char hex2asc(unsigned char c)
+ {
+-	return c + '0' + ((unsigned)(9 - c) >> 4 & 0x27);
++	return c + '0' + ((unsigned int)(9 - c) >> 4 & 0x27);
+ }
+ 
+ static void crypt_status(struct dm_target *ti, status_type_t type,
+-			 unsigned status_flags, char *result, unsigned maxlen)
++			 unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	struct crypt_config *cc = ti->private;
+-	unsigned i, sz = 0;
++	unsigned int i, sz = 0;
+ 	int num_feature_args = 0;
+ 
+ 	switch (type) {
+@@ -3568,8 +3568,8 @@ static void crypt_resume(struct dm_target *ti)
+  *	key set <key>
+  *	key wipe
+  */
+-static int crypt_message(struct dm_target *ti, unsigned argc, char **argv,
+-			 char *result, unsigned maxlen)
++static int crypt_message(struct dm_target *ti, unsigned int argc, char **argv,
++			 char *result, unsigned int maxlen)
+ {
+ 	struct crypt_config *cc = ti->private;
+ 	int key_size, ret = -EINVAL;
+@@ -3630,10 +3630,10 @@ static void crypt_io_hints(struct dm_target *ti, struct queue_limits *limits)
+ 	limits->max_segment_size = PAGE_SIZE;
+ 
+ 	limits->logical_block_size =
+-		max_t(unsigned, limits->logical_block_size, cc->sector_size);
++		max_t(unsigned int, limits->logical_block_size, cc->sector_size);
+ 	limits->physical_block_size =
+-		max_t(unsigned, limits->physical_block_size, cc->sector_size);
+-	limits->io_min = max_t(unsigned, limits->io_min, cc->sector_size);
++		max_t(unsigned int, limits->physical_block_size, cc->sector_size);
++	limits->io_min = max_t(unsigned int, limits->io_min, cc->sector_size);
+ 	limits->dma_alignment = limits->logical_block_size - 1;
+ }
+ 
+diff --git a/drivers/md/dm-delay.c b/drivers/md/dm-delay.c
+index 869afef5654ae..02b8f4e818276 100644
+--- a/drivers/md/dm-delay.c
++++ b/drivers/md/dm-delay.c
+@@ -20,8 +20,8 @@
+ struct delay_class {
+ 	struct dm_dev *dev;
+ 	sector_t start;
+-	unsigned delay;
+-	unsigned ops;
++	unsigned int delay;
++	unsigned int ops;
+ };
+ 
+ struct delay_c {
+@@ -305,7 +305,7 @@ static int delay_map(struct dm_target *ti, struct bio *bio)
+ 	DMEMIT("%s %llu %u", (c)->dev->name, (unsigned long long)(c)->start, (c)->delay)
+ 
+ static void delay_status(struct dm_target *ti, status_type_t type,
+-			 unsigned status_flags, char *result, unsigned maxlen)
++			 unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	struct delay_c *dc = ti->private;
+ 	int sz = 0;
+diff --git a/drivers/md/dm-ebs-target.c b/drivers/md/dm-ebs-target.c
+index 512cc6cea095a..7606c6695a0e2 100644
+--- a/drivers/md/dm-ebs-target.c
++++ b/drivers/md/dm-ebs-target.c
+@@ -390,7 +390,7 @@ static int ebs_map(struct dm_target *ti, struct bio *bio)
+ }
+ 
+ static void ebs_status(struct dm_target *ti, status_type_t type,
+-		       unsigned status_flags, char *result, unsigned maxlen)
++		       unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	struct ebs_c *ec = ti->private;
+ 
+diff --git a/drivers/md/dm-era-target.c b/drivers/md/dm-era-target.c
+index e92c1afc3677f..a96290103cca8 100644
+--- a/drivers/md/dm-era-target.c
++++ b/drivers/md/dm-era-target.c
+@@ -51,7 +51,7 @@ static void writeset_free(struct writeset *ws)
+ }
+ 
+ static int setup_on_disk_bitset(struct dm_disk_bitset *info,
+-				unsigned nr_bits, dm_block_t *root)
++				unsigned int nr_bits, dm_block_t *root)
+ {
+ 	int r;
+ 
+@@ -62,7 +62,7 @@ static int setup_on_disk_bitset(struct dm_disk_bitset *info,
+ 	return dm_bitset_resize(info, *root, 0, nr_bits, false, root);
+ }
+ 
+-static size_t bitset_size(unsigned nr_bits)
++static size_t bitset_size(unsigned int nr_bits)
+ {
+ 	return sizeof(unsigned long) * dm_div_up(nr_bits, BITS_PER_LONG);
+ }
+@@ -323,10 +323,10 @@ static int superblock_lock(struct era_metadata *md,
+ static int superblock_all_zeroes(struct dm_block_manager *bm, bool *result)
+ {
+ 	int r;
+-	unsigned i;
++	unsigned int i;
+ 	struct dm_block *b;
+ 	__le64 *data_le, zero = cpu_to_le64(0);
+-	unsigned sb_block_size = dm_bm_block_size(bm) / sizeof(__le64);
++	unsigned int sb_block_size = dm_bm_block_size(bm) / sizeof(__le64);
+ 
+ 	/*
+ 	 * We can't use a validator here - it may be all zeroes.
+@@ -363,12 +363,12 @@ static void ws_unpack(const struct writeset_disk *disk, struct writeset_metadata
+ 	core->root = le64_to_cpu(disk->root);
+ }
+ 
+-static void ws_inc(void *context, const void *value, unsigned count)
++static void ws_inc(void *context, const void *value, unsigned int count)
+ {
+ 	struct era_metadata *md = context;
+ 	struct writeset_disk ws_d;
+ 	dm_block_t b;
+-	unsigned i;
++	unsigned int i;
+ 
+ 	for (i = 0; i < count; i++) {
+ 		memcpy(&ws_d, value + (i * sizeof(ws_d)), sizeof(ws_d));
+@@ -377,12 +377,12 @@ static void ws_inc(void *context, const void *value, unsigned count)
+ 	}
+ }
+ 
+-static void ws_dec(void *context, const void *value, unsigned count)
++static void ws_dec(void *context, const void *value, unsigned int count)
+ {
+ 	struct era_metadata *md = context;
+ 	struct writeset_disk ws_d;
+ 	dm_block_t b;
+-	unsigned i;
++	unsigned int i;
+ 
+ 	for (i = 0; i < count; i++) {
+ 		memcpy(&ws_d, value + (i * sizeof(ws_d)), sizeof(ws_d));
+@@ -667,7 +667,7 @@ static void swap_writeset(struct era_metadata *md, struct writeset *new_writeset
+  *--------------------------------------------------------------*/
+ struct digest {
+ 	uint32_t era;
+-	unsigned nr_bits, current_bit;
++	unsigned int nr_bits, current_bit;
+ 	struct writeset_metadata writeset;
+ 	__le32 value;
+ 	struct dm_disk_bitset info;
+@@ -702,7 +702,7 @@ static int metadata_digest_transcribe_writeset(struct era_metadata *md,
+ {
+ 	int r;
+ 	bool marked;
+-	unsigned b, e = min(d->current_bit + INSERTS_PER_STEP, d->nr_bits);
++	unsigned int b, e = min(d->current_bit + INSERTS_PER_STEP, d->nr_bits);
+ 
+ 	for (b = d->current_bit; b < e; b++) {
+ 		r = writeset_marked_on_disk(&d->info, &d->writeset, b, &marked);
+@@ -1439,7 +1439,7 @@ static bool valid_block_size(dm_block_t block_size)
+ /*
+  * <metadata dev> <data dev> <data block size (sectors)>
+  */
+-static int era_ctr(struct dm_target *ti, unsigned argc, char **argv)
++static int era_ctr(struct dm_target *ti, unsigned int argc, char **argv)
+ {
+ 	int r;
+ 	char dummy;
+@@ -1618,7 +1618,7 @@ static int era_preresume(struct dm_target *ti)
+  * <current era> <held metadata root | '-'>
+  */
+ static void era_status(struct dm_target *ti, status_type_t type,
+-		       unsigned status_flags, char *result, unsigned maxlen)
++		       unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	int r;
+ 	struct era *era = ti->private;
+@@ -1633,10 +1633,10 @@ static void era_status(struct dm_target *ti, status_type_t type,
+ 			goto err;
+ 
+ 		DMEMIT("%u %llu/%llu %u",
+-		       (unsigned) (DM_ERA_METADATA_BLOCK_SIZE >> SECTOR_SHIFT),
++		       (unsigned int) (DM_ERA_METADATA_BLOCK_SIZE >> SECTOR_SHIFT),
+ 		       (unsigned long long) stats.used,
+ 		       (unsigned long long) stats.total,
+-		       (unsigned) stats.era);
++		       (unsigned int) stats.era);
+ 
+ 		if (stats.snap != SUPERBLOCK_LOCATION)
+ 			DMEMIT(" %llu", stats.snap);
+@@ -1662,8 +1662,8 @@ err:
+ 	DMEMIT("Error");
+ }
+ 
+-static int era_message(struct dm_target *ti, unsigned argc, char **argv,
+-		       char *result, unsigned maxlen)
++static int era_message(struct dm_target *ti, unsigned int argc, char **argv,
++		       char *result, unsigned int maxlen)
+ {
+ 	struct era *era = ti->private;
+ 
+diff --git a/drivers/md/dm-exception-store.c b/drivers/md/dm-exception-store.c
+index 3997f34cfebc6..cc3987c97eb94 100644
+--- a/drivers/md/dm-exception-store.c
++++ b/drivers/md/dm-exception-store.c
+@@ -142,7 +142,7 @@ EXPORT_SYMBOL(dm_exception_store_type_unregister);
+ static int set_chunk_size(struct dm_exception_store *store,
+ 			  const char *chunk_size_arg, char **error)
+ {
+-	unsigned chunk_size;
++	unsigned int chunk_size;
+ 
+ 	if (kstrtouint(chunk_size_arg, 10, &chunk_size)) {
+ 		*error = "Invalid chunk size";
+@@ -158,7 +158,7 @@ static int set_chunk_size(struct dm_exception_store *store,
+ }
+ 
+ int dm_exception_store_set_chunk_size(struct dm_exception_store *store,
+-				      unsigned chunk_size,
++				      unsigned int chunk_size,
+ 				      char **error)
+ {
+ 	/* Check chunk_size is a power of 2 */
+@@ -190,7 +190,7 @@ int dm_exception_store_set_chunk_size(struct dm_exception_store *store,
+ 
+ int dm_exception_store_create(struct dm_target *ti, int argc, char **argv,
+ 			      struct dm_snapshot *snap,
+-			      unsigned *args_used,
++			      unsigned int *args_used,
+ 			      struct dm_exception_store **store)
+ {
+ 	int r = 0;
+diff --git a/drivers/md/dm-exception-store.h b/drivers/md/dm-exception-store.h
+index b5f20eba36415..862df68a7db04 100644
+--- a/drivers/md/dm-exception-store.h
++++ b/drivers/md/dm-exception-store.h
+@@ -96,9 +96,9 @@ struct dm_exception_store_type {
+ 	 */
+ 	void (*drop_snapshot) (struct dm_exception_store *store);
+ 
+-	unsigned (*status) (struct dm_exception_store *store,
+-			    status_type_t status, char *result,
+-			    unsigned maxlen);
++	unsigned int (*status) (struct dm_exception_store *store,
++				status_type_t status, char *result,
++				unsigned int maxlen);
+ 
+ 	/*
+ 	 * Return how full the snapshot is.
+@@ -118,9 +118,9 @@ struct dm_exception_store {
+ 	struct dm_snapshot *snap;
+ 
+ 	/* Size of data blocks saved - must be a power of 2 */
+-	unsigned chunk_size;
+-	unsigned chunk_mask;
+-	unsigned chunk_shift;
++	unsigned int chunk_size;
++	unsigned int chunk_mask;
++	unsigned int chunk_shift;
+ 
+ 	void *context;
+ 
+@@ -144,7 +144,7 @@ static inline chunk_t dm_chunk_number(chunk_t chunk)
+ 	return chunk & (chunk_t)((1ULL << DM_CHUNK_NUMBER_BITS) - 1ULL);
+ }
+ 
+-static inline unsigned dm_consecutive_chunk_count(struct dm_exception *e)
++static inline unsigned int dm_consecutive_chunk_count(struct dm_exception *e)
+ {
+ 	return e->new_chunk >> DM_CHUNK_NUMBER_BITS;
+ }
+@@ -181,12 +181,12 @@ int dm_exception_store_type_register(struct dm_exception_store_type *type);
+ int dm_exception_store_type_unregister(struct dm_exception_store_type *type);
+ 
+ int dm_exception_store_set_chunk_size(struct dm_exception_store *store,
+-				      unsigned chunk_size,
++				      unsigned int chunk_size,
+ 				      char **error);
+ 
+ int dm_exception_store_create(struct dm_target *ti, int argc, char **argv,
+ 			      struct dm_snapshot *snap,
+-			      unsigned *args_used,
++			      unsigned int *args_used,
+ 			      struct dm_exception_store **store);
+ void dm_exception_store_destroy(struct dm_exception_store *store);
+ 
+diff --git a/drivers/md/dm-flakey.c b/drivers/md/dm-flakey.c
+index 335684a1aeaa5..7efbdb42cf3b4 100644
+--- a/drivers/md/dm-flakey.c
++++ b/drivers/md/dm-flakey.c
+@@ -26,12 +26,12 @@ struct flakey_c {
+ 	struct dm_dev *dev;
+ 	unsigned long start_time;
+ 	sector_t start;
+-	unsigned up_interval;
+-	unsigned down_interval;
++	unsigned int up_interval;
++	unsigned int down_interval;
+ 	unsigned long flags;
+-	unsigned corrupt_bio_byte;
+-	unsigned corrupt_bio_rw;
+-	unsigned corrupt_bio_value;
++	unsigned int corrupt_bio_byte;
++	unsigned int corrupt_bio_rw;
++	unsigned int corrupt_bio_value;
+ 	blk_opf_t corrupt_bio_flags;
+ };
+ 
+@@ -48,7 +48,7 @@ static int parse_features(struct dm_arg_set *as, struct flakey_c *fc,
+ 			  struct dm_target *ti)
+ {
+ 	int r;
+-	unsigned argc;
++	unsigned int argc;
+ 	const char *arg_name;
+ 
+ 	static const struct dm_arg _args[] = {
+@@ -148,7 +148,7 @@ static int parse_features(struct dm_arg_set *as, struct flakey_c *fc,
+ 			BUILD_BUG_ON(sizeof(fc->corrupt_bio_flags) !=
+ 				     sizeof(unsigned int));
+ 			r = dm_read_arg(_args + 3, as,
+-				(__force unsigned *)&fc->corrupt_bio_flags,
++				(__force unsigned int *)&fc->corrupt_bio_flags,
+ 				&ti->error);
+ 			if (r)
+ 				return r;
+@@ -324,7 +324,7 @@ static void corrupt_bio_data(struct bio *bio, struct flakey_c *fc)
+ static int flakey_map(struct dm_target *ti, struct bio *bio)
+ {
+ 	struct flakey_c *fc = ti->private;
+-	unsigned elapsed;
++	unsigned int elapsed;
+ 	struct per_bio_data *pb = dm_per_bio_data(bio, sizeof(struct per_bio_data));
+ 	pb->bio_submitted = false;
+ 
+@@ -417,11 +417,11 @@ static int flakey_end_io(struct dm_target *ti, struct bio *bio,
+ }
+ 
+ static void flakey_status(struct dm_target *ti, status_type_t type,
+-			  unsigned status_flags, char *result, unsigned maxlen)
++			  unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+-	unsigned sz = 0;
++	unsigned int sz = 0;
+ 	struct flakey_c *fc = ti->private;
+-	unsigned drop_writes, error_writes;
++	unsigned int drop_writes, error_writes;
+ 
+ 	switch (type) {
+ 	case STATUSTYPE_INFO:
+diff --git a/drivers/md/dm-integrity.c b/drivers/md/dm-integrity.c
+index 1388ee35571e0..53f9f765df9fd 100644
+--- a/drivers/md/dm-integrity.c
++++ b/drivers/md/dm-integrity.c
+@@ -157,13 +157,13 @@ struct alg_spec {
+ 	char *alg_string;
+ 	char *key_string;
+ 	__u8 *key;
+-	unsigned key_size;
++	unsigned int key_size;
+ };
+ 
+ struct dm_integrity_c {
+ 	struct dm_dev *dev;
+ 	struct dm_dev *meta_dev;
+-	unsigned tag_size;
++	unsigned int tag_size;
+ 	__s8 log2_tag_size;
+ 	sector_t start;
+ 	mempool_t journal_io_mempool;
+@@ -171,8 +171,8 @@ struct dm_integrity_c {
+ 	struct dm_bufio_client *bufio;
+ 	struct workqueue_struct *metadata_wq;
+ 	struct superblock *sb;
+-	unsigned journal_pages;
+-	unsigned n_bitmap_blocks;
++	unsigned int journal_pages;
++	unsigned int n_bitmap_blocks;
+ 
+ 	struct page_list *journal;
+ 	struct page_list *journal_io;
+@@ -180,7 +180,7 @@ struct dm_integrity_c {
+ 	struct page_list *recalc_bitmap;
+ 	struct page_list *may_write_bitmap;
+ 	struct bitmap_block_status *bbs;
+-	unsigned bitmap_flush_interval;
++	unsigned int bitmap_flush_interval;
+ 	int synchronous_mode;
+ 	struct bio_list synchronous_bios;
+ 	struct delayed_work bitmap_flush_work;
+@@ -201,12 +201,12 @@ struct dm_integrity_c {
+ 	unsigned char journal_entries_per_sector;
+ 	unsigned char journal_section_entries;
+ 	unsigned short journal_section_sectors;
+-	unsigned journal_sections;
+-	unsigned journal_entries;
++	unsigned int journal_sections;
++	unsigned int journal_entries;
+ 	sector_t data_device_sectors;
+ 	sector_t meta_device_sectors;
+-	unsigned initial_sectors;
+-	unsigned metadata_run;
++	unsigned int initial_sectors;
++	unsigned int metadata_run;
+ 	__s8 log2_metadata_run;
+ 	__u8 log2_buffer_sectors;
+ 	__u8 sectors_per_block;
+@@ -230,17 +230,17 @@ struct dm_integrity_c {
+ 	unsigned char commit_seq;
+ 	commit_id_t commit_ids[N_COMMIT_IDS];
+ 
+-	unsigned committed_section;
+-	unsigned n_committed_sections;
++	unsigned int committed_section;
++	unsigned int n_committed_sections;
+ 
+-	unsigned uncommitted_section;
+-	unsigned n_uncommitted_sections;
++	unsigned int uncommitted_section;
++	unsigned int n_uncommitted_sections;
+ 
+-	unsigned free_section;
++	unsigned int free_section;
+ 	unsigned char free_section_entry;
+-	unsigned free_sectors;
++	unsigned int free_sectors;
+ 
+-	unsigned free_sectors_threshold;
++	unsigned int free_sectors_threshold;
+ 
+ 	struct workqueue_struct *commit_wq;
+ 	struct work_struct commit_work;
+@@ -257,7 +257,7 @@ struct dm_integrity_c {
+ 
+ 	unsigned long autocommit_jiffies;
+ 	struct timer_list autocommit_timer;
+-	unsigned autocommit_msec;
++	unsigned int autocommit_msec;
+ 
+ 	wait_queue_head_t copy_to_journal_wait;
+ 
+@@ -305,7 +305,7 @@ struct dm_integrity_io {
+ 	struct dm_integrity_range range;
+ 
+ 	sector_t metadata_block;
+-	unsigned metadata_offset;
++	unsigned int metadata_offset;
+ 
+ 	atomic_t in_flight;
+ 	blk_status_t bi_status;
+@@ -329,7 +329,7 @@ struct journal_io {
+ struct bitmap_block_status {
+ 	struct work_struct work;
+ 	struct dm_integrity_c *ic;
+-	unsigned idx;
++	unsigned int idx;
+ 	unsigned long *bitmap;
+ 	struct bio_list bio_queue;
+ 	spinlock_t bio_queue_lock;
+@@ -410,8 +410,8 @@ static bool dm_integrity_disable_recalculate(struct dm_integrity_c *ic)
+ 	return false;
+ }
+ 
+-static commit_id_t dm_integrity_commit_id(struct dm_integrity_c *ic, unsigned i,
+-					  unsigned j, unsigned char seq)
++static commit_id_t dm_integrity_commit_id(struct dm_integrity_c *ic, unsigned int i,
++					  unsigned int j, unsigned char seq)
+ {
+ 	/*
+ 	 * Xor the number with section and sector, so that if a piece of
+@@ -426,7 +426,7 @@ static void get_area_and_offset(struct dm_integrity_c *ic, sector_t data_sector,
+ 	if (!ic->meta_dev) {
+ 		__u8 log2_interleave_sectors = ic->sb->log2_interleave_sectors;
+ 		*area = data_sector >> log2_interleave_sectors;
+-		*offset = (unsigned)data_sector & ((1U << log2_interleave_sectors) - 1);
++		*offset = (unsigned int)data_sector & ((1U << log2_interleave_sectors) - 1);
+ 	} else {
+ 		*area = 0;
+ 		*offset = data_sector;
+@@ -435,15 +435,15 @@ static void get_area_and_offset(struct dm_integrity_c *ic, sector_t data_sector,
+ 
+ #define sector_to_block(ic, n)						\
+ do {									\
+-	BUG_ON((n) & (unsigned)((ic)->sectors_per_block - 1));		\
++	BUG_ON((n) & (unsigned int)((ic)->sectors_per_block - 1));		\
+ 	(n) >>= (ic)->sb->log2_sectors_per_block;			\
+ } while (0)
+ 
+ static __u64 get_metadata_sector_and_offset(struct dm_integrity_c *ic, sector_t area,
+-					    sector_t offset, unsigned *metadata_offset)
++					    sector_t offset, unsigned int *metadata_offset)
+ {
+ 	__u64 ms;
+-	unsigned mo;
++	unsigned int mo;
+ 
+ 	ms = area << ic->sb->log2_interleave_sectors;
+ 	if (likely(ic->log2_metadata_run >= 0))
+@@ -484,7 +484,7 @@ static sector_t get_data_sector(struct dm_integrity_c *ic, sector_t area, sector
+ 	return result;
+ }
+ 
+-static void wraparound_section(struct dm_integrity_c *ic, unsigned *sec_ptr)
++static void wraparound_section(struct dm_integrity_c *ic, unsigned int *sec_ptr)
+ {
+ 	if (unlikely(*sec_ptr >= ic->journal_sections))
+ 		*sec_ptr -= ic->journal_sections;
+@@ -508,7 +508,7 @@ static int sb_mac(struct dm_integrity_c *ic, bool wr)
+ {
+ 	SHASH_DESC_ON_STACK(desc, ic->journal_mac);
+ 	int r;
+-	unsigned size = crypto_shash_digestsize(ic->journal_mac);
++	unsigned int size = crypto_shash_digestsize(ic->journal_mac);
+ 
+ 	if (sizeof(struct superblock) + size > 1 << SECTOR_SHIFT) {
+ 		dm_integrity_io_error(ic, "digest is too long", -EINVAL);
+@@ -704,8 +704,8 @@ repeat:
+ 
+ static void block_bitmap_copy(struct dm_integrity_c *ic, struct page_list *dst, struct page_list *src)
+ {
+-	unsigned n_bitmap_pages = DIV_ROUND_UP(ic->n_bitmap_blocks, PAGE_SIZE / BITMAP_BLOCK_SIZE);
+-	unsigned i;
++	unsigned int n_bitmap_pages = DIV_ROUND_UP(ic->n_bitmap_blocks, PAGE_SIZE / BITMAP_BLOCK_SIZE);
++	unsigned int i;
+ 
+ 	for (i = 0; i < n_bitmap_pages; i++) {
+ 		unsigned long *dst_data = lowmem_page_address(dst[i].page);
+@@ -716,18 +716,18 @@ static void block_bitmap_copy(struct dm_integrity_c *ic, struct page_list *dst,
+ 
+ static struct bitmap_block_status *sector_to_bitmap_block(struct dm_integrity_c *ic, sector_t sector)
+ {
+-	unsigned bit = sector >> (ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit);
+-	unsigned bitmap_block = bit / (BITMAP_BLOCK_SIZE * 8);
++	unsigned int bit = sector >> (ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit);
++	unsigned int bitmap_block = bit / (BITMAP_BLOCK_SIZE * 8);
+ 
+ 	BUG_ON(bitmap_block >= ic->n_bitmap_blocks);
+ 	return &ic->bbs[bitmap_block];
+ }
+ 
+-static void access_journal_check(struct dm_integrity_c *ic, unsigned section, unsigned offset,
++static void access_journal_check(struct dm_integrity_c *ic, unsigned int section, unsigned int offset,
+ 				 bool e, const char *function)
+ {
+ #if defined(CONFIG_DM_DEBUG) || defined(INTERNAL_VERIFY)
+-	unsigned limit = e ? ic->journal_section_entries : ic->journal_section_sectors;
++	unsigned int limit = e ? ic->journal_section_entries : ic->journal_section_sectors;
+ 
+ 	if (unlikely(section >= ic->journal_sections) ||
+ 	    unlikely(offset >= limit)) {
+@@ -738,10 +738,10 @@ static void access_journal_check(struct dm_integrity_c *ic, unsigned section, un
+ #endif
+ }
+ 
+-static void page_list_location(struct dm_integrity_c *ic, unsigned section, unsigned offset,
+-			       unsigned *pl_index, unsigned *pl_offset)
++static void page_list_location(struct dm_integrity_c *ic, unsigned int section, unsigned int offset,
++			       unsigned int *pl_index, unsigned int *pl_offset)
+ {
+-	unsigned sector;
++	unsigned int sector;
+ 
+ 	access_journal_check(ic, section, offset, false, "page_list_location");
+ 
+@@ -752,9 +752,9 @@ static void page_list_location(struct dm_integrity_c *ic, unsigned section, unsi
+ }
+ 
+ static struct journal_sector *access_page_list(struct dm_integrity_c *ic, struct page_list *pl,
+-					       unsigned section, unsigned offset, unsigned *n_sectors)
++					       unsigned int section, unsigned int offset, unsigned int *n_sectors)
+ {
+-	unsigned pl_index, pl_offset;
++	unsigned int pl_index, pl_offset;
+ 	char *va;
+ 
+ 	page_list_location(ic, section, offset, &pl_index, &pl_offset);
+@@ -767,14 +767,14 @@ static struct journal_sector *access_page_list(struct dm_integrity_c *ic, struct
+ 	return (struct journal_sector *)(va + pl_offset);
+ }
+ 
+-static struct journal_sector *access_journal(struct dm_integrity_c *ic, unsigned section, unsigned offset)
++static struct journal_sector *access_journal(struct dm_integrity_c *ic, unsigned int section, unsigned int offset)
+ {
+ 	return access_page_list(ic, ic->journal, section, offset, NULL);
+ }
+ 
+-static struct journal_entry *access_journal_entry(struct dm_integrity_c *ic, unsigned section, unsigned n)
++static struct journal_entry *access_journal_entry(struct dm_integrity_c *ic, unsigned int section, unsigned int n)
+ {
+-	unsigned rel_sector, offset;
++	unsigned int rel_sector, offset;
+ 	struct journal_sector *js;
+ 
+ 	access_journal_check(ic, section, n, true, "access_journal_entry");
+@@ -786,7 +786,7 @@ static struct journal_entry *access_journal_entry(struct dm_integrity_c *ic, uns
+ 	return (struct journal_entry *)((char *)js + offset * ic->journal_entry_size);
+ }
+ 
+-static struct journal_sector *access_journal_data(struct dm_integrity_c *ic, unsigned section, unsigned n)
++static struct journal_sector *access_journal_data(struct dm_integrity_c *ic, unsigned int section, unsigned int n)
+ {
+ 	n <<= ic->sb->log2_sectors_per_block;
+ 
+@@ -797,11 +797,11 @@ static struct journal_sector *access_journal_data(struct dm_integrity_c *ic, uns
+ 	return access_journal(ic, section, n);
+ }
+ 
+-static void section_mac(struct dm_integrity_c *ic, unsigned section, __u8 result[JOURNAL_MAC_SIZE])
++static void section_mac(struct dm_integrity_c *ic, unsigned int section, __u8 result[JOURNAL_MAC_SIZE])
+ {
+ 	SHASH_DESC_ON_STACK(desc, ic->journal_mac);
+ 	int r;
+-	unsigned j, size;
++	unsigned int j, size;
+ 
+ 	desc->tfm = ic->journal_mac;
+ 
+@@ -866,10 +866,10 @@ err:
+ 	memset(result, 0, JOURNAL_MAC_SIZE);
+ }
+ 
+-static void rw_section_mac(struct dm_integrity_c *ic, unsigned section, bool wr)
++static void rw_section_mac(struct dm_integrity_c *ic, unsigned int section, bool wr)
+ {
+ 	__u8 result[JOURNAL_MAC_SIZE];
+-	unsigned j;
++	unsigned int j;
+ 
+ 	if (!ic->journal_mac)
+ 		return;
+@@ -898,12 +898,12 @@ static void complete_journal_op(void *context)
+ 		complete(&comp->comp);
+ }
+ 
+-static void xor_journal(struct dm_integrity_c *ic, bool encrypt, unsigned section,
+-			unsigned n_sections, struct journal_completion *comp)
++static void xor_journal(struct dm_integrity_c *ic, bool encrypt, unsigned int section,
++			unsigned int n_sections, struct journal_completion *comp)
+ {
+ 	struct async_submit_ctl submit;
+ 	size_t n_bytes = (size_t)(n_sections * ic->journal_section_sectors) << SECTOR_SHIFT;
+-	unsigned pl_index, pl_offset, section_index;
++	unsigned int pl_index, pl_offset, section_index;
+ 	struct page_list *source_pl, *target_pl;
+ 
+ 	if (likely(encrypt)) {
+@@ -928,7 +928,7 @@ static void xor_journal(struct dm_integrity_c *ic, bool encrypt, unsigned sectio
+ 		struct page *dst_page;
+ 
+ 		while (unlikely(pl_index == section_index)) {
+-			unsigned dummy;
++			unsigned int dummy;
+ 			if (likely(encrypt))
+ 				rw_section_mac(ic, section, true);
+ 			section++;
+@@ -990,8 +990,8 @@ static bool do_crypt(bool encrypt, struct skcipher_request *req, struct journal_
+ 	return false;
+ }
+ 
+-static void crypt_journal(struct dm_integrity_c *ic, bool encrypt, unsigned section,
+-			  unsigned n_sections, struct journal_completion *comp)
++static void crypt_journal(struct dm_integrity_c *ic, bool encrypt, unsigned int section,
++			  unsigned int n_sections, struct journal_completion *comp)
+ {
+ 	struct scatterlist **source_sg;
+ 	struct scatterlist **target_sg;
+@@ -1008,7 +1008,7 @@ static void crypt_journal(struct dm_integrity_c *ic, bool encrypt, unsigned sect
+ 
+ 	do {
+ 		struct skcipher_request *req;
+-		unsigned ivsize;
++		unsigned int ivsize;
+ 		char *iv;
+ 
+ 		if (likely(encrypt))
+@@ -1034,8 +1034,8 @@ static void crypt_journal(struct dm_integrity_c *ic, bool encrypt, unsigned sect
+ 	complete_journal_op(comp);
+ }
+ 
+-static void encrypt_journal(struct dm_integrity_c *ic, bool encrypt, unsigned section,
+-			    unsigned n_sections, struct journal_completion *comp)
++static void encrypt_journal(struct dm_integrity_c *ic, bool encrypt, unsigned int section,
++			    unsigned int n_sections, struct journal_completion *comp)
+ {
+ 	if (ic->journal_xor)
+ 		return xor_journal(ic, encrypt, section, n_sections, comp);
+@@ -1052,12 +1052,12 @@ static void complete_journal_io(unsigned long error, void *context)
+ }
+ 
+ static void rw_journal_sectors(struct dm_integrity_c *ic, blk_opf_t opf,
+-			       unsigned sector, unsigned n_sectors,
++			       unsigned int sector, unsigned int n_sectors,
+ 			       struct journal_completion *comp)
+ {
+ 	struct dm_io_request io_req;
+ 	struct dm_io_region io_loc;
+-	unsigned pl_index, pl_offset;
++	unsigned int pl_index, pl_offset;
+ 	int r;
+ 
+ 	if (unlikely(dm_integrity_failed(ic))) {
+@@ -1099,10 +1099,10 @@ static void rw_journal_sectors(struct dm_integrity_c *ic, blk_opf_t opf,
+ }
+ 
+ static void rw_journal(struct dm_integrity_c *ic, blk_opf_t opf,
+-		       unsigned section, unsigned n_sections,
++		       unsigned int section, unsigned int n_sections,
+ 		       struct journal_completion *comp)
+ {
+-	unsigned sector, n_sectors;
++	unsigned int sector, n_sectors;
+ 
+ 	sector = section * ic->journal_section_sectors;
+ 	n_sectors = n_sections * ic->journal_section_sectors;
+@@ -1110,12 +1110,12 @@ static void rw_journal(struct dm_integrity_c *ic, blk_opf_t opf,
+ 	rw_journal_sectors(ic, opf, sector, n_sectors, comp);
+ }
+ 
+-static void write_journal(struct dm_integrity_c *ic, unsigned commit_start, unsigned commit_sections)
++static void write_journal(struct dm_integrity_c *ic, unsigned int commit_start, unsigned int commit_sections)
+ {
+ 	struct journal_completion io_comp;
+ 	struct journal_completion crypt_comp_1;
+ 	struct journal_completion crypt_comp_2;
+-	unsigned i;
++	unsigned int i;
+ 
+ 	io_comp.ic = ic;
+ 	init_completion(&io_comp.comp);
+@@ -1135,7 +1135,7 @@ static void write_journal(struct dm_integrity_c *ic, unsigned commit_start, unsi
+ 		rw_journal(ic, REQ_OP_WRITE | REQ_FUA | REQ_SYNC, commit_start,
+ 			   commit_sections, &io_comp);
+ 	} else {
+-		unsigned to_end;
++		unsigned int to_end;
+ 		io_comp.in_flight = (atomic_t)ATOMIC_INIT(2);
+ 		to_end = ic->journal_sections - commit_start;
+ 		if (ic->journal_io) {
+@@ -1172,15 +1172,15 @@ static void write_journal(struct dm_integrity_c *ic, unsigned commit_start, unsi
+ 	wait_for_completion_io(&io_comp.comp);
+ }
+ 
+-static void copy_from_journal(struct dm_integrity_c *ic, unsigned section, unsigned offset,
+-			      unsigned n_sectors, sector_t target, io_notify_fn fn, void *data)
++static void copy_from_journal(struct dm_integrity_c *ic, unsigned int section, unsigned int offset,
++			      unsigned int n_sectors, sector_t target, io_notify_fn fn, void *data)
+ {
+ 	struct dm_io_request io_req;
+ 	struct dm_io_region io_loc;
+ 	int r;
+-	unsigned sector, pl_index, pl_offset;
++	unsigned int sector, pl_index, pl_offset;
+ 
+-	BUG_ON((target | n_sectors | offset) & (unsigned)(ic->sectors_per_block - 1));
++	BUG_ON((target | n_sectors | offset) & (unsigned int)(ic->sectors_per_block - 1));
+ 
+ 	if (unlikely(dm_integrity_failed(ic))) {
+ 		fn(-1UL, data);
+@@ -1221,7 +1221,7 @@ static bool add_new_range(struct dm_integrity_c *ic, struct dm_integrity_range *
+ 	struct rb_node **n = &ic->in_progress.rb_node;
+ 	struct rb_node *parent;
+ 
+-	BUG_ON((new_range->logical_sector | new_range->n_sectors) & (unsigned)(ic->sectors_per_block - 1));
++	BUG_ON((new_range->logical_sector | new_range->n_sectors) & (unsigned int)(ic->sectors_per_block - 1));
+ 
+ 	if (likely(check_waiting)) {
+ 		struct dm_integrity_range *range;
+@@ -1339,10 +1339,10 @@ static void remove_journal_node(struct dm_integrity_c *ic, struct journal_node *
+ 
+ #define NOT_FOUND	(-1U)
+ 
+-static unsigned find_journal_node(struct dm_integrity_c *ic, sector_t sector, sector_t *next_sector)
++static unsigned int find_journal_node(struct dm_integrity_c *ic, sector_t sector, sector_t *next_sector)
+ {
+ 	struct rb_node *n = ic->journal_tree_root.rb_node;
+-	unsigned found = NOT_FOUND;
++	unsigned int found = NOT_FOUND;
+ 	*next_sector = (sector_t)-1;
+ 	while (n) {
+ 		struct journal_node *j = container_of(n, struct journal_node, node);
+@@ -1360,7 +1360,7 @@ static unsigned find_journal_node(struct dm_integrity_c *ic, sector_t sector, se
+ 	return found;
+ }
+ 
+-static bool test_journal_node(struct dm_integrity_c *ic, unsigned pos, sector_t sector)
++static bool test_journal_node(struct dm_integrity_c *ic, unsigned int pos, sector_t sector)
+ {
+ 	struct journal_node *node, *next_node;
+ 	struct rb_node *next;
+@@ -1385,7 +1385,7 @@ static bool find_newer_committed_node(struct dm_integrity_c *ic, struct journal_
+ {
+ 	struct rb_node *next;
+ 	struct journal_node *next_node;
+-	unsigned next_section;
++	unsigned int next_section;
+ 
+ 	BUG_ON(RB_EMPTY_NODE(&node->node));
+ 
+@@ -1398,7 +1398,7 @@ static bool find_newer_committed_node(struct dm_integrity_c *ic, struct journal_
+ 	if (next_node->sector != node->sector)
+ 		return false;
+ 
+-	next_section = (unsigned)(next_node - ic->journal_tree) / ic->journal_section_entries;
++	next_section = (unsigned int)(next_node - ic->journal_tree) / ic->journal_section_entries;
+ 	if (next_section >= ic->committed_section &&
+ 	    next_section < ic->committed_section + ic->n_committed_sections)
+ 		return true;
+@@ -1413,17 +1413,17 @@ static bool find_newer_committed_node(struct dm_integrity_c *ic, struct journal_
+ #define TAG_CMP		2
+ 
+ static int dm_integrity_rw_tag(struct dm_integrity_c *ic, unsigned char *tag, sector_t *metadata_block,
+-			       unsigned *metadata_offset, unsigned total_size, int op)
++			       unsigned int *metadata_offset, unsigned int total_size, int op)
+ {
+ #define MAY_BE_FILLER		1
+ #define MAY_BE_HASH		2
+-	unsigned hash_offset = 0;
+-	unsigned may_be = MAY_BE_HASH | (ic->discard ? MAY_BE_FILLER : 0);
++	unsigned int hash_offset = 0;
++	unsigned int may_be = MAY_BE_HASH | (ic->discard ? MAY_BE_FILLER : 0);
+ 
+ 	do {
+ 		unsigned char *data, *dp;
+ 		struct dm_buffer *b;
+-		unsigned to_copy;
++		unsigned int to_copy;
+ 		int r;
+ 
+ 		r = dm_integrity_failed(ic);
+@@ -1453,7 +1453,7 @@ static int dm_integrity_rw_tag(struct dm_integrity_c *ic, unsigned char *tag, se
+ 						goto thorough_test;
+ 				}
+ 			} else {
+-				unsigned i, ts;
++				unsigned int i, ts;
+ thorough_test:
+ 				ts = total_size;
+ 
+@@ -1652,7 +1652,7 @@ static void integrity_sector_checksum(struct dm_integrity_c *ic, sector_t sector
+ 	__le64 sector_le = cpu_to_le64(sector);
+ 	SHASH_DESC_ON_STACK(req, ic->internal_hash);
+ 	int r;
+-	unsigned digest_size;
++	unsigned int digest_size;
+ 
+ 	req->tfm = ic->internal_hash;
+ 
+@@ -1709,13 +1709,13 @@ static void integrity_metadata(struct work_struct *w)
+ 	if (ic->internal_hash) {
+ 		struct bvec_iter iter;
+ 		struct bio_vec bv;
+-		unsigned digest_size = crypto_shash_digestsize(ic->internal_hash);
++		unsigned int digest_size = crypto_shash_digestsize(ic->internal_hash);
+ 		struct bio *bio = dm_bio_from_per_bio_data(dio, sizeof(struct dm_integrity_io));
+ 		char *checksums;
+-		unsigned extra_space = unlikely(digest_size > ic->tag_size) ? digest_size - ic->tag_size : 0;
++		unsigned int extra_space = unlikely(digest_size > ic->tag_size) ? digest_size - ic->tag_size : 0;
+ 		char checksums_onstack[max((size_t)HASH_MAX_DIGESTSIZE, MAX_TAG_SIZE)];
+ 		sector_t sector;
+-		unsigned sectors_to_process;
++		unsigned int sectors_to_process;
+ 
+ 		if (unlikely(ic->mode == 'R'))
+ 			goto skip_io;
+@@ -1735,14 +1735,13 @@ static void integrity_metadata(struct work_struct *w)
+ 		}
+ 
+ 		if (unlikely(dio->op == REQ_OP_DISCARD)) {
+-			sector_t bi_sector = dio->bio_details.bi_iter.bi_sector;
+-			unsigned bi_size = dio->bio_details.bi_iter.bi_size;
+-			unsigned max_size = likely(checksums != checksums_onstack) ? PAGE_SIZE : HASH_MAX_DIGESTSIZE;
+-			unsigned max_blocks = max_size / ic->tag_size;
++			unsigned int bi_size = dio->bio_details.bi_iter.bi_size;
++			unsigned int max_size = likely(checksums != checksums_onstack) ? PAGE_SIZE : HASH_MAX_DIGESTSIZE;
++			unsigned int max_blocks = max_size / ic->tag_size;
+ 			memset(checksums, DISCARD_FILLER, max_size);
+ 
+ 			while (bi_size) {
+-				unsigned this_step_blocks = bi_size >> (SECTOR_SHIFT + ic->sb->log2_sectors_per_block);
++				unsigned int this_step_blocks = bi_size >> (SECTOR_SHIFT + ic->sb->log2_sectors_per_block);
+ 				this_step_blocks = min(this_step_blocks, max_blocks);
+ 				r = dm_integrity_rw_tag(ic, checksums, &dio->metadata_block, &dio->metadata_offset,
+ 							this_step_blocks * ic->tag_size, TAG_WRITE);
+@@ -1752,13 +1751,7 @@ static void integrity_metadata(struct work_struct *w)
+ 					goto error;
+ 				}
+ 
+-				/*if (bi_size < this_step_blocks << (SECTOR_SHIFT + ic->sb->log2_sectors_per_block)) {
+-					printk("BUGG: bi_sector: %llx, bi_size: %u\n", bi_sector, bi_size);
+-					printk("BUGG: this_step_blocks: %u\n", this_step_blocks);
+-					BUG();
+-				}*/
+ 				bi_size -= this_step_blocks << (SECTOR_SHIFT + ic->sb->log2_sectors_per_block);
+-				bi_sector += this_step_blocks << ic->sb->log2_sectors_per_block;
+ 			}
+ 
+ 			if (likely(checksums != checksums_onstack))
+@@ -1770,7 +1763,7 @@ static void integrity_metadata(struct work_struct *w)
+ 		sectors_to_process = dio->range.n_sectors;
+ 
+ 		__bio_for_each_segment(bv, bio, iter, dio->bio_details.bi_iter) {
+-			unsigned pos;
++			unsigned int pos;
+ 			char *mem, *checksums_ptr;
+ 
+ again:
+@@ -1823,13 +1816,13 @@ again:
+ 		if (bip) {
+ 			struct bio_vec biv;
+ 			struct bvec_iter iter;
+-			unsigned data_to_process = dio->range.n_sectors;
++			unsigned int data_to_process = dio->range.n_sectors;
+ 			sector_to_block(ic, data_to_process);
+ 			data_to_process *= ic->tag_size;
+ 
+ 			bip_for_each_vec(biv, bip, iter) {
+ 				unsigned char *tag;
+-				unsigned this_len;
++				unsigned int this_len;
+ 
+ 				BUG_ON(PageHighMem(biv.bv_page));
+ 				tag = bvec_virt(&biv);
+@@ -1867,7 +1860,7 @@ static int dm_integrity_map(struct dm_target *ti, struct bio *bio)
+ 	if (unlikely(dio->op == REQ_OP_DISCARD)) {
+ 		if (ti->max_io_len) {
+ 			sector_t sec = dm_target_offset(ti, bio->bi_iter.bi_sector);
+-			unsigned log2_max_io_len = __fls(ti->max_io_len);
++			unsigned int log2_max_io_len = __fls(ti->max_io_len);
+ 			sector_t start_boundary = sec >> log2_max_io_len;
+ 			sector_t end_boundary = (sec + bio_sectors(bio) - 1) >> log2_max_io_len;
+ 			if (start_boundary < end_boundary) {
+@@ -1897,7 +1890,7 @@ static int dm_integrity_map(struct dm_target *ti, struct bio *bio)
+ 		      ic->provided_data_sectors);
+ 		return DM_MAPIO_KILL;
+ 	}
+-	if (unlikely((dio->range.logical_sector | bio_sectors(bio)) & (unsigned)(ic->sectors_per_block - 1))) {
++	if (unlikely((dio->range.logical_sector | bio_sectors(bio)) & (unsigned int)(ic->sectors_per_block - 1))) {
+ 		DMERR("Bio not aligned on %u sectors: 0x%llx, 0x%x",
+ 		      ic->sectors_per_block,
+ 		      dio->range.logical_sector, bio_sectors(bio));
+@@ -1919,7 +1912,7 @@ static int dm_integrity_map(struct dm_target *ti, struct bio *bio)
+ 	bip = bio_integrity(bio);
+ 	if (!ic->internal_hash) {
+ 		if (bip) {
+-			unsigned wanted_tag_size = bio_sectors(bio) >> ic->sb->log2_sectors_per_block;
++			unsigned int wanted_tag_size = bio_sectors(bio) >> ic->sb->log2_sectors_per_block;
+ 			if (ic->log2_tag_size >= 0)
+ 				wanted_tag_size <<= ic->log2_tag_size;
+ 			else
+@@ -1949,11 +1942,11 @@ static int dm_integrity_map(struct dm_target *ti, struct bio *bio)
+ }
+ 
+ static bool __journal_read_write(struct dm_integrity_io *dio, struct bio *bio,
+-				 unsigned journal_section, unsigned journal_entry)
++				 unsigned int journal_section, unsigned int journal_entry)
+ {
+ 	struct dm_integrity_c *ic = dio->ic;
+ 	sector_t logical_sector;
+-	unsigned n_sectors;
++	unsigned int n_sectors;
+ 
+ 	logical_sector = dio->range.logical_sector;
+ 	n_sectors = dio->range.n_sectors;
+@@ -1976,7 +1969,7 @@ retry_kmap:
+ 			if (unlikely(dio->op == REQ_OP_READ)) {
+ 				struct journal_sector *js;
+ 				char *mem_ptr;
+-				unsigned s;
++				unsigned int s;
+ 
+ 				if (unlikely(journal_entry_is_inprogress(je))) {
+ 					flush_dcache_page(bv.bv_page);
+@@ -2013,12 +2006,12 @@ retry_kmap:
+ 
+ 			if (!ic->internal_hash) {
+ 				struct bio_integrity_payload *bip = bio_integrity(bio);
+-				unsigned tag_todo = ic->tag_size;
++				unsigned int tag_todo = ic->tag_size;
+ 				char *tag_ptr = journal_entry_tag(ic, je);
+ 
+ 				if (bip) do {
+ 					struct bio_vec biv = bvec_iter_bvec(bip->bip_vec, bip->bip_iter);
+-					unsigned tag_now = min(biv.bv_len, tag_todo);
++					unsigned int tag_now = min(biv.bv_len, tag_todo);
+ 					char *tag_addr;
+ 					BUG_ON(PageHighMem(biv.bv_page));
+ 					tag_addr = bvec_virt(&biv);
+@@ -2037,7 +2030,7 @@ retry_kmap:
+ 
+ 			if (likely(dio->op == REQ_OP_WRITE)) {
+ 				struct journal_sector *js;
+-				unsigned s;
++				unsigned int s;
+ 
+ 				js = access_journal_data(ic, journal_section, journal_entry);
+ 				memcpy(js, mem + bv.bv_offset, ic->sectors_per_block << SECTOR_SHIFT);
+@@ -2048,7 +2041,7 @@ retry_kmap:
+ 				} while (++s < ic->sectors_per_block);
+ 
+ 				if (ic->internal_hash) {
+-					unsigned digest_size = crypto_shash_digestsize(ic->internal_hash);
++					unsigned int digest_size = crypto_shash_digestsize(ic->internal_hash);
+ 					if (unlikely(digest_size > ic->tag_size)) {
+ 						char checksums_onstack[HASH_MAX_DIGESTSIZE];
+ 						integrity_sector_checksum(ic, logical_sector, (char *)js, checksums_onstack);
+@@ -2105,8 +2098,8 @@ static void dm_integrity_map_continue(struct dm_integrity_io *dio, bool from_map
+ {
+ 	struct dm_integrity_c *ic = dio->ic;
+ 	struct bio *bio = dm_bio_from_per_bio_data(dio, sizeof(struct dm_integrity_io));
+-	unsigned journal_section, journal_entry;
+-	unsigned journal_read_pos;
++	unsigned int journal_section, journal_entry;
++	unsigned int journal_read_pos;
+ 	struct completion read_comp;
+ 	bool discard_retried = false;
+ 	bool need_sync_io = ic->internal_hash && dio->op == REQ_OP_READ;
+@@ -2131,8 +2124,8 @@ retry:
+ 	journal_read_pos = NOT_FOUND;
+ 	if (ic->mode == 'J' && likely(dio->op != REQ_OP_DISCARD)) {
+ 		if (dio->op == REQ_OP_WRITE) {
+-			unsigned next_entry, i, pos;
+-			unsigned ws, we, range_sectors;
++			unsigned int next_entry, i, pos;
++			unsigned int ws, we, range_sectors;
+ 
+ 			dio->range.n_sectors = min(dio->range.n_sectors,
+ 						   (sector_t)ic->free_sectors << ic->sb->log2_sectors_per_block);
+@@ -2185,8 +2178,8 @@ retry:
+ 				if (unlikely(dio->range.n_sectors > next_sector - dio->range.logical_sector))
+ 					dio->range.n_sectors = next_sector - dio->range.logical_sector;
+ 			} else {
+-				unsigned i;
+-				unsigned jp = journal_read_pos + 1;
++				unsigned int i;
++				unsigned int jp = journal_read_pos + 1;
+ 				for (i = ic->sectors_per_block; i < dio->range.n_sectors; i += ic->sectors_per_block, jp++) {
+ 					if (!test_journal_node(ic, jp, dio->range.logical_sector + i))
+ 						break;
+@@ -2218,7 +2211,7 @@ offload_to_thread:
+ 		 */
+ 		if (journal_read_pos != NOT_FOUND) {
+ 			sector_t next_sector;
+-			unsigned new_pos = find_journal_node(ic, dio->range.logical_sector, &next_sector);
++			unsigned int new_pos = find_journal_node(ic, dio->range.logical_sector, &next_sector);
+ 			if (unlikely(new_pos != journal_read_pos)) {
+ 				remove_range_unlocked(ic, &dio->range);
+ 				goto retry;
+@@ -2227,7 +2220,7 @@ offload_to_thread:
+ 	}
+ 	if (ic->mode == 'J' && likely(dio->op == REQ_OP_DISCARD) && !discard_retried) {
+ 		sector_t next_sector;
+-		unsigned new_pos = find_journal_node(ic, dio->range.logical_sector, &next_sector);
++		unsigned int new_pos = find_journal_node(ic, dio->range.logical_sector, &next_sector);
+ 		if (unlikely(new_pos != NOT_FOUND) ||
+ 		    unlikely(next_sector < dio->range.logical_sector - dio->range.n_sectors)) {
+ 			remove_range_unlocked(ic, &dio->range);
+@@ -2354,8 +2347,8 @@ static void pad_uncommitted(struct dm_integrity_c *ic)
+ static void integrity_commit(struct work_struct *w)
+ {
+ 	struct dm_integrity_c *ic = container_of(w, struct dm_integrity_c, commit_work);
+-	unsigned commit_start, commit_sections;
+-	unsigned i, j, n;
++	unsigned int commit_start, commit_sections;
++	unsigned int i, j, n;
+ 	struct bio *flushes;
+ 
+ 	del_timer(&ic->autocommit_timer);
+@@ -2433,17 +2426,17 @@ static void complete_copy_from_journal(unsigned long error, void *context)
+ static void restore_last_bytes(struct dm_integrity_c *ic, struct journal_sector *js,
+ 			       struct journal_entry *je)
+ {
+-	unsigned s = 0;
++	unsigned int s = 0;
+ 	do {
+ 		js->commit_id = je->last_bytes[s];
+ 		js++;
+ 	} while (++s < ic->sectors_per_block);
+ }
+ 
+-static void do_journal_write(struct dm_integrity_c *ic, unsigned write_start,
+-			     unsigned write_sections, bool from_replay)
++static void do_journal_write(struct dm_integrity_c *ic, unsigned int write_start,
++			     unsigned int write_sections, bool from_replay)
+ {
+-	unsigned i, j, n;
++	unsigned int i, j, n;
+ 	struct journal_completion comp;
+ 	struct blk_plug plug;
+ 
+@@ -2462,9 +2455,9 @@ static void do_journal_write(struct dm_integrity_c *ic, unsigned write_start,
+ 		for (j = 0; j < ic->journal_section_entries; j++) {
+ 			struct journal_entry *je = access_journal_entry(ic, i, j);
+ 			sector_t sec, area, offset;
+-			unsigned k, l, next_loop;
++			unsigned int k, l, next_loop;
+ 			sector_t metadata_block;
+-			unsigned metadata_offset;
++			unsigned int metadata_offset;
+ 			struct journal_io *io;
+ 
+ 			if (journal_entry_is_unused(je))
+@@ -2472,7 +2465,7 @@ static void do_journal_write(struct dm_integrity_c *ic, unsigned write_start,
+ 			BUG_ON(unlikely(journal_entry_is_inprogress(je)) && !from_replay);
+ 			sec = journal_entry_get_sector(je);
+ 			if (unlikely(from_replay)) {
+-				if (unlikely(sec & (unsigned)(ic->sectors_per_block - 1))) {
++				if (unlikely(sec & (unsigned int)(ic->sectors_per_block - 1))) {
+ 					dm_integrity_io_error(ic, "invalid sector in journal", -EIO);
+ 					sec &= ~(sector_t)(ic->sectors_per_block - 1);
+ 				}
+@@ -2590,9 +2583,9 @@ skip_io:
+ static void integrity_writer(struct work_struct *w)
+ {
+ 	struct dm_integrity_c *ic = container_of(w, struct dm_integrity_c, writer_work);
+-	unsigned write_start, write_sections;
++	unsigned int write_start, write_sections;
+ 
+-	unsigned prev_free_sectors;
++	unsigned int prev_free_sectors;
+ 
+ 	spin_lock_irq(&ic->endio_wait.lock);
+ 	write_start = ic->committed_section;
+@@ -2639,12 +2632,12 @@ static void integrity_recalc(struct work_struct *w)
+ 	struct dm_io_region io_loc;
+ 	sector_t area, offset;
+ 	sector_t metadata_block;
+-	unsigned metadata_offset;
++	unsigned int metadata_offset;
+ 	sector_t logical_sector, n_sectors;
+ 	__u8 *t;
+-	unsigned i;
++	unsigned int i;
+ 	int r;
+-	unsigned super_counter = 0;
++	unsigned int super_counter = 0;
+ 
+ 	DEBUG_print("start recalculation... (position %llx)\n", le64_to_cpu(ic->sb->recalc_sector));
+ 
+@@ -2668,7 +2661,7 @@ next_chunk:
+ 	get_area_and_offset(ic, range.logical_sector, &area, &offset);
+ 	range.n_sectors = min((sector_t)RECALC_SECTORS, ic->provided_data_sectors - range.logical_sector);
+ 	if (!ic->meta_dev)
+-		range.n_sectors = min(range.n_sectors, ((sector_t)1U << ic->sb->log2_interleave_sectors) - (unsigned)offset);
++		range.n_sectors = min(range.n_sectors, ((sector_t)1U << ic->sb->log2_interleave_sectors) - (unsigned int)offset);
+ 
+ 	add_new_range_and_wait(ic, &range);
+ 	spin_unlock_irq(&ic->endio_wait.lock);
+@@ -2859,10 +2852,10 @@ static void bitmap_flush_work(struct work_struct *work)
+ }
+ 
+ 
+-static void init_journal(struct dm_integrity_c *ic, unsigned start_section,
+-			 unsigned n_sections, unsigned char commit_seq)
++static void init_journal(struct dm_integrity_c *ic, unsigned int start_section,
++			 unsigned int n_sections, unsigned char commit_seq)
+ {
+-	unsigned i, j, n;
++	unsigned int i, j, n;
+ 
+ 	if (!n_sections)
+ 		return;
+@@ -2885,7 +2878,7 @@ static void init_journal(struct dm_integrity_c *ic, unsigned start_section,
+ 	write_journal(ic, start_section, n_sections);
+ }
+ 
+-static int find_commit_seq(struct dm_integrity_c *ic, unsigned i, unsigned j, commit_id_t id)
++static int find_commit_seq(struct dm_integrity_c *ic, unsigned int i, unsigned int j, commit_id_t id)
+ {
+ 	unsigned char k;
+ 	for (k = 0; k < N_COMMIT_IDS; k++) {
+@@ -2898,11 +2891,11 @@ static int find_commit_seq(struct dm_integrity_c *ic, unsigned i, unsigned j, co
+ 
+ static void replay_journal(struct dm_integrity_c *ic)
+ {
+-	unsigned i, j;
++	unsigned int i, j;
+ 	bool used_commit_ids[N_COMMIT_IDS];
+-	unsigned max_commit_id_sections[N_COMMIT_IDS];
+-	unsigned write_start, write_sections;
+-	unsigned continue_section;
++	unsigned int max_commit_id_sections[N_COMMIT_IDS];
++	unsigned int write_start, write_sections;
++	unsigned int continue_section;
+ 	bool journal_empty;
+ 	unsigned char unused, last_used, want_commit_seq;
+ 
+@@ -3020,7 +3013,7 @@ brk:
+ 		ic->commit_seq = want_commit_seq;
+ 		DEBUG_print("continuing from section %u, commit seq %d\n", write_start, ic->commit_seq);
+ 	} else {
+-		unsigned s;
++		unsigned int s;
+ 		unsigned char erase_seq;
+ clear_journal:
+ 		DEBUG_print("clearing journal\n");
+@@ -3252,10 +3245,10 @@ static void dm_integrity_resume(struct dm_target *ti)
+ }
+ 
+ static void dm_integrity_status(struct dm_target *ti, status_type_t type,
+-				unsigned status_flags, char *result, unsigned maxlen)
++				unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	struct dm_integrity_c *ic = (struct dm_integrity_c *)ti->private;
+-	unsigned arg_count;
++	unsigned int arg_count;
+ 	size_t sz = 0;
+ 
+ 	switch (type) {
+@@ -3305,7 +3298,7 @@ static void dm_integrity_status(struct dm_target *ti, status_type_t type,
+ 		DMEMIT(" interleave_sectors:%u", 1U << ic->sb->log2_interleave_sectors);
+ 		DMEMIT(" buffer_sectors:%u", 1U << ic->log2_buffer_sectors);
+ 		if (ic->mode == 'J') {
+-			DMEMIT(" journal_watermark:%u", (unsigned)watermark_percentage);
++			DMEMIT(" journal_watermark:%u", (unsigned int)watermark_percentage);
+ 			DMEMIT(" commit_time:%u", ic->autocommit_msec);
+ 		}
+ 		if (ic->mode == 'B') {
+@@ -3384,7 +3377,7 @@ static void dm_integrity_io_hints(struct dm_target *ti, struct queue_limits *lim
+ 
+ static void calculate_journal_section_size(struct dm_integrity_c *ic)
+ {
+-	unsigned sector_space = JOURNAL_SECTOR_DATA;
++	unsigned int sector_space = JOURNAL_SECTOR_DATA;
+ 
+ 	ic->journal_sections = le32_to_cpu(ic->sb->journal_sections);
+ 	ic->journal_entry_size = roundup(offsetof(struct journal_entry, last_bytes[ic->sectors_per_block]) + ic->tag_size,
+@@ -3461,9 +3454,10 @@ static void get_provided_data_sectors(struct dm_integrity_c *ic)
+ 	}
+ }
+ 
+-static int initialize_superblock(struct dm_integrity_c *ic, unsigned journal_sectors, unsigned interleave_sectors)
++static int initialize_superblock(struct dm_integrity_c *ic,
++				 unsigned int journal_sectors, unsigned int interleave_sectors)
+ {
+-	unsigned journal_sections;
++	unsigned int journal_sections;
+ 	int test_bit;
+ 
+ 	memset(ic->sb, 0, SB_SECTORS << SECTOR_SHIFT);
+@@ -3548,7 +3542,7 @@ static void dm_integrity_set(struct dm_target *ti, struct dm_integrity_c *ic)
+ 
+ static void dm_integrity_free_page_list(struct page_list *pl)
+ {
+-	unsigned i;
++	unsigned int i;
+ 
+ 	if (!pl)
+ 		return;
+@@ -3557,10 +3551,10 @@ static void dm_integrity_free_page_list(struct page_list *pl)
+ 	kvfree(pl);
+ }
+ 
+-static struct page_list *dm_integrity_alloc_page_list(unsigned n_pages)
++static struct page_list *dm_integrity_alloc_page_list(unsigned int n_pages)
+ {
+ 	struct page_list *pl;
+-	unsigned i;
++	unsigned int i;
+ 
+ 	pl = kvmalloc_array(n_pages + 1, sizeof(struct page_list), GFP_KERNEL | __GFP_ZERO);
+ 	if (!pl)
+@@ -3583,7 +3577,7 @@ static struct page_list *dm_integrity_alloc_page_list(unsigned n_pages)
+ 
+ static void dm_integrity_free_journal_scatterlist(struct dm_integrity_c *ic, struct scatterlist **sl)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	for (i = 0; i < ic->journal_sections; i++)
+ 		kvfree(sl[i]);
+ 	kvfree(sl);
+@@ -3593,7 +3587,7 @@ static struct scatterlist **dm_integrity_alloc_journal_scatterlist(struct dm_int
+ 								   struct page_list *pl)
+ {
+ 	struct scatterlist **sl;
+-	unsigned i;
++	unsigned int i;
+ 
+ 	sl = kvmalloc_array(ic->journal_sections,
+ 			    sizeof(struct scatterlist *),
+@@ -3603,10 +3597,10 @@ static struct scatterlist **dm_integrity_alloc_journal_scatterlist(struct dm_int
+ 
+ 	for (i = 0; i < ic->journal_sections; i++) {
+ 		struct scatterlist *s;
+-		unsigned start_index, start_offset;
+-		unsigned end_index, end_offset;
+-		unsigned n_pages;
+-		unsigned idx;
++		unsigned int start_index, start_offset;
++		unsigned int end_index, end_offset;
++		unsigned int n_pages;
++		unsigned int idx;
+ 
+ 		page_list_location(ic, i, 0, &start_index, &start_offset);
+ 		page_list_location(ic, i, ic->journal_section_sectors - 1,
+@@ -3624,7 +3618,7 @@ static struct scatterlist **dm_integrity_alloc_journal_scatterlist(struct dm_int
+ 		sg_init_table(s, n_pages);
+ 		for (idx = start_index; idx <= end_index; idx++) {
+ 			char *va = lowmem_page_address(pl[idx].page);
+-			unsigned start = 0, end = PAGE_SIZE;
++			unsigned int start = 0, end = PAGE_SIZE;
+ 			if (idx == start_index)
+ 				start = start_offset;
+ 			if (idx == end_index)
+@@ -3711,7 +3705,7 @@ static int get_mac(struct crypto_shash **hash, struct alg_spec *a, char **error,
+ static int create_journal(struct dm_integrity_c *ic, char **error)
+ {
+ 	int r = 0;
+-	unsigned i;
++	unsigned int i;
+ 	__u64 journal_pages, journal_desc_size, journal_tree_size;
+ 	unsigned char *crypt_data = NULL, *crypt_iv = NULL;
+ 	struct skcipher_request *req = NULL;
+@@ -3738,7 +3732,7 @@ static int create_journal(struct dm_integrity_c *ic, char **error)
+ 		goto bad;
+ 	}
+ 	if (ic->journal_crypt_alg.alg_string) {
+-		unsigned ivsize, blocksize;
++		unsigned int ivsize, blocksize;
+ 		struct journal_completion comp;
+ 
+ 		comp.ic = ic;
+@@ -3827,7 +3821,7 @@ static int create_journal(struct dm_integrity_c *ic, char **error)
+ 			crypto_free_skcipher(ic->journal_crypt);
+ 			ic->journal_crypt = NULL;
+ 		} else {
+-			unsigned crypt_len = roundup(ivsize, blocksize);
++			unsigned int crypt_len = roundup(ivsize, blocksize);
+ 
+ 			req = skcipher_request_alloc(ic->journal_crypt, GFP_KERNEL);
+ 			if (!req) {
+@@ -3915,7 +3909,7 @@ static int create_journal(struct dm_integrity_c *ic, char **error)
+ 	}
+ 
+ 	for (i = 0; i < N_COMMIT_IDS; i++) {
+-		unsigned j;
++		unsigned int j;
+ retest_commit_id:
+ 		for (j = 0; j < i; j++) {
+ 			if (ic->commit_ids[j] == ic->commit_ids[i]) {
+@@ -3969,17 +3963,17 @@ bad:
+  *		journal_mac
+  *		recalculate
+  */
+-static int dm_integrity_ctr(struct dm_target *ti, unsigned argc, char **argv)
++static int dm_integrity_ctr(struct dm_target *ti, unsigned int argc, char **argv)
+ {
+ 	struct dm_integrity_c *ic;
+ 	char dummy;
+ 	int r;
+-	unsigned extra_args;
++	unsigned int extra_args;
+ 	struct dm_arg_set as;
+ 	static const struct dm_arg _args[] = {
+ 		{0, 18, "Invalid number of feature args"},
+ 	};
+-	unsigned journal_sectors, interleave_sectors, buffer_sectors, journal_watermark, sync_msec;
++	unsigned int journal_sectors, interleave_sectors, buffer_sectors, journal_watermark, sync_msec;
+ 	bool should_write_sb;
+ 	__u64 threshold;
+ 	unsigned long long start;
+@@ -4058,7 +4052,7 @@ static int dm_integrity_ctr(struct dm_target *ti, unsigned argc, char **argv)
+ 
+ 	while (extra_args--) {
+ 		const char *opt_string;
+-		unsigned val;
++		unsigned int val;
+ 		unsigned long long llval;
+ 		opt_string = dm_shift_arg(&as);
+ 		if (!opt_string) {
+@@ -4391,7 +4385,7 @@ try_smaller_buffer:
+ 	DEBUG_print("	journal_entries_per_sector %u\n", ic->journal_entries_per_sector);
+ 	DEBUG_print("	journal_section_entries %u\n", ic->journal_section_entries);
+ 	DEBUG_print("	journal_section_sectors %u\n", ic->journal_section_sectors);
+-	DEBUG_print("	journal_sections %u\n", (unsigned)le32_to_cpu(ic->sb->journal_sections));
++	DEBUG_print("	journal_sections %u\n", (unsigned int)le32_to_cpu(ic->sb->journal_sections));
+ 	DEBUG_print("	journal_entries %u\n", ic->journal_entries);
+ 	DEBUG_print("	log2_interleave_sectors %d\n", ic->sb->log2_interleave_sectors);
+ 	DEBUG_print("	data_device_sectors 0x%llx\n", bdev_nr_sectors(ic->dev->bdev));
+@@ -4465,8 +4459,8 @@ try_smaller_buffer:
+ 	}
+ 
+ 	if (ic->mode == 'B') {
+-		unsigned i;
+-		unsigned n_bitmap_pages = DIV_ROUND_UP(ic->n_bitmap_blocks, PAGE_SIZE / BITMAP_BLOCK_SIZE);
++		unsigned int i;
++		unsigned int n_bitmap_pages = DIV_ROUND_UP(ic->n_bitmap_blocks, PAGE_SIZE / BITMAP_BLOCK_SIZE);
+ 
+ 		ic->recalc_bitmap = dm_integrity_alloc_page_list(n_bitmap_pages);
+ 		if (!ic->recalc_bitmap) {
+@@ -4486,7 +4480,7 @@ try_smaller_buffer:
+ 		INIT_DELAYED_WORK(&ic->bitmap_flush_work, bitmap_flush_work);
+ 		for (i = 0; i < ic->n_bitmap_blocks; i++) {
+ 			struct bitmap_block_status *bbs = &ic->bbs[i];
+-			unsigned sector, pl_index, pl_offset;
++			unsigned int sector, pl_index, pl_offset;
+ 
+ 			INIT_WORK(&bbs->work, bitmap_block_work);
+ 			bbs->ic = ic;
+@@ -4523,7 +4517,7 @@ try_smaller_buffer:
+ 			goto bad;
+ 	}
+ 	if (ic->mode == 'B') {
+-		unsigned max_io_len = ((sector_t)ic->sectors_per_block << ic->log2_blocks_per_bitmap_bit) * (BITMAP_BLOCK_SIZE * 8);
++		unsigned int max_io_len = ((sector_t)ic->sectors_per_block << ic->log2_blocks_per_bitmap_bit) * (BITMAP_BLOCK_SIZE * 8);
+ 		if (!max_io_len)
+ 			max_io_len = 1U << 31;
+ 		DEBUG_print("max_io_len: old %u, new %u\n", ti->max_io_len, max_io_len);
+@@ -4594,7 +4588,7 @@ static void dm_integrity_dtr(struct dm_target *ti)
+ 	if (ic->journal_io_scatterlist)
+ 		dm_integrity_free_journal_scatterlist(ic, ic->journal_io_scatterlist);
+ 	if (ic->sk_requests) {
+-		unsigned i;
++		unsigned int i;
+ 
+ 		for (i = 0; i < ic->journal_sections; i++) {
+ 			struct skcipher_request *req = ic->sk_requests[i];
+diff --git a/drivers/md/dm-io-rewind.c b/drivers/md/dm-io-rewind.c
+index 0db53ccb94ba7..773c4cff8b89f 100644
+--- a/drivers/md/dm-io-rewind.c
++++ b/drivers/md/dm-io-rewind.c
+@@ -57,7 +57,7 @@ static void dm_bio_integrity_rewind(struct bio *bio, unsigned int bytes_done)
+ {
+ 	struct bio_integrity_payload *bip = bio_integrity(bio);
+ 	struct blk_integrity *bi = blk_get_integrity(bio->bi_bdev->bd_disk);
+-	unsigned bytes = bio_integrity_bytes(bi, bytes_done >> 9);
++	unsigned int bytes = bio_integrity_bytes(bi, bytes_done >> 9);
+ 
+ 	bip->bip_iter.bi_sector -= bio_integrity_intervals(bi, bytes_done >> 9);
+ 	dm_bvec_iter_rewind(bip->bip_vec, &bip->bip_iter, bytes);
+@@ -131,7 +131,7 @@ static inline void dm_bio_rewind_iter(const struct bio *bio,
+  * rewinding from end of bio and restoring its original position.
+  * Caller is also responsibile for restoring bio's size.
+  */
+-static void dm_bio_rewind(struct bio *bio, unsigned bytes)
++static void dm_bio_rewind(struct bio *bio, unsigned int bytes)
+ {
+ 	if (bio_integrity(bio))
+ 		dm_bio_integrity_rewind(bio, bytes);
+diff --git a/drivers/md/dm-io.c b/drivers/md/dm-io.c
+index 7835645334593..e488b05e35fa3 100644
+--- a/drivers/md/dm-io.c
++++ b/drivers/md/dm-io.c
+@@ -48,7 +48,7 @@ static struct kmem_cache *_dm_io_cache;
+ struct dm_io_client *dm_io_client_create(void)
+ {
+ 	struct dm_io_client *client;
+-	unsigned min_ios = dm_get_reserved_bio_based_ios();
++	unsigned int min_ios = dm_get_reserved_bio_based_ios();
+ 	int ret;
+ 
+ 	client = kzalloc(sizeof(*client), GFP_KERNEL);
+@@ -88,7 +88,7 @@ EXPORT_SYMBOL(dm_io_client_destroy);
+  * bi_private.
+  *---------------------------------------------------------------*/
+ static void store_io_and_region_in_bio(struct bio *bio, struct io *io,
+-				       unsigned region)
++				       unsigned int region)
+ {
+ 	if (unlikely(!IS_ALIGNED((unsigned long)io, DM_IO_MAX_REGIONS))) {
+ 		DMCRIT("Unaligned struct io pointer %p", io);
+@@ -99,7 +99,7 @@ static void store_io_and_region_in_bio(struct bio *bio, struct io *io,
+ }
+ 
+ static void retrieve_io_and_region_from_bio(struct bio *bio, struct io **io,
+-				       unsigned *region)
++				       unsigned int *region)
+ {
+ 	unsigned long val = (unsigned long)bio->bi_private;
+ 
+@@ -137,7 +137,7 @@ static void dec_count(struct io *io, unsigned int region, blk_status_t error)
+ static void endio(struct bio *bio)
+ {
+ 	struct io *io;
+-	unsigned region;
++	unsigned int region;
+ 	blk_status_t error;
+ 
+ 	if (bio->bi_status && bio_data_dir(bio) == READ)
+@@ -160,11 +160,11 @@ static void endio(struct bio *bio)
+  *---------------------------------------------------------------*/
+ struct dpages {
+ 	void (*get_page)(struct dpages *dp,
+-			 struct page **p, unsigned long *len, unsigned *offset);
++			 struct page **p, unsigned long *len, unsigned int *offset);
+ 	void (*next_page)(struct dpages *dp);
+ 
+ 	union {
+-		unsigned context_u;
++		unsigned int context_u;
+ 		struct bvec_iter context_bi;
+ 	};
+ 	void *context_ptr;
+@@ -177,9 +177,9 @@ struct dpages {
+  * Functions for getting the pages from a list.
+  */
+ static void list_get_page(struct dpages *dp,
+-		  struct page **p, unsigned long *len, unsigned *offset)
++		  struct page **p, unsigned long *len, unsigned int *offset)
+ {
+-	unsigned o = dp->context_u;
++	unsigned int o = dp->context_u;
+ 	struct page_list *pl = (struct page_list *) dp->context_ptr;
+ 
+ 	*p = pl->page;
+@@ -194,7 +194,7 @@ static void list_next_page(struct dpages *dp)
+ 	dp->context_u = 0;
+ }
+ 
+-static void list_dp_init(struct dpages *dp, struct page_list *pl, unsigned offset)
++static void list_dp_init(struct dpages *dp, struct page_list *pl, unsigned int offset)
+ {
+ 	dp->get_page = list_get_page;
+ 	dp->next_page = list_next_page;
+@@ -206,7 +206,7 @@ static void list_dp_init(struct dpages *dp, struct page_list *pl, unsigned offse
+  * Functions for getting the pages from a bvec.
+  */
+ static void bio_get_page(struct dpages *dp, struct page **p,
+-			 unsigned long *len, unsigned *offset)
++			 unsigned long *len, unsigned int *offset)
+ {
+ 	struct bio_vec bvec = bvec_iter_bvec((struct bio_vec *)dp->context_ptr,
+ 					     dp->context_bi);
+@@ -244,7 +244,7 @@ static void bio_dp_init(struct dpages *dp, struct bio *bio)
+  * Functions for getting the pages from a VMA.
+  */
+ static void vm_get_page(struct dpages *dp,
+-		 struct page **p, unsigned long *len, unsigned *offset)
++		 struct page **p, unsigned long *len, unsigned int *offset)
+ {
+ 	*p = vmalloc_to_page(dp->context_ptr);
+ 	*offset = dp->context_u;
+@@ -269,7 +269,7 @@ static void vm_dp_init(struct dpages *dp, void *data)
+  * Functions for getting the pages from kernel memory.
+  */
+ static void km_get_page(struct dpages *dp, struct page **p, unsigned long *len,
+-			unsigned *offset)
++			unsigned int *offset)
+ {
+ 	*p = virt_to_page(dp->context_ptr);
+ 	*offset = dp->context_u;
+@@ -293,15 +293,15 @@ static void km_dp_init(struct dpages *dp, void *data)
+ /*-----------------------------------------------------------------
+  * IO routines that accept a list of pages.
+  *---------------------------------------------------------------*/
+-static void do_region(const blk_opf_t opf, unsigned region,
++static void do_region(const blk_opf_t opf, unsigned int region,
+ 		      struct dm_io_region *where, struct dpages *dp,
+ 		      struct io *io)
+ {
+ 	struct bio *bio;
+ 	struct page *page;
+ 	unsigned long len;
+-	unsigned offset;
+-	unsigned num_bvecs;
++	unsigned int offset;
++	unsigned int num_bvecs;
+ 	sector_t remaining = where->count;
+ 	struct request_queue *q = bdev_get_queue(where->bdev);
+ 	sector_t num_sectors;
+@@ -508,7 +508,7 @@ static int dp_init(struct dm_io_request *io_req, struct dpages *dp,
+ 	return 0;
+ }
+ 
+-int dm_io(struct dm_io_request *io_req, unsigned num_regions,
++int dm_io(struct dm_io_request *io_req, unsigned int num_regions,
+ 	  struct dm_io_region *where, unsigned long *sync_error_bits)
+ {
+ 	int r;
+diff --git a/drivers/md/dm-ioctl.c b/drivers/md/dm-ioctl.c
+index fdb7846a97a40..41d55218b0764 100644
+--- a/drivers/md/dm-ioctl.c
++++ b/drivers/md/dm-ioctl.c
+@@ -31,7 +31,7 @@ struct dm_file {
+ 	 * poll will wait until the global event number is greater than
+ 	 * this value.
+ 	 */
+-	volatile unsigned global_event_nr;
++	volatile unsigned int global_event_nr;
+ };
+ 
+ /*-----------------------------------------------------------------
+@@ -413,7 +413,7 @@ static struct mapped_device *dm_hash_rename(struct dm_ioctl *param,
+ 	struct hash_cell *hc;
+ 	struct dm_table *table;
+ 	struct mapped_device *md;
+-	unsigned change_uuid = (param->flags & DM_UUID_FLAG) ? 1 : 0;
++	unsigned int change_uuid = (param->flags & DM_UUID_FLAG) ? 1 : 0;
+ 	int srcu_idx;
+ 
+ 	/*
+@@ -1021,7 +1021,7 @@ static int dev_rename(struct file *filp, struct dm_ioctl *param, size_t param_si
+ 	int r;
+ 	char *new_data = (char *) param + param->data_start;
+ 	struct mapped_device *md;
+-	unsigned change_uuid = (param->flags & DM_UUID_FLAG) ? 1 : 0;
++	unsigned int change_uuid = (param->flags & DM_UUID_FLAG) ? 1 : 0;
+ 
+ 	if (new_data < param->data ||
+ 	    invalid_str(new_data, (void *) param + param_size) || !*new_data ||
+@@ -1096,7 +1096,7 @@ out:
+ static int do_suspend(struct dm_ioctl *param)
+ {
+ 	int r = 0;
+-	unsigned suspend_flags = DM_SUSPEND_LOCKFS_FLAG;
++	unsigned int suspend_flags = DM_SUSPEND_LOCKFS_FLAG;
+ 	struct mapped_device *md;
+ 
+ 	md = find_device(param);
+@@ -1125,7 +1125,7 @@ out:
+ static int do_resume(struct dm_ioctl *param)
+ {
+ 	int r = 0;
+-	unsigned suspend_flags = DM_SUSPEND_LOCKFS_FLAG;
++	unsigned int suspend_flags = DM_SUSPEND_LOCKFS_FLAG;
+ 	struct hash_cell *hc;
+ 	struct mapped_device *md;
+ 	struct dm_table *new_map, *old_map = NULL;
+@@ -1243,7 +1243,7 @@ static void retrieve_status(struct dm_table *table,
+ 	char *outbuf, *outptr;
+ 	status_type_t type;
+ 	size_t remaining, len, used = 0;
+-	unsigned status_flags = 0;
++	unsigned int status_flags = 0;
+ 
+ 	outptr = outbuf = get_result_buffer(param, param_size, &len);
+ 
+@@ -1648,8 +1648,8 @@ static int table_status(struct file *filp, struct dm_ioctl *param, size_t param_
+  * Returns a number <= 1 if message was processed by device mapper.
+  * Returns 2 if message should be delivered to the target.
+  */
+-static int message_for_md(struct mapped_device *md, unsigned argc, char **argv,
+-			  char *result, unsigned maxlen)
++static int message_for_md(struct mapped_device *md, unsigned int argc, char **argv,
++			  char *result, unsigned int maxlen)
+ {
+ 	int r;
+ 
+@@ -1859,7 +1859,7 @@ static int copy_params(struct dm_ioctl __user *user, struct dm_ioctl *param_kern
+ 	struct dm_ioctl *dmi;
+ 	int secure_data;
+ 	const size_t minimum_data_size = offsetof(struct dm_ioctl, data);
+-	unsigned noio_flag;
++	unsigned int noio_flag;
+ 
+ 	if (copy_from_user(param_kernel, user, minimum_data_size))
+ 		return -EFAULT;
+diff --git a/drivers/md/dm-kcopyd.c b/drivers/md/dm-kcopyd.c
+index 4d3bbbea2e9a8..0ef78e56aa88c 100644
+--- a/drivers/md/dm-kcopyd.c
++++ b/drivers/md/dm-kcopyd.c
+@@ -34,14 +34,14 @@
+ #define DEFAULT_SUB_JOB_SIZE_KB 512
+ #define MAX_SUB_JOB_SIZE_KB     1024
+ 
+-static unsigned kcopyd_subjob_size_kb = DEFAULT_SUB_JOB_SIZE_KB;
++static unsigned int kcopyd_subjob_size_kb = DEFAULT_SUB_JOB_SIZE_KB;
+ 
+ module_param(kcopyd_subjob_size_kb, uint, S_IRUGO | S_IWUSR);
+ MODULE_PARM_DESC(kcopyd_subjob_size_kb, "Sub-job size for dm-kcopyd clients");
+ 
+-static unsigned dm_get_kcopyd_subjob_size(void)
++static unsigned int dm_get_kcopyd_subjob_size(void)
+ {
+-	unsigned sub_job_size_kb;
++	unsigned int sub_job_size_kb;
+ 
+ 	sub_job_size_kb = __dm_get_module_param(&kcopyd_subjob_size_kb,
+ 						DEFAULT_SUB_JOB_SIZE_KB,
+@@ -56,9 +56,9 @@ static unsigned dm_get_kcopyd_subjob_size(void)
+  *---------------------------------------------------------------*/
+ struct dm_kcopyd_client {
+ 	struct page_list *pages;
+-	unsigned nr_reserved_pages;
+-	unsigned nr_free_pages;
+-	unsigned sub_job_size;
++	unsigned int nr_reserved_pages;
++	unsigned int nr_free_pages;
++	unsigned int sub_job_size;
+ 
+ 	struct dm_io_client *io_client;
+ 
+@@ -119,7 +119,7 @@ static DEFINE_SPINLOCK(throttle_spinlock);
+ 
+ static void io_job_start(struct dm_kcopyd_throttle *t)
+ {
+-	unsigned throttle, now, difference;
++	unsigned int throttle, now, difference;
+ 	int slept = 0, skew;
+ 
+ 	if (unlikely(!t))
+@@ -182,7 +182,7 @@ static void io_job_finish(struct dm_kcopyd_throttle *t)
+ 		goto skip_limit;
+ 
+ 	if (!t->num_io_jobs) {
+-		unsigned now, difference;
++		unsigned int now, difference;
+ 
+ 		now = jiffies;
+ 		difference = now - t->last_jiffies;
+@@ -303,9 +303,9 @@ static void drop_pages(struct page_list *pl)
+ /*
+  * Allocate and reserve nr_pages for the use of a specific client.
+  */
+-static int client_reserve_pages(struct dm_kcopyd_client *kc, unsigned nr_pages)
++static int client_reserve_pages(struct dm_kcopyd_client *kc, unsigned int nr_pages)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	struct page_list *pl = NULL, *next;
+ 
+ 	for (i = 0; i < nr_pages; i++) {
+@@ -341,7 +341,7 @@ static void client_free_pages(struct dm_kcopyd_client *kc)
+ struct kcopyd_job {
+ 	struct dm_kcopyd_client *kc;
+ 	struct list_head list;
+-	unsigned flags;
++	unsigned int flags;
+ 
+ 	/*
+ 	 * Error state of the job.
+@@ -582,7 +582,7 @@ static int run_io_job(struct kcopyd_job *job)
+ static int run_pages_job(struct kcopyd_job *job)
+ {
+ 	int r;
+-	unsigned nr_pages = dm_div_up(job->dests[0].count, PAGE_SIZE >> 9);
++	unsigned int nr_pages = dm_div_up(job->dests[0].count, PAGE_SIZE >> 9);
+ 
+ 	r = kcopyd_get_pages(job->kc, nr_pages, &job->pages);
+ 	if (!r) {
+@@ -849,8 +849,8 @@ void dm_kcopyd_copy(struct dm_kcopyd_client *kc, struct dm_io_region *from,
+ EXPORT_SYMBOL(dm_kcopyd_copy);
+ 
+ void dm_kcopyd_zero(struct dm_kcopyd_client *kc,
+-		    unsigned num_dests, struct dm_io_region *dests,
+-		    unsigned flags, dm_kcopyd_notify_fn fn, void *context)
++		    unsigned int num_dests, struct dm_io_region *dests,
++		    unsigned int flags, dm_kcopyd_notify_fn fn, void *context)
+ {
+ 	dm_kcopyd_copy(kc, NULL, num_dests, dests, flags, fn, context);
+ }
+@@ -906,7 +906,7 @@ int kcopyd_cancel(struct kcopyd_job *job, int block)
+ struct dm_kcopyd_client *dm_kcopyd_client_create(struct dm_kcopyd_throttle *throttle)
+ {
+ 	int r;
+-	unsigned reserve_pages;
++	unsigned int reserve_pages;
+ 	struct dm_kcopyd_client *kc;
+ 
+ 	kc = kzalloc(sizeof(*kc), GFP_KERNEL);
+diff --git a/drivers/md/dm-linear.c b/drivers/md/dm-linear.c
+index 3212ef6aa81bb..26b1af6461771 100644
+--- a/drivers/md/dm-linear.c
++++ b/drivers/md/dm-linear.c
+@@ -95,7 +95,7 @@ static int linear_map(struct dm_target *ti, struct bio *bio)
+ }
+ 
+ static void linear_status(struct dm_target *ti, status_type_t type,
+-			  unsigned status_flags, char *result, unsigned maxlen)
++			  unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	struct linear_c *lc = (struct linear_c *) ti->private;
+ 	size_t sz = 0;
+diff --git a/drivers/md/dm-log-userspace-base.c b/drivers/md/dm-log-userspace-base.c
+index 9ab93ebea8895..9fc69382692bd 100644
+--- a/drivers/md/dm-log-userspace-base.c
++++ b/drivers/md/dm-log-userspace-base.c
+@@ -123,7 +123,7 @@ retry:
+ }
+ 
+ static int build_constructor_string(struct dm_target *ti,
+-				    unsigned argc, char **argv,
++				    unsigned int argc, char **argv,
+ 				    char **ctr_str)
+ {
+ 	int i, str_size;
+@@ -188,7 +188,7 @@ static void do_flush(struct work_struct *work)
+  * to the userspace ctr function.
+  */
+ static int userspace_ctr(struct dm_dirty_log *log, struct dm_target *ti,
+-			 unsigned argc, char **argv)
++			 unsigned int argc, char **argv)
+ {
+ 	int r = 0;
+ 	int str_size;
+@@ -792,7 +792,7 @@ static region_t userspace_get_sync_count(struct dm_dirty_log *log)
+  * Returns: amount of space consumed
+  */
+ static int userspace_status(struct dm_dirty_log *log, status_type_t status_type,
+-			    char *result, unsigned maxlen)
++			    char *result, unsigned int maxlen)
+ {
+ 	int r = 0;
+ 	char *table_args;
+diff --git a/drivers/md/dm-log-userspace-transfer.c b/drivers/md/dm-log-userspace-transfer.c
+index fdf8ec304f8d2..072559b709edd 100644
+--- a/drivers/md/dm-log-userspace-transfer.c
++++ b/drivers/md/dm-log-userspace-transfer.c
+@@ -142,7 +142,7 @@ static void cn_ulog_callback(struct cn_msg *msg, struct netlink_skb_parms *nsp)
+ 		fill_pkg(msg, NULL);
+ 	else if (msg->len < sizeof(*tfr))
+ 		DMERR("Incomplete message received (expected %u, got %u): [%u]",
+-		      (unsigned)sizeof(*tfr), msg->len, msg->seq);
++		      (unsigned int)sizeof(*tfr), msg->len, msg->seq);
+ 	else
+ 		fill_pkg(NULL, tfr);
+ 	spin_unlock(&receiving_list_lock);
+diff --git a/drivers/md/dm-log-writes.c b/drivers/md/dm-log-writes.c
+index 178e13a5b059f..efdfb2e1868a4 100644
+--- a/drivers/md/dm-log-writes.c
++++ b/drivers/md/dm-log-writes.c
+@@ -792,10 +792,10 @@ static int normal_end_io(struct dm_target *ti, struct bio *bio,
+  * INFO format: <logged entries> <highest allocated sector>
+  */
+ static void log_writes_status(struct dm_target *ti, status_type_t type,
+-			      unsigned status_flags, char *result,
+-			      unsigned maxlen)
++			      unsigned int status_flags, char *result,
++			      unsigned int maxlen)
+ {
+-	unsigned sz = 0;
++	unsigned int sz = 0;
+ 	struct log_writes_c *lc = ti->private;
+ 
+ 	switch (type) {
+@@ -844,8 +844,8 @@ static int log_writes_iterate_devices(struct dm_target *ti,
+  * Messages supported:
+  *   mark <mark data> - specify the marked data.
+  */
+-static int log_writes_message(struct dm_target *ti, unsigned argc, char **argv,
+-			      char *result, unsigned maxlen)
++static int log_writes_message(struct dm_target *ti, unsigned int argc, char **argv,
++			      char *result, unsigned int maxlen)
+ {
+ 	int r = -EINVAL;
+ 	struct log_writes_c *lc = ti->private;
+diff --git a/drivers/md/dm-log.c b/drivers/md/dm-log.c
+index cf10fa6677972..159f2c05dfd3c 100644
+--- a/drivers/md/dm-log.c
++++ b/drivers/md/dm-log.c
+@@ -223,7 +223,7 @@ struct log_c {
+ 	unsigned int region_count;
+ 	region_t sync_count;
+ 
+-	unsigned bitset_uint32_count;
++	unsigned int bitset_uint32_count;
+ 	uint32_t *clean_bits;
+ 	uint32_t *sync_bits;
+ 	uint32_t *recovering_bits;	/* FIXME: this seems excessive */
+@@ -255,20 +255,20 @@ struct log_c {
+  * The touched member needs to be updated every time we access
+  * one of the bitsets.
+  */
+-static inline int log_test_bit(uint32_t *bs, unsigned bit)
++static inline int log_test_bit(uint32_t *bs, unsigned int bit)
+ {
+ 	return test_bit_le(bit, bs) ? 1 : 0;
+ }
+ 
+ static inline void log_set_bit(struct log_c *l,
+-			       uint32_t *bs, unsigned bit)
++			       uint32_t *bs, unsigned int bit)
+ {
+ 	__set_bit_le(bit, bs);
+ 	l->touched_cleaned = 1;
+ }
+ 
+ static inline void log_clear_bit(struct log_c *l,
+-				 uint32_t *bs, unsigned bit)
++				 uint32_t *bs, unsigned int bit)
+ {
+ 	__clear_bit_le(bit, bs);
+ 	l->touched_dirtied = 1;
+@@ -582,7 +582,7 @@ static void fail_log_device(struct log_c *lc)
+ static int disk_resume(struct dm_dirty_log *log)
+ {
+ 	int r;
+-	unsigned i;
++	unsigned int i;
+ 	struct log_c *lc = (struct log_c *) log->context;
+ 	size_t size = lc->bitset_uint32_count * sizeof(uint32_t);
+ 
+diff --git a/drivers/md/dm-mpath.c b/drivers/md/dm-mpath.c
+index 0e325469a252a..91c25ad8eed84 100644
+--- a/drivers/md/dm-mpath.c
++++ b/drivers/md/dm-mpath.c
+@@ -29,7 +29,7 @@
+ 
+ #define DM_MSG_PREFIX "multipath"
+ #define DM_PG_INIT_DELAY_MSECS 2000
+-#define DM_PG_INIT_DELAY_DEFAULT ((unsigned) -1)
++#define DM_PG_INIT_DELAY_DEFAULT ((unsigned int) -1)
+ #define QUEUE_IF_NO_PATH_TIMEOUT_DEFAULT 0
+ 
+ static unsigned long queue_if_no_path_timeout_secs = QUEUE_IF_NO_PATH_TIMEOUT_DEFAULT;
+@@ -39,7 +39,7 @@ struct pgpath {
+ 	struct list_head list;
+ 
+ 	struct priority_group *pg;	/* Owning PG */
+-	unsigned fail_count;		/* Cumulative failure count */
++	unsigned int fail_count;		/* Cumulative failure count */
+ 
+ 	struct dm_path path;
+ 	struct delayed_work activate_path;
+@@ -59,8 +59,8 @@ struct priority_group {
+ 	struct multipath *m;		/* Owning multipath instance */
+ 	struct path_selector ps;
+ 
+-	unsigned pg_num;		/* Reference number */
+-	unsigned nr_pgpaths;		/* Number of paths in PG */
++	unsigned int pg_num;		/* Reference number */
++	unsigned int nr_pgpaths;		/* Number of paths in PG */
+ 	struct list_head pgpaths;
+ 
+ 	bool bypassed:1;		/* Temporarily bypass this PG? */
+@@ -78,14 +78,14 @@ struct multipath {
+ 	struct priority_group *next_pg;	/* Switch to this PG if set */
+ 
+ 	atomic_t nr_valid_paths;	/* Total number of usable paths */
+-	unsigned nr_priority_groups;
++	unsigned int nr_priority_groups;
+ 	struct list_head priority_groups;
+ 
+ 	const char *hw_handler_name;
+ 	char *hw_handler_params;
+ 	wait_queue_head_t pg_init_wait;	/* Wait for pg_init completion */
+-	unsigned pg_init_retries;	/* Number of times to retry pg_init */
+-	unsigned pg_init_delay_msecs;	/* Number of msecs before pg_init retry */
++	unsigned int pg_init_retries;	/* Number of times to retry pg_init */
++	unsigned int pg_init_delay_msecs;	/* Number of msecs before pg_init retry */
+ 	atomic_t pg_init_in_progress;	/* Only one pg_init allowed at once */
+ 	atomic_t pg_init_count;		/* Number of times pg_init called */
+ 
+@@ -397,7 +397,7 @@ static struct pgpath *choose_pgpath(struct multipath *m, size_t nr_bytes)
+ 	unsigned long flags;
+ 	struct priority_group *pg;
+ 	struct pgpath *pgpath;
+-	unsigned bypassed = 1;
++	unsigned int bypassed = 1;
+ 
+ 	if (!atomic_read(&m->nr_valid_paths)) {
+ 		spin_lock_irqsave(&m->lock, flags);
+@@ -840,7 +840,7 @@ static int parse_path_selector(struct dm_arg_set *as, struct priority_group *pg,
+ {
+ 	int r;
+ 	struct path_selector_type *pst;
+-	unsigned ps_argc;
++	unsigned int ps_argc;
+ 
+ 	static const struct dm_arg _args[] = {
+ 		{0, 1024, "invalid number of path selector args"},
+@@ -983,7 +983,7 @@ static struct priority_group *parse_priority_group(struct dm_arg_set *as,
+ 	};
+ 
+ 	int r;
+-	unsigned i, nr_selector_args, nr_args;
++	unsigned int i, nr_selector_args, nr_args;
+ 	struct priority_group *pg;
+ 	struct dm_target *ti = m->ti;
+ 
+@@ -1049,7 +1049,7 @@ static struct priority_group *parse_priority_group(struct dm_arg_set *as,
+ 
+ static int parse_hw_handler(struct dm_arg_set *as, struct multipath *m)
+ {
+-	unsigned hw_argc;
++	unsigned int hw_argc;
+ 	int ret;
+ 	struct dm_target *ti = m->ti;
+ 
+@@ -1101,7 +1101,7 @@ fail:
+ static int parse_features(struct dm_arg_set *as, struct multipath *m)
+ {
+ 	int r;
+-	unsigned argc;
++	unsigned int argc;
+ 	struct dm_target *ti = m->ti;
+ 	const char *arg_name;
+ 
+@@ -1170,7 +1170,7 @@ static int parse_features(struct dm_arg_set *as, struct multipath *m)
+ 	return r;
+ }
+ 
+-static int multipath_ctr(struct dm_target *ti, unsigned argc, char **argv)
++static int multipath_ctr(struct dm_target *ti, unsigned int argc, char **argv)
+ {
+ 	/* target arguments */
+ 	static const struct dm_arg _args[] = {
+@@ -1181,8 +1181,8 @@ static int multipath_ctr(struct dm_target *ti, unsigned argc, char **argv)
+ 	int r;
+ 	struct multipath *m;
+ 	struct dm_arg_set as;
+-	unsigned pg_count = 0;
+-	unsigned next_pg_num;
++	unsigned int pg_count = 0;
++	unsigned int next_pg_num;
+ 	unsigned long flags;
+ 
+ 	as.argc = argc;
+@@ -1224,7 +1224,7 @@ static int multipath_ctr(struct dm_target *ti, unsigned argc, char **argv)
+ 	/* parse the priority groups */
+ 	while (as.argc) {
+ 		struct priority_group *pg;
+-		unsigned nr_valid_paths = atomic_read(&m->nr_valid_paths);
++		unsigned int nr_valid_paths = atomic_read(&m->nr_valid_paths);
+ 
+ 		pg = parse_priority_group(&as, m);
+ 		if (IS_ERR(pg)) {
+@@ -1365,7 +1365,7 @@ static int reinstate_path(struct pgpath *pgpath)
+ 	int r = 0, run_queue = 0;
+ 	unsigned long flags;
+ 	struct multipath *m = pgpath->pg->m;
+-	unsigned nr_valid_paths;
++	unsigned int nr_valid_paths;
+ 
+ 	spin_lock_irqsave(&m->lock, flags);
+ 
+@@ -1454,7 +1454,7 @@ static void bypass_pg(struct multipath *m, struct priority_group *pg,
+ static int switch_pg_num(struct multipath *m, const char *pgstr)
+ {
+ 	struct priority_group *pg;
+-	unsigned pgnum;
++	unsigned int pgnum;
+ 	unsigned long flags;
+ 	char dummy;
+ 
+@@ -1487,7 +1487,7 @@ static int switch_pg_num(struct multipath *m, const char *pgstr)
+ static int bypass_pg_num(struct multipath *m, const char *pgstr, bool bypassed)
+ {
+ 	struct priority_group *pg;
+-	unsigned pgnum;
++	unsigned int pgnum;
+ 	char dummy;
+ 
+ 	if (!pgstr || (sscanf(pgstr, "%u%c", &pgnum, &dummy) != 1) || !pgnum ||
+@@ -1789,14 +1789,14 @@ static void multipath_resume(struct dm_target *ti)
+  *      num_paths num_selector_args [path_dev [selector_args]* ]+ ]+
+  */
+ static void multipath_status(struct dm_target *ti, status_type_t type,
+-			     unsigned status_flags, char *result, unsigned maxlen)
++			     unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	int sz = 0, pg_counter, pgpath_counter;
+ 	unsigned long flags;
+ 	struct multipath *m = ti->private;
+ 	struct priority_group *pg;
+ 	struct pgpath *p;
+-	unsigned pg_num;
++	unsigned int pg_num;
+ 	char state;
+ 
+ 	spin_lock_irqsave(&m->lock, flags);
+@@ -1948,8 +1948,8 @@ static void multipath_status(struct dm_target *ti, status_type_t type,
+ 	spin_unlock_irqrestore(&m->lock, flags);
+ }
+ 
+-static int multipath_message(struct dm_target *ti, unsigned argc, char **argv,
+-			     char *result, unsigned maxlen)
++static int multipath_message(struct dm_target *ti, unsigned int argc, char **argv,
++			     char *result, unsigned int maxlen)
+ {
+ 	int r = -EINVAL;
+ 	struct dm_dev *dev;
+diff --git a/drivers/md/dm-mpath.h b/drivers/md/dm-mpath.h
+index e230f71962596..5343698fe5f1b 100644
+--- a/drivers/md/dm-mpath.h
++++ b/drivers/md/dm-mpath.h
+@@ -17,6 +17,6 @@ struct dm_path {
+ };
+ 
+ /* Callback for hwh_pg_init_fn to use when complete */
+-void dm_pg_init_complete(struct dm_path *path, unsigned err_flags);
++void dm_pg_init_complete(struct dm_path *path, unsigned int err_flags);
+ 
+ #endif
+diff --git a/drivers/md/dm-path-selector.h b/drivers/md/dm-path-selector.h
+index 83cac2b04b668..0f2b37af87662 100644
+--- a/drivers/md/dm-path-selector.h
++++ b/drivers/md/dm-path-selector.h
+@@ -52,7 +52,7 @@ struct path_selector_type {
+ 	/*
+ 	 * Constructs a path selector object, takes custom arguments
+ 	 */
+-	int (*create) (struct path_selector *ps, unsigned argc, char **argv);
++	int (*create) (struct path_selector *ps, unsigned int argc, char **argv);
+ 	void (*destroy) (struct path_selector *ps);
+ 
+ 	/*
+diff --git a/drivers/md/dm-ps-io-affinity.c b/drivers/md/dm-ps-io-affinity.c
+index f74501e65a8ed..76ce4ce872229 100644
+--- a/drivers/md/dm-ps-io-affinity.c
++++ b/drivers/md/dm-ps-io-affinity.c
+@@ -108,7 +108,7 @@ free_pi:
+ 	return ret;
+ }
+ 
+-static int ioa_create(struct path_selector *ps, unsigned argc, char **argv)
++static int ioa_create(struct path_selector *ps, unsigned int argc, char **argv)
+ {
+ 	struct selector *s;
+ 
+@@ -138,7 +138,7 @@ free_selector:
+ static void ioa_destroy(struct path_selector *ps)
+ {
+ 	struct selector *s = ps->context;
+-	unsigned cpu;
++	unsigned int cpu;
+ 
+ 	for_each_cpu(cpu, s->path_mask)
+ 		ioa_free_path(s, cpu);
+diff --git a/drivers/md/dm-ps-queue-length.c b/drivers/md/dm-ps-queue-length.c
+index cef70657bbbc2..6fbec9fc242d9 100644
+--- a/drivers/md/dm-ps-queue-length.c
++++ b/drivers/md/dm-ps-queue-length.c
+@@ -35,7 +35,7 @@ struct selector {
+ struct path_info {
+ 	struct list_head	list;
+ 	struct dm_path		*path;
+-	unsigned		repeat_count;
++	unsigned int		repeat_count;
+ 	atomic_t		qlen;	/* the number of in-flight I/Os */
+ };
+ 
+@@ -52,7 +52,7 @@ static struct selector *alloc_selector(void)
+ 	return s;
+ }
+ 
+-static int ql_create(struct path_selector *ps, unsigned argc, char **argv)
++static int ql_create(struct path_selector *ps, unsigned int argc, char **argv)
+ {
+ 	struct selector *s = alloc_selector();
+ 
+@@ -84,9 +84,9 @@ static void ql_destroy(struct path_selector *ps)
+ }
+ 
+ static int ql_status(struct path_selector *ps, struct dm_path *path,
+-		     status_type_t type, char *result, unsigned maxlen)
++		     status_type_t type, char *result, unsigned int maxlen)
+ {
+-	unsigned sz = 0;
++	unsigned int sz = 0;
+ 	struct path_info *pi;
+ 
+ 	/* When called with NULL path, return selector status/args. */
+@@ -116,7 +116,7 @@ static int ql_add_path(struct path_selector *ps, struct dm_path *path,
+ {
+ 	struct selector *s = ps->context;
+ 	struct path_info *pi;
+-	unsigned repeat_count = QL_MIN_IO;
++	unsigned int repeat_count = QL_MIN_IO;
+ 	char dummy;
+ 	unsigned long flags;
+ 
+diff --git a/drivers/md/dm-ps-round-robin.c b/drivers/md/dm-ps-round-robin.c
+index 27f44c5fa04e8..1d07392b5ed48 100644
+--- a/drivers/md/dm-ps-round-robin.c
++++ b/drivers/md/dm-ps-round-robin.c
+@@ -26,7 +26,7 @@
+ struct path_info {
+ 	struct list_head list;
+ 	struct dm_path *path;
+-	unsigned repeat_count;
++	unsigned int repeat_count;
+ };
+ 
+ static void free_paths(struct list_head *paths)
+@@ -62,7 +62,7 @@ static struct selector *alloc_selector(void)
+ 	return s;
+ }
+ 
+-static int rr_create(struct path_selector *ps, unsigned argc, char **argv)
++static int rr_create(struct path_selector *ps, unsigned int argc, char **argv)
+ {
+ 	struct selector *s;
+ 
+@@ -119,7 +119,7 @@ static int rr_add_path(struct path_selector *ps, struct dm_path *path,
+ {
+ 	struct selector *s = ps->context;
+ 	struct path_info *pi;
+-	unsigned repeat_count = RR_MIN_IO;
++	unsigned int repeat_count = RR_MIN_IO;
+ 	char dummy;
+ 	unsigned long flags;
+ 
+diff --git a/drivers/md/dm-ps-service-time.c b/drivers/md/dm-ps-service-time.c
+index 3ec9c33265c52..84d26234dc053 100644
+--- a/drivers/md/dm-ps-service-time.c
++++ b/drivers/md/dm-ps-service-time.c
+@@ -30,8 +30,8 @@ struct selector {
+ struct path_info {
+ 	struct list_head list;
+ 	struct dm_path *path;
+-	unsigned repeat_count;
+-	unsigned relative_throughput;
++	unsigned int repeat_count;
++	unsigned int relative_throughput;
+ 	atomic_t in_flight_size;	/* Total size of in-flight I/Os */
+ };
+ 
+@@ -48,7 +48,7 @@ static struct selector *alloc_selector(void)
+ 	return s;
+ }
+ 
+-static int st_create(struct path_selector *ps, unsigned argc, char **argv)
++static int st_create(struct path_selector *ps, unsigned int argc, char **argv)
+ {
+ 	struct selector *s = alloc_selector();
+ 
+@@ -80,9 +80,9 @@ static void st_destroy(struct path_selector *ps)
+ }
+ 
+ static int st_status(struct path_selector *ps, struct dm_path *path,
+-		     status_type_t type, char *result, unsigned maxlen)
++		     status_type_t type, char *result, unsigned int maxlen)
+ {
+-	unsigned sz = 0;
++	unsigned int sz = 0;
+ 	struct path_info *pi;
+ 
+ 	if (!path)
+@@ -113,8 +113,8 @@ static int st_add_path(struct path_selector *ps, struct dm_path *path,
+ {
+ 	struct selector *s = ps->context;
+ 	struct path_info *pi;
+-	unsigned repeat_count = ST_MIN_IO;
+-	unsigned relative_throughput = 1;
++	unsigned int repeat_count = ST_MIN_IO;
++	unsigned int relative_throughput = 1;
+ 	char dummy;
+ 	unsigned long flags;
+ 
+diff --git a/drivers/md/dm-raid.c b/drivers/md/dm-raid.c
+index 54263679a7b14..b26c12856b1db 100644
+--- a/drivers/md/dm-raid.c
++++ b/drivers/md/dm-raid.c
+@@ -3712,7 +3712,7 @@ static void raid_status(struct dm_target *ti, status_type_t type,
+ }
+ 
+ static int raid_message(struct dm_target *ti, unsigned int argc, char **argv,
+-			char *result, unsigned maxlen)
++			char *result, unsigned int maxlen)
+ {
+ 	struct raid_set *rs = ti->private;
+ 	struct mddev *mddev = &rs->md;
+diff --git a/drivers/md/dm-raid1.c b/drivers/md/dm-raid1.c
+index 06a38dc320253..8bd7e87d3538e 100644
+--- a/drivers/md/dm-raid1.c
++++ b/drivers/md/dm-raid1.c
+@@ -82,7 +82,7 @@ struct mirror_set {
+ 
+ 	struct work_struct trigger_event;
+ 
+-	unsigned nr_mirrors;
++	unsigned int nr_mirrors;
+ 	struct mirror mirror[];
+ };
+ 
+@@ -327,7 +327,7 @@ static void recovery_complete(int read_err, unsigned long write_err,
+ 
+ static void recover(struct mirror_set *ms, struct dm_region *reg)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	struct dm_io_region from, to[DM_KCOPYD_MAX_REGIONS], *dest;
+ 	struct mirror *m;
+ 	unsigned long flags = 0;
+@@ -593,7 +593,7 @@ static void do_reads(struct mirror_set *ms, struct bio_list *reads)
+ 
+ static void write_callback(unsigned long error, void *context)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	struct bio *bio = (struct bio *) context;
+ 	struct mirror_set *ms;
+ 	int should_wake = 0;
+@@ -963,10 +963,10 @@ static int get_mirror(struct mirror_set *ms, struct dm_target *ti,
+  * Create dirty log: log_type #log_params <log_params>
+  */
+ static struct dm_dirty_log *create_dirty_log(struct dm_target *ti,
+-					     unsigned argc, char **argv,
+-					     unsigned *args_used)
++					     unsigned int argc, char **argv,
++					     unsigned int *args_used)
+ {
+-	unsigned param_count;
++	unsigned int param_count;
+ 	struct dm_dirty_log *dl;
+ 	char dummy;
+ 
+@@ -997,10 +997,10 @@ static struct dm_dirty_log *create_dirty_log(struct dm_target *ti,
+ 	return dl;
+ }
+ 
+-static int parse_features(struct mirror_set *ms, unsigned argc, char **argv,
+-			  unsigned *args_used)
++static int parse_features(struct mirror_set *ms, unsigned int argc, char **argv,
++			  unsigned int *args_used)
+ {
+-	unsigned num_features;
++	unsigned int num_features;
+ 	struct dm_target *ti = ms->ti;
+ 	char dummy;
+ 	int i;
+@@ -1389,7 +1389,7 @@ static char device_status_char(struct mirror *m)
+ 
+ 
+ static void mirror_status(struct dm_target *ti, status_type_t type,
+-			  unsigned status_flags, char *result, unsigned maxlen)
++			  unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	unsigned int m, sz = 0;
+ 	int num_feature_args = 0;
+@@ -1458,7 +1458,7 @@ static int mirror_iterate_devices(struct dm_target *ti,
+ {
+ 	struct mirror_set *ms = ti->private;
+ 	int ret = 0;
+-	unsigned i;
++	unsigned int i;
+ 
+ 	for (i = 0; !ret && i < ms->nr_mirrors; i++)
+ 		ret = fn(ti, ms->mirror[i].dev,
+diff --git a/drivers/md/dm-region-hash.c b/drivers/md/dm-region-hash.c
+index 1f760451e6f48..adbdb4b671372 100644
+--- a/drivers/md/dm-region-hash.c
++++ b/drivers/md/dm-region-hash.c
+@@ -56,17 +56,17 @@
+  *---------------------------------------------------------------*/
+ struct dm_region_hash {
+ 	uint32_t region_size;
+-	unsigned region_shift;
++	unsigned int region_shift;
+ 
+ 	/* holds persistent region state */
+ 	struct dm_dirty_log *log;
+ 
+ 	/* hash table */
+ 	rwlock_t hash_lock;
+-	unsigned mask;
+-	unsigned nr_buckets;
+-	unsigned prime;
+-	unsigned shift;
++	unsigned int mask;
++	unsigned int nr_buckets;
++	unsigned int prime;
++	unsigned int shift;
+ 	struct list_head *buckets;
+ 
+ 	/*
+@@ -74,7 +74,7 @@ struct dm_region_hash {
+ 	 */
+ 	int flush_failure;
+ 
+-	unsigned max_recovery; /* Max # of regions to recover in parallel */
++	unsigned int max_recovery; /* Max # of regions to recover in parallel */
+ 
+ 	spinlock_t region_lock;
+ 	atomic_t recovery_in_flight;
+@@ -163,12 +163,12 @@ struct dm_region_hash *dm_region_hash_create(
+ 						     struct bio_list *bios),
+ 		void (*wakeup_workers)(void *context),
+ 		void (*wakeup_all_recovery_waiters)(void *context),
+-		sector_t target_begin, unsigned max_recovery,
++		sector_t target_begin, unsigned int max_recovery,
+ 		struct dm_dirty_log *log, uint32_t region_size,
+ 		region_t nr_regions)
+ {
+ 	struct dm_region_hash *rh;
+-	unsigned nr_buckets, max_buckets;
++	unsigned int nr_buckets, max_buckets;
+ 	size_t i;
+ 	int ret;
+ 
+@@ -236,7 +236,7 @@ EXPORT_SYMBOL_GPL(dm_region_hash_create);
+ 
+ void dm_region_hash_destroy(struct dm_region_hash *rh)
+ {
+-	unsigned h;
++	unsigned int h;
+ 	struct dm_region *reg, *nreg;
+ 
+ 	BUG_ON(!list_empty(&rh->quiesced_regions));
+@@ -263,9 +263,9 @@ struct dm_dirty_log *dm_rh_dirty_log(struct dm_region_hash *rh)
+ }
+ EXPORT_SYMBOL_GPL(dm_rh_dirty_log);
+ 
+-static unsigned rh_hash(struct dm_region_hash *rh, region_t region)
++static unsigned int rh_hash(struct dm_region_hash *rh, region_t region)
+ {
+-	return (unsigned) ((region * rh->prime) >> rh->shift) & rh->mask;
++	return (unsigned int) ((region * rh->prime) >> rh->shift) & rh->mask;
+ }
+ 
+ static struct dm_region *__rh_lookup(struct dm_region_hash *rh, region_t region)
+diff --git a/drivers/md/dm-rq.c b/drivers/md/dm-rq.c
+index a41209a43506c..80f46e01bca44 100644
+--- a/drivers/md/dm-rq.c
++++ b/drivers/md/dm-rq.c
+@@ -23,33 +23,33 @@ struct dm_rq_target_io {
+ 	union map_info info;
+ 	struct dm_stats_aux stats_aux;
+ 	unsigned long duration_jiffies;
+-	unsigned n_sectors;
+-	unsigned completed;
++	unsigned int n_sectors;
++	unsigned int completed;
+ };
+ 
+ #define DM_MQ_NR_HW_QUEUES 1
+ #define DM_MQ_QUEUE_DEPTH 2048
+-static unsigned dm_mq_nr_hw_queues = DM_MQ_NR_HW_QUEUES;
+-static unsigned dm_mq_queue_depth = DM_MQ_QUEUE_DEPTH;
++static unsigned int dm_mq_nr_hw_queues = DM_MQ_NR_HW_QUEUES;
++static unsigned int dm_mq_queue_depth = DM_MQ_QUEUE_DEPTH;
+ 
+ /*
+  * Request-based DM's mempools' reserved IOs set by the user.
+  */
+ #define RESERVED_REQUEST_BASED_IOS	256
+-static unsigned reserved_rq_based_ios = RESERVED_REQUEST_BASED_IOS;
++static unsigned int reserved_rq_based_ios = RESERVED_REQUEST_BASED_IOS;
+ 
+-unsigned dm_get_reserved_rq_based_ios(void)
++unsigned int dm_get_reserved_rq_based_ios(void)
+ {
+ 	return __dm_get_module_param(&reserved_rq_based_ios,
+ 				     RESERVED_REQUEST_BASED_IOS, DM_RESERVED_MAX_IOS);
+ }
+ 
+-static unsigned dm_get_blk_mq_nr_hw_queues(void)
++static unsigned int dm_get_blk_mq_nr_hw_queues(void)
+ {
+ 	return __dm_get_module_param(&dm_mq_nr_hw_queues, 1, 32);
+ }
+ 
+-static unsigned dm_get_blk_mq_queue_depth(void)
++static unsigned int dm_get_blk_mq_queue_depth(void)
+ {
+ 	return __dm_get_module_param(&dm_mq_queue_depth,
+ 				     DM_MQ_QUEUE_DEPTH, BLK_MQ_MAX_DEPTH);
+diff --git a/drivers/md/dm-rq.h b/drivers/md/dm-rq.h
+index 1eea0da641db5..2c97ad1451400 100644
+--- a/drivers/md/dm-rq.h
++++ b/drivers/md/dm-rq.h
+@@ -38,7 +38,7 @@ void dm_stop_queue(struct request_queue *q);
+ 
+ void dm_mq_kick_requeue_list(struct mapped_device *md);
+ 
+-unsigned dm_get_reserved_rq_based_ios(void);
++unsigned int dm_get_reserved_rq_based_ios(void);
+ 
+ ssize_t dm_attr_rq_based_seq_io_merge_deadline_show(struct mapped_device *md, char *buf);
+ ssize_t dm_attr_rq_based_seq_io_merge_deadline_store(struct mapped_device *md,
+diff --git a/drivers/md/dm-snap-persistent.c b/drivers/md/dm-snap-persistent.c
+index 680cc05ec6542..5176810f5d243 100644
+--- a/drivers/md/dm-snap-persistent.c
++++ b/drivers/md/dm-snap-persistent.c
+@@ -303,7 +303,7 @@ static int read_header(struct pstore *ps, int *new_snapshot)
+ {
+ 	int r;
+ 	struct disk_header *dh;
+-	unsigned chunk_size;
++	unsigned int chunk_size;
+ 	int chunk_size_supplied = 1;
+ 	char *chunk_err;
+ 
+@@ -895,11 +895,11 @@ err_workqueue:
+ 	return r;
+ }
+ 
+-static unsigned persistent_status(struct dm_exception_store *store,
++static unsigned int persistent_status(struct dm_exception_store *store,
+ 				  status_type_t status, char *result,
+-				  unsigned maxlen)
++				  unsigned int maxlen)
+ {
+-	unsigned sz = 0;
++	unsigned int sz = 0;
+ 
+ 	switch (status) {
+ 	case STATUSTYPE_INFO:
+diff --git a/drivers/md/dm-snap-transient.c b/drivers/md/dm-snap-transient.c
+index 0e0ae4c36b374..d83a0565bd101 100644
+--- a/drivers/md/dm-snap-transient.c
++++ b/drivers/md/dm-snap-transient.c
+@@ -84,11 +84,11 @@ static int transient_ctr(struct dm_exception_store *store, char *options)
+ 	return 0;
+ }
+ 
+-static unsigned transient_status(struct dm_exception_store *store,
++static unsigned int transient_status(struct dm_exception_store *store,
+ 				 status_type_t status, char *result,
+-				 unsigned maxlen)
++				 unsigned int maxlen)
+ {
+-	unsigned sz = 0;
++	unsigned int sz = 0;
+ 
+ 	switch (status) {
+ 	case STATUSTYPE_INFO:
+diff --git a/drivers/md/dm-snap.c b/drivers/md/dm-snap.c
+index d1c2f84d27e36..c64d987c544d7 100644
+--- a/drivers/md/dm-snap.c
++++ b/drivers/md/dm-snap.c
+@@ -41,7 +41,7 @@ static const char dm_snapshot_merge_target_name[] = "snapshot-merge";
+ 
+ struct dm_exception_table {
+ 	uint32_t hash_mask;
+-	unsigned hash_shift;
++	unsigned int hash_shift;
+ 	struct hlist_bl_head *table;
+ };
+ 
+@@ -106,7 +106,7 @@ struct dm_snapshot {
+ 	/* The on disk metadata handler */
+ 	struct dm_exception_store *store;
+ 
+-	unsigned in_progress;
++	unsigned int in_progress;
+ 	struct wait_queue_head in_progress_wait;
+ 
+ 	struct dm_kcopyd_client *kcopyd_client;
+@@ -161,7 +161,7 @@ struct dm_snapshot {
+  */
+ #define DEFAULT_COW_THRESHOLD 2048
+ 
+-static unsigned cow_threshold = DEFAULT_COW_THRESHOLD;
++static unsigned int cow_threshold = DEFAULT_COW_THRESHOLD;
+ module_param_named(snapshot_cow_threshold, cow_threshold, uint, 0644);
+ MODULE_PARM_DESC(snapshot_cow_threshold, "Maximum number of chunks being copied on write");
+ 
+@@ -324,7 +324,7 @@ struct origin {
+ struct dm_origin {
+ 	struct dm_dev *dev;
+ 	struct dm_target *ti;
+-	unsigned split_boundary;
++	unsigned int split_boundary;
+ 	struct list_head hash_list;
+ };
+ 
+@@ -377,7 +377,7 @@ static void exit_origin_hash(void)
+ 	kfree(_dm_origins);
+ }
+ 
+-static unsigned origin_hash(struct block_device *bdev)
++static unsigned int origin_hash(struct block_device *bdev)
+ {
+ 	return bdev->bd_dev & ORIGIN_MASK;
+ }
+@@ -652,7 +652,7 @@ static void dm_exception_table_unlock(struct dm_exception_table_lock *lock)
+ }
+ 
+ static int dm_exception_table_init(struct dm_exception_table *et,
+-				   uint32_t size, unsigned hash_shift)
++				   uint32_t size, unsigned int hash_shift)
+ {
+ 	unsigned int i;
+ 
+@@ -850,7 +850,7 @@ static int dm_add_exception(void *context, chunk_t old, chunk_t new)
+ static uint32_t __minimum_chunk_size(struct origin *o)
+ {
+ 	struct dm_snapshot *snap;
+-	unsigned chunk_size = rounddown_pow_of_two(UINT_MAX);
++	unsigned int chunk_size = rounddown_pow_of_two(UINT_MAX);
+ 
+ 	if (o)
+ 		list_for_each_entry(snap, &o->snapshots, list)
+@@ -1010,7 +1010,7 @@ out:
+ }
+ 
+ static int origin_write_extent(struct dm_snapshot *merging_snap,
+-			       sector_t sector, unsigned chunk_size);
++			       sector_t sector, unsigned int chunk_size);
+ 
+ static void merge_callback(int read_err, unsigned long write_err,
+ 			   void *context);
+@@ -1183,7 +1183,7 @@ static int parse_snapshot_features(struct dm_arg_set *as, struct dm_snapshot *s,
+ 				   struct dm_target *ti)
+ {
+ 	int r;
+-	unsigned argc;
++	unsigned int argc;
+ 	const char *arg_name;
+ 
+ 	static const struct dm_arg _args[] = {
+@@ -1241,7 +1241,7 @@ static int snapshot_ctr(struct dm_target *ti, unsigned int argc, char **argv)
+ 	int r = -EINVAL;
+ 	char *origin_path, *cow_path;
+ 	dev_t origin_dev, cow_dev;
+-	unsigned args_used, num_flush_bios = 1;
++	unsigned int args_used, num_flush_bios = 1;
+ 	fmode_t origin_mode = FMODE_READ;
+ 
+ 	if (argc < 4) {
+@@ -2315,11 +2315,11 @@ static void snapshot_merge_resume(struct dm_target *ti)
+ }
+ 
+ static void snapshot_status(struct dm_target *ti, status_type_t type,
+-			    unsigned status_flags, char *result, unsigned maxlen)
++			    unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+-	unsigned sz = 0;
++	unsigned int sz = 0;
+ 	struct dm_snapshot *snap = ti->private;
+-	unsigned num_features;
++	unsigned int num_features;
+ 
+ 	switch (type) {
+ 	case STATUSTYPE_INFO:
+@@ -2592,7 +2592,7 @@ again:
+  * size must be a multiple of merging_snap's chunk_size.
+  */
+ static int origin_write_extent(struct dm_snapshot *merging_snap,
+-			       sector_t sector, unsigned size)
++			       sector_t sector, unsigned int size)
+ {
+ 	int must_wait = 0;
+ 	sector_t n;
+@@ -2668,7 +2668,7 @@ static void origin_dtr(struct dm_target *ti)
+ static int origin_map(struct dm_target *ti, struct bio *bio)
+ {
+ 	struct dm_origin *o = ti->private;
+-	unsigned available_sectors;
++	unsigned int available_sectors;
+ 
+ 	bio_set_dev(bio, o->dev->bdev);
+ 
+@@ -2679,7 +2679,7 @@ static int origin_map(struct dm_target *ti, struct bio *bio)
+ 		return DM_MAPIO_REMAPPED;
+ 
+ 	available_sectors = o->split_boundary -
+-		((unsigned)bio->bi_iter.bi_sector & (o->split_boundary - 1));
++		((unsigned int)bio->bi_iter.bi_sector & (o->split_boundary - 1));
+ 
+ 	if (bio_sectors(bio) > available_sectors)
+ 		dm_accept_partial_bio(bio, available_sectors);
+@@ -2713,7 +2713,7 @@ static void origin_postsuspend(struct dm_target *ti)
+ }
+ 
+ static void origin_status(struct dm_target *ti, status_type_t type,
+-			  unsigned status_flags, char *result, unsigned maxlen)
++			  unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	struct dm_origin *o = ti->private;
+ 
+diff --git a/drivers/md/dm-stats.c b/drivers/md/dm-stats.c
+index d12ba9bce145d..7eeb3c2a2492b 100644
+--- a/drivers/md/dm-stats.c
++++ b/drivers/md/dm-stats.c
+@@ -42,12 +42,12 @@ struct dm_stat_shared {
+ struct dm_stat {
+ 	struct list_head list_entry;
+ 	int id;
+-	unsigned stat_flags;
++	unsigned int stat_flags;
+ 	size_t n_entries;
+ 	sector_t start;
+ 	sector_t end;
+ 	sector_t step;
+-	unsigned n_histogram_entries;
++	unsigned int n_histogram_entries;
+ 	unsigned long long *histogram_boundaries;
+ 	const char *program_id;
+ 	const char *aux_data;
+@@ -63,7 +63,7 @@ struct dm_stat {
+ 
+ struct dm_stats_last_position {
+ 	sector_t last_sector;
+-	unsigned last_rw;
++	unsigned int last_rw;
+ };
+ 
+ /*
+@@ -255,8 +255,8 @@ static void dm_stats_recalc_precise_timestamps(struct dm_stats *stats)
+ }
+ 
+ static int dm_stats_create(struct dm_stats *stats, sector_t start, sector_t end,
+-			   sector_t step, unsigned stat_flags,
+-			   unsigned n_histogram_entries,
++			   sector_t step, unsigned int stat_flags,
++			   unsigned int n_histogram_entries,
+ 			   unsigned long long *histogram_boundaries,
+ 			   const char *program_id, const char *aux_data,
+ 			   void (*suspend_callback)(struct mapped_device *),
+@@ -475,11 +475,11 @@ do_sync_free:
+ }
+ 
+ static int dm_stats_list(struct dm_stats *stats, const char *program,
+-			 char *result, unsigned maxlen)
++			 char *result, unsigned int maxlen)
+ {
+ 	struct dm_stat *s;
+ 	sector_t len;
+-	unsigned sz = 0;
++	unsigned int sz = 0;
+ 
+ 	/*
+ 	 * Output format:
+@@ -499,7 +499,7 @@ static int dm_stats_list(struct dm_stats *stats, const char *program,
+ 			if (s->stat_flags & STAT_PRECISE_TIMESTAMPS)
+ 				DMEMIT(" precise_timestamps");
+ 			if (s->n_histogram_entries) {
+-				unsigned i;
++				unsigned int i;
+ 				DMEMIT(" histogram:");
+ 				for (i = 0; i < s->n_histogram_entries; i++) {
+ 					if (i)
+@@ -523,7 +523,7 @@ static void dm_stat_round(struct dm_stat *s, struct dm_stat_shared *shared,
+ 	 * This is racy, but so is part_round_stats_single.
+ 	 */
+ 	unsigned long long now, difference;
+-	unsigned in_flight_read, in_flight_write;
++	unsigned int in_flight_read, in_flight_write;
+ 
+ 	if (likely(!(s->stat_flags & STAT_PRECISE_TIMESTAMPS)))
+ 		now = jiffies;
+@@ -534,8 +534,8 @@ static void dm_stat_round(struct dm_stat *s, struct dm_stat_shared *shared,
+ 	if (!difference)
+ 		return;
+ 
+-	in_flight_read = (unsigned)atomic_read(&shared->in_flight[READ]);
+-	in_flight_write = (unsigned)atomic_read(&shared->in_flight[WRITE]);
++	in_flight_read = (unsigned int)atomic_read(&shared->in_flight[READ]);
++	in_flight_write = (unsigned int)atomic_read(&shared->in_flight[WRITE]);
+ 	if (in_flight_read)
+ 		p->io_ticks[READ] += difference;
+ 	if (in_flight_write)
+@@ -596,9 +596,9 @@ static void dm_stat_for_entry(struct dm_stat *s, size_t entry,
+ 			duration = stats_aux->duration_ns;
+ 		}
+ 		if (s->n_histogram_entries) {
+-			unsigned lo = 0, hi = s->n_histogram_entries + 1;
++			unsigned int lo = 0, hi = s->n_histogram_entries + 1;
+ 			while (lo + 1 < hi) {
+-				unsigned mid = (lo + hi) / 2;
++				unsigned int mid = (lo + hi) / 2;
+ 				if (s->histogram_boundaries[mid - 1] > duration) {
+ 					hi = mid;
+ 				} else {
+@@ -656,7 +656,7 @@ static void __dm_stat_bio(struct dm_stat *s, int bi_rw,
+ }
+ 
+ void dm_stats_account_io(struct dm_stats *stats, unsigned long bi_rw,
+-			 sector_t bi_sector, unsigned bi_sectors, bool end,
++			 sector_t bi_sector, unsigned int bi_sectors, bool end,
+ 			 unsigned long start_time,
+ 			 struct dm_stats_aux *stats_aux)
+ {
+@@ -745,7 +745,7 @@ static void __dm_stat_init_temporary_percpu_totals(struct dm_stat_shared *shared
+ 		shared->tmp.io_ticks_total += READ_ONCE(p->io_ticks_total);
+ 		shared->tmp.time_in_queue += READ_ONCE(p->time_in_queue);
+ 		if (s->n_histogram_entries) {
+-			unsigned i;
++			unsigned int i;
+ 			for (i = 0; i < s->n_histogram_entries + 1; i++)
+ 				shared->tmp.histogram[i] += READ_ONCE(p->histogram[i]);
+ 		}
+@@ -779,7 +779,7 @@ static void __dm_stat_clear(struct dm_stat *s, size_t idx_start, size_t idx_end,
+ 		p->time_in_queue -= shared->tmp.time_in_queue;
+ 		local_irq_enable();
+ 		if (s->n_histogram_entries) {
+-			unsigned i;
++			unsigned int i;
+ 			for (i = 0; i < s->n_histogram_entries + 1; i++) {
+ 				local_irq_disable();
+ 				p = &s->stat_percpu[smp_processor_id()][x];
+@@ -816,7 +816,7 @@ static int dm_stats_clear(struct dm_stats *stats, int id)
+ static unsigned long long dm_jiffies_to_msec64(struct dm_stat *s, unsigned long long j)
+ {
+ 	unsigned long long result;
+-	unsigned mult;
++	unsigned int mult;
+ 
+ 	if (s->stat_flags & STAT_PRECISE_TIMESTAMPS)
+ 		return j;
+@@ -836,9 +836,9 @@ static unsigned long long dm_jiffies_to_msec64(struct dm_stat *s, unsigned long
+ 
+ static int dm_stats_print(struct dm_stats *stats, int id,
+ 			  size_t idx_start, size_t idx_len,
+-			  bool clear, char *result, unsigned maxlen)
++			  bool clear, char *result, unsigned int maxlen)
+ {
+-	unsigned sz = 0;
++	unsigned int sz = 0;
+ 	struct dm_stat *s;
+ 	size_t x;
+ 	sector_t start, end, step;
+@@ -894,7 +894,7 @@ static int dm_stats_print(struct dm_stats *stats, int id,
+ 		       dm_jiffies_to_msec64(s, shared->tmp.io_ticks[READ]),
+ 		       dm_jiffies_to_msec64(s, shared->tmp.io_ticks[WRITE]));
+ 		if (s->n_histogram_entries) {
+-			unsigned i;
++			unsigned int i;
+ 			for (i = 0; i < s->n_histogram_entries + 1; i++) {
+ 				DMEMIT("%s%llu", !i ? " " : ":", shared->tmp.histogram[i]);
+ 			}
+@@ -943,11 +943,11 @@ static int dm_stats_set_aux(struct dm_stats *stats, int id, const char *aux_data
+ 	return 0;
+ }
+ 
+-static int parse_histogram(const char *h, unsigned *n_histogram_entries,
++static int parse_histogram(const char *h, unsigned int *n_histogram_entries,
+ 			   unsigned long long **histogram_boundaries)
+ {
+ 	const char *q;
+-	unsigned n;
++	unsigned int n;
+ 	unsigned long long last;
+ 
+ 	*n_histogram_entries = 1;
+@@ -982,23 +982,23 @@ static int parse_histogram(const char *h, unsigned *n_histogram_entries,
+ }
+ 
+ static int message_stats_create(struct mapped_device *md,
+-				unsigned argc, char **argv,
+-				char *result, unsigned maxlen)
++				unsigned int argc, char **argv,
++				char *result, unsigned int maxlen)
+ {
+ 	int r;
+ 	int id;
+ 	char dummy;
+ 	unsigned long long start, end, len, step;
+-	unsigned divisor;
++	unsigned int divisor;
+ 	const char *program_id, *aux_data;
+-	unsigned stat_flags = 0;
++	unsigned int stat_flags = 0;
+ 
+-	unsigned n_histogram_entries = 0;
++	unsigned int n_histogram_entries = 0;
+ 	unsigned long long *histogram_boundaries = NULL;
+ 
+ 	struct dm_arg_set as, as_backup;
+ 	const char *a;
+-	unsigned feature_args;
++	unsigned int feature_args;
+ 
+ 	/*
+ 	 * Input format:
+@@ -1107,7 +1107,7 @@ ret:
+ }
+ 
+ static int message_stats_delete(struct mapped_device *md,
+-				unsigned argc, char **argv)
++				unsigned int argc, char **argv)
+ {
+ 	int id;
+ 	char dummy;
+@@ -1122,7 +1122,7 @@ static int message_stats_delete(struct mapped_device *md,
+ }
+ 
+ static int message_stats_clear(struct mapped_device *md,
+-			       unsigned argc, char **argv)
++			       unsigned int argc, char **argv)
+ {
+ 	int id;
+ 	char dummy;
+@@ -1137,8 +1137,8 @@ static int message_stats_clear(struct mapped_device *md,
+ }
+ 
+ static int message_stats_list(struct mapped_device *md,
+-			      unsigned argc, char **argv,
+-			      char *result, unsigned maxlen)
++			      unsigned int argc, char **argv,
++			      char *result, unsigned int maxlen)
+ {
+ 	int r;
+ 	const char *program = NULL;
+@@ -1160,8 +1160,8 @@ static int message_stats_list(struct mapped_device *md,
+ }
+ 
+ static int message_stats_print(struct mapped_device *md,
+-			       unsigned argc, char **argv, bool clear,
+-			       char *result, unsigned maxlen)
++			       unsigned int argc, char **argv, bool clear,
++			       char *result, unsigned int maxlen)
+ {
+ 	int id;
+ 	char dummy;
+@@ -1187,7 +1187,7 @@ static int message_stats_print(struct mapped_device *md,
+ }
+ 
+ static int message_stats_set_aux(struct mapped_device *md,
+-				 unsigned argc, char **argv)
++				 unsigned int argc, char **argv)
+ {
+ 	int id;
+ 	char dummy;
+@@ -1201,8 +1201,8 @@ static int message_stats_set_aux(struct mapped_device *md,
+ 	return dm_stats_set_aux(dm_get_stats(md), id, argv[2]);
+ }
+ 
+-int dm_stats_message(struct mapped_device *md, unsigned argc, char **argv,
+-		     char *result, unsigned maxlen)
++int dm_stats_message(struct mapped_device *md, unsigned int argc, char **argv,
++		     char *result, unsigned int maxlen)
+ {
+ 	int r;
+ 
+diff --git a/drivers/md/dm-stats.h b/drivers/md/dm-stats.h
+index ee32b099f1cf7..c6728c8b41594 100644
+--- a/drivers/md/dm-stats.h
++++ b/drivers/md/dm-stats.h
+@@ -26,11 +26,11 @@ void dm_stats_cleanup(struct dm_stats *st);
+ 
+ struct mapped_device;
+ 
+-int dm_stats_message(struct mapped_device *md, unsigned argc, char **argv,
+-		     char *result, unsigned maxlen);
++int dm_stats_message(struct mapped_device *md, unsigned int argc, char **argv,
++		     char *result, unsigned int maxlen);
+ 
+ void dm_stats_account_io(struct dm_stats *stats, unsigned long bi_rw,
+-			 sector_t bi_sector, unsigned bi_sectors, bool end,
++			 sector_t bi_sector, unsigned int bi_sectors, bool end,
+ 			 unsigned long start_time,
+ 			 struct dm_stats_aux *aux);
+ 
+diff --git a/drivers/md/dm-stripe.c b/drivers/md/dm-stripe.c
+index baa085cc67bde..a81ed080730a7 100644
+--- a/drivers/md/dm-stripe.c
++++ b/drivers/md/dm-stripe.c
+@@ -273,7 +273,7 @@ static int stripe_map(struct dm_target *ti, struct bio *bio)
+ {
+ 	struct stripe_c *sc = ti->private;
+ 	uint32_t stripe;
+-	unsigned target_bio_nr;
++	unsigned int target_bio_nr;
+ 
+ 	if (bio->bi_opf & REQ_PREFLUSH) {
+ 		target_bio_nr = dm_bio_get_target_bio_nr(bio);
+@@ -359,7 +359,7 @@ static size_t stripe_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff,
+  */
+ 
+ static void stripe_status(struct dm_target *ti, status_type_t type,
+-			  unsigned status_flags, char *result, unsigned maxlen)
++			  unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	struct stripe_c *sc = (struct stripe_c *) ti->private;
+ 	unsigned int sz = 0;
+@@ -406,7 +406,7 @@ static void stripe_status(struct dm_target *ti, status_type_t type,
+ static int stripe_end_io(struct dm_target *ti, struct bio *bio,
+ 		blk_status_t *error)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	char major_minor[16];
+ 	struct stripe_c *sc = ti->private;
+ 
+@@ -444,7 +444,7 @@ static int stripe_iterate_devices(struct dm_target *ti,
+ {
+ 	struct stripe_c *sc = ti->private;
+ 	int ret = 0;
+-	unsigned i = 0;
++	unsigned int i = 0;
+ 
+ 	do {
+ 		ret = fn(ti, sc->stripe[i].dev,
+@@ -459,7 +459,7 @@ static void stripe_io_hints(struct dm_target *ti,
+ 			    struct queue_limits *limits)
+ {
+ 	struct stripe_c *sc = ti->private;
+-	unsigned chunk_size = sc->chunk_size << SECTOR_SHIFT;
++	unsigned int chunk_size = sc->chunk_size << SECTOR_SHIFT;
+ 
+ 	blk_limits_io_min(limits, chunk_size);
+ 	blk_limits_io_opt(limits, chunk_size * sc->stripes);
+diff --git a/drivers/md/dm-switch.c b/drivers/md/dm-switch.c
+index 534dc2ca8bb06..f734b5a097443 100644
+--- a/drivers/md/dm-switch.c
++++ b/drivers/md/dm-switch.c
+@@ -38,9 +38,9 @@ struct switch_path {
+ struct switch_ctx {
+ 	struct dm_target *ti;
+ 
+-	unsigned nr_paths;		/* Number of paths in path_list. */
++	unsigned int nr_paths;		/* Number of paths in path_list. */
+ 
+-	unsigned region_size;		/* Region size in 512-byte sectors */
++	unsigned int region_size;		/* Region size in 512-byte sectors */
+ 	unsigned long nr_regions;	/* Number of regions making up the device */
+ 	signed char region_size_bits;	/* log2 of region_size or -1 */
+ 
+@@ -56,8 +56,8 @@ struct switch_ctx {
+ 	struct switch_path path_list[];
+ };
+ 
+-static struct switch_ctx *alloc_switch_ctx(struct dm_target *ti, unsigned nr_paths,
+-					   unsigned region_size)
++static struct switch_ctx *alloc_switch_ctx(struct dm_target *ti, unsigned int nr_paths,
++					   unsigned int region_size)
+ {
+ 	struct switch_ctx *sctx;
+ 
+@@ -73,7 +73,7 @@ static struct switch_ctx *alloc_switch_ctx(struct dm_target *ti, unsigned nr_pat
+ 	return sctx;
+ }
+ 
+-static int alloc_region_table(struct dm_target *ti, unsigned nr_paths)
++static int alloc_region_table(struct dm_target *ti, unsigned int nr_paths)
+ {
+ 	struct switch_ctx *sctx = ti->private;
+ 	sector_t nr_regions = ti->len;
+@@ -124,7 +124,7 @@ static int alloc_region_table(struct dm_target *ti, unsigned nr_paths)
+ }
+ 
+ static void switch_get_position(struct switch_ctx *sctx, unsigned long region_nr,
+-				unsigned long *region_index, unsigned *bit)
++				unsigned long *region_index, unsigned int *bit)
+ {
+ 	if (sctx->region_entries_per_slot_bits >= 0) {
+ 		*region_index = region_nr >> sctx->region_entries_per_slot_bits;
+@@ -137,10 +137,10 @@ static void switch_get_position(struct switch_ctx *sctx, unsigned long region_nr
+ 	*bit *= sctx->region_table_entry_bits;
+ }
+ 
+-static unsigned switch_region_table_read(struct switch_ctx *sctx, unsigned long region_nr)
++static unsigned int switch_region_table_read(struct switch_ctx *sctx, unsigned long region_nr)
+ {
+ 	unsigned long region_index;
+-	unsigned bit;
++	unsigned int bit;
+ 
+ 	switch_get_position(sctx, region_nr, &region_index, &bit);
+ 
+@@ -151,9 +151,9 @@ static unsigned switch_region_table_read(struct switch_ctx *sctx, unsigned long
+ /*
+  * Find which path to use at given offset.
+  */
+-static unsigned switch_get_path_nr(struct switch_ctx *sctx, sector_t offset)
++static unsigned int switch_get_path_nr(struct switch_ctx *sctx, sector_t offset)
+ {
+-	unsigned path_nr;
++	unsigned int path_nr;
+ 	sector_t p;
+ 
+ 	p = offset;
+@@ -172,10 +172,10 @@ static unsigned switch_get_path_nr(struct switch_ctx *sctx, sector_t offset)
+ }
+ 
+ static void switch_region_table_write(struct switch_ctx *sctx, unsigned long region_nr,
+-				      unsigned value)
++				      unsigned int value)
+ {
+ 	unsigned long region_index;
+-	unsigned bit;
++	unsigned int bit;
+ 	region_table_slot_t pte;
+ 
+ 	switch_get_position(sctx, region_nr, &region_index, &bit);
+@@ -191,7 +191,7 @@ static void switch_region_table_write(struct switch_ctx *sctx, unsigned long reg
+  */
+ static void initialise_region_table(struct switch_ctx *sctx)
+ {
+-	unsigned path_nr = 0;
++	unsigned int path_nr = 0;
+ 	unsigned long region_nr;
+ 
+ 	for (region_nr = 0; region_nr < sctx->nr_regions; region_nr++) {
+@@ -249,7 +249,7 @@ static void switch_dtr(struct dm_target *ti)
+  * Optional args are to allow for future extension: currently this
+  * parameter must be 0.
+  */
+-static int switch_ctr(struct dm_target *ti, unsigned argc, char **argv)
++static int switch_ctr(struct dm_target *ti, unsigned int argc, char **argv)
+ {
+ 	static const struct dm_arg _args[] = {
+ 		{1, (KMALLOC_MAX_SIZE - sizeof(struct switch_ctx)) / sizeof(struct switch_path), "Invalid number of paths"},
+@@ -259,7 +259,7 @@ static int switch_ctr(struct dm_target *ti, unsigned argc, char **argv)
+ 
+ 	struct switch_ctx *sctx;
+ 	struct dm_arg_set as;
+-	unsigned nr_paths, region_size, nr_optional_args;
++	unsigned int nr_paths, region_size, nr_optional_args;
+ 	int r;
+ 
+ 	as.argc = argc;
+@@ -320,7 +320,7 @@ static int switch_map(struct dm_target *ti, struct bio *bio)
+ {
+ 	struct switch_ctx *sctx = ti->private;
+ 	sector_t offset = dm_target_offset(ti, bio->bi_iter.bi_sector);
+-	unsigned path_nr = switch_get_path_nr(sctx, offset);
++	unsigned int path_nr = switch_get_path_nr(sctx, offset);
+ 
+ 	bio_set_dev(bio, sctx->path_list[path_nr].dmdev->bdev);
+ 	bio->bi_iter.bi_sector = sctx->path_list[path_nr].start + offset;
+@@ -371,9 +371,9 @@ static __always_inline unsigned long parse_hex(const char **string)
+ }
+ 
+ static int process_set_region_mappings(struct switch_ctx *sctx,
+-				       unsigned argc, char **argv)
++				       unsigned int argc, char **argv)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	unsigned long region_index = 0;
+ 
+ 	for (i = 1; i < argc; i++) {
+@@ -466,8 +466,8 @@ static int process_set_region_mappings(struct switch_ctx *sctx,
+  *
+  * Only set_region_mappings is supported.
+  */
+-static int switch_message(struct dm_target *ti, unsigned argc, char **argv,
+-			  char *result, unsigned maxlen)
++static int switch_message(struct dm_target *ti, unsigned int argc, char **argv,
++			  char *result, unsigned int maxlen)
+ {
+ 	static DEFINE_MUTEX(message_mutex);
+ 
+@@ -487,10 +487,10 @@ static int switch_message(struct dm_target *ti, unsigned argc, char **argv,
+ }
+ 
+ static void switch_status(struct dm_target *ti, status_type_t type,
+-			  unsigned status_flags, char *result, unsigned maxlen)
++			  unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	struct switch_ctx *sctx = ti->private;
+-	unsigned sz = 0;
++	unsigned int sz = 0;
+ 	int path_nr;
+ 
+ 	switch (type) {
+@@ -519,7 +519,7 @@ static void switch_status(struct dm_target *ti, status_type_t type,
+ static int switch_prepare_ioctl(struct dm_target *ti, struct block_device **bdev)
+ {
+ 	struct switch_ctx *sctx = ti->private;
+-	unsigned path_nr;
++	unsigned int path_nr;
+ 
+ 	path_nr = switch_get_path_nr(sctx, 0);
+ 
+diff --git a/drivers/md/dm-table.c b/drivers/md/dm-table.c
+index 078da18bb86d8..32b2d3b99d786 100644
+--- a/drivers/md/dm-table.c
++++ b/drivers/md/dm-table.c
+@@ -126,7 +126,7 @@ static int alloc_targets(struct dm_table *t, unsigned int num)
+ }
+ 
+ int dm_table_create(struct dm_table **result, fmode_t mode,
+-		    unsigned num_targets, struct mapped_device *md)
++		    unsigned int num_targets, struct mapped_device *md)
+ {
+ 	struct dm_table *t = kzalloc(sizeof(*t), GFP_KERNEL);
+ 
+@@ -470,10 +470,10 @@ static int adjoin(struct dm_table *t, struct dm_target *ti)
+  * On the other hand, dm-switch needs to process bulk data using messages and
+  * excessive use of GFP_NOIO could cause trouble.
+  */
+-static char **realloc_argv(unsigned *size, char **old_argv)
++static char **realloc_argv(unsigned int *size, char **old_argv)
+ {
+ 	char **argv;
+-	unsigned new_size;
++	unsigned int new_size;
+ 	gfp_t gfp;
+ 
+ 	if (*size) {
+@@ -499,7 +499,7 @@ static char **realloc_argv(unsigned *size, char **old_argv)
+ int dm_split_args(int *argc, char ***argvp, char *input)
+ {
+ 	char *start, *end = input, *out, **argv = NULL;
+-	unsigned array_size = 0;
++	unsigned int array_size = 0;
+ 
+ 	*argc = 0;
+ 
+@@ -732,9 +732,8 @@ int dm_table_add_target(struct dm_table *t, const char *type,
+ /*
+  * Target argument parsing helpers.
+  */
+-static int validate_next_arg(const struct dm_arg *arg,
+-			     struct dm_arg_set *arg_set,
+-			     unsigned *value, char **error, unsigned grouped)
++static int validate_next_arg(const struct dm_arg *arg, struct dm_arg_set *arg_set,
++			     unsigned int *value, char **error, unsigned int grouped)
+ {
+ 	const char *arg_str = dm_shift_arg(arg_set);
+ 	char dummy;
+@@ -752,14 +751,14 @@ static int validate_next_arg(const struct dm_arg *arg,
+ }
+ 
+ int dm_read_arg(const struct dm_arg *arg, struct dm_arg_set *arg_set,
+-		unsigned *value, char **error)
++		unsigned int *value, char **error)
+ {
+ 	return validate_next_arg(arg, arg_set, value, error, 0);
+ }
+ EXPORT_SYMBOL(dm_read_arg);
+ 
+ int dm_read_arg_group(const struct dm_arg *arg, struct dm_arg_set *arg_set,
+-		      unsigned *value, char **error)
++		      unsigned int *value, char **error)
+ {
+ 	return validate_next_arg(arg, arg_set, value, error, 1);
+ }
+@@ -780,7 +779,7 @@ const char *dm_shift_arg(struct dm_arg_set *as)
+ }
+ EXPORT_SYMBOL(dm_shift_arg);
+ 
+-void dm_consume_args(struct dm_arg_set *as, unsigned num_args)
++void dm_consume_args(struct dm_arg_set *as, unsigned int num_args)
+ {
+ 	BUG_ON(as->argc < num_args);
+ 	as->argc -= num_args;
+@@ -856,7 +855,7 @@ static int device_is_rq_stackable(struct dm_target *ti, struct dm_dev *dev,
+ 
+ static int dm_table_determine_type(struct dm_table *t)
+ {
+-	unsigned bio_based = 0, request_based = 0, hybrid = 0;
++	unsigned int bio_based = 0, request_based = 0, hybrid = 0;
+ 	struct dm_target *ti;
+ 	struct list_head *devices = dm_table_get_devices(t);
+ 	enum dm_queue_mode live_md_type = dm_get_md_type(t->md);
+@@ -1535,7 +1534,7 @@ static bool dm_table_any_dev_attr(struct dm_table *t,
+ static int count_device(struct dm_target *ti, struct dm_dev *dev,
+ 			sector_t start, sector_t len, void *data)
+ {
+-	unsigned *num_devices = data;
++	unsigned int *num_devices = data;
+ 
+ 	(*num_devices)++;
+ 
+@@ -1565,7 +1564,7 @@ bool dm_table_has_no_data_devices(struct dm_table *t)
+ {
+ 	for (unsigned int i = 0; i < t->num_targets; i++) {
+ 		struct dm_target *ti = dm_table_get_target(t, i);
+-		unsigned num_devices = 0;
++		unsigned int num_devices = 0;
+ 
+ 		if (!ti->type->iterate_devices)
+ 			return false;
+diff --git a/drivers/md/dm-thin-metadata.c b/drivers/md/dm-thin-metadata.c
+index 6bcc4c4786d89..80545ec541210 100644
+--- a/drivers/md/dm-thin-metadata.c
++++ b/drivers/md/dm-thin-metadata.c
+@@ -318,12 +318,12 @@ static void unpack_block_time(uint64_t v, dm_block_t *b, uint32_t *t)
+  */
+ typedef int (*run_fn)(struct dm_space_map *, dm_block_t, dm_block_t);
+ 
+-static void with_runs(struct dm_space_map *sm, const __le64 *value_le, unsigned count, run_fn fn)
++static void with_runs(struct dm_space_map *sm, const __le64 *value_le, unsigned int count, run_fn fn)
+ {
+ 	uint64_t b, begin, end;
+ 	uint32_t t;
+ 	bool in_run = false;
+-	unsigned i;
++	unsigned int i;
+ 
+ 	for (i = 0; i < count; i++, value_le++) {
+ 		/* We know value_le is 8 byte aligned */
+@@ -348,13 +348,13 @@ static void with_runs(struct dm_space_map *sm, const __le64 *value_le, unsigned
+ 		fn(sm, begin, end);
+ }
+ 
+-static void data_block_inc(void *context, const void *value_le, unsigned count)
++static void data_block_inc(void *context, const void *value_le, unsigned int count)
+ {
+ 	with_runs((struct dm_space_map *) context,
+ 		  (const __le64 *) value_le, count, dm_sm_inc_blocks);
+ }
+ 
+-static void data_block_dec(void *context, const void *value_le, unsigned count)
++static void data_block_dec(void *context, const void *value_le, unsigned int count)
+ {
+ 	with_runs((struct dm_space_map *) context,
+ 		  (const __le64 *) value_le, count, dm_sm_dec_blocks);
+@@ -374,21 +374,21 @@ static int data_block_equal(void *context, const void *value1_le, const void *va
+ 	return b1 == b2;
+ }
+ 
+-static void subtree_inc(void *context, const void *value, unsigned count)
++static void subtree_inc(void *context, const void *value, unsigned int count)
+ {
+ 	struct dm_btree_info *info = context;
+ 	const __le64 *root_le = value;
+-	unsigned i;
++	unsigned int i;
+ 
+ 	for (i = 0; i < count; i++, root_le++)
+ 		dm_tm_inc(info->tm, le64_to_cpu(*root_le));
+ }
+ 
+-static void subtree_dec(void *context, const void *value, unsigned count)
++static void subtree_dec(void *context, const void *value, unsigned int count)
+ {
+ 	struct dm_btree_info *info = context;
+ 	const __le64 *root_le = value;
+-	unsigned i;
++	unsigned int i;
+ 
+ 	for (i = 0; i < count; i++, root_le++)
+ 		if (dm_btree_del(info, le64_to_cpu(*root_le)))
+@@ -448,10 +448,10 @@ static int superblock_lock(struct dm_pool_metadata *pmd,
+ static int __superblock_all_zeroes(struct dm_block_manager *bm, int *result)
+ {
+ 	int r;
+-	unsigned i;
++	unsigned int i;
+ 	struct dm_block *b;
+ 	__le64 *data_le, zero = cpu_to_le64(0);
+-	unsigned block_size = dm_bm_block_size(bm) / sizeof(__le64);
++	unsigned int block_size = dm_bm_block_size(bm) / sizeof(__le64);
+ 
+ 	/*
+ 	 * We can't use a validator here - it may be all zeroes.
+@@ -971,7 +971,7 @@ struct dm_pool_metadata *dm_pool_metadata_open(struct block_device *bdev,
+ int dm_pool_metadata_close(struct dm_pool_metadata *pmd)
+ {
+ 	int r;
+-	unsigned open_devices = 0;
++	unsigned int open_devices = 0;
+ 	struct dm_thin_device *td, *tmp;
+ 
+ 	down_read(&pmd->root_lock);
+@@ -1679,7 +1679,7 @@ int dm_thin_insert_block(struct dm_thin_device *td, dm_block_t block,
+ static int __remove_range(struct dm_thin_device *td, dm_block_t begin, dm_block_t end)
+ {
+ 	int r;
+-	unsigned count, total_count = 0;
++	unsigned int count, total_count = 0;
+ 	struct dm_pool_metadata *pmd = td->pmd;
+ 	dm_block_t keys[1] = { td->id };
+ 	__le64 value;
+diff --git a/drivers/md/dm-thin.c b/drivers/md/dm-thin.c
+index 90df6c3da77a2..d12aff50974a9 100644
+--- a/drivers/md/dm-thin.c
++++ b/drivers/md/dm-thin.c
+@@ -32,7 +32,7 @@
+ #define COMMIT_PERIOD HZ
+ #define NO_SPACE_TIMEOUT_SECS 60
+ 
+-static unsigned no_space_timeout_secs = NO_SPACE_TIMEOUT_SECS;
++static unsigned int no_space_timeout_secs = NO_SPACE_TIMEOUT_SECS;
+ 
+ DECLARE_DM_KCOPYD_THROTTLE_WITH_MODULE_PARM(snapshot_copy_throttle,
+ 		"A percentage of time allocated for copy on write");
+@@ -254,7 +254,7 @@ struct pool {
+ 	struct delayed_work no_space_timeout;
+ 
+ 	unsigned long last_commit_jiffies;
+-	unsigned ref_count;
++	unsigned int ref_count;
+ 
+ 	spinlock_t lock;
+ 	struct bio_list deferred_flush_bios;
+@@ -2159,7 +2159,7 @@ static void process_thin_deferred_bios(struct thin_c *tc)
+ 	struct bio *bio;
+ 	struct bio_list bios;
+ 	struct blk_plug plug;
+-	unsigned count = 0;
++	unsigned int count = 0;
+ 
+ 	if (tc->requeue_mode) {
+ 		error_thin_bio_list(tc, &tc->deferred_bio_list,
+@@ -2229,9 +2229,9 @@ static int cmp_cells(const void *lhs, const void *rhs)
+ 	return 0;
+ }
+ 
+-static unsigned sort_cells(struct pool *pool, struct list_head *cells)
++static unsigned int sort_cells(struct pool *pool, struct list_head *cells)
+ {
+-	unsigned count = 0;
++	unsigned int count = 0;
+ 	struct dm_bio_prison_cell *cell, *tmp;
+ 
+ 	list_for_each_entry_safe(cell, tmp, cells, user_list) {
+@@ -2252,7 +2252,7 @@ static void process_thin_deferred_cells(struct thin_c *tc)
+ 	struct pool *pool = tc->pool;
+ 	struct list_head cells;
+ 	struct dm_bio_prison_cell *cell;
+-	unsigned i, j, count;
++	unsigned int i, j, count;
+ 
+ 	INIT_LIST_HEAD(&cells);
+ 
+@@ -3115,7 +3115,7 @@ static int parse_pool_features(struct dm_arg_set *as, struct pool_features *pf,
+ 			       struct dm_target *ti)
+ {
+ 	int r;
+-	unsigned argc;
++	unsigned int argc;
+ 	const char *arg_name;
+ 
+ 	static const struct dm_arg _args[] = {
+@@ -3252,7 +3252,7 @@ static dm_block_t calc_metadata_threshold(struct pool_c *pt)
+  *	     read_only: Don't allow any changes to be made to the pool metadata.
+  *	     error_if_no_space: error IOs, instead of queueing, if no space.
+  */
+-static int pool_ctr(struct dm_target *ti, unsigned argc, char **argv)
++static int pool_ctr(struct dm_target *ti, unsigned int argc, char **argv)
+ {
+ 	int r, pool_created = 0;
+ 	struct pool_c *pt;
+@@ -3648,7 +3648,7 @@ static void pool_postsuspend(struct dm_target *ti)
+ 	(void) commit(pool);
+ }
+ 
+-static int check_arg_count(unsigned argc, unsigned args_required)
++static int check_arg_count(unsigned int argc, unsigned int args_required)
+ {
+ 	if (argc != args_required) {
+ 		DMWARN("Message received with %u arguments instead of %u.",
+@@ -3671,7 +3671,7 @@ static int read_dev_id(char *arg, dm_thin_id *dev_id, int warning)
+ 	return -EINVAL;
+ }
+ 
+-static int process_create_thin_mesg(unsigned argc, char **argv, struct pool *pool)
++static int process_create_thin_mesg(unsigned int argc, char **argv, struct pool *pool)
+ {
+ 	dm_thin_id dev_id;
+ 	int r;
+@@ -3694,7 +3694,7 @@ static int process_create_thin_mesg(unsigned argc, char **argv, struct pool *poo
+ 	return 0;
+ }
+ 
+-static int process_create_snap_mesg(unsigned argc, char **argv, struct pool *pool)
++static int process_create_snap_mesg(unsigned int argc, char **argv, struct pool *pool)
+ {
+ 	dm_thin_id dev_id;
+ 	dm_thin_id origin_dev_id;
+@@ -3722,7 +3722,7 @@ static int process_create_snap_mesg(unsigned argc, char **argv, struct pool *poo
+ 	return 0;
+ }
+ 
+-static int process_delete_mesg(unsigned argc, char **argv, struct pool *pool)
++static int process_delete_mesg(unsigned int argc, char **argv, struct pool *pool)
+ {
+ 	dm_thin_id dev_id;
+ 	int r;
+@@ -3742,7 +3742,7 @@ static int process_delete_mesg(unsigned argc, char **argv, struct pool *pool)
+ 	return r;
+ }
+ 
+-static int process_set_transaction_id_mesg(unsigned argc, char **argv, struct pool *pool)
++static int process_set_transaction_id_mesg(unsigned int argc, char **argv, struct pool *pool)
+ {
+ 	dm_thin_id old_id, new_id;
+ 	int r;
+@@ -3771,7 +3771,7 @@ static int process_set_transaction_id_mesg(unsigned argc, char **argv, struct po
+ 	return 0;
+ }
+ 
+-static int process_reserve_metadata_snap_mesg(unsigned argc, char **argv, struct pool *pool)
++static int process_reserve_metadata_snap_mesg(unsigned int argc, char **argv, struct pool *pool)
+ {
+ 	int r;
+ 
+@@ -3788,7 +3788,7 @@ static int process_reserve_metadata_snap_mesg(unsigned argc, char **argv, struct
+ 	return r;
+ }
+ 
+-static int process_release_metadata_snap_mesg(unsigned argc, char **argv, struct pool *pool)
++static int process_release_metadata_snap_mesg(unsigned int argc, char **argv, struct pool *pool)
+ {
+ 	int r;
+ 
+@@ -3812,8 +3812,8 @@ static int process_release_metadata_snap_mesg(unsigned argc, char **argv, struct
+  *   reserve_metadata_snap
+  *   release_metadata_snap
+  */
+-static int pool_message(struct dm_target *ti, unsigned argc, char **argv,
+-			char *result, unsigned maxlen)
++static int pool_message(struct dm_target *ti, unsigned int argc, char **argv,
++			char *result, unsigned int maxlen)
+ {
+ 	int r = -EINVAL;
+ 	struct pool_c *pt = ti->private;
+@@ -3853,9 +3853,9 @@ static int pool_message(struct dm_target *ti, unsigned argc, char **argv,
+ }
+ 
+ static void emit_flags(struct pool_features *pf, char *result,
+-		       unsigned sz, unsigned maxlen)
++		       unsigned int sz, unsigned int maxlen)
+ {
+-	unsigned count = !pf->zero_new_blocks + !pf->discard_enabled +
++	unsigned int count = !pf->zero_new_blocks + !pf->discard_enabled +
+ 		!pf->discard_passdown + (pf->mode == PM_READ_ONLY) +
+ 		pf->error_if_no_space;
+ 	DMEMIT("%u ", count);
+@@ -3883,10 +3883,10 @@ static void emit_flags(struct pool_features *pf, char *result,
+  *    <pool mode> <discard config> <no space config> <needs_check>
+  */
+ static void pool_status(struct dm_target *ti, status_type_t type,
+-			unsigned status_flags, char *result, unsigned maxlen)
++			unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	int r;
+-	unsigned sz = 0;
++	unsigned int sz = 0;
+ 	uint64_t transaction_id;
+ 	dm_block_t nr_free_blocks_data;
+ 	dm_block_t nr_free_blocks_metadata;
+@@ -4148,7 +4148,7 @@ static void thin_dtr(struct dm_target *ti)
+  * If the pool device has discards disabled, they get disabled for the thin
+  * device as well.
+  */
+-static int thin_ctr(struct dm_target *ti, unsigned argc, char **argv)
++static int thin_ctr(struct dm_target *ti, unsigned int argc, char **argv)
+ {
+ 	int r;
+ 	struct thin_c *tc;
+@@ -4371,7 +4371,7 @@ static int thin_preresume(struct dm_target *ti)
+  * <nr mapped sectors> <highest mapped sector>
+  */
+ static void thin_status(struct dm_target *ti, status_type_t type,
+-			unsigned status_flags, char *result, unsigned maxlen)
++			unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	int r;
+ 	ssize_t sz = 0;
+diff --git a/drivers/md/dm-uevent.c b/drivers/md/dm-uevent.c
+index 8671267200d88..a02b3f6ea47a8 100644
+--- a/drivers/md/dm-uevent.c
++++ b/drivers/md/dm-uevent.c
+@@ -60,7 +60,7 @@ static struct dm_uevent *dm_build_path_uevent(struct mapped_device *md,
+ 					      enum kobject_action action,
+ 					      const char *dm_action,
+ 					      const char *path,
+-					      unsigned nr_valid_paths)
++					      unsigned int nr_valid_paths)
+ {
+ 	struct dm_uevent *event;
+ 
+@@ -168,7 +168,7 @@ EXPORT_SYMBOL_GPL(dm_send_uevents);
+  *
+  */
+ void dm_path_uevent(enum dm_uevent_type event_type, struct dm_target *ti,
+-		   const char *path, unsigned nr_valid_paths)
++		   const char *path, unsigned int nr_valid_paths)
+ {
+ 	struct mapped_device *md = dm_table_get_md(ti->table);
+ 	struct dm_uevent *event;
+diff --git a/drivers/md/dm-uevent.h b/drivers/md/dm-uevent.h
+index d30d226f2a181..2c9ba561fd8e9 100644
+--- a/drivers/md/dm-uevent.h
++++ b/drivers/md/dm-uevent.h
+@@ -20,7 +20,7 @@ extern void dm_uevent_exit(void);
+ extern void dm_send_uevents(struct list_head *events, struct kobject *kobj);
+ extern void dm_path_uevent(enum dm_uevent_type event_type,
+ 			   struct dm_target *ti, const char *path,
+-			   unsigned nr_valid_paths);
++			   unsigned int nr_valid_paths);
+ 
+ #else
+ 
+@@ -37,7 +37,7 @@ static inline void dm_send_uevents(struct list_head *events,
+ }
+ static inline void dm_path_uevent(enum dm_uevent_type event_type,
+ 				  struct dm_target *ti, const char *path,
+-				  unsigned nr_valid_paths)
++				  unsigned int nr_valid_paths)
+ {
+ }
+ 
+diff --git a/drivers/md/dm-verity-fec.c b/drivers/md/dm-verity-fec.c
+index 23cffce564035..962fc32c947c5 100644
+--- a/drivers/md/dm-verity-fec.c
++++ b/drivers/md/dm-verity-fec.c
+@@ -59,14 +59,14 @@ static int fec_decode_rs8(struct dm_verity *v, struct dm_verity_fec_io *fio,
+  * to the data block. Caller is responsible for releasing buf.
+  */
+ static u8 *fec_read_parity(struct dm_verity *v, u64 rsb, int index,
+-			   unsigned *offset, struct dm_buffer **buf)
++			   unsigned int *offset, struct dm_buffer **buf)
+ {
+ 	u64 position, block, rem;
+ 	u8 *res;
+ 
+ 	position = (index + rsb) * v->fec->roots;
+ 	block = div64_u64_rem(position, v->fec->io_size, &rem);
+-	*offset = (unsigned)rem;
++	*offset = (unsigned int)rem;
+ 
+ 	res = dm_bufio_read(v->fec->bufio, block, buf);
+ 	if (IS_ERR(res)) {
+@@ -102,7 +102,7 @@ static u8 *fec_read_parity(struct dm_verity *v, u64 rsb, int index,
+  */
+ static inline u8 *fec_buffer_rs_block(struct dm_verity *v,
+ 				      struct dm_verity_fec_io *fio,
+-				      unsigned i, unsigned j)
++				      unsigned int i, unsigned int j)
+ {
+ 	return &fio->bufs[i][j * v->fec->rsn];
+ }
+@@ -111,7 +111,7 @@ static inline u8 *fec_buffer_rs_block(struct dm_verity *v,
+  * Return an index to the current RS block when called inside
+  * fec_for_each_buffer_rs_block.
+  */
+-static inline unsigned fec_buffer_rs_index(unsigned i, unsigned j)
++static inline unsigned int fec_buffer_rs_index(unsigned int i, unsigned int j)
+ {
+ 	return (i << DM_VERITY_FEC_BUF_RS_BITS) + j;
+ }
+@@ -121,12 +121,12 @@ static inline unsigned fec_buffer_rs_index(unsigned i, unsigned j)
+  * starting from block_offset.
+  */
+ static int fec_decode_bufs(struct dm_verity *v, struct dm_verity_fec_io *fio,
+-			   u64 rsb, int byte_index, unsigned block_offset,
++			   u64 rsb, int byte_index, unsigned int block_offset,
+ 			   int neras)
+ {
+ 	int r, corrected = 0, res;
+ 	struct dm_buffer *buf;
+-	unsigned n, i, offset;
++	unsigned int n, i, offset;
+ 	u8 *par, *block;
+ 
+ 	par = fec_read_parity(v, rsb, block_offset, &offset, &buf);
+@@ -197,7 +197,7 @@ static int fec_is_erasure(struct dm_verity *v, struct dm_verity_io *io,
+  * fits into buffers. Check for erasure locations if @neras is non-NULL.
+  */
+ static int fec_read_bufs(struct dm_verity *v, struct dm_verity_io *io,
+-			 u64 rsb, u64 target, unsigned block_offset,
++			 u64 rsb, u64 target, unsigned int block_offset,
+ 			 int *neras)
+ {
+ 	bool is_zero;
+@@ -208,7 +208,7 @@ static int fec_read_bufs(struct dm_verity *v, struct dm_verity_io *io,
+ 	u64 block, ileaved;
+ 	u8 *bbuf, *rs_block;
+ 	u8 want_digest[HASH_MAX_DIGESTSIZE];
+-	unsigned n, k;
++	unsigned int n, k;
+ 
+ 	if (neras)
+ 		*neras = 0;
+@@ -304,7 +304,7 @@ done:
+  */
+ static int fec_alloc_bufs(struct dm_verity *v, struct dm_verity_fec_io *fio)
+ {
+-	unsigned n;
++	unsigned int n;
+ 
+ 	if (!fio->rs)
+ 		fio->rs = mempool_alloc(&v->fec->rs_pool, GFP_NOIO);
+@@ -344,7 +344,7 @@ static int fec_alloc_bufs(struct dm_verity *v, struct dm_verity_fec_io *fio)
+  */
+ static void fec_init_bufs(struct dm_verity *v, struct dm_verity_fec_io *fio)
+ {
+-	unsigned n;
++	unsigned int n;
+ 
+ 	fec_for_each_buffer(fio, n)
+ 		memset(fio->bufs[n], 0, v->fec->rsn << DM_VERITY_FEC_BUF_RS_BITS);
+@@ -362,7 +362,7 @@ static int fec_decode_rsb(struct dm_verity *v, struct dm_verity_io *io,
+ 			  bool use_erasures)
+ {
+ 	int r, neras = 0;
+-	unsigned pos;
++	unsigned int pos;
+ 
+ 	r = fec_alloc_bufs(v, fio);
+ 	if (unlikely(r < 0))
+@@ -484,7 +484,7 @@ done:
+  */
+ void verity_fec_finish_io(struct dm_verity_io *io)
+ {
+-	unsigned n;
++	unsigned int n;
+ 	struct dm_verity_fec *f = io->v->fec;
+ 	struct dm_verity_fec_io *fio = fec_io(io);
+ 
+@@ -522,8 +522,8 @@ void verity_fec_init_io(struct dm_verity_io *io)
+ /*
+  * Append feature arguments and values to the status table.
+  */
+-unsigned verity_fec_status_table(struct dm_verity *v, unsigned sz,
+-				 char *result, unsigned maxlen)
++unsigned int verity_fec_status_table(struct dm_verity *v, unsigned int sz,
++				 char *result, unsigned int maxlen)
+ {
+ 	if (!verity_fec_is_enabled(v))
+ 		return sz;
+@@ -589,7 +589,7 @@ bool verity_is_fec_opt_arg(const char *arg_name)
+ }
+ 
+ int verity_fec_parse_opt_args(struct dm_arg_set *as, struct dm_verity *v,
+-			      unsigned *argc, const char *arg_name)
++			      unsigned int *argc, const char *arg_name)
+ {
+ 	int r;
+ 	struct dm_target *ti = v->ti;
+diff --git a/drivers/md/dm-verity-fec.h b/drivers/md/dm-verity-fec.h
+index 3c46c8d618833..8454070d28242 100644
+--- a/drivers/md/dm-verity-fec.h
++++ b/drivers/md/dm-verity-fec.h
+@@ -55,10 +55,10 @@ struct dm_verity_fec_io {
+ 	struct rs_control *rs;	/* Reed-Solomon state */
+ 	int erasures[DM_VERITY_FEC_MAX_RSN];	/* erasures for decode_rs8 */
+ 	u8 *bufs[DM_VERITY_FEC_BUF_MAX];	/* bufs for deinterleaving */
+-	unsigned nbufs;		/* number of buffers allocated */
++	unsigned int nbufs;		/* number of buffers allocated */
+ 	u8 *output;		/* buffer for corrected output */
+ 	size_t output_pos;
+-	unsigned level;		/* recursion level */
++	unsigned int level;		/* recursion level */
+ };
+ 
+ #ifdef CONFIG_DM_VERITY_FEC
+@@ -72,15 +72,15 @@ extern int verity_fec_decode(struct dm_verity *v, struct dm_verity_io *io,
+ 			     enum verity_block_type type, sector_t block,
+ 			     u8 *dest, struct bvec_iter *iter);
+ 
+-extern unsigned verity_fec_status_table(struct dm_verity *v, unsigned sz,
+-					char *result, unsigned maxlen);
++extern unsigned int verity_fec_status_table(struct dm_verity *v, unsigned int sz,
++					char *result, unsigned int maxlen);
+ 
+ extern void verity_fec_finish_io(struct dm_verity_io *io);
+ extern void verity_fec_init_io(struct dm_verity_io *io);
+ 
+ extern bool verity_is_fec_opt_arg(const char *arg_name);
+ extern int verity_fec_parse_opt_args(struct dm_arg_set *as,
+-				     struct dm_verity *v, unsigned *argc,
++				     struct dm_verity *v, unsigned int *argc,
+ 				     const char *arg_name);
+ 
+ extern void verity_fec_dtr(struct dm_verity *v);
+@@ -106,9 +106,9 @@ static inline int verity_fec_decode(struct dm_verity *v,
+ 	return -EOPNOTSUPP;
+ }
+ 
+-static inline unsigned verity_fec_status_table(struct dm_verity *v,
+-					       unsigned sz, char *result,
+-					       unsigned maxlen)
++static inline unsigned int verity_fec_status_table(struct dm_verity *v,
++					       unsigned int sz, char *result,
++					       unsigned int maxlen)
+ {
+ 	return sz;
+ }
+@@ -128,7 +128,7 @@ static inline bool verity_is_fec_opt_arg(const char *arg_name)
+ 
+ static inline int verity_fec_parse_opt_args(struct dm_arg_set *as,
+ 					    struct dm_verity *v,
+-					    unsigned *argc,
++					    unsigned int *argc,
+ 					    const char *arg_name)
+ {
+ 	return -EINVAL;
+diff --git a/drivers/md/dm-verity-target.c b/drivers/md/dm-verity-target.c
+index ccf5b852fbf7a..64e8ac429984d 100644
+--- a/drivers/md/dm-verity-target.c
++++ b/drivers/md/dm-verity-target.c
+@@ -41,7 +41,7 @@
+ #define DM_VERITY_OPTS_MAX		(4 + DM_VERITY_OPTS_FEC + \
+ 					 DM_VERITY_ROOT_HASH_VERIFICATION_OPTS)
+ 
+-static unsigned dm_verity_prefetch_cluster = DM_VERITY_DEFAULT_PREFETCH_SIZE;
++static unsigned int dm_verity_prefetch_cluster = DM_VERITY_DEFAULT_PREFETCH_SIZE;
+ 
+ module_param_named(prefetch_cluster, dm_verity_prefetch_cluster, uint, S_IRUGO | S_IWUSR);
+ 
+@@ -51,7 +51,7 @@ struct dm_verity_prefetch_work {
+ 	struct work_struct work;
+ 	struct dm_verity *v;
+ 	sector_t block;
+-	unsigned n_blocks;
++	unsigned int n_blocks;
+ };
+ 
+ /*
+@@ -196,10 +196,10 @@ out:
+ }
+ 
+ static void verity_hash_at_level(struct dm_verity *v, sector_t block, int level,
+-				 sector_t *hash_block, unsigned *offset)
++				 sector_t *hash_block, unsigned int *offset)
+ {
+ 	sector_t position = verity_position_at_level(v, block, level);
+-	unsigned idx;
++	unsigned int idx;
+ 
+ 	*hash_block = v->hash_level_block[level] + (position >> v->hash_per_block_bits);
+ 
+@@ -287,7 +287,7 @@ static int verity_verify_level(struct dm_verity *v, struct dm_verity_io *io,
+ 	u8 *data;
+ 	int r;
+ 	sector_t hash_block;
+-	unsigned offset;
++	unsigned int offset;
+ 
+ 	verity_hash_at_level(v, block, level, &hash_block, &offset);
+ 
+@@ -445,13 +445,13 @@ int verity_for_bv_block(struct dm_verity *v, struct dm_verity_io *io,
+ 				       struct dm_verity_io *io, u8 *data,
+ 				       size_t len))
+ {
+-	unsigned todo = 1 << v->data_dev_block_bits;
++	unsigned int todo = 1 << v->data_dev_block_bits;
+ 	struct bio *bio = dm_bio_from_per_bio_data(io, v->ti->per_io_data_size);
+ 
+ 	do {
+ 		int r;
+ 		u8 *page;
+-		unsigned len;
++		unsigned int len;
+ 		struct bio_vec bv = bio_iter_iovec(bio, *iter);
+ 
+ 		page = bvec_kmap_local(&bv);
+@@ -688,7 +688,7 @@ static void verity_prefetch_io(struct work_struct *work)
+ 		verity_hash_at_level(v, pw->block, i, &hash_block_start, NULL);
+ 		verity_hash_at_level(v, pw->block + pw->n_blocks - 1, i, &hash_block_end, NULL);
+ 		if (!i) {
+-			unsigned cluster = READ_ONCE(dm_verity_prefetch_cluster);
++			unsigned int cluster = READ_ONCE(dm_verity_prefetch_cluster);
+ 
+ 			cluster >>= v->data_dev_block_bits;
+ 			if (unlikely(!cluster))
+@@ -753,7 +753,7 @@ static int verity_map(struct dm_target *ti, struct bio *bio)
+ 	bio_set_dev(bio, v->data_dev->bdev);
+ 	bio->bi_iter.bi_sector = verity_map_sector(v, bio->bi_iter.bi_sector);
+ 
+-	if (((unsigned)bio->bi_iter.bi_sector | bio_sectors(bio)) &
++	if (((unsigned int)bio->bi_iter.bi_sector | bio_sectors(bio)) &
+ 	    ((1 << (v->data_dev_block_bits - SECTOR_SHIFT)) - 1)) {
+ 		DMERR_LIMIT("unaligned io");
+ 		return DM_MAPIO_KILL;
+@@ -789,12 +789,12 @@ static int verity_map(struct dm_target *ti, struct bio *bio)
+  * Status: V (valid) or C (corruption found)
+  */
+ static void verity_status(struct dm_target *ti, status_type_t type,
+-			  unsigned status_flags, char *result, unsigned maxlen)
++			  unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	struct dm_verity *v = ti->private;
+-	unsigned args = 0;
+-	unsigned sz = 0;
+-	unsigned x;
++	unsigned int args = 0;
++	unsigned int sz = 0;
++	unsigned int x;
+ 
+ 	switch (type) {
+ 	case STATUSTYPE_INFO:
+@@ -1054,7 +1054,7 @@ static int verity_parse_opt_args(struct dm_arg_set *as, struct dm_verity *v,
+ 				 bool only_modifier_opts)
+ {
+ 	int r = 0;
+-	unsigned argc;
++	unsigned int argc;
+ 	struct dm_target *ti = v->ti;
+ 	const char *arg_name;
+ 
+@@ -1156,7 +1156,7 @@ static int verity_parse_opt_args(struct dm_arg_set *as, struct dm_verity *v,
+  *	<digest>
+  *	<salt>		Hex string or "-" if no salt.
+  */
+-static int verity_ctr(struct dm_target *ti, unsigned argc, char **argv)
++static int verity_ctr(struct dm_target *ti, unsigned int argc, char **argv)
+ {
+ 	struct dm_verity *v;
+ 	struct dm_verity_sig_opts verify_args = {0};
+diff --git a/drivers/md/dm-verity.h b/drivers/md/dm-verity.h
+index 98f306ec6a33d..2f555b4203679 100644
+--- a/drivers/md/dm-verity.h
++++ b/drivers/md/dm-verity.h
+@@ -42,7 +42,7 @@ struct dm_verity {
+ 	u8 *root_digest;	/* digest of the root block */
+ 	u8 *salt;		/* salt: its size is salt_size */
+ 	u8 *zero_digest;	/* digest for a zero block */
+-	unsigned salt_size;
++	unsigned int salt_size;
+ 	sector_t data_start;	/* data offset in 512-byte sectors */
+ 	sector_t hash_start;	/* hash start in blocks */
+ 	sector_t data_blocks;	/* the number of data blocks */
+@@ -54,10 +54,10 @@ struct dm_verity {
+ 	unsigned char version;
+ 	bool hash_failed:1;	/* set if hash of any block failed */
+ 	bool use_tasklet:1;	/* try to verify in tasklet before work-queue */
+-	unsigned digest_size;	/* digest size for the current hash algorithm */
++	unsigned int digest_size;	/* digest size for the current hash algorithm */
+ 	unsigned int ahash_reqsize;/* the size of temporary space for crypto */
+ 	enum verity_mode mode;	/* mode for handling verification errors */
+-	unsigned corrupted_errs;/* Number of errors for corrupted blocks */
++	unsigned int corrupted_errs;/* Number of errors for corrupted blocks */
+ 
+ 	struct workqueue_struct *verify_wq;
+ 
+@@ -77,7 +77,7 @@ struct dm_verity_io {
+ 	bio_end_io_t *orig_bi_end_io;
+ 
+ 	sector_t block;
+-	unsigned n_blocks;
++	unsigned int n_blocks;
+ 	bool in_tasklet;
+ 
+ 	struct bvec_iter iter;
+diff --git a/drivers/md/dm-writecache.c b/drivers/md/dm-writecache.c
+index 96a003eb73234..431c84595ddb7 100644
+--- a/drivers/md/dm-writecache.c
++++ b/drivers/md/dm-writecache.c
+@@ -128,9 +128,9 @@ struct dm_writecache {
+ 	unsigned long max_age;
+ 	unsigned long pause;
+ 
+-	unsigned uncommitted_blocks;
+-	unsigned autocommit_blocks;
+-	unsigned max_writeback_jobs;
++	unsigned int uncommitted_blocks;
++	unsigned int autocommit_blocks;
++	unsigned int max_writeback_jobs;
+ 
+ 	int error;
+ 
+@@ -155,7 +155,7 @@ struct dm_writecache {
+ 	sector_t data_device_sectors;
+ 	void *block_start;
+ 	struct wc_entry *entries;
+-	unsigned block_size;
++	unsigned int block_size;
+ 	unsigned char block_size_bits;
+ 
+ 	bool pmem_mode:1;
+@@ -178,13 +178,13 @@ struct dm_writecache {
+ 	bool metadata_only:1;
+ 	bool pause_set:1;
+ 
+-	unsigned high_wm_percent_value;
+-	unsigned low_wm_percent_value;
+-	unsigned autocommit_time_value;
+-	unsigned max_age_value;
+-	unsigned pause_value;
++	unsigned int high_wm_percent_value;
++	unsigned int low_wm_percent_value;
++	unsigned int autocommit_time_value;
++	unsigned int max_age_value;
++	unsigned int pause_value;
+ 
+-	unsigned writeback_all;
++	unsigned int writeback_all;
+ 	struct workqueue_struct *writeback_wq;
+ 	struct work_struct writeback_work;
+ 	struct work_struct flush_work;
+@@ -202,7 +202,7 @@ struct dm_writecache {
+ 
+ 	struct dm_kcopyd_client *dm_kcopyd;
+ 	unsigned long *dirty_bitmap;
+-	unsigned dirty_bitmap_size;
++	unsigned int dirty_bitmap_size;
+ 
+ 	struct bio_set bio_set;
+ 	mempool_t copy_pool;
+@@ -227,7 +227,7 @@ struct writeback_struct {
+ 	struct list_head endio_entry;
+ 	struct dm_writecache *wc;
+ 	struct wc_entry **wc_list;
+-	unsigned wc_list_n;
++	unsigned int wc_list_n;
+ 	struct wc_entry *wc_list_inline[WB_LIST_INLINE];
+ 	struct bio bio;
+ };
+@@ -236,7 +236,7 @@ struct copy_struct {
+ 	struct list_head endio_entry;
+ 	struct dm_writecache *wc;
+ 	struct wc_entry *e;
+-	unsigned n_entries;
++	unsigned int n_entries;
+ 	int error;
+ };
+ 
+@@ -369,7 +369,7 @@ static struct page *persistent_memory_page(void *addr)
+ 		return virt_to_page(addr);
+ }
+ 
+-static unsigned persistent_memory_page_offset(void *addr)
++static unsigned int persistent_memory_page_offset(void *addr)
+ {
+ 	return (unsigned long)addr & (PAGE_SIZE - 1);
+ }
+@@ -502,11 +502,11 @@ static void ssd_commit_flushed(struct dm_writecache *wc, bool wait_for_ios)
+ 		COMPLETION_INITIALIZER_ONSTACK(endio.c),
+ 		ATOMIC_INIT(1),
+ 	};
+-	unsigned bitmap_bits = wc->dirty_bitmap_size * 8;
+-	unsigned i = 0;
++	unsigned int bitmap_bits = wc->dirty_bitmap_size * 8;
++	unsigned int i = 0;
+ 
+ 	while (1) {
+-		unsigned j;
++		unsigned int j;
+ 		i = find_next_bit(wc->dirty_bitmap, bitmap_bits, i);
+ 		if (unlikely(i == bitmap_bits))
+ 			break;
+@@ -1100,7 +1100,7 @@ erase_this:
+ 	wc_unlock(wc);
+ }
+ 
+-static int process_flush_mesg(unsigned argc, char **argv, struct dm_writecache *wc)
++static int process_flush_mesg(unsigned int argc, char **argv, struct dm_writecache *wc)
+ {
+ 	if (argc != 1)
+ 		return -EINVAL;
+@@ -1133,7 +1133,7 @@ static int process_flush_mesg(unsigned argc, char **argv, struct dm_writecache *
+ 	return 0;
+ }
+ 
+-static int process_flush_on_suspend_mesg(unsigned argc, char **argv, struct dm_writecache *wc)
++static int process_flush_on_suspend_mesg(unsigned int argc, char **argv, struct dm_writecache *wc)
+ {
+ 	if (argc != 1)
+ 		return -EINVAL;
+@@ -1153,7 +1153,7 @@ static void activate_cleaner(struct dm_writecache *wc)
+ 	wc->freelist_low_watermark = wc->n_blocks;
+ }
+ 
+-static int process_cleaner_mesg(unsigned argc, char **argv, struct dm_writecache *wc)
++static int process_cleaner_mesg(unsigned int argc, char **argv, struct dm_writecache *wc)
+ {
+ 	if (argc != 1)
+ 		return -EINVAL;
+@@ -1167,7 +1167,7 @@ static int process_cleaner_mesg(unsigned argc, char **argv, struct dm_writecache
+ 	return 0;
+ }
+ 
+-static int process_clear_stats_mesg(unsigned argc, char **argv, struct dm_writecache *wc)
++static int process_clear_stats_mesg(unsigned int argc, char **argv, struct dm_writecache *wc)
+ {
+ 	if (argc != 1)
+ 		return -EINVAL;
+@@ -1179,8 +1179,8 @@ static int process_clear_stats_mesg(unsigned argc, char **argv, struct dm_writec
+ 	return 0;
+ }
+ 
+-static int writecache_message(struct dm_target *ti, unsigned argc, char **argv,
+-			      char *result, unsigned maxlen)
++static int writecache_message(struct dm_target *ti, unsigned int argc, char **argv,
++			      char *result, unsigned int maxlen)
+ {
+ 	int r = -EINVAL;
+ 	struct dm_writecache *wc = ti->private;
+@@ -1238,9 +1238,9 @@ static void memcpy_flushcache_optimized(void *dest, void *source, size_t size)
+ static void bio_copy_block(struct dm_writecache *wc, struct bio *bio, void *data)
+ {
+ 	void *buf;
+-	unsigned size;
++	unsigned int size;
+ 	int rw = bio_data_dir(bio);
+-	unsigned remaining_size = wc->block_size;
++	unsigned int remaining_size = wc->block_size;
+ 
+ 	do {
+ 		struct bio_vec bv = bio_iter_iovec(bio, bio->bi_iter);
+@@ -1371,7 +1371,7 @@ read_next_block:
+ static void writecache_bio_copy_ssd(struct dm_writecache *wc, struct bio *bio,
+ 				    struct wc_entry *e, bool search_used)
+ {
+-	unsigned bio_size = wc->block_size;
++	unsigned int bio_size = wc->block_size;
+ 	sector_t start_cache_sec = cache_sector(wc, e);
+ 	sector_t current_cache_sec = start_cache_sec + (bio_size >> SECTOR_SHIFT);
+ 
+@@ -1540,7 +1540,7 @@ static int writecache_map(struct dm_target *ti, struct bio *bio)
+ 
+ 	bio->bi_iter.bi_sector = dm_target_offset(ti, bio->bi_iter.bi_sector);
+ 
+-	if (unlikely((((unsigned)bio->bi_iter.bi_sector | bio_sectors(bio)) &
++	if (unlikely((((unsigned int)bio->bi_iter.bi_sector | bio_sectors(bio)) &
+ 				(wc->block_size / 512 - 1)) != 0)) {
+ 		DMERR("I/O is not aligned, sector %llu, size %u, block size %u",
+ 		      (unsigned long long)bio->bi_iter.bi_sector,
+@@ -1666,7 +1666,7 @@ static void writecache_copy_endio(int read_err, unsigned long write_err, void *p
+ 
+ static void __writecache_endio_pmem(struct dm_writecache *wc, struct list_head *list)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	struct writeback_struct *wb;
+ 	struct wc_entry *e;
+ 	unsigned long n_walked = 0;
+@@ -1782,7 +1782,7 @@ pop_from_list:
+ static bool wc_add_block(struct writeback_struct *wb, struct wc_entry *e)
+ {
+ 	struct dm_writecache *wc = wb->wc;
+-	unsigned block_size = wc->block_size;
++	unsigned int block_size = wc->block_size;
+ 	void *address = memory_data(wc, e);
+ 
+ 	persistent_memory_flush_cache(address, block_size);
+@@ -1817,7 +1817,7 @@ static void __writecache_writeback_pmem(struct dm_writecache *wc, struct writeba
+ 	struct wc_entry *e, *f;
+ 	struct bio *bio;
+ 	struct writeback_struct *wb;
+-	unsigned max_pages;
++	unsigned int max_pages;
+ 
+ 	while (wbl->size) {
+ 		wbl->size--;
+@@ -1880,7 +1880,7 @@ static void __writecache_writeback_ssd(struct dm_writecache *wc, struct writebac
+ 	struct copy_struct *c;
+ 
+ 	while (wbl->size) {
+-		unsigned n_sectors;
++		unsigned int n_sectors;
+ 
+ 		wbl->size--;
+ 		e = container_of(wbl->list.prev, struct wc_entry, lru);
+@@ -2092,7 +2092,7 @@ restart:
+ 	}
+ }
+ 
+-static int calculate_memory_size(uint64_t device_size, unsigned block_size,
++static int calculate_memory_size(uint64_t device_size, unsigned int block_size,
+ 				 size_t *n_blocks_p, size_t *n_metadata_blocks_p)
+ {
+ 	uint64_t n_blocks, offset;
+@@ -2207,12 +2207,12 @@ static void writecache_dtr(struct dm_target *ti)
+ 	kfree(wc);
+ }
+ 
+-static int writecache_ctr(struct dm_target *ti, unsigned argc, char **argv)
++static int writecache_ctr(struct dm_target *ti, unsigned int argc, char **argv)
+ {
+ 	struct dm_writecache *wc;
+ 	struct dm_arg_set as;
+ 	const char *string;
+-	unsigned opt_params;
++	unsigned int opt_params;
+ 	size_t offset, data_size;
+ 	int i, r;
+ 	char dummy;
+@@ -2419,7 +2419,7 @@ static int writecache_ctr(struct dm_target *ti, unsigned argc, char **argv)
+ 				goto invalid_optional;
+ 			wc->autocommit_blocks_set = true;
+ 		} else if (!strcasecmp(string, "autocommit_time") && opt_params >= 1) {
+-			unsigned autocommit_msecs;
++			unsigned int autocommit_msecs;
+ 			string = dm_shift_arg(&as), opt_params--;
+ 			if (sscanf(string, "%u%c", &autocommit_msecs, &dummy) != 1)
+ 				goto invalid_optional;
+@@ -2429,7 +2429,7 @@ static int writecache_ctr(struct dm_target *ti, unsigned argc, char **argv)
+ 			wc->autocommit_time_value = autocommit_msecs;
+ 			wc->autocommit_time_set = true;
+ 		} else if (!strcasecmp(string, "max_age") && opt_params >= 1) {
+-			unsigned max_age_msecs;
++			unsigned int max_age_msecs;
+ 			string = dm_shift_arg(&as), opt_params--;
+ 			if (sscanf(string, "%u%c", &max_age_msecs, &dummy) != 1)
+ 				goto invalid_optional;
+@@ -2454,7 +2454,7 @@ static int writecache_ctr(struct dm_target *ti, unsigned argc, char **argv)
+ 		} else if (!strcasecmp(string, "metadata_only")) {
+ 			wc->metadata_only = true;
+ 		} else if (!strcasecmp(string, "pause_writeback") && opt_params >= 1) {
+-			unsigned pause_msecs;
++			unsigned int pause_msecs;
+ 			if (WC_MODE_PMEM(wc))
+ 				goto invalid_optional;
+ 			string = dm_shift_arg(&as), opt_params--;
+@@ -2653,11 +2653,11 @@ bad:
+ }
+ 
+ static void writecache_status(struct dm_target *ti, status_type_t type,
+-			      unsigned status_flags, char *result, unsigned maxlen)
++			      unsigned int status_flags, char *result, unsigned int maxlen)
+ {
+ 	struct dm_writecache *wc = ti->private;
+-	unsigned extra_args;
+-	unsigned sz = 0;
++	unsigned int extra_args;
++	unsigned int sz = 0;
+ 
+ 	switch (type) {
+ 	case STATUSTYPE_INFO:
+diff --git a/drivers/md/dm.c b/drivers/md/dm.c
+index 38d8aa21ef7ac..24284d22f15bc 100644
+--- a/drivers/md/dm.c
++++ b/drivers/md/dm.c
+@@ -83,7 +83,7 @@ struct clone_info {
+ 	struct bio *bio;
+ 	struct dm_io *io;
+ 	sector_t sector;
+-	unsigned sector_count;
++	unsigned int sector_count;
+ 	bool is_abnormal_io:1;
+ 	bool submit_as_polled:1;
+ };
+@@ -111,7 +111,7 @@ struct bio *dm_bio_from_per_bio_data(void *data, size_t data_size)
+ }
+ EXPORT_SYMBOL_GPL(dm_bio_from_per_bio_data);
+ 
+-unsigned dm_bio_get_target_bio_nr(const struct bio *bio)
++unsigned int dm_bio_get_target_bio_nr(const struct bio *bio)
+ {
+ 	return container_of(bio, struct dm_target_io, clone)->target_bio_nr;
+ }
+@@ -142,7 +142,7 @@ struct table_device {
+  * Bio-based DM's mempools' reserved IOs set by the user.
+  */
+ #define RESERVED_BIO_BASED_IOS		16
+-static unsigned reserved_bio_based_ios = RESERVED_BIO_BASED_IOS;
++static unsigned int reserved_bio_based_ios = RESERVED_BIO_BASED_IOS;
+ 
+ static int __dm_get_module_param_int(int *module_param, int min, int max)
+ {
+@@ -165,11 +165,10 @@ static int __dm_get_module_param_int(int *module_param, int min, int max)
+ 	return param;
+ }
+ 
+-unsigned __dm_get_module_param(unsigned *module_param,
+-			       unsigned def, unsigned max)
++unsigned int __dm_get_module_param(unsigned int *module_param, unsigned int def, unsigned int max)
+ {
+-	unsigned param = READ_ONCE(*module_param);
+-	unsigned modified_param = 0;
++	unsigned int param = READ_ONCE(*module_param);
++	unsigned int modified_param = 0;
+ 
+ 	if (!param)
+ 		modified_param = def;
+@@ -184,14 +183,14 @@ unsigned __dm_get_module_param(unsigned *module_param,
+ 	return param;
+ }
+ 
+-unsigned dm_get_reserved_bio_based_ios(void)
++unsigned int dm_get_reserved_bio_based_ios(void)
+ {
+ 	return __dm_get_module_param(&reserved_bio_based_ios,
+ 				     RESERVED_BIO_BASED_IOS, DM_RESERVED_MAX_IOS);
+ }
+ EXPORT_SYMBOL_GPL(dm_get_reserved_bio_based_ios);
+ 
+-static unsigned dm_get_numa_node(void)
++static unsigned int dm_get_numa_node(void)
+ {
+ 	return __dm_get_module_param_int(&dm_numa_node,
+ 					 DM_NUMA_NODE, num_online_nodes() - 1);
+@@ -603,7 +602,7 @@ static void free_io(struct dm_io *io)
+ }
+ 
+ static struct bio *alloc_tio(struct clone_info *ci, struct dm_target *ti,
+-			     unsigned target_bio_nr, unsigned *len, gfp_t gfp_mask)
++			     unsigned int target_bio_nr, unsigned int *len, gfp_t gfp_mask)
+ {
+ 	struct mapped_device *md = ci->io->md;
+ 	struct dm_target_io *tio;
+@@ -1327,11 +1326,11 @@ out:
+  * the partially processed part (the sum of regions 1+2) must be the same for all
+  * copies of the bio.
+  */
+-void dm_accept_partial_bio(struct bio *bio, unsigned n_sectors)
++void dm_accept_partial_bio(struct bio *bio, unsigned int n_sectors)
+ {
+ 	struct dm_target_io *tio = clone_to_tio(bio);
+ 	struct dm_io *io = tio->io;
+-	unsigned bio_sectors = bio_sectors(bio);
++	unsigned int bio_sectors = bio_sectors(bio);
+ 
+ 	BUG_ON(dm_tio_flagged(tio, DM_TIO_IS_DUPLICATE_BIO));
+ 	BUG_ON(op_is_zone_mgmt(bio_op(bio)));
+@@ -1460,7 +1459,7 @@ static void __map_bio(struct bio *clone)
+ 	}
+ }
+ 
+-static void setup_split_accounting(struct clone_info *ci, unsigned len)
++static void setup_split_accounting(struct clone_info *ci, unsigned int len)
+ {
+ 	struct dm_io *io = ci->io;
+ 
+@@ -1476,7 +1475,8 @@ static void setup_split_accounting(struct clone_info *ci, unsigned len)
+ }
+ 
+ static void alloc_multiple_bios(struct bio_list *blist, struct clone_info *ci,
+-				struct dm_target *ti, unsigned num_bios)
++				struct dm_target *ti, unsigned int num_bios,
++				unsigned *len)
+ {
+ 	struct bio *bio;
+ 	int try;
+@@ -1487,7 +1487,7 @@ static void alloc_multiple_bios(struct bio_list *blist, struct clone_info *ci,
+ 		if (try)
+ 			mutex_lock(&ci->io->md->table_devices_lock);
+ 		for (bio_nr = 0; bio_nr < num_bios; bio_nr++) {
+-			bio = alloc_tio(ci, ti, bio_nr, NULL,
++			bio = alloc_tio(ci, ti, bio_nr, len,
+ 					try ? GFP_NOIO : GFP_NOWAIT);
+ 			if (!bio)
+ 				break;
+@@ -1505,7 +1505,7 @@ static void alloc_multiple_bios(struct bio_list *blist, struct clone_info *ci,
+ }
+ 
+ static int __send_duplicate_bios(struct clone_info *ci, struct dm_target *ti,
+-				 unsigned int num_bios, unsigned *len)
++				 unsigned int num_bios, unsigned int *len)
+ {
+ 	struct bio_list blist = BIO_EMPTY_LIST;
+ 	struct bio *clone;
+@@ -1525,7 +1525,7 @@ static int __send_duplicate_bios(struct clone_info *ci, struct dm_target *ti,
+ 		if (len)
+ 			setup_split_accounting(ci, *len);
+ 		/* dm_accept_partial_bio() is not supported with shared tio->len_ptr */
+-		alloc_multiple_bios(&blist, ci, ti, num_bios);
++		alloc_multiple_bios(&blist, ci, ti, num_bios, len);
+ 		while ((clone = bio_list_pop(&blist))) {
+ 			dm_tio_set_flag(clone_to_tio(clone), DM_TIO_IS_DUPLICATE_BIO);
+ 			__map_bio(clone);
+@@ -1573,10 +1573,9 @@ static void __send_empty_flush(struct clone_info *ci)
+ }
+ 
+ static void __send_changing_extent_only(struct clone_info *ci, struct dm_target *ti,
+-					unsigned num_bios)
++					unsigned int num_bios)
+ {
+-	unsigned len;
+-	unsigned int bios;
++	unsigned int len, bios;
+ 
+ 	len = min_t(sector_t, ci->sector_count,
+ 		    max_io_len_target_boundary(ti, dm_target_offset(ti, ci->sector)));
+@@ -1614,7 +1613,7 @@ static bool is_abnormal_io(struct bio *bio)
+ static blk_status_t __process_abnormal_io(struct clone_info *ci,
+ 					  struct dm_target *ti)
+ {
+-	unsigned num_bios = 0;
++	unsigned int num_bios = 0;
+ 
+ 	switch (bio_op(ci->bio)) {
+ 	case REQ_OP_DISCARD:
+@@ -1692,7 +1691,7 @@ static blk_status_t __split_and_process_bio(struct clone_info *ci)
+ {
+ 	struct bio *clone;
+ 	struct dm_target *ti;
+-	unsigned len;
++	unsigned int len;
+ 
+ 	ti = dm_table_find_target(ci->map, ci->sector);
+ 	if (unlikely(!ti))
+@@ -2389,7 +2388,7 @@ out_undo_holders:
+ struct mapped_device *dm_get_md(dev_t dev)
+ {
+ 	struct mapped_device *md;
+-	unsigned minor = MINOR(dev);
++	unsigned int minor = MINOR(dev);
+ 
+ 	if (MAJOR(dev) != _major || minor >= (1 << MINORBITS))
+ 		return NULL;
+@@ -2672,7 +2671,7 @@ static void unlock_fs(struct mapped_device *md)
+  * are being added to md->deferred list.
+  */
+ static int __dm_suspend(struct mapped_device *md, struct dm_table *map,
+-			unsigned suspend_flags, unsigned int task_state,
++			unsigned int suspend_flags, unsigned int task_state,
+ 			int dmf_suspended_flag)
+ {
+ 	bool do_lockfs = suspend_flags & DM_SUSPEND_LOCKFS_FLAG;
+@@ -2779,7 +2778,7 @@ static int __dm_suspend(struct mapped_device *md, struct dm_table *map,
+  *
+  * To abort suspend, start the request_queue.
+  */
+-int dm_suspend(struct mapped_device *md, unsigned suspend_flags)
++int dm_suspend(struct mapped_device *md, unsigned int suspend_flags)
+ {
+ 	struct dm_table *map = NULL;
+ 	int r = 0;
+@@ -2881,7 +2880,7 @@ out:
+  * It may be used only from the kernel.
+  */
+ 
+-static void __dm_internal_suspend(struct mapped_device *md, unsigned suspend_flags)
++static void __dm_internal_suspend(struct mapped_device *md, unsigned int suspend_flags)
+ {
+ 	struct dm_table *map = NULL;
+ 
+@@ -2983,10 +2982,10 @@ EXPORT_SYMBOL_GPL(dm_internal_resume_fast);
+  * Event notification.
+  *---------------------------------------------------------------*/
+ int dm_kobject_uevent(struct mapped_device *md, enum kobject_action action,
+-		      unsigned cookie, bool need_resize_uevent)
++		      unsigned int cookie, bool need_resize_uevent)
+ {
+ 	int r;
+-	unsigned noio_flag;
++	unsigned int noio_flag;
+ 	char udev_cookie[DM_COOKIE_LENGTH];
+ 	char *envp[3] = { NULL, NULL, NULL };
+ 	char **envpp = envp;
+diff --git a/drivers/md/dm.h b/drivers/md/dm.h
+index a9a3ffcad084c..a7917df09cafb 100644
+--- a/drivers/md/dm.h
++++ b/drivers/md/dm.h
+@@ -203,7 +203,7 @@ int dm_get_table_device(struct mapped_device *md, dev_t dev, fmode_t mode,
+ void dm_put_table_device(struct mapped_device *md, struct dm_dev *d);
+ 
+ int dm_kobject_uevent(struct mapped_device *md, enum kobject_action action,
+-		      unsigned cookie, bool need_resize_uevent);
++		      unsigned int cookie, bool need_resize_uevent);
+ 
+ void dm_internal_suspend(struct mapped_device *md);
+ void dm_internal_resume(struct mapped_device *md);
+@@ -222,6 +222,6 @@ void dm_free_md_mempools(struct dm_md_mempools *pools);
+ /*
+  * Various helpers
+  */
+-unsigned dm_get_reserved_bio_based_ios(void);
++unsigned int dm_get_reserved_bio_based_ios(void);
+ 
+ #endif
+diff --git a/drivers/md/persistent-data/dm-array.c b/drivers/md/persistent-data/dm-array.c
+index 3a963d783a865..eff9b41869f29 100644
+--- a/drivers/md/persistent-data/dm-array.c
++++ b/drivers/md/persistent-data/dm-array.c
+@@ -68,8 +68,8 @@ static int array_block_check(struct dm_block_validator *v,
+ 					       CSUM_XOR));
+ 	if (csum_disk != bh_le->csum) {
+ 		DMERR_LIMIT("array_block_check failed: csum %u != wanted %u",
+-			    (unsigned) le32_to_cpu(csum_disk),
+-			    (unsigned) le32_to_cpu(bh_le->csum));
++			    (unsigned int) le32_to_cpu(csum_disk),
++			    (unsigned int) le32_to_cpu(bh_le->csum));
+ 		return -EILSEQ;
+ 	}
+ 
+@@ -94,7 +94,7 @@ static struct dm_block_validator array_validator = {
+  * index - The index into _this_ specific block.
+  */
+ static void *element_at(struct dm_array_info *info, struct array_block *ab,
+-			unsigned index)
++			unsigned int index)
+ {
+ 	unsigned char *entry = (unsigned char *) (ab + 1);
+ 
+@@ -108,9 +108,9 @@ static void *element_at(struct dm_array_info *info, struct array_block *ab,
+  * in an array block.
+  */
+ static void on_entries(struct dm_array_info *info, struct array_block *ab,
+-		       void (*fn)(void *, const void *, unsigned))
++		       void (*fn)(void *, const void *, unsigned int))
+ {
+-	unsigned nr_entries = le32_to_cpu(ab->nr_entries);
++	unsigned int nr_entries = le32_to_cpu(ab->nr_entries);
+ 	fn(info->value_type.context, element_at(info, ab, 0), nr_entries);
+ }
+ 
+@@ -171,7 +171,7 @@ static int alloc_ablock(struct dm_array_info *info, size_t size_of_block,
+  * the current number of entries.
+  */
+ static void fill_ablock(struct dm_array_info *info, struct array_block *ab,
+-			const void *value, unsigned new_nr)
++			const void *value, unsigned int new_nr)
+ {
+ 	uint32_t nr_entries, delta, i;
+ 	struct dm_btree_value_type *vt = &info->value_type;
+@@ -194,7 +194,7 @@ static void fill_ablock(struct dm_array_info *info, struct array_block *ab,
+  * entries.
+  */
+ static void trim_ablock(struct dm_array_info *info, struct array_block *ab,
+-			unsigned new_nr)
++			unsigned int new_nr)
+ {
+ 	uint32_t nr_entries, delta;
+ 	struct dm_btree_value_type *vt = &info->value_type;
+@@ -247,7 +247,7 @@ static void unlock_ablock(struct dm_array_info *info, struct dm_block *block)
+  * / max_entries).
+  */
+ static int lookup_ablock(struct dm_array_info *info, dm_block_t root,
+-			 unsigned index, struct dm_block **block,
++			 unsigned int index, struct dm_block **block,
+ 			 struct array_block **ab)
+ {
+ 	int r;
+@@ -295,7 +295,7 @@ static int __shadow_ablock(struct dm_array_info *info, dm_block_t b,
+  * The shadow op will often be a noop.  Only insert if it really
+  * copied data.
+  */
+-static int __reinsert_ablock(struct dm_array_info *info, unsigned index,
++static int __reinsert_ablock(struct dm_array_info *info, unsigned int index,
+ 			     struct dm_block *block, dm_block_t b,
+ 			     dm_block_t *root)
+ {
+@@ -321,7 +321,7 @@ static int __reinsert_ablock(struct dm_array_info *info, unsigned index,
+  * for both the current root block, and the new one.
+  */
+ static int shadow_ablock(struct dm_array_info *info, dm_block_t *root,
+-			 unsigned index, struct dm_block **block,
++			 unsigned int index, struct dm_block **block,
+ 			 struct array_block **ab)
+ {
+ 	int r;
+@@ -346,7 +346,7 @@ static int shadow_ablock(struct dm_array_info *info, dm_block_t *root,
+  */
+ static int insert_new_ablock(struct dm_array_info *info, size_t size_of_block,
+ 			     uint32_t max_entries,
+-			     unsigned block_index, uint32_t nr,
++			     unsigned int block_index, uint32_t nr,
+ 			     const void *value, dm_block_t *root)
+ {
+ 	int r;
+@@ -365,8 +365,8 @@ static int insert_new_ablock(struct dm_array_info *info, size_t size_of_block,
+ }
+ 
+ static int insert_full_ablocks(struct dm_array_info *info, size_t size_of_block,
+-			       unsigned begin_block, unsigned end_block,
+-			       unsigned max_entries, const void *value,
++			       unsigned int begin_block, unsigned int end_block,
++			       unsigned int max_entries, const void *value,
+ 			       dm_block_t *root)
+ {
+ 	int r = 0;
+@@ -402,20 +402,20 @@ struct resize {
+ 	/*
+ 	 * Maximum nr entries in an array block.
+ 	 */
+-	unsigned max_entries;
++	unsigned int max_entries;
+ 
+ 	/*
+ 	 * nr of completely full blocks in the array.
+ 	 *
+ 	 * 'old' refers to before the resize, 'new' after.
+ 	 */
+-	unsigned old_nr_full_blocks, new_nr_full_blocks;
++	unsigned int old_nr_full_blocks, new_nr_full_blocks;
+ 
+ 	/*
+ 	 * Number of entries in the final block.  0 iff only full blocks in
+ 	 * the array.
+ 	 */
+-	unsigned old_nr_entries_in_last_block, new_nr_entries_in_last_block;
++	unsigned int old_nr_entries_in_last_block, new_nr_entries_in_last_block;
+ 
+ 	/*
+ 	 * The default value used when growing the array.
+@@ -430,8 +430,8 @@ struct resize {
+  * begin_index - the index of the first array block to remove.
+  * end_index - the one-past-the-end value.  ie. this block is not removed.
+  */
+-static int drop_blocks(struct resize *resize, unsigned begin_index,
+-		       unsigned end_index)
++static int drop_blocks(struct resize *resize, unsigned int begin_index,
++		       unsigned int end_index)
+ {
+ 	int r;
+ 
+@@ -449,8 +449,8 @@ static int drop_blocks(struct resize *resize, unsigned begin_index,
+ /*
+  * Calculates how many blocks are needed for the array.
+  */
+-static unsigned total_nr_blocks_needed(unsigned nr_full_blocks,
+-				       unsigned nr_entries_in_last_block)
++static unsigned int total_nr_blocks_needed(unsigned int nr_full_blocks,
++				       unsigned int nr_entries_in_last_block)
+ {
+ 	return nr_full_blocks + (nr_entries_in_last_block ? 1 : 0);
+ }
+@@ -461,7 +461,7 @@ static unsigned total_nr_blocks_needed(unsigned nr_full_blocks,
+ static int shrink(struct resize *resize)
+ {
+ 	int r;
+-	unsigned begin, end;
++	unsigned int begin, end;
+ 	struct dm_block *block;
+ 	struct array_block *ab;
+ 
+@@ -527,7 +527,7 @@ static int grow_add_tail_block(struct resize *resize)
+ static int grow_needs_more_blocks(struct resize *resize)
+ {
+ 	int r;
+-	unsigned old_nr_blocks = resize->old_nr_full_blocks;
++	unsigned int old_nr_blocks = resize->old_nr_full_blocks;
+ 
+ 	if (resize->old_nr_entries_in_last_block > 0) {
+ 		old_nr_blocks++;
+@@ -569,11 +569,11 @@ static int grow(struct resize *resize)
+  * These are the value_type functions for the btree elements, which point
+  * to array blocks.
+  */
+-static void block_inc(void *context, const void *value, unsigned count)
++static void block_inc(void *context, const void *value, unsigned int count)
+ {
+ 	const __le64 *block_le = value;
+ 	struct dm_array_info *info = context;
+-	unsigned i;
++	unsigned int i;
+ 
+ 	for (i = 0; i < count; i++, block_le++)
+ 		dm_tm_inc(info->btree_info.tm, le64_to_cpu(*block_le));
+@@ -618,9 +618,9 @@ static void __block_dec(void *context, const void *value)
+ 	dm_tm_dec(info->btree_info.tm, b);
+ }
+ 
+-static void block_dec(void *context, const void *value, unsigned count)
++static void block_dec(void *context, const void *value, unsigned int count)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	for (i = 0; i < count; i++, value += sizeof(__le64))
+ 		__block_dec(context, value);
+ }
+@@ -700,10 +700,11 @@ int dm_array_resize(struct dm_array_info *info, dm_block_t root,
+ EXPORT_SYMBOL_GPL(dm_array_resize);
+ 
+ static int populate_ablock_with_values(struct dm_array_info *info, struct array_block *ab,
+-				       value_fn fn, void *context, unsigned base, unsigned new_nr)
++				       value_fn fn, void *context,
++				       unsigned int base, unsigned int new_nr)
+ {
+ 	int r;
+-	unsigned i;
++	unsigned int i;
+ 	struct dm_btree_value_type *vt = &info->value_type;
+ 
+ 	BUG_ON(le32_to_cpu(ab->nr_entries));
+@@ -728,7 +729,7 @@ int dm_array_new(struct dm_array_info *info, dm_block_t *root,
+ 	int r;
+ 	struct dm_block *block;
+ 	struct array_block *ab;
+-	unsigned block_index, end_block, size_of_block, max_entries;
++	unsigned int block_index, end_block, size_of_block, max_entries;
+ 
+ 	r = dm_array_empty(info, root);
+ 	if (r)
+@@ -776,7 +777,7 @@ int dm_array_get_value(struct dm_array_info *info, dm_block_t root,
+ 	struct dm_block *block;
+ 	struct array_block *ab;
+ 	size_t size_of_block;
+-	unsigned entry, max_entries;
++	unsigned int entry, max_entries;
+ 
+ 	size_of_block = dm_bm_block_size(dm_tm_get_bm(info->btree_info.tm));
+ 	max_entries = calc_max_entries(info->value_type.size, size_of_block);
+@@ -804,8 +805,8 @@ static int array_set_value(struct dm_array_info *info, dm_block_t root,
+ 	struct dm_block *block;
+ 	struct array_block *ab;
+ 	size_t size_of_block;
+-	unsigned max_entries;
+-	unsigned entry;
++	unsigned int max_entries;
++	unsigned int entry;
+ 	void *old_value;
+ 	struct dm_btree_value_type *vt = &info->value_type;
+ 
+@@ -861,9 +862,9 @@ static int walk_ablock(void *context, uint64_t *keys, void *leaf)
+ 	struct walk_info *wi = context;
+ 
+ 	int r;
+-	unsigned i;
++	unsigned int i;
+ 	__le64 block_le;
+-	unsigned nr_entries, max_entries;
++	unsigned int nr_entries, max_entries;
+ 	struct dm_block *block;
+ 	struct array_block *ab;
+ 
+diff --git a/drivers/md/persistent-data/dm-array.h b/drivers/md/persistent-data/dm-array.h
+index d7d2d579c662c..b6c7077c73591 100644
+--- a/drivers/md/persistent-data/dm-array.h
++++ b/drivers/md/persistent-data/dm-array.h
+@@ -198,7 +198,7 @@ struct dm_array_cursor {
+ 
+ 	struct dm_block *block;
+ 	struct array_block *ab;
+-	unsigned index;
++	unsigned int index;
+ };
+ 
+ int dm_array_cursor_begin(struct dm_array_info *info,
+diff --git a/drivers/md/persistent-data/dm-bitset.c b/drivers/md/persistent-data/dm-bitset.c
+index b7208d82e748a..625d93498cddb 100644
+--- a/drivers/md/persistent-data/dm-bitset.c
++++ b/drivers/md/persistent-data/dm-bitset.c
+@@ -41,7 +41,7 @@ EXPORT_SYMBOL_GPL(dm_bitset_empty);
+ 
+ struct packer_context {
+ 	bit_value_fn fn;
+-	unsigned nr_bits;
++	unsigned int nr_bits;
+ 	void *context;
+ };
+ 
+@@ -49,7 +49,7 @@ static int pack_bits(uint32_t index, void *value, void *context)
+ {
+ 	int r;
+ 	struct packer_context *p = context;
+-	unsigned bit, nr = min(64u, p->nr_bits - (index * 64));
++	unsigned int bit, nr = min(64u, p->nr_bits - (index * 64));
+ 	uint64_t word = 0;
+ 	bool bv;
+ 
+@@ -147,7 +147,7 @@ static int get_array_entry(struct dm_disk_bitset *info, dm_block_t root,
+ 			   uint32_t index, dm_block_t *new_root)
+ {
+ 	int r;
+-	unsigned array_index = index / BITS_PER_ARRAY_ENTRY;
++	unsigned int array_index = index / BITS_PER_ARRAY_ENTRY;
+ 
+ 	if (info->current_index_set) {
+ 		if (info->current_index == array_index)
+@@ -165,7 +165,7 @@ int dm_bitset_set_bit(struct dm_disk_bitset *info, dm_block_t root,
+ 		      uint32_t index, dm_block_t *new_root)
+ {
+ 	int r;
+-	unsigned b = index % BITS_PER_ARRAY_ENTRY;
++	unsigned int b = index % BITS_PER_ARRAY_ENTRY;
+ 
+ 	r = get_array_entry(info, root, index, new_root);
+ 	if (r)
+@@ -182,7 +182,7 @@ int dm_bitset_clear_bit(struct dm_disk_bitset *info, dm_block_t root,
+ 			uint32_t index, dm_block_t *new_root)
+ {
+ 	int r;
+-	unsigned b = index % BITS_PER_ARRAY_ENTRY;
++	unsigned int b = index % BITS_PER_ARRAY_ENTRY;
+ 
+ 	r = get_array_entry(info, root, index, new_root);
+ 	if (r)
+@@ -199,7 +199,7 @@ int dm_bitset_test_bit(struct dm_disk_bitset *info, dm_block_t root,
+ 		       uint32_t index, dm_block_t *new_root, bool *result)
+ {
+ 	int r;
+-	unsigned b = index % BITS_PER_ARRAY_ENTRY;
++	unsigned int b = index % BITS_PER_ARRAY_ENTRY;
+ 
+ 	r = get_array_entry(info, root, index, new_root);
+ 	if (r)
+diff --git a/drivers/md/persistent-data/dm-block-manager.c b/drivers/md/persistent-data/dm-block-manager.c
+index 11935864f50f5..1f40100908d7c 100644
+--- a/drivers/md/persistent-data/dm-block-manager.c
++++ b/drivers/md/persistent-data/dm-block-manager.c
+@@ -57,10 +57,10 @@ struct waiter {
+ 	int wants_write;
+ };
+ 
+-static unsigned __find_holder(struct block_lock *lock,
++static unsigned int __find_holder(struct block_lock *lock,
+ 			      struct task_struct *task)
+ {
+-	unsigned i;
++	unsigned int i;
+ 
+ 	for (i = 0; i < MAX_HOLDERS; i++)
+ 		if (lock->holders[i] == task)
+@@ -73,7 +73,7 @@ static unsigned __find_holder(struct block_lock *lock,
+ /* call this *after* you increment lock->count */
+ static void __add_holder(struct block_lock *lock, struct task_struct *task)
+ {
+-	unsigned h = __find_holder(lock, NULL);
++	unsigned int h = __find_holder(lock, NULL);
+ #ifdef CONFIG_DM_DEBUG_BLOCK_STACK_TRACING
+ 	struct stack_store *t;
+ #endif
+@@ -90,14 +90,14 @@ static void __add_holder(struct block_lock *lock, struct task_struct *task)
+ /* call this *before* you decrement lock->count */
+ static void __del_holder(struct block_lock *lock, struct task_struct *task)
+ {
+-	unsigned h = __find_holder(lock, task);
++	unsigned int h = __find_holder(lock, task);
+ 	lock->holders[h] = NULL;
+ 	put_task_struct(task);
+ }
+ 
+ static int __check_holder(struct block_lock *lock)
+ {
+-	unsigned i;
++	unsigned int i;
+ 
+ 	for (i = 0; i < MAX_HOLDERS; i++) {
+ 		if (lock->holders[i] == current) {
+@@ -376,8 +376,8 @@ struct dm_block_manager {
+ };
+ 
+ struct dm_block_manager *dm_block_manager_create(struct block_device *bdev,
+-						 unsigned block_size,
+-						 unsigned max_held_per_thread)
++						 unsigned int block_size,
++						 unsigned int max_held_per_thread)
+ {
+ 	int r;
+ 	struct dm_block_manager *bm;
+@@ -415,7 +415,7 @@ void dm_block_manager_destroy(struct dm_block_manager *bm)
+ }
+ EXPORT_SYMBOL_GPL(dm_block_manager_destroy);
+ 
+-unsigned dm_bm_block_size(struct dm_block_manager *bm)
++unsigned int dm_bm_block_size(struct dm_block_manager *bm)
+ {
+ 	return dm_bufio_get_block_size(bm->bufio);
+ }
+diff --git a/drivers/md/persistent-data/dm-block-manager.h b/drivers/md/persistent-data/dm-block-manager.h
+index e728937f376a3..58a23b8ec1902 100644
+--- a/drivers/md/persistent-data/dm-block-manager.h
++++ b/drivers/md/persistent-data/dm-block-manager.h
+@@ -32,11 +32,11 @@ void *dm_block_data(struct dm_block *b);
+  */
+ struct dm_block_manager;
+ struct dm_block_manager *dm_block_manager_create(
+-	struct block_device *bdev, unsigned block_size,
+-	unsigned max_held_per_thread);
++	struct block_device *bdev, unsigned int block_size,
++	unsigned int max_held_per_thread);
+ void dm_block_manager_destroy(struct dm_block_manager *bm);
+ 
+-unsigned dm_bm_block_size(struct dm_block_manager *bm);
++unsigned int dm_bm_block_size(struct dm_block_manager *bm);
+ dm_block_t dm_bm_nr_blocks(struct dm_block_manager *bm);
+ 
+ /*----------------------------------------------------------------*/
+diff --git a/drivers/md/persistent-data/dm-btree-remove.c b/drivers/md/persistent-data/dm-btree-remove.c
+index 4ead31e0d8ce5..ac213138b0217 100644
+--- a/drivers/md/persistent-data/dm-btree-remove.c
++++ b/drivers/md/persistent-data/dm-btree-remove.c
+@@ -124,10 +124,10 @@ static int node_copy(struct btree_node *left, struct btree_node *right, int shif
+ /*
+  * Delete a specific entry from a leaf node.
+  */
+-static void delete_at(struct btree_node *n, unsigned index)
++static void delete_at(struct btree_node *n, unsigned int index)
+ {
+-	unsigned nr_entries = le32_to_cpu(n->header.nr_entries);
+-	unsigned nr_to_copy = nr_entries - (index + 1);
++	unsigned int nr_entries = le32_to_cpu(n->header.nr_entries);
++	unsigned int nr_to_copy = nr_entries - (index + 1);
+ 	uint32_t value_size = le32_to_cpu(n->header.value_size);
+ 	BUG_ON(index >= nr_entries);
+ 
+@@ -144,20 +144,20 @@ static void delete_at(struct btree_node *n, unsigned index)
+ 	n->header.nr_entries = cpu_to_le32(nr_entries - 1);
+ }
+ 
+-static unsigned merge_threshold(struct btree_node *n)
++static unsigned int merge_threshold(struct btree_node *n)
+ {
+ 	return le32_to_cpu(n->header.max_entries) / 3;
+ }
+ 
+ struct child {
+-	unsigned index;
++	unsigned int index;
+ 	struct dm_block *block;
+ 	struct btree_node *n;
+ };
+ 
+ static int init_child(struct dm_btree_info *info, struct dm_btree_value_type *vt,
+ 		      struct btree_node *parent,
+-		      unsigned index, struct child *result)
++		      unsigned int index, struct child *result)
+ {
+ 	int r, inc;
+ 	dm_block_t root;
+@@ -263,7 +263,7 @@ static int __rebalance2(struct dm_btree_info *info, struct btree_node *parent,
+ 		/*
+ 		 * Rebalance.
+ 		 */
+-		unsigned target_left = (nr_left + nr_right) / 2;
++		unsigned int target_left = (nr_left + nr_right) / 2;
+ 		ret = shift(left, right, nr_left - target_left);
+ 		if (ret)
+ 			return ret;
+@@ -273,7 +273,7 @@ static int __rebalance2(struct dm_btree_info *info, struct btree_node *parent,
+ }
+ 
+ static int rebalance2(struct shadow_spine *s, struct dm_btree_info *info,
+-		      struct dm_btree_value_type *vt, unsigned left_index)
++		      struct dm_btree_value_type *vt, unsigned int left_index)
+ {
+ 	int r;
+ 	struct btree_node *parent;
+@@ -310,7 +310,7 @@ static int delete_center_node(struct dm_btree_info *info, struct btree_node *par
+ 			      uint32_t nr_left, uint32_t nr_center, uint32_t nr_right)
+ {
+ 	uint32_t max_entries = le32_to_cpu(left->header.max_entries);
+-	unsigned shift = min(max_entries - nr_left, nr_center);
++	unsigned int shift = min(max_entries - nr_left, nr_center);
+ 
+ 	if (nr_left + shift > max_entries) {
+ 		DMERR("node shift out of bounds");
+@@ -351,10 +351,10 @@ static int redistribute3(struct dm_btree_info *info, struct btree_node *parent,
+ {
+ 	int s, ret;
+ 	uint32_t max_entries = le32_to_cpu(left->header.max_entries);
+-	unsigned total = nr_left + nr_center + nr_right;
+-	unsigned target_right = total / 3;
+-	unsigned remainder = (target_right * 3) != total;
+-	unsigned target_left = target_right + remainder;
++	unsigned int total = nr_left + nr_center + nr_right;
++	unsigned int target_right = total / 3;
++	unsigned int remainder = (target_right * 3) != total;
++	unsigned int target_left = target_right + remainder;
+ 
+ 	BUG_ON(target_left > max_entries);
+ 	BUG_ON(target_right > max_entries);
+@@ -422,7 +422,7 @@ static int __rebalance3(struct dm_btree_info *info, struct btree_node *parent,
+ 	uint32_t nr_center = le32_to_cpu(center->header.nr_entries);
+ 	uint32_t nr_right = le32_to_cpu(right->header.nr_entries);
+ 
+-	unsigned threshold = merge_threshold(left) * 4 + 1;
++	unsigned int threshold = merge_threshold(left) * 4 + 1;
+ 
+ 	if ((left->header.max_entries != center->header.max_entries) ||
+ 	    (center->header.max_entries != right->header.max_entries)) {
+@@ -440,7 +440,7 @@ static int __rebalance3(struct dm_btree_info *info, struct btree_node *parent,
+ }
+ 
+ static int rebalance3(struct shadow_spine *s, struct dm_btree_info *info,
+-		      struct dm_btree_value_type *vt, unsigned left_index)
++		      struct dm_btree_value_type *vt, unsigned int left_index)
+ {
+ 	int r;
+ 	struct btree_node *parent = dm_block_data(shadow_current(s));
+@@ -519,7 +519,7 @@ static int rebalance_children(struct shadow_spine *s,
+ 	return r;
+ }
+ 
+-static int do_leaf(struct btree_node *n, uint64_t key, unsigned *index)
++static int do_leaf(struct btree_node *n, uint64_t key, unsigned int *index)
+ {
+ 	int i = lower_bound(n, key);
+ 
+@@ -539,7 +539,7 @@ static int do_leaf(struct btree_node *n, uint64_t key, unsigned *index)
+  */
+ static int remove_raw(struct shadow_spine *s, struct dm_btree_info *info,
+ 		      struct dm_btree_value_type *vt, dm_block_t root,
+-		      uint64_t key, unsigned *index)
++		      uint64_t key, unsigned int *index)
+ {
+ 	int i = *index, r;
+ 	struct btree_node *n;
+@@ -589,7 +589,7 @@ static int remove_raw(struct shadow_spine *s, struct dm_btree_info *info,
+ int dm_btree_remove(struct dm_btree_info *info, dm_block_t root,
+ 		    uint64_t *keys, dm_block_t *new_root)
+ {
+-	unsigned level, last_level = info->levels - 1;
++	unsigned int level, last_level = info->levels - 1;
+ 	int index = 0, r = 0;
+ 	struct shadow_spine spine;
+ 	struct btree_node *n;
+@@ -601,7 +601,7 @@ int dm_btree_remove(struct dm_btree_info *info, dm_block_t root,
+ 		r = remove_raw(&spine, info,
+ 			       (level == last_level ?
+ 				&info->value_type : &le64_vt),
+-			       root, keys[level], (unsigned *)&index);
++			       root, keys[level], (unsigned int *)&index);
+ 		if (r < 0)
+ 			break;
+ 
+@@ -685,9 +685,9 @@ static int remove_nearest(struct shadow_spine *s, struct dm_btree_info *info,
+ 
+ static int remove_one(struct dm_btree_info *info, dm_block_t root,
+ 		      uint64_t *keys, uint64_t end_key,
+-		      dm_block_t *new_root, unsigned *nr_removed)
++		      dm_block_t *new_root, unsigned int *nr_removed)
+ {
+-	unsigned level, last_level = info->levels - 1;
++	unsigned int level, last_level = info->levels - 1;
+ 	int index = 0, r = 0;
+ 	struct shadow_spine spine;
+ 	struct btree_node *n;
+@@ -698,7 +698,7 @@ static int remove_one(struct dm_btree_info *info, dm_block_t root,
+ 	init_shadow_spine(&spine, info);
+ 	for (level = 0; level < last_level; level++) {
+ 		r = remove_raw(&spine, info, &le64_vt,
+-			       root, keys[level], (unsigned *) &index);
++			       root, keys[level], (unsigned int *) &index);
+ 		if (r < 0)
+ 			goto out;
+ 
+@@ -742,7 +742,7 @@ out:
+ 
+ int dm_btree_remove_leaves(struct dm_btree_info *info, dm_block_t root,
+ 			   uint64_t *first_key, uint64_t end_key,
+-			   dm_block_t *new_root, unsigned *nr_removed)
++			   dm_block_t *new_root, unsigned int *nr_removed)
+ {
+ 	int r;
+ 
+diff --git a/drivers/md/persistent-data/dm-btree-spine.c b/drivers/md/persistent-data/dm-btree-spine.c
+index e653458888a7c..45a39d4f1c10f 100644
+--- a/drivers/md/persistent-data/dm-btree-spine.c
++++ b/drivers/md/persistent-data/dm-btree-spine.c
+@@ -234,12 +234,12 @@ dm_block_t shadow_root(struct shadow_spine *s)
+ 	return s->root;
+ }
+ 
+-static void le64_inc(void *context, const void *value_le, unsigned count)
++static void le64_inc(void *context, const void *value_le, unsigned int count)
+ {
+ 	dm_tm_with_runs(context, value_le, count, dm_tm_inc_range);
+ }
+ 
+-static void le64_dec(void *context, const void *value_le, unsigned count)
++static void le64_dec(void *context, const void *value_le, unsigned int count)
+ {
+ 	dm_tm_with_runs(context, value_le, count, dm_tm_dec_range);
+ }
+diff --git a/drivers/md/persistent-data/dm-btree.c b/drivers/md/persistent-data/dm-btree.c
+index 5ce64e93aae74..1cc783d7030d8 100644
+--- a/drivers/md/persistent-data/dm-btree.c
++++ b/drivers/md/persistent-data/dm-btree.c
+@@ -23,8 +23,8 @@ static void memcpy_disk(void *dest, const void *src, size_t len)
+ 	__dm_unbless_for_disk(src);
+ }
+ 
+-static void array_insert(void *base, size_t elt_size, unsigned nr_elts,
+-			 unsigned index, void *elt)
++static void array_insert(void *base, size_t elt_size, unsigned int nr_elts,
++			 unsigned int index, void *elt)
+ 	__dm_written_to_disk(elt)
+ {
+ 	if (index < nr_elts)
+@@ -80,7 +80,7 @@ void inc_children(struct dm_transaction_manager *tm, struct btree_node *n,
+ 		vt->inc(vt->context, value_ptr(n, 0), nr_entries);
+ }
+ 
+-static int insert_at(size_t value_size, struct btree_node *node, unsigned index,
++static int insert_at(size_t value_size, struct btree_node *node, unsigned int index,
+ 		     uint64_t key, void *value)
+ 	__dm_written_to_disk(value)
+ {
+@@ -162,9 +162,9 @@ EXPORT_SYMBOL_GPL(dm_btree_empty);
+ struct frame {
+ 	struct dm_block *b;
+ 	struct btree_node *n;
+-	unsigned level;
+-	unsigned nr_children;
+-	unsigned current_child;
++	unsigned int level;
++	unsigned int nr_children;
++	unsigned int current_child;
+ };
+ 
+ struct del_stack {
+@@ -193,7 +193,7 @@ static int unprocessed_frames(struct del_stack *s)
+ 
+ static void prefetch_children(struct del_stack *s, struct frame *f)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	struct dm_block_manager *bm = dm_tm_get_bm(s->tm);
+ 
+ 	for (i = 0; i < f->nr_children; i++)
+@@ -205,7 +205,7 @@ static bool is_internal_level(struct dm_btree_info *info, struct frame *f)
+ 	return f->level < (info->levels - 1);
+ }
+ 
+-static int push_frame(struct del_stack *s, dm_block_t b, unsigned level)
++static int push_frame(struct del_stack *s, dm_block_t b, unsigned int level)
+ {
+ 	int r;
+ 	uint32_t ref_count;
+@@ -371,7 +371,7 @@ static int btree_lookup_raw(struct ro_spine *s, dm_block_t block, uint64_t key,
+ int dm_btree_lookup(struct dm_btree_info *info, dm_block_t root,
+ 		    uint64_t *keys, void *value_le)
+ {
+-	unsigned level, last_level = info->levels - 1;
++	unsigned int level, last_level = info->levels - 1;
+ 	int r = -ENODATA;
+ 	uint64_t rkey;
+ 	__le64 internal_value_le;
+@@ -467,7 +467,7 @@ out:
+ int dm_btree_lookup_next(struct dm_btree_info *info, dm_block_t root,
+ 			 uint64_t *keys, uint64_t *rkey, void *value_le)
+ {
+-	unsigned level;
++	unsigned int level;
+ 	int r = -ENODATA;
+ 	__le64 internal_value_le;
+ 	struct ro_spine spine;
+@@ -502,9 +502,9 @@ EXPORT_SYMBOL_GPL(dm_btree_lookup_next);
+  * Copies entries from one region of a btree node to another.  The regions
+  * must not overlap.
+  */
+-static void copy_entries(struct btree_node *dest, unsigned dest_offset,
+-			 struct btree_node *src, unsigned src_offset,
+-			 unsigned count)
++static void copy_entries(struct btree_node *dest, unsigned int dest_offset,
++			 struct btree_node *src, unsigned int src_offset,
++			 unsigned int count)
+ {
+ 	size_t value_size = le32_to_cpu(dest->header.value_size);
+ 	memcpy(dest->keys + dest_offset, src->keys + src_offset, count * sizeof(uint64_t));
+@@ -515,9 +515,9 @@ static void copy_entries(struct btree_node *dest, unsigned dest_offset,
+  * Moves entries from one region fo a btree node to another.  The regions
+  * may overlap.
+  */
+-static void move_entries(struct btree_node *dest, unsigned dest_offset,
+-			 struct btree_node *src, unsigned src_offset,
+-			 unsigned count)
++static void move_entries(struct btree_node *dest, unsigned int dest_offset,
++			 struct btree_node *src, unsigned int src_offset,
++			 unsigned int count)
+ {
+ 	size_t value_size = le32_to_cpu(dest->header.value_size);
+ 	memmove(dest->keys + dest_offset, src->keys + src_offset, count * sizeof(uint64_t));
+@@ -528,7 +528,7 @@ static void move_entries(struct btree_node *dest, unsigned dest_offset,
+  * Erases the first 'count' entries of a btree node, shifting following
+  * entries down into their place.
+  */
+-static void shift_down(struct btree_node *n, unsigned count)
++static void shift_down(struct btree_node *n, unsigned int count)
+ {
+ 	move_entries(n, 0, n, count, le32_to_cpu(n->header.nr_entries) - count);
+ }
+@@ -537,7 +537,7 @@ static void shift_down(struct btree_node *n, unsigned count)
+  * Moves entries in a btree node up 'count' places, making space for
+  * new entries at the start of the node.
+  */
+-static void shift_up(struct btree_node *n, unsigned count)
++static void shift_up(struct btree_node *n, unsigned int count)
+ {
+ 	move_entries(n, count, n, 0, le32_to_cpu(n->header.nr_entries));
+ }
+@@ -548,18 +548,18 @@ static void shift_up(struct btree_node *n, unsigned count)
+  */
+ static void redistribute2(struct btree_node *left, struct btree_node *right)
+ {
+-	unsigned nr_left = le32_to_cpu(left->header.nr_entries);
+-	unsigned nr_right = le32_to_cpu(right->header.nr_entries);
+-	unsigned total = nr_left + nr_right;
+-	unsigned target_left = total / 2;
+-	unsigned target_right = total - target_left;
++	unsigned int nr_left = le32_to_cpu(left->header.nr_entries);
++	unsigned int nr_right = le32_to_cpu(right->header.nr_entries);
++	unsigned int total = nr_left + nr_right;
++	unsigned int target_left = total / 2;
++	unsigned int target_right = total - target_left;
+ 
+ 	if (nr_left < target_left) {
+-		unsigned delta = target_left - nr_left;
++		unsigned int delta = target_left - nr_left;
+ 		copy_entries(left, nr_left, right, 0, delta);
+ 		shift_down(right, delta);
+ 	} else if (nr_left > target_left) {
+-		unsigned delta = nr_left - target_left;
++		unsigned int delta = nr_left - target_left;
+ 		if (nr_right)
+ 			shift_up(right, delta);
+ 		copy_entries(right, 0, left, target_left, delta);
+@@ -576,10 +576,10 @@ static void redistribute2(struct btree_node *left, struct btree_node *right)
+ static void redistribute3(struct btree_node *left, struct btree_node *center,
+ 			  struct btree_node *right)
+ {
+-	unsigned nr_left = le32_to_cpu(left->header.nr_entries);
+-	unsigned nr_center = le32_to_cpu(center->header.nr_entries);
+-	unsigned nr_right = le32_to_cpu(right->header.nr_entries);
+-	unsigned total, target_left, target_center, target_right;
++	unsigned int nr_left = le32_to_cpu(left->header.nr_entries);
++	unsigned int nr_center = le32_to_cpu(center->header.nr_entries);
++	unsigned int nr_right = le32_to_cpu(right->header.nr_entries);
++	unsigned int total, target_left, target_center, target_right;
+ 
+ 	BUG_ON(nr_center);
+ 
+@@ -589,19 +589,19 @@ static void redistribute3(struct btree_node *left, struct btree_node *center,
+ 	target_right = (total - target_left - target_center);
+ 
+ 	if (nr_left < target_left) {
+-		unsigned left_short = target_left - nr_left;
++		unsigned int left_short = target_left - nr_left;
+ 		copy_entries(left, nr_left, right, 0, left_short);
+ 		copy_entries(center, 0, right, left_short, target_center);
+ 		shift_down(right, nr_right - target_right);
+ 
+ 	} else if (nr_left < (target_left + target_center)) {
+-		unsigned left_to_center = nr_left - target_left;
++		unsigned int left_to_center = nr_left - target_left;
+ 		copy_entries(center, 0, left, target_left, left_to_center);
+ 		copy_entries(center, left_to_center, right, 0, target_center - left_to_center);
+ 		shift_down(right, nr_right - target_right);
+ 
+ 	} else {
+-		unsigned right_short = target_right - nr_right;
++		unsigned int right_short = target_right - nr_right;
+ 		shift_up(right, right_short);
+ 		copy_entries(right, 0, left, nr_left - right_short, right_short);
+ 		copy_entries(center, 0, left, target_left, nr_left - target_left);
+@@ -642,7 +642,7 @@ static void redistribute3(struct btree_node *left, struct btree_node *center,
+  *
+  * Where A* is a shadow of A.
+  */
+-static int split_one_into_two(struct shadow_spine *s, unsigned parent_index,
++static int split_one_into_two(struct shadow_spine *s, unsigned int parent_index,
+ 			      struct dm_btree_value_type *vt, uint64_t key)
+ {
+ 	int r;
+@@ -696,7 +696,7 @@ static int split_one_into_two(struct shadow_spine *s, unsigned parent_index,
+  * to the new shadow.
+  */
+ static int shadow_child(struct dm_btree_info *info, struct dm_btree_value_type *vt,
+-			struct btree_node *parent, unsigned index,
++			struct btree_node *parent, unsigned int index,
+ 			struct dm_block **result)
+ {
+ 	int r, inc;
+@@ -725,11 +725,11 @@ static int shadow_child(struct dm_btree_info *info, struct dm_btree_value_type *
+  * Splits two nodes into three.  This is more work, but results in fuller
+  * nodes, so saves metadata space.
+  */
+-static int split_two_into_three(struct shadow_spine *s, unsigned parent_index,
++static int split_two_into_three(struct shadow_spine *s, unsigned int parent_index,
+                                 struct dm_btree_value_type *vt, uint64_t key)
+ {
+ 	int r;
+-	unsigned middle_index;
++	unsigned int middle_index;
+ 	struct dm_block *left, *middle, *right, *parent;
+ 	struct btree_node *ln, *rn, *mn, *pn;
+ 	__le64 location;
+@@ -830,7 +830,7 @@ static int btree_split_beneath(struct shadow_spine *s, uint64_t key)
+ {
+ 	int r;
+ 	size_t size;
+-	unsigned nr_left, nr_right;
++	unsigned int nr_left, nr_right;
+ 	struct dm_block *left, *right, *new_parent;
+ 	struct btree_node *pn, *ln, *rn;
+ 	__le64 val;
+@@ -904,7 +904,7 @@ static int btree_split_beneath(struct shadow_spine *s, uint64_t key)
+  * Redistributes a node's entries with its left sibling.
+  */
+ static int rebalance_left(struct shadow_spine *s, struct dm_btree_value_type *vt,
+-			  unsigned parent_index, uint64_t key)
++			  unsigned int parent_index, uint64_t key)
+ {
+ 	int r;
+ 	struct dm_block *sib;
+@@ -933,7 +933,7 @@ static int rebalance_left(struct shadow_spine *s, struct dm_btree_value_type *vt
+  * Redistributes a nodes entries with its right sibling.
+  */
+ static int rebalance_right(struct shadow_spine *s, struct dm_btree_value_type *vt,
+-			   unsigned parent_index, uint64_t key)
++			   unsigned int parent_index, uint64_t key)
+ {
+ 	int r;
+ 	struct dm_block *sib;
+@@ -961,10 +961,10 @@ static int rebalance_right(struct shadow_spine *s, struct dm_btree_value_type *v
+ /*
+  * Returns the number of spare entries in a node.
+  */
+-static int get_node_free_space(struct dm_btree_info *info, dm_block_t b, unsigned *space)
++static int get_node_free_space(struct dm_btree_info *info, dm_block_t b, unsigned int *space)
+ {
+ 	int r;
+-	unsigned nr_entries;
++	unsigned int nr_entries;
+ 	struct dm_block *block;
+ 	struct btree_node *node;
+ 
+@@ -990,12 +990,12 @@ static int get_node_free_space(struct dm_btree_info *info, dm_block_t b, unsigne
+  */
+ #define SPACE_THRESHOLD 8
+ static int rebalance_or_split(struct shadow_spine *s, struct dm_btree_value_type *vt,
+-			      unsigned parent_index, uint64_t key)
++			      unsigned int parent_index, uint64_t key)
+ {
+ 	int r;
+ 	struct btree_node *parent = dm_block_data(shadow_parent(s));
+-	unsigned nr_parent = le32_to_cpu(parent->header.nr_entries);
+-	unsigned free_space;
++	unsigned int nr_parent = le32_to_cpu(parent->header.nr_entries);
++	unsigned int free_space;
+ 	int left_shared = 0, right_shared = 0;
+ 
+ 	/* Should we move entries to the left sibling? */
+@@ -1080,7 +1080,7 @@ static bool has_space_for_insert(struct btree_node *node, uint64_t key)
+ 
+ static int btree_insert_raw(struct shadow_spine *s, dm_block_t root,
+ 			    struct dm_btree_value_type *vt,
+-			    uint64_t key, unsigned *index)
++			    uint64_t key, unsigned int *index)
+ {
+ 	int r, i = *index, top = 1;
+ 	struct btree_node *node;
+@@ -1214,7 +1214,7 @@ int btree_get_overwrite_leaf(struct dm_btree_info *info, dm_block_t root,
+ }
+ 
+ static bool need_insert(struct btree_node *node, uint64_t *keys,
+-			unsigned level, unsigned index)
++			unsigned int level, unsigned int index)
+ {
+         return ((index >= le32_to_cpu(node->header.nr_entries)) ||
+ 		(le64_to_cpu(node->keys[index]) != keys[level]));
+@@ -1226,7 +1226,7 @@ static int insert(struct dm_btree_info *info, dm_block_t root,
+ 		  __dm_written_to_disk(value)
+ {
+ 	int r;
+-	unsigned level, index = -1, last_level = info->levels - 1;
++	unsigned int level, index = -1, last_level = info->levels - 1;
+ 	dm_block_t block = root;
+ 	struct shadow_spine spine;
+ 	struct btree_node *n;
+@@ -1412,7 +1412,7 @@ static int walk_node(struct dm_btree_info *info, dm_block_t block,
+ 		     void *context)
+ {
+ 	int r;
+-	unsigned i, nr;
++	unsigned int i, nr;
+ 	struct dm_block *node;
+ 	struct btree_node *n;
+ 	uint64_t keys;
+@@ -1455,7 +1455,7 @@ EXPORT_SYMBOL_GPL(dm_btree_walk);
+ 
+ static void prefetch_values(struct dm_btree_cursor *c)
+ {
+-	unsigned i, nr;
++	unsigned int i, nr;
+ 	__le64 value_le;
+ 	struct cursor_node *n = c->nodes + c->depth - 1;
+ 	struct btree_node *bn = dm_block_data(n->b);
+diff --git a/drivers/md/persistent-data/dm-btree.h b/drivers/md/persistent-data/dm-btree.h
+index d2ae5aa4d00b6..5566e7c32e829 100644
+--- a/drivers/md/persistent-data/dm-btree.h
++++ b/drivers/md/persistent-data/dm-btree.h
+@@ -58,14 +58,14 @@ struct dm_btree_value_type {
+ 	 * somewhere.) This method is _not_ called for insertion of a new
+ 	 * value: It is assumed the ref count is already 1.
+ 	 */
+-	void (*inc)(void *context, const void *value, unsigned count);
++	void (*inc)(void *context, const void *value, unsigned int count);
+ 
+ 	/*
+ 	 * These values are being deleted.  The btree takes care of freeing
+ 	 * the memory pointed to by @value.  Often the del function just
+ 	 * needs to decrement a reference counts somewhere.
+ 	 */
+-	void (*dec)(void *context, const void *value, unsigned count);
++	void (*dec)(void *context, const void *value, unsigned int count);
+ 
+ 	/*
+ 	 * A test for equality between two values.  When a value is
+@@ -84,7 +84,7 @@ struct dm_btree_info {
+ 	/*
+ 	 * Number of nested btrees. (Not the depth of a single tree.)
+ 	 */
+-	unsigned levels;
++	unsigned int levels;
+ 	struct dm_btree_value_type value_type;
+ };
+ 
+@@ -149,7 +149,7 @@ int dm_btree_remove(struct dm_btree_info *info, dm_block_t root,
+  */
+ int dm_btree_remove_leaves(struct dm_btree_info *info, dm_block_t root,
+ 			   uint64_t *keys, uint64_t end_key,
+-			   dm_block_t *new_root, unsigned *nr_removed);
++			   dm_block_t *new_root, unsigned int *nr_removed);
+ 
+ /*
+  * Returns < 0 on failure.  Otherwise the number of key entries that have
+@@ -188,7 +188,7 @@ int dm_btree_walk(struct dm_btree_info *info, dm_block_t root,
+ 
+ struct cursor_node {
+ 	struct dm_block *b;
+-	unsigned index;
++	unsigned int index;
+ };
+ 
+ struct dm_btree_cursor {
+@@ -196,7 +196,7 @@ struct dm_btree_cursor {
+ 	dm_block_t root;
+ 
+ 	bool prefetch_leaves;
+-	unsigned depth;
++	unsigned int depth;
+ 	struct cursor_node nodes[DM_BTREE_CURSOR_MAX_DEPTH];
+ };
+ 
+diff --git a/drivers/md/persistent-data/dm-persistent-data-internal.h b/drivers/md/persistent-data/dm-persistent-data-internal.h
+index c49e26fff36c8..b945a2be93fb2 100644
+--- a/drivers/md/persistent-data/dm-persistent-data-internal.h
++++ b/drivers/md/persistent-data/dm-persistent-data-internal.h
+@@ -9,11 +9,11 @@
+ 
+ #include "dm-block-manager.h"
+ 
+-static inline unsigned dm_hash_block(dm_block_t b, unsigned hash_mask)
++static inline unsigned int dm_hash_block(dm_block_t b, unsigned int hash_mask)
+ {
+-	const unsigned BIG_PRIME = 4294967291UL;
++	const unsigned int BIG_PRIME = 4294967291UL;
+ 
+-	return (((unsigned) b) * BIG_PRIME) & hash_mask;
++	return (((unsigned int) b) * BIG_PRIME) & hash_mask;
+ }
+ 
+ #endif	/* _PERSISTENT_DATA_INTERNAL_H */
+diff --git a/drivers/md/persistent-data/dm-space-map-common.c b/drivers/md/persistent-data/dm-space-map-common.c
+index bfbfa750e0160..af800efed9f3c 100644
+--- a/drivers/md/persistent-data/dm-space-map-common.c
++++ b/drivers/md/persistent-data/dm-space-map-common.c
+@@ -126,7 +126,7 @@ static void *dm_bitmap_data(struct dm_block *b)
+ 
+ #define WORD_MASK_HIGH 0xAAAAAAAAAAAAAAAAULL
+ 
+-static unsigned dm_bitmap_word_used(void *addr, unsigned b)
++static unsigned int dm_bitmap_word_used(void *addr, unsigned int b)
+ {
+ 	__le64 *words_le = addr;
+ 	__le64 *w_le = words_le + (b >> ENTRIES_SHIFT);
+@@ -137,11 +137,11 @@ static unsigned dm_bitmap_word_used(void *addr, unsigned b)
+ 	return !(~bits & mask);
+ }
+ 
+-static unsigned sm_lookup_bitmap(void *addr, unsigned b)
++static unsigned int sm_lookup_bitmap(void *addr, unsigned int b)
+ {
+ 	__le64 *words_le = addr;
+ 	__le64 *w_le = words_le + (b >> ENTRIES_SHIFT);
+-	unsigned hi, lo;
++	unsigned int hi, lo;
+ 
+ 	b = (b & (ENTRIES_PER_WORD - 1)) << 1;
+ 	hi = !!test_bit_le(b, (void *) w_le);
+@@ -149,7 +149,7 @@ static unsigned sm_lookup_bitmap(void *addr, unsigned b)
+ 	return (hi << 1) | lo;
+ }
+ 
+-static void sm_set_bitmap(void *addr, unsigned b, unsigned val)
++static void sm_set_bitmap(void *addr, unsigned int b, unsigned int val)
+ {
+ 	__le64 *words_le = addr;
+ 	__le64 *w_le = words_le + (b >> ENTRIES_SHIFT);
+@@ -167,8 +167,8 @@ static void sm_set_bitmap(void *addr, unsigned b, unsigned val)
+ 		__clear_bit_le(b + 1, (void *) w_le);
+ }
+ 
+-static int sm_find_free(void *addr, unsigned begin, unsigned end,
+-			unsigned *result)
++static int sm_find_free(void *addr, unsigned int begin, unsigned int end,
++			unsigned int *result)
+ {
+ 	while (begin < end) {
+ 		if (!(begin & (ENTRIES_PER_WORD - 1)) &&
+@@ -237,7 +237,7 @@ int sm_ll_extend(struct ll_disk *ll, dm_block_t extra_blocks)
+ {
+ 	int r;
+ 	dm_block_t i, nr_blocks, nr_indexes;
+-	unsigned old_blocks, blocks;
++	unsigned int old_blocks, blocks;
+ 
+ 	nr_blocks = ll->nr_blocks + extra_blocks;
+ 	old_blocks = dm_sector_div_up(ll->nr_blocks, ll->entries_per_block);
+@@ -351,7 +351,7 @@ int sm_ll_find_free_block(struct ll_disk *ll, dm_block_t begin,
+ 
+ 	for (i = index_begin; i < index_end; i++, begin = 0) {
+ 		struct dm_block *blk;
+-		unsigned position;
++		unsigned int position;
+ 		uint32_t bit_end;
+ 
+ 		r = ll->load_ie(ll, i, &ie_disk);
+@@ -369,7 +369,7 @@ int sm_ll_find_free_block(struct ll_disk *ll, dm_block_t begin,
+ 		bit_end = (i == index_end - 1) ?  end : ll->entries_per_block;
+ 
+ 		r = sm_find_free(dm_bitmap_data(blk),
+-				 max_t(unsigned, begin, le32_to_cpu(ie_disk.none_free_before)),
++				 max_t(unsigned int, begin, le32_to_cpu(ie_disk.none_free_before)),
+ 				 bit_end, &position);
+ 		if (r == -ENOSPC) {
+ 			/*
+@@ -1097,7 +1097,7 @@ static inline int ie_cache_writeback(struct ll_disk *ll, struct ie_cache *iec)
+ 			       &iec->index, &iec->ie, &ll->bitmap_root);
+ }
+ 
+-static inline unsigned hash_index(dm_block_t index)
++static inline unsigned int hash_index(dm_block_t index)
+ {
+ 	return dm_hash_block(index, IE_CACHE_MASK);
+ }
+@@ -1106,7 +1106,7 @@ static int disk_ll_load_ie(struct ll_disk *ll, dm_block_t index,
+ 			   struct disk_index_entry *ie)
+ {
+ 	int r;
+-	unsigned h = hash_index(index);
++	unsigned int h = hash_index(index);
+ 	struct ie_cache *iec = ll->ie_cache + h;
+ 
+ 	if (iec->valid) {
+@@ -1137,7 +1137,7 @@ static int disk_ll_save_ie(struct ll_disk *ll, dm_block_t index,
+ 			   struct disk_index_entry *ie)
+ {
+ 	int r;
+-	unsigned h = hash_index(index);
++	unsigned int h = hash_index(index);
+ 	struct ie_cache *iec = ll->ie_cache + h;
+ 
+ 	ll->bitmap_index_changed = true;
+@@ -1164,7 +1164,7 @@ static int disk_ll_save_ie(struct ll_disk *ll, dm_block_t index,
+ 
+ static int disk_ll_init_index(struct ll_disk *ll)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	for (i = 0; i < IE_CACHE_SIZE; i++) {
+ 		struct ie_cache *iec = ll->ie_cache + i;
+ 		iec->valid = false;
+@@ -1186,7 +1186,7 @@ static dm_block_t disk_ll_max_entries(struct ll_disk *ll)
+ static int disk_ll_commit(struct ll_disk *ll)
+ {
+ 	int r = 0;
+-	unsigned i;
++	unsigned int i;
+ 
+ 	for (i = 0; i < IE_CACHE_SIZE; i++) {
+ 		struct ie_cache *iec = ll->ie_cache + i;
+diff --git a/drivers/md/persistent-data/dm-space-map-metadata.c b/drivers/md/persistent-data/dm-space-map-metadata.c
+index 392ae26134a4e..0d1fcdf29c835 100644
+--- a/drivers/md/persistent-data/dm-space-map-metadata.c
++++ b/drivers/md/persistent-data/dm-space-map-metadata.c
+@@ -94,8 +94,8 @@ struct block_op {
+ };
+ 
+ struct bop_ring_buffer {
+-	unsigned begin;
+-	unsigned end;
++	unsigned int begin;
++	unsigned int end;
+ 	struct block_op bops[MAX_RECURSIVE_ALLOCATIONS + 1];
+ };
+ 
+@@ -110,9 +110,9 @@ static bool brb_empty(struct bop_ring_buffer *brb)
+ 	return brb->begin == brb->end;
+ }
+ 
+-static unsigned brb_next(struct bop_ring_buffer *brb, unsigned old)
++static unsigned int brb_next(struct bop_ring_buffer *brb, unsigned int old)
+ {
+-	unsigned r = old + 1;
++	unsigned int r = old + 1;
+ 	return r >= ARRAY_SIZE(brb->bops) ? 0 : r;
+ }
+ 
+@@ -120,7 +120,7 @@ static int brb_push(struct bop_ring_buffer *brb,
+ 		    enum block_op_type type, dm_block_t b, dm_block_t e)
+ {
+ 	struct block_op *bop;
+-	unsigned next = brb_next(brb, brb->end);
++	unsigned int next = brb_next(brb, brb->end);
+ 
+ 	/*
+ 	 * We don't allow the last bop to be filled, this way we can
+@@ -171,8 +171,8 @@ struct sm_metadata {
+ 
+ 	dm_block_t begin;
+ 
+-	unsigned recursion_count;
+-	unsigned allocated_this_transaction;
++	unsigned int recursion_count;
++	unsigned int allocated_this_transaction;
+ 	struct bop_ring_buffer uncommitted;
+ 
+ 	struct threshold threshold;
+@@ -300,9 +300,9 @@ static int sm_metadata_get_count(struct dm_space_map *sm, dm_block_t b,
+ 				 uint32_t *result)
+ {
+ 	int r;
+-	unsigned i;
++	unsigned int i;
+ 	struct sm_metadata *smm = container_of(sm, struct sm_metadata, sm);
+-	unsigned adjustment = 0;
++	unsigned int adjustment = 0;
+ 
+ 	/*
+ 	 * We may have some uncommitted adjustments to add.  This list
+@@ -340,7 +340,7 @@ static int sm_metadata_count_is_more_than_one(struct dm_space_map *sm,
+ 					      dm_block_t b, int *result)
+ {
+ 	int r, adjustment = 0;
+-	unsigned i;
++	unsigned int i;
+ 	struct sm_metadata *smm = container_of(sm, struct sm_metadata, sm);
+ 	uint32_t rc;
+ 
+diff --git a/drivers/md/persistent-data/dm-transaction-manager.c b/drivers/md/persistent-data/dm-transaction-manager.c
+index 16643fc974e84..39885f8355847 100644
+--- a/drivers/md/persistent-data/dm-transaction-manager.c
++++ b/drivers/md/persistent-data/dm-transaction-manager.c
+@@ -28,14 +28,14 @@ struct prefetch_set {
+ 	dm_block_t blocks[PREFETCH_SIZE];
+ };
+ 
+-static unsigned prefetch_hash(dm_block_t b)
++static unsigned int prefetch_hash(dm_block_t b)
+ {
+ 	return hash_64(b, PREFETCH_BITS);
+ }
+ 
+ static void prefetch_wipe(struct prefetch_set *p)
+ {
+-	unsigned i;
++	unsigned int i;
+ 	for (i = 0; i < PREFETCH_SIZE; i++)
+ 		p->blocks[i] = PREFETCH_SENTINEL;
+ }
+@@ -48,7 +48,7 @@ static void prefetch_init(struct prefetch_set *p)
+ 
+ static void prefetch_add(struct prefetch_set *p, dm_block_t b)
+ {
+-	unsigned h = prefetch_hash(b);
++	unsigned int h = prefetch_hash(b);
+ 
+ 	mutex_lock(&p->lock);
+ 	if (p->blocks[h] == PREFETCH_SENTINEL)
+@@ -59,7 +59,7 @@ static void prefetch_add(struct prefetch_set *p, dm_block_t b)
+ 
+ static void prefetch_issue(struct prefetch_set *p, struct dm_block_manager *bm)
+ {
+-	unsigned i;
++	unsigned int i;
+ 
+ 	mutex_lock(&p->lock);
+ 
+@@ -103,7 +103,7 @@ struct dm_transaction_manager {
+ static int is_shadow(struct dm_transaction_manager *tm, dm_block_t b)
+ {
+ 	int r = 0;
+-	unsigned bucket = dm_hash_block(b, DM_HASH_MASK);
++	unsigned int bucket = dm_hash_block(b, DM_HASH_MASK);
+ 	struct shadow_info *si;
+ 
+ 	spin_lock(&tm->lock);
+@@ -123,7 +123,7 @@ static int is_shadow(struct dm_transaction_manager *tm, dm_block_t b)
+  */
+ static void insert_shadow(struct dm_transaction_manager *tm, dm_block_t b)
+ {
+-	unsigned bucket;
++	unsigned int bucket;
+ 	struct shadow_info *si;
+ 
+ 	si = kmalloc(sizeof(*si), GFP_NOIO);
+@@ -393,11 +393,11 @@ void dm_tm_dec_range(struct dm_transaction_manager *tm, dm_block_t b, dm_block_t
+ EXPORT_SYMBOL_GPL(dm_tm_dec_range);
+ 
+ void dm_tm_with_runs(struct dm_transaction_manager *tm,
+-		     const __le64 *value_le, unsigned count, dm_tm_run_fn fn)
++		     const __le64 *value_le, unsigned int count, dm_tm_run_fn fn)
+ {
+ 	uint64_t b, begin, end;
+ 	bool in_run = false;
+-	unsigned i;
++	unsigned int i;
+ 
+ 	for (i = 0; i < count; i++, value_le++) {
+ 		b = le64_to_cpu(*value_le);
+diff --git a/drivers/md/persistent-data/dm-transaction-manager.h b/drivers/md/persistent-data/dm-transaction-manager.h
+index 906c02ed0365b..0f573a4a01aeb 100644
+--- a/drivers/md/persistent-data/dm-transaction-manager.h
++++ b/drivers/md/persistent-data/dm-transaction-manager.h
+@@ -111,7 +111,7 @@ void dm_tm_dec_range(struct dm_transaction_manager *tm, dm_block_t b, dm_block_t
+  */
+ typedef void (*dm_tm_run_fn)(struct dm_transaction_manager *, dm_block_t, dm_block_t);
+ void dm_tm_with_runs(struct dm_transaction_manager *tm,
+-		     const __le64 *value_le, unsigned count, dm_tm_run_fn fn);
++		     const __le64 *value_le, unsigned int count, dm_tm_run_fn fn);
+ 
+ int dm_tm_ref(struct dm_transaction_manager *tm, dm_block_t b, uint32_t *result);
+ 
+diff --git a/drivers/net/dsa/mv88e6xxx/chip.c b/drivers/net/dsa/mv88e6xxx/chip.c
+index 8cf27e2654fcf..f1d9ee2a78b0f 100644
+--- a/drivers/net/dsa/mv88e6xxx/chip.c
++++ b/drivers/net/dsa/mv88e6xxx/chip.c
+@@ -5514,7 +5514,7 @@ static const struct mv88e6xxx_ops mv88e6393x_ops = {
+ 	 * .port_set_upstream_port method.
+ 	 */
+ 	.set_egress_port = mv88e6393x_set_egress_port,
+-	.watchdog_ops = &mv88e6390_watchdog_ops,
++	.watchdog_ops = &mv88e6393x_watchdog_ops,
+ 	.mgmt_rsvd2cpu = mv88e6393x_port_mgmt_rsvd2cpu,
+ 	.pot_clear = mv88e6xxx_g2_pot_clear,
+ 	.reset = mv88e6352_g1_reset,
+diff --git a/drivers/net/dsa/mv88e6xxx/global2.c b/drivers/net/dsa/mv88e6xxx/global2.c
+index fa65ecd9cb853..ec49939968fac 100644
+--- a/drivers/net/dsa/mv88e6xxx/global2.c
++++ b/drivers/net/dsa/mv88e6xxx/global2.c
+@@ -931,6 +931,26 @@ const struct mv88e6xxx_irq_ops mv88e6390_watchdog_ops = {
+ 	.irq_free = mv88e6390_watchdog_free,
+ };
+ 
++static int mv88e6393x_watchdog_action(struct mv88e6xxx_chip *chip, int irq)
++{
++	mv88e6390_watchdog_action(chip, irq);
++
++	/* Fix for clearing the force WD event bit.
++	 * Unreleased erratum on mv88e6393x.
++	 */
++	mv88e6xxx_g2_write(chip, MV88E6390_G2_WDOG_CTL,
++			   MV88E6390_G2_WDOG_CTL_UPDATE |
++			   MV88E6390_G2_WDOG_CTL_PTR_EVENT);
++
++	return IRQ_HANDLED;
++}
++
++const struct mv88e6xxx_irq_ops mv88e6393x_watchdog_ops = {
++	.irq_action = mv88e6393x_watchdog_action,
++	.irq_setup = mv88e6390_watchdog_setup,
++	.irq_free = mv88e6390_watchdog_free,
++};
++
+ static irqreturn_t mv88e6xxx_g2_watchdog_thread_fn(int irq, void *dev_id)
+ {
+ 	struct mv88e6xxx_chip *chip = dev_id;
+diff --git a/drivers/net/dsa/mv88e6xxx/global2.h b/drivers/net/dsa/mv88e6xxx/global2.h
+index 7536b8b0ad011..c05fad5c9f19d 100644
+--- a/drivers/net/dsa/mv88e6xxx/global2.h
++++ b/drivers/net/dsa/mv88e6xxx/global2.h
+@@ -363,6 +363,7 @@ int mv88e6xxx_g2_device_mapping_write(struct mv88e6xxx_chip *chip, int target,
+ extern const struct mv88e6xxx_irq_ops mv88e6097_watchdog_ops;
+ extern const struct mv88e6xxx_irq_ops mv88e6250_watchdog_ops;
+ extern const struct mv88e6xxx_irq_ops mv88e6390_watchdog_ops;
++extern const struct mv88e6xxx_irq_ops mv88e6393x_watchdog_ops;
+ 
+ extern const struct mv88e6xxx_avb_ops mv88e6165_avb_ops;
+ extern const struct mv88e6xxx_avb_ops mv88e6352_avb_ops;
+diff --git a/drivers/net/ethernet/google/gve/gve.h b/drivers/net/ethernet/google/gve/gve.h
+index 160735484465a..458149a77ebe6 100644
+--- a/drivers/net/ethernet/google/gve/gve.h
++++ b/drivers/net/ethernet/google/gve/gve.h
+@@ -47,6 +47,8 @@
+ 
+ #define GVE_RX_BUFFER_SIZE_DQO 2048
+ 
++#define GVE_GQ_TX_MIN_PKT_DESC_BYTES 182
++
+ /* Each slot in the desc ring has a 1:1 mapping to a slot in the data ring */
+ struct gve_rx_desc_queue {
+ 	struct gve_rx_desc *desc_ring; /* the descriptor ring */
+diff --git a/drivers/net/ethernet/google/gve/gve_tx.c b/drivers/net/ethernet/google/gve/gve_tx.c
+index 4888bf05fbedb..5e11b82367545 100644
+--- a/drivers/net/ethernet/google/gve/gve_tx.c
++++ b/drivers/net/ethernet/google/gve/gve_tx.c
+@@ -284,8 +284,8 @@ static inline int gve_skb_fifo_bytes_required(struct gve_tx_ring *tx,
+ 	int bytes;
+ 	int hlen;
+ 
+-	hlen = skb_is_gso(skb) ? skb_checksum_start_offset(skb) +
+-				 tcp_hdrlen(skb) : skb_headlen(skb);
++	hlen = skb_is_gso(skb) ? skb_checksum_start_offset(skb) + tcp_hdrlen(skb) :
++				 min_t(int, GVE_GQ_TX_MIN_PKT_DESC_BYTES, skb->len);
+ 
+ 	pad_bytes = gve_tx_fifo_pad_alloc_one_frag(&tx->tx_fifo,
+ 						   hlen);
+@@ -454,13 +454,11 @@ static int gve_tx_add_skb_copy(struct gve_priv *priv, struct gve_tx_ring *tx, st
+ 	pkt_desc = &tx->desc[idx];
+ 
+ 	l4_hdr_offset = skb_checksum_start_offset(skb);
+-	/* If the skb is gso, then we want the tcp header in the first segment
+-	 * otherwise we want the linear portion of the skb (which will contain
+-	 * the checksum because skb->csum_start and skb->csum_offset are given
+-	 * relative to skb->head) in the first segment.
++	/* If the skb is gso, then we want the tcp header alone in the first segment
++	 * otherwise we want the minimum required by the gVNIC spec.
+ 	 */
+ 	hlen = is_gso ? l4_hdr_offset + tcp_hdrlen(skb) :
+-			skb_headlen(skb);
++			min_t(int, GVE_GQ_TX_MIN_PKT_DESC_BYTES, skb->len);
+ 
+ 	info->skb =  skb;
+ 	/* We don't want to split the header, so if necessary, pad to the end
+diff --git a/drivers/net/ethernet/intel/ice/ice_virtchnl_fdir.c b/drivers/net/ethernet/intel/ice/ice_virtchnl_fdir.c
+index a2645ff3100e4..7f72604079723 100644
+--- a/drivers/net/ethernet/intel/ice/ice_virtchnl_fdir.c
++++ b/drivers/net/ethernet/intel/ice/ice_virtchnl_fdir.c
+@@ -541,6 +541,21 @@ static void ice_vc_fdir_rem_prof_all(struct ice_vf *vf)
+ 	}
+ }
+ 
++/**
++ * ice_vc_fdir_reset_cnt_all - reset all FDIR counters for this VF FDIR
++ * @fdir: pointer to the VF FDIR structure
++ */
++static void ice_vc_fdir_reset_cnt_all(struct ice_vf_fdir *fdir)
++{
++	enum ice_fltr_ptype flow;
++
++	for (flow = ICE_FLTR_PTYPE_NONF_NONE;
++	     flow < ICE_FLTR_PTYPE_MAX; flow++) {
++		fdir->fdir_fltr_cnt[flow][0] = 0;
++		fdir->fdir_fltr_cnt[flow][1] = 0;
++	}
++}
++
+ /**
+  * ice_vc_fdir_has_prof_conflict
+  * @vf: pointer to the VF structure
+@@ -1871,7 +1886,7 @@ int ice_vc_add_fdir_fltr(struct ice_vf *vf, u8 *msg)
+ 		v_ret = VIRTCHNL_STATUS_SUCCESS;
+ 		stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
+ 		dev_dbg(dev, "VF %d: set FDIR context failed\n", vf->vf_id);
+-		goto err_free_conf;
++		goto err_rem_entry;
+ 	}
+ 
+ 	ret = ice_vc_fdir_write_fltr(vf, conf, true, is_tun);
+@@ -1880,15 +1895,16 @@ int ice_vc_add_fdir_fltr(struct ice_vf *vf, u8 *msg)
+ 		stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
+ 		dev_err(dev, "VF %d: writing FDIR rule failed, ret:%d\n",
+ 			vf->vf_id, ret);
+-		goto err_rem_entry;
++		goto err_clr_irq;
+ 	}
+ 
+ exit:
+ 	kfree(stat);
+ 	return ret;
+ 
+-err_rem_entry:
++err_clr_irq:
+ 	ice_vc_fdir_clear_irq_ctx(vf);
++err_rem_entry:
+ 	ice_vc_fdir_remove_entry(vf, conf, conf->flow_id);
+ err_free_conf:
+ 	devm_kfree(dev, conf);
+@@ -1997,6 +2013,7 @@ void ice_vf_fdir_init(struct ice_vf *vf)
+ 	spin_lock_init(&fdir->ctx_lock);
+ 	fdir->ctx_irq.flags = 0;
+ 	fdir->ctx_done.flags = 0;
++	ice_vc_fdir_reset_cnt_all(fdir);
+ }
+ 
+ /**
+diff --git a/drivers/net/ethernet/stmicro/stmmac/dwmac-intel.c b/drivers/net/ethernet/stmicro/stmmac/dwmac-intel.c
+index 13aa919633b47..ab9f876b6df7e 100644
+--- a/drivers/net/ethernet/stmicro/stmmac/dwmac-intel.c
++++ b/drivers/net/ethernet/stmicro/stmmac/dwmac-intel.c
+@@ -251,7 +251,6 @@ static void intel_speed_mode_2500(struct net_device *ndev, void *intel_data)
+ 		priv->plat->mdio_bus_data->xpcs_an_inband = false;
+ 	} else {
+ 		priv->plat->max_speed = 1000;
+-		priv->plat->mdio_bus_data->xpcs_an_inband = true;
+ 	}
+ }
+ 
+diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
+index 93321437f0933..fa3ce3b0d9a56 100644
+--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
++++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
+@@ -1132,20 +1132,26 @@ static void stmmac_check_pcs_mode(struct stmmac_priv *priv)
+ static int stmmac_init_phy(struct net_device *dev)
+ {
+ 	struct stmmac_priv *priv = netdev_priv(dev);
++	struct fwnode_handle *phy_fwnode;
+ 	struct fwnode_handle *fwnode;
+ 	int ret;
+ 
++	if (!phylink_expects_phy(priv->phylink))
++		return 0;
++
+ 	fwnode = of_fwnode_handle(priv->plat->phylink_node);
+ 	if (!fwnode)
+ 		fwnode = dev_fwnode(priv->device);
+ 
+ 	if (fwnode)
+-		ret = phylink_fwnode_phy_connect(priv->phylink, fwnode, 0);
++		phy_fwnode = fwnode_get_phy_node(fwnode);
++	else
++		phy_fwnode = NULL;
+ 
+ 	/* Some DT bindings do not set-up the PHY handle. Let's try to
+ 	 * manually parse it
+ 	 */
+-	if (!fwnode || ret) {
++	if (!phy_fwnode || IS_ERR(phy_fwnode)) {
+ 		int addr = priv->plat->phy_addr;
+ 		struct phy_device *phydev;
+ 
+@@ -1161,6 +1167,9 @@ static int stmmac_init_phy(struct net_device *dev)
+ 		}
+ 
+ 		ret = phylink_connect_phy(priv->phylink, phydev);
++	} else {
++		fwnode_handle_put(phy_fwnode);
++		ret = phylink_fwnode_phy_connect(priv->phylink, fwnode, 0);
+ 	}
+ 
+ 	if (!priv->plat->pmt) {
+@@ -6618,6 +6627,8 @@ int stmmac_xdp_open(struct net_device *dev)
+ 		goto init_error;
+ 	}
+ 
++	stmmac_reset_queues_param(priv);
++
+ 	/* DMA CSR Channel configuration */
+ 	for (chan = 0; chan < dma_csr_ch; chan++) {
+ 		stmmac_init_chan(priv, priv->ioaddr, priv->plat->dma_cfg, chan);
+@@ -6944,7 +6955,7 @@ static void stmmac_napi_del(struct net_device *dev)
+ int stmmac_reinit_queues(struct net_device *dev, u32 rx_cnt, u32 tx_cnt)
+ {
+ 	struct stmmac_priv *priv = netdev_priv(dev);
+-	int ret = 0;
++	int ret = 0, i;
+ 
+ 	if (netif_running(dev))
+ 		stmmac_release(dev);
+@@ -6953,6 +6964,10 @@ int stmmac_reinit_queues(struct net_device *dev, u32 rx_cnt, u32 tx_cnt)
+ 
+ 	priv->plat->rx_queues_to_use = rx_cnt;
+ 	priv->plat->tx_queues_to_use = tx_cnt;
++	if (!netif_is_rxfh_configured(dev))
++		for (i = 0; i < ARRAY_SIZE(priv->rss.table); i++)
++			priv->rss.table[i] = ethtool_rxfh_indir_default(i,
++									rx_cnt);
+ 
+ 	stmmac_napi_add(dev);
+ 
+diff --git a/drivers/net/ethernet/ti/am65-cpsw-nuss.c b/drivers/net/ethernet/ti/am65-cpsw-nuss.c
+index 8ff1c84a23ce7..25466cbdc16bd 100644
+--- a/drivers/net/ethernet/ti/am65-cpsw-nuss.c
++++ b/drivers/net/ethernet/ti/am65-cpsw-nuss.c
+@@ -2819,7 +2819,8 @@ err_free_phylink:
+ 	am65_cpsw_nuss_phylink_cleanup(common);
+ 	am65_cpts_release(common->cpts);
+ err_of_clear:
+-	of_platform_device_destroy(common->mdio_dev, NULL);
++	if (common->mdio_dev)
++		of_platform_device_destroy(common->mdio_dev, NULL);
+ err_pm_clear:
+ 	pm_runtime_put_sync(dev);
+ 	pm_runtime_disable(dev);
+@@ -2848,7 +2849,8 @@ static int am65_cpsw_nuss_remove(struct platform_device *pdev)
+ 	am65_cpsw_nuss_phylink_cleanup(common);
+ 	am65_cpts_release(common->cpts);
+ 
+-	of_platform_device_destroy(common->mdio_dev, NULL);
++	if (common->mdio_dev)
++		of_platform_device_destroy(common->mdio_dev, NULL);
+ 
+ 	pm_runtime_put_sync(&pdev->dev);
+ 	pm_runtime_disable(&pdev->dev);
+diff --git a/drivers/net/phy/phylink.c b/drivers/net/phy/phylink.c
+index a202ce6611fde..4073e8243df3f 100644
+--- a/drivers/net/phy/phylink.c
++++ b/drivers/net/phy/phylink.c
+@@ -1552,6 +1552,25 @@ void phylink_destroy(struct phylink *pl)
+ }
+ EXPORT_SYMBOL_GPL(phylink_destroy);
+ 
++/**
++ * phylink_expects_phy() - Determine if phylink expects a phy to be attached
++ * @pl: a pointer to a &struct phylink returned from phylink_create()
++ *
++ * When using fixed-link mode, or in-band mode with 1000base-X or 2500base-X,
++ * no PHY is needed.
++ *
++ * Returns true if phylink will be expecting a PHY.
++ */
++bool phylink_expects_phy(struct phylink *pl)
++{
++	if (pl->cfg_link_an_mode == MLO_AN_FIXED ||
++	    (pl->cfg_link_an_mode == MLO_AN_INBAND &&
++	     phy_interface_mode_is_8023z(pl->link_config.interface)))
++		return false;
++	return true;
++}
++EXPORT_SYMBOL_GPL(phylink_expects_phy);
++
+ static void phylink_phy_change(struct phy_device *phydev, bool up)
+ {
+ 	struct phylink *pl = phydev->phylink;
+diff --git a/drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c b/drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c
+index d0daef674e728..e300278ea38c6 100644
+--- a/drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c
++++ b/drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c
+@@ -991,15 +991,34 @@ static const struct sdio_device_id brcmf_sdmmc_ids[] = {
+ MODULE_DEVICE_TABLE(sdio, brcmf_sdmmc_ids);
+ 
+ 
+-static void brcmf_sdiod_acpi_set_power_manageable(struct device *dev,
+-						  int val)
++static void brcmf_sdiod_acpi_save_power_manageable(struct brcmf_sdio_dev *sdiodev)
+ {
+ #if IS_ENABLED(CONFIG_ACPI)
+ 	struct acpi_device *adev;
+ 
+-	adev = ACPI_COMPANION(dev);
++	adev = ACPI_COMPANION(&sdiodev->func1->dev);
+ 	if (adev)
+-		adev->flags.power_manageable = 0;
++		sdiodev->func1_power_manageable = adev->flags.power_manageable;
++
++	adev = ACPI_COMPANION(&sdiodev->func2->dev);
++	if (adev)
++		sdiodev->func2_power_manageable = adev->flags.power_manageable;
++#endif
++}
++
++static void brcmf_sdiod_acpi_set_power_manageable(struct brcmf_sdio_dev *sdiodev,
++						  int enable)
++{
++#if IS_ENABLED(CONFIG_ACPI)
++	struct acpi_device *adev;
++
++	adev = ACPI_COMPANION(&sdiodev->func1->dev);
++	if (adev)
++		adev->flags.power_manageable = enable ? sdiodev->func1_power_manageable : 0;
++
++	adev = ACPI_COMPANION(&sdiodev->func2->dev);
++	if (adev)
++		adev->flags.power_manageable = enable ? sdiodev->func2_power_manageable : 0;
+ #endif
+ }
+ 
+@@ -1009,7 +1028,6 @@ static int brcmf_ops_sdio_probe(struct sdio_func *func,
+ 	int err;
+ 	struct brcmf_sdio_dev *sdiodev;
+ 	struct brcmf_bus *bus_if;
+-	struct device *dev;
+ 
+ 	brcmf_dbg(SDIO, "Enter\n");
+ 	brcmf_dbg(SDIO, "Class=%x\n", func->class);
+@@ -1017,14 +1035,9 @@ static int brcmf_ops_sdio_probe(struct sdio_func *func,
+ 	brcmf_dbg(SDIO, "sdio device ID: 0x%04x\n", func->device);
+ 	brcmf_dbg(SDIO, "Function#: %d\n", func->num);
+ 
+-	dev = &func->dev;
+-
+ 	/* Set MMC_QUIRK_LENIENT_FN0 for this card */
+ 	func->card->quirks |= MMC_QUIRK_LENIENT_FN0;
+ 
+-	/* prohibit ACPI power management for this device */
+-	brcmf_sdiod_acpi_set_power_manageable(dev, 0);
+-
+ 	/* Consume func num 1 but dont do anything with it. */
+ 	if (func->num == 1)
+ 		return 0;
+@@ -1055,6 +1068,7 @@ static int brcmf_ops_sdio_probe(struct sdio_func *func,
+ 	dev_set_drvdata(&sdiodev->func1->dev, bus_if);
+ 	sdiodev->dev = &sdiodev->func1->dev;
+ 
++	brcmf_sdiod_acpi_save_power_manageable(sdiodev);
+ 	brcmf_sdiod_change_state(sdiodev, BRCMF_SDIOD_DOWN);
+ 
+ 	brcmf_dbg(SDIO, "F2 found, calling brcmf_sdiod_probe...\n");
+@@ -1120,6 +1134,8 @@ void brcmf_sdio_wowl_config(struct device *dev, bool enabled)
+ 
+ 	if (sdiodev->settings->bus.sdio.oob_irq_supported ||
+ 	    pm_caps & MMC_PM_WAKE_SDIO_IRQ) {
++		/* Stop ACPI from turning off the device when wowl is enabled */
++		brcmf_sdiod_acpi_set_power_manageable(sdiodev, !enabled);
+ 		sdiodev->wowl_enabled = enabled;
+ 		brcmf_dbg(SDIO, "Configuring WOWL, enabled=%d\n", enabled);
+ 		return;
+diff --git a/drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.h b/drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.h
+index b76d34d36bde6..0d18ed15b4032 100644
+--- a/drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.h
++++ b/drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.h
+@@ -188,6 +188,8 @@ struct brcmf_sdio_dev {
+ 	char nvram_name[BRCMF_FW_NAME_LEN];
+ 	char clm_name[BRCMF_FW_NAME_LEN];
+ 	bool wowl_enabled;
++	bool func1_power_manageable;
++	bool func2_power_manageable;
+ 	enum brcmf_sdiod_state state;
+ 	struct brcmf_sdiod_freezer *freezer;
+ 	const struct firmware *clm_fw;
+diff --git a/drivers/net/wireless/mediatek/mt76/mt7603/main.c b/drivers/net/wireless/mediatek/mt76/mt7603/main.c
+index ca50feb0b3a9d..1b1358c6bb464 100644
+--- a/drivers/net/wireless/mediatek/mt76/mt7603/main.c
++++ b/drivers/net/wireless/mediatek/mt76/mt7603/main.c
+@@ -512,15 +512,15 @@ mt7603_set_key(struct ieee80211_hw *hw, enum set_key_cmd cmd,
+ 	    !(key->flags & IEEE80211_KEY_FLAG_PAIRWISE))
+ 		return -EOPNOTSUPP;
+ 
+-	if (cmd == SET_KEY) {
+-		key->hw_key_idx = wcid->idx;
+-		wcid->hw_key_idx = idx;
+-	} else {
++	if (cmd != SET_KEY) {
+ 		if (idx == wcid->hw_key_idx)
+ 			wcid->hw_key_idx = -1;
+ 
+-		key = NULL;
++		return 0;
+ 	}
++
++	key->hw_key_idx = wcid->idx;
++	wcid->hw_key_idx = idx;
+ 	mt76_wcid_key_setup(&dev->mt76, wcid, key);
+ 
+ 	return mt7603_wtbl_set_key(dev, wcid->idx, key);
+diff --git a/drivers/net/wireless/mediatek/mt76/mt7615/mac.c b/drivers/net/wireless/mediatek/mt76/mt7615/mac.c
+index 2ce1705c0f433..a0412a29fb491 100644
+--- a/drivers/net/wireless/mediatek/mt76/mt7615/mac.c
++++ b/drivers/net/wireless/mediatek/mt76/mt7615/mac.c
+@@ -1180,8 +1180,7 @@ EXPORT_SYMBOL_GPL(mt7615_mac_set_rates);
+ static int
+ mt7615_mac_wtbl_update_key(struct mt7615_dev *dev, struct mt76_wcid *wcid,
+ 			   struct ieee80211_key_conf *key,
+-			   enum mt76_cipher_type cipher, u16 cipher_mask,
+-			   enum set_key_cmd cmd)
++			   enum mt76_cipher_type cipher, u16 cipher_mask)
+ {
+ 	u32 addr = mt7615_mac_wtbl_addr(dev, wcid->idx) + 30 * 4;
+ 	u8 data[32] = {};
+@@ -1190,27 +1189,18 @@ mt7615_mac_wtbl_update_key(struct mt7615_dev *dev, struct mt76_wcid *wcid,
+ 		return -EINVAL;
+ 
+ 	mt76_rr_copy(dev, addr, data, sizeof(data));
+-	if (cmd == SET_KEY) {
+-		if (cipher == MT_CIPHER_TKIP) {
+-			/* Rx/Tx MIC keys are swapped */
+-			memcpy(data, key->key, 16);
+-			memcpy(data + 16, key->key + 24, 8);
+-			memcpy(data + 24, key->key + 16, 8);
+-		} else {
+-			if (cipher_mask == BIT(cipher))
+-				memcpy(data, key->key, key->keylen);
+-			else if (cipher != MT_CIPHER_BIP_CMAC_128)
+-				memcpy(data, key->key, 16);
+-			if (cipher == MT_CIPHER_BIP_CMAC_128)
+-				memcpy(data + 16, key->key, 16);
+-		}
++	if (cipher == MT_CIPHER_TKIP) {
++		/* Rx/Tx MIC keys are swapped */
++		memcpy(data, key->key, 16);
++		memcpy(data + 16, key->key + 24, 8);
++		memcpy(data + 24, key->key + 16, 8);
+ 	} else {
++		if (cipher_mask == BIT(cipher))
++			memcpy(data, key->key, key->keylen);
++		else if (cipher != MT_CIPHER_BIP_CMAC_128)
++			memcpy(data, key->key, 16);
+ 		if (cipher == MT_CIPHER_BIP_CMAC_128)
+-			memset(data + 16, 0, 16);
+-		else if (cipher_mask)
+-			memset(data, 0, 16);
+-		if (!cipher_mask)
+-			memset(data, 0, sizeof(data));
++			memcpy(data + 16, key->key, 16);
+ 	}
+ 
+ 	mt76_wr_copy(dev, addr, data, sizeof(data));
+@@ -1221,7 +1211,7 @@ mt7615_mac_wtbl_update_key(struct mt7615_dev *dev, struct mt76_wcid *wcid,
+ static int
+ mt7615_mac_wtbl_update_pk(struct mt7615_dev *dev, struct mt76_wcid *wcid,
+ 			  enum mt76_cipher_type cipher, u16 cipher_mask,
+-			  int keyidx, enum set_key_cmd cmd)
++			  int keyidx)
+ {
+ 	u32 addr = mt7615_mac_wtbl_addr(dev, wcid->idx), w0, w1;
+ 
+@@ -1240,9 +1230,7 @@ mt7615_mac_wtbl_update_pk(struct mt7615_dev *dev, struct mt76_wcid *wcid,
+ 	else
+ 		w0 &= ~MT_WTBL_W0_RX_IK_VALID;
+ 
+-	if (cmd == SET_KEY &&
+-	    (cipher != MT_CIPHER_BIP_CMAC_128 ||
+-	     cipher_mask == BIT(cipher))) {
++	if (cipher != MT_CIPHER_BIP_CMAC_128 || cipher_mask == BIT(cipher)) {
+ 		w0 &= ~MT_WTBL_W0_KEY_IDX;
+ 		w0 |= FIELD_PREP(MT_WTBL_W0_KEY_IDX, keyidx);
+ 	}
+@@ -1259,19 +1247,10 @@ mt7615_mac_wtbl_update_pk(struct mt7615_dev *dev, struct mt76_wcid *wcid,
+ 
+ static void
+ mt7615_mac_wtbl_update_cipher(struct mt7615_dev *dev, struct mt76_wcid *wcid,
+-			      enum mt76_cipher_type cipher, u16 cipher_mask,
+-			      enum set_key_cmd cmd)
++			      enum mt76_cipher_type cipher, u16 cipher_mask)
+ {
+ 	u32 addr = mt7615_mac_wtbl_addr(dev, wcid->idx);
+ 
+-	if (!cipher_mask) {
+-		mt76_clear(dev, addr + 2 * 4, MT_WTBL_W2_KEY_TYPE);
+-		return;
+-	}
+-
+-	if (cmd != SET_KEY)
+-		return;
+-
+ 	if (cipher == MT_CIPHER_BIP_CMAC_128 &&
+ 	    cipher_mask & ~BIT(MT_CIPHER_BIP_CMAC_128))
+ 		return;
+@@ -1282,8 +1261,7 @@ mt7615_mac_wtbl_update_cipher(struct mt7615_dev *dev, struct mt76_wcid *wcid,
+ 
+ int __mt7615_mac_wtbl_set_key(struct mt7615_dev *dev,
+ 			      struct mt76_wcid *wcid,
+-			      struct ieee80211_key_conf *key,
+-			      enum set_key_cmd cmd)
++			      struct ieee80211_key_conf *key)
+ {
+ 	enum mt76_cipher_type cipher;
+ 	u16 cipher_mask = wcid->cipher;
+@@ -1293,19 +1271,14 @@ int __mt7615_mac_wtbl_set_key(struct mt7615_dev *dev,
+ 	if (cipher == MT_CIPHER_NONE)
+ 		return -EOPNOTSUPP;
+ 
+-	if (cmd == SET_KEY)
+-		cipher_mask |= BIT(cipher);
+-	else
+-		cipher_mask &= ~BIT(cipher);
+-
+-	mt7615_mac_wtbl_update_cipher(dev, wcid, cipher, cipher_mask, cmd);
+-	err = mt7615_mac_wtbl_update_key(dev, wcid, key, cipher, cipher_mask,
+-					 cmd);
++	cipher_mask |= BIT(cipher);
++	mt7615_mac_wtbl_update_cipher(dev, wcid, cipher, cipher_mask);
++	err = mt7615_mac_wtbl_update_key(dev, wcid, key, cipher, cipher_mask);
+ 	if (err < 0)
+ 		return err;
+ 
+ 	err = mt7615_mac_wtbl_update_pk(dev, wcid, cipher, cipher_mask,
+-					key->keyidx, cmd);
++					key->keyidx);
+ 	if (err < 0)
+ 		return err;
+ 
+@@ -1316,13 +1289,12 @@ int __mt7615_mac_wtbl_set_key(struct mt7615_dev *dev,
+ 
+ int mt7615_mac_wtbl_set_key(struct mt7615_dev *dev,
+ 			    struct mt76_wcid *wcid,
+-			    struct ieee80211_key_conf *key,
+-			    enum set_key_cmd cmd)
++			    struct ieee80211_key_conf *key)
+ {
+ 	int err;
+ 
+ 	spin_lock_bh(&dev->mt76.lock);
+-	err = __mt7615_mac_wtbl_set_key(dev, wcid, key, cmd);
++	err = __mt7615_mac_wtbl_set_key(dev, wcid, key);
+ 	spin_unlock_bh(&dev->mt76.lock);
+ 
+ 	return err;
+diff --git a/drivers/net/wireless/mediatek/mt76/mt7615/main.c b/drivers/net/wireless/mediatek/mt76/mt7615/main.c
+index 8d4733f87cda9..a3b0d4e9146a6 100644
+--- a/drivers/net/wireless/mediatek/mt76/mt7615/main.c
++++ b/drivers/net/wireless/mediatek/mt76/mt7615/main.c
+@@ -391,18 +391,17 @@ static int mt7615_set_key(struct ieee80211_hw *hw, enum set_key_cmd cmd,
+ 
+ 	if (cmd == SET_KEY)
+ 		*wcid_keyidx = idx;
+-	else if (idx == *wcid_keyidx)
+-		*wcid_keyidx = -1;
+-	else
++	else {
++		if (idx == *wcid_keyidx)
++			*wcid_keyidx = -1;
+ 		goto out;
++	}
+ 
+-	mt76_wcid_key_setup(&dev->mt76, wcid,
+-			    cmd == SET_KEY ? key : NULL);
+-
++	mt76_wcid_key_setup(&dev->mt76, wcid, key);
+ 	if (mt76_is_mmio(&dev->mt76))
+-		err = mt7615_mac_wtbl_set_key(dev, wcid, key, cmd);
++		err = mt7615_mac_wtbl_set_key(dev, wcid, key);
+ 	else
+-		err = __mt7615_mac_wtbl_set_key(dev, wcid, key, cmd);
++		err = __mt7615_mac_wtbl_set_key(dev, wcid, key);
+ 
+ out:
+ 	mt7615_mutex_release(dev);
+diff --git a/drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h b/drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h
+index 060d52c81d9e9..c0e62082b62df 100644
+--- a/drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h
++++ b/drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h
+@@ -482,11 +482,9 @@ int mt7615_mac_write_txwi(struct mt7615_dev *dev, __le32 *txwi,
+ void mt7615_mac_set_timing(struct mt7615_phy *phy);
+ int __mt7615_mac_wtbl_set_key(struct mt7615_dev *dev,
+ 			      struct mt76_wcid *wcid,
+-			      struct ieee80211_key_conf *key,
+-			      enum set_key_cmd cmd);
++			      struct ieee80211_key_conf *key);
+ int mt7615_mac_wtbl_set_key(struct mt7615_dev *dev, struct mt76_wcid *wcid,
+-			    struct ieee80211_key_conf *key,
+-			    enum set_key_cmd cmd);
++			    struct ieee80211_key_conf *key);
+ void mt7615_mac_reset_work(struct work_struct *work);
+ u32 mt7615_mac_get_sta_tid_sn(struct mt7615_dev *dev, int wcid, u8 tid);
+ 
+diff --git a/drivers/net/wireless/mediatek/mt76/mt76x02_util.c b/drivers/net/wireless/mediatek/mt76/mt76x02_util.c
+index 604ddcc211231..324535a0dd6d4 100644
+--- a/drivers/net/wireless/mediatek/mt76/mt76x02_util.c
++++ b/drivers/net/wireless/mediatek/mt76/mt76x02_util.c
+@@ -455,20 +455,20 @@ int mt76x02_set_key(struct ieee80211_hw *hw, enum set_key_cmd cmd,
+ 	msta = sta ? (struct mt76x02_sta *)sta->drv_priv : NULL;
+ 	wcid = msta ? &msta->wcid : &mvif->group_wcid;
+ 
+-	if (cmd == SET_KEY) {
+-		key->hw_key_idx = wcid->idx;
+-		wcid->hw_key_idx = idx;
+-		if (key->flags & IEEE80211_KEY_FLAG_RX_MGMT) {
+-			key->flags |= IEEE80211_KEY_FLAG_SW_MGMT_TX;
+-			wcid->sw_iv = true;
+-		}
+-	} else {
++	if (cmd != SET_KEY) {
+ 		if (idx == wcid->hw_key_idx) {
+ 			wcid->hw_key_idx = -1;
+ 			wcid->sw_iv = false;
+ 		}
+ 
+-		key = NULL;
++		return 0;
++	}
++
++	key->hw_key_idx = wcid->idx;
++	wcid->hw_key_idx = idx;
++	if (key->flags & IEEE80211_KEY_FLAG_RX_MGMT) {
++		key->flags |= IEEE80211_KEY_FLAG_SW_MGMT_TX;
++		wcid->sw_iv = true;
+ 	}
+ 	mt76_wcid_key_setup(&dev->mt76, wcid, key);
+ 
+diff --git a/drivers/net/wireless/mediatek/mt76/mt7915/main.c b/drivers/net/wireless/mediatek/mt76/mt7915/main.c
+index 060cb88e82e30..bda26bd62412e 100644
+--- a/drivers/net/wireless/mediatek/mt76/mt7915/main.c
++++ b/drivers/net/wireless/mediatek/mt76/mt7915/main.c
+@@ -393,16 +393,15 @@ static int mt7915_set_key(struct ieee80211_hw *hw, enum set_key_cmd cmd,
+ 		mt7915_mcu_add_bss_info(phy, vif, true);
+ 	}
+ 
+-	if (cmd == SET_KEY)
++	if (cmd == SET_KEY) {
+ 		*wcid_keyidx = idx;
+-	else if (idx == *wcid_keyidx)
+-		*wcid_keyidx = -1;
+-	else
++	} else {
++		if (idx == *wcid_keyidx)
++			*wcid_keyidx = -1;
+ 		goto out;
++	}
+ 
+-	mt76_wcid_key_setup(&dev->mt76, wcid,
+-			    cmd == SET_KEY ? key : NULL);
+-
++	mt76_wcid_key_setup(&dev->mt76, wcid, key);
+ 	err = mt76_connac_mcu_add_key(&dev->mt76, vif, &msta->bip,
+ 				      key, MCU_EXT_CMD(STA_REC_UPDATE),
+ 				      &msta->wcid, cmd);
+diff --git a/drivers/net/wireless/mediatek/mt76/mt7921/main.c b/drivers/net/wireless/mediatek/mt76/mt7921/main.c
+index 111d9221b94f5..60fbbd1ac2f78 100644
+--- a/drivers/net/wireless/mediatek/mt76/mt7921/main.c
++++ b/drivers/net/wireless/mediatek/mt76/mt7921/main.c
+@@ -464,16 +464,15 @@ static int mt7921_set_key(struct ieee80211_hw *hw, enum set_key_cmd cmd,
+ 
+ 	mt7921_mutex_acquire(dev);
+ 
+-	if (cmd == SET_KEY)
++	if (cmd == SET_KEY) {
+ 		*wcid_keyidx = idx;
+-	else if (idx == *wcid_keyidx)
+-		*wcid_keyidx = -1;
+-	else
++	} else {
++		if (idx == *wcid_keyidx)
++			*wcid_keyidx = -1;
+ 		goto out;
++	}
+ 
+-	mt76_wcid_key_setup(&dev->mt76, wcid,
+-			    cmd == SET_KEY ? key : NULL);
+-
++	mt76_wcid_key_setup(&dev->mt76, wcid, key);
+ 	err = mt76_connac_mcu_add_key(&dev->mt76, vif, &msta->bip,
+ 				      key, MCU_UNI_CMD(STA_REC_UPDATE),
+ 				      &msta->wcid, cmd);
+diff --git a/drivers/nvme/host/core.c b/drivers/nvme/host/core.c
+index a95e48b51da66..cb71ce3413c2d 100644
+--- a/drivers/nvme/host/core.c
++++ b/drivers/nvme/host/core.c
+@@ -1711,6 +1711,9 @@ static void nvme_config_discard(struct gendisk *disk, struct nvme_ns *ns)
+ 	struct request_queue *queue = disk->queue;
+ 	u32 size = queue_logical_block_size(queue);
+ 
++	if (ctrl->dmrsl && ctrl->dmrsl <= nvme_sect_to_lba(ns, UINT_MAX))
++		ctrl->max_discard_sectors = nvme_lba_to_sect(ns, ctrl->dmrsl);
++
+ 	if (ctrl->max_discard_sectors == 0) {
+ 		blk_queue_max_discard_sectors(queue, 0);
+ 		return;
+@@ -1725,9 +1728,6 @@ static void nvme_config_discard(struct gendisk *disk, struct nvme_ns *ns)
+ 	if (queue->limits.max_discard_sectors)
+ 		return;
+ 
+-	if (ctrl->dmrsl && ctrl->dmrsl <= nvme_sect_to_lba(ns, UINT_MAX))
+-		ctrl->max_discard_sectors = nvme_lba_to_sect(ns, ctrl->dmrsl);
+-
+ 	blk_queue_max_discard_sectors(queue, ctrl->max_discard_sectors);
+ 	blk_queue_max_discard_segments(queue, ctrl->max_discard_segments);
+ 
+diff --git a/drivers/pci/doe.c b/drivers/pci/doe.c
+index 66d9ab2886468..e5e9b287b9766 100644
+--- a/drivers/pci/doe.c
++++ b/drivers/pci/doe.c
+@@ -128,7 +128,7 @@ static int pci_doe_send_req(struct pci_doe_mb *doe_mb,
+ 		return -EIO;
+ 
+ 	/* Length is 2 DW of header + length of payload in DW */
+-	length = 2 + task->request_pl_sz / sizeof(u32);
++	length = 2 + task->request_pl_sz / sizeof(__le32);
+ 	if (length > PCI_DOE_MAX_LENGTH)
+ 		return -EIO;
+ 	if (length == PCI_DOE_MAX_LENGTH)
+@@ -141,9 +141,9 @@ static int pci_doe_send_req(struct pci_doe_mb *doe_mb,
+ 	pci_write_config_dword(pdev, offset + PCI_DOE_WRITE,
+ 			       FIELD_PREP(PCI_DOE_DATA_OBJECT_HEADER_2_LENGTH,
+ 					  length));
+-	for (i = 0; i < task->request_pl_sz / sizeof(u32); i++)
++	for (i = 0; i < task->request_pl_sz / sizeof(__le32); i++)
+ 		pci_write_config_dword(pdev, offset + PCI_DOE_WRITE,
+-				       task->request_pl[i]);
++				       le32_to_cpu(task->request_pl[i]));
+ 
+ 	pci_doe_write_ctrl(doe_mb, PCI_DOE_CTRL_GO);
+ 
+@@ -195,11 +195,11 @@ static int pci_doe_recv_resp(struct pci_doe_mb *doe_mb, struct pci_doe_task *tas
+ 
+ 	/* First 2 dwords have already been read */
+ 	length -= 2;
+-	payload_length = min(length, task->response_pl_sz / sizeof(u32));
++	payload_length = min(length, task->response_pl_sz / sizeof(__le32));
+ 	/* Read the rest of the response payload */
+ 	for (i = 0; i < payload_length; i++) {
+-		pci_read_config_dword(pdev, offset + PCI_DOE_READ,
+-				      &task->response_pl[i]);
++		pci_read_config_dword(pdev, offset + PCI_DOE_READ, &val);
++		task->response_pl[i] = cpu_to_le32(val);
+ 		/* Prior to the last ack, ensure Data Object Ready */
+ 		if (i == (payload_length - 1) && !pci_doe_data_obj_ready(doe_mb))
+ 			return -EIO;
+@@ -217,13 +217,14 @@ static int pci_doe_recv_resp(struct pci_doe_mb *doe_mb, struct pci_doe_task *tas
+ 	if (FIELD_GET(PCI_DOE_STATUS_ERROR, val))
+ 		return -EIO;
+ 
+-	return min(length, task->response_pl_sz / sizeof(u32)) * sizeof(u32);
++	return min(length, task->response_pl_sz / sizeof(__le32)) * sizeof(__le32);
+ }
+ 
+ static void signal_task_complete(struct pci_doe_task *task, int rv)
+ {
+ 	task->rv = rv;
+ 	task->complete(task);
++	destroy_work_on_stack(&task->work);
+ }
+ 
+ static void signal_task_abort(struct pci_doe_task *task, int rv)
+@@ -317,14 +318,16 @@ static int pci_doe_discovery(struct pci_doe_mb *doe_mb, u8 *index, u16 *vid,
+ {
+ 	u32 request_pl = FIELD_PREP(PCI_DOE_DATA_OBJECT_DISC_REQ_3_INDEX,
+ 				    *index);
++	__le32 request_pl_le = cpu_to_le32(request_pl);
++	__le32 response_pl_le;
+ 	u32 response_pl;
+ 	DECLARE_COMPLETION_ONSTACK(c);
+ 	struct pci_doe_task task = {
+ 		.prot.vid = PCI_VENDOR_ID_PCI_SIG,
+ 		.prot.type = PCI_DOE_PROTOCOL_DISCOVERY,
+-		.request_pl = &request_pl,
++		.request_pl = &request_pl_le,
+ 		.request_pl_sz = sizeof(request_pl),
+-		.response_pl = &response_pl,
++		.response_pl = &response_pl_le,
+ 		.response_pl_sz = sizeof(response_pl),
+ 		.complete = pci_doe_task_complete,
+ 		.private = &c,
+@@ -340,6 +343,7 @@ static int pci_doe_discovery(struct pci_doe_mb *doe_mb, u8 *index, u16 *vid,
+ 	if (task.rv != sizeof(response_pl))
+ 		return -EIO;
+ 
++	response_pl = le32_to_cpu(response_pl_le);
+ 	*vid = FIELD_GET(PCI_DOE_DATA_OBJECT_DISC_RSP_3_VID, response_pl);
+ 	*protocol = FIELD_GET(PCI_DOE_DATA_OBJECT_DISC_RSP_3_PROTOCOL,
+ 			      response_pl);
+@@ -520,6 +524,8 @@ EXPORT_SYMBOL_GPL(pci_doe_supports_prot);
+  * task->complete will be called when the state machine is done processing this
+  * task.
+  *
++ * @task must be allocated on the stack.
++ *
+  * Excess data will be discarded.
+  *
+  * RETURNS: 0 when task has been successfully queued, -ERRNO on error
+@@ -533,15 +539,15 @@ int pci_doe_submit_task(struct pci_doe_mb *doe_mb, struct pci_doe_task *task)
+ 	 * DOE requests must be a whole number of DW and the response needs to
+ 	 * be big enough for at least 1 DW
+ 	 */
+-	if (task->request_pl_sz % sizeof(u32) ||
+-	    task->response_pl_sz < sizeof(u32))
++	if (task->request_pl_sz % sizeof(__le32) ||
++	    task->response_pl_sz < sizeof(__le32))
+ 		return -EINVAL;
+ 
+ 	if (test_bit(PCI_DOE_FLAG_DEAD, &doe_mb->flags))
+ 		return -EIO;
+ 
+ 	task->doe_mb = doe_mb;
+-	INIT_WORK(&task->work, doe_statemachine_work);
++	INIT_WORK_ONSTACK(&task->work, doe_statemachine_work);
+ 	queue_work(doe_mb->work_queue, &task->work);
+ 	return 0;
+ }
+diff --git a/drivers/platform/x86/think-lmi.c b/drivers/platform/x86/think-lmi.c
+index 74af3e593b2ca..336b9029d1515 100644
+--- a/drivers/platform/x86/think-lmi.c
++++ b/drivers/platform/x86/think-lmi.c
+@@ -920,7 +920,7 @@ static ssize_t display_name_show(struct kobject *kobj, struct kobj_attribute *at
+ static ssize_t current_value_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf)
+ {
+ 	struct tlmi_attr_setting *setting = to_tlmi_attr_setting(kobj);
+-	char *item, *value;
++	char *item, *value, *p;
+ 	int ret;
+ 
+ 	ret = tlmi_setting(setting->index, &item, LENOVO_BIOS_SETTING_GUID);
+@@ -930,10 +930,15 @@ static ssize_t current_value_show(struct kobject *kobj, struct kobj_attribute *a
+ 	/* validate and split from `item,value` -> `value` */
+ 	value = strpbrk(item, ",");
+ 	if (!value || value == item || !strlen(value + 1))
+-		return -EINVAL;
+-
+-	ret = sysfs_emit(buf, "%s\n", value + 1);
++		ret = -EINVAL;
++	else {
++		/* On Workstations remove the Options part after the value */
++		p = strchrnul(value, ';');
++		*p = '\0';
++		ret = sysfs_emit(buf, "%s\n", value + 1);
++	}
+ 	kfree(item);
++
+ 	return ret;
+ }
+ 
+@@ -1457,10 +1462,10 @@ static int tlmi_analyze(void)
+ 			 * name string.
+ 			 * Try and pull that out if it's available.
+ 			 */
+-			char *item, *optstart, *optend;
++			char *optitem, *optstart, *optend;
+ 
+-			if (!tlmi_setting(setting->index, &item, LENOVO_BIOS_SETTING_GUID)) {
+-				optstart = strstr(item, "[Optional:");
++			if (!tlmi_setting(setting->index, &optitem, LENOVO_BIOS_SETTING_GUID)) {
++				optstart = strstr(optitem, "[Optional:");
+ 				if (optstart) {
+ 					optstart += strlen("[Optional:");
+ 					optend = strstr(optstart, "]");
+@@ -1469,6 +1474,7 @@ static int tlmi_analyze(void)
+ 							kstrndup(optstart, optend - optstart,
+ 									GFP_KERNEL);
+ 				}
++				kfree(optitem);
+ 			}
+ 		}
+ 		/*
+diff --git a/drivers/pwm/pwm-atmel.c b/drivers/pwm/pwm-atmel.c
+index 8e00a4286145b..cdbc23649032c 100644
+--- a/drivers/pwm/pwm-atmel.c
++++ b/drivers/pwm/pwm-atmel.c
+@@ -356,8 +356,8 @@ static int atmel_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	return 0;
+ }
+ 
+-static void atmel_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-				struct pwm_state *state)
++static int atmel_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++			       struct pwm_state *state)
+ {
+ 	struct atmel_pwm_chip *atmel_pwm = to_atmel_pwm_chip(chip);
+ 	u32 sr, cmr;
+@@ -396,6 +396,8 @@ static void atmel_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 		state->polarity = PWM_POLARITY_INVERSED;
+ 	else
+ 		state->polarity = PWM_POLARITY_NORMAL;
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops atmel_pwm_ops = {
+diff --git a/drivers/pwm/pwm-bcm-iproc.c b/drivers/pwm/pwm-bcm-iproc.c
+index 7251037d4dd56..97ec131eb7c1b 100644
+--- a/drivers/pwm/pwm-bcm-iproc.c
++++ b/drivers/pwm/pwm-bcm-iproc.c
+@@ -68,8 +68,8 @@ static void iproc_pwmc_disable(struct iproc_pwmc *ip, unsigned int channel)
+ 	ndelay(400);
+ }
+ 
+-static void iproc_pwmc_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-				 struct pwm_state *state)
++static int iproc_pwmc_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++				struct pwm_state *state)
+ {
+ 	struct iproc_pwmc *ip = to_iproc_pwmc(chip);
+ 	u64 tmp, multi, rate;
+@@ -91,7 +91,7 @@ static void iproc_pwmc_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	if (rate == 0) {
+ 		state->period = 0;
+ 		state->duty_cycle = 0;
+-		return;
++		return 0;
+ 	}
+ 
+ 	value = readl(ip->base + IPROC_PWM_PRESCALE_OFFSET);
+@@ -107,6 +107,8 @@ static void iproc_pwmc_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	value = readl(ip->base + IPROC_PWM_DUTY_CYCLE_OFFSET(pwm->hwpwm));
+ 	tmp = (value & IPROC_PWM_PERIOD_MAX) * multi;
+ 	state->duty_cycle = div64_u64(tmp, rate);
++
++	return 0;
+ }
+ 
+ static int iproc_pwmc_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+diff --git a/drivers/pwm/pwm-crc.c b/drivers/pwm/pwm-crc.c
+index 7b357d1cf6421..4099850117ba4 100644
+--- a/drivers/pwm/pwm-crc.c
++++ b/drivers/pwm/pwm-crc.c
+@@ -121,8 +121,8 @@ static int crc_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	return 0;
+ }
+ 
+-static void crc_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-			      struct pwm_state *state)
++static int crc_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++			     struct pwm_state *state)
+ {
+ 	struct crystalcove_pwm *crc_pwm = to_crc_pwm(chip);
+ 	struct device *dev = crc_pwm->chip.dev;
+@@ -132,13 +132,13 @@ static void crc_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	error = regmap_read(crc_pwm->regmap, PWM0_CLK_DIV, &clk_div_reg);
+ 	if (error) {
+ 		dev_err(dev, "Error reading PWM0_CLK_DIV %d\n", error);
+-		return;
++		return 0;
+ 	}
+ 
+ 	error = regmap_read(crc_pwm->regmap, PWM0_DUTY_CYCLE, &duty_cycle_reg);
+ 	if (error) {
+ 		dev_err(dev, "Error reading PWM0_DUTY_CYCLE %d\n", error);
+-		return;
++		return 0;
+ 	}
+ 
+ 	clk_div = (clk_div_reg & ~PWM_OUTPUT_ENABLE) + 1;
+@@ -149,6 +149,8 @@ static void crc_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 		DIV_ROUND_UP_ULL(duty_cycle_reg * state->period, PWM_MAX_LEVEL);
+ 	state->polarity = PWM_POLARITY_NORMAL;
+ 	state->enabled = !!(clk_div_reg & PWM_OUTPUT_ENABLE);
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops crc_pwm_ops = {
+diff --git a/drivers/pwm/pwm-cros-ec.c b/drivers/pwm/pwm-cros-ec.c
+index 7f10f56c3eb66..e55bc36ed4972 100644
+--- a/drivers/pwm/pwm-cros-ec.c
++++ b/drivers/pwm/pwm-cros-ec.c
+@@ -183,8 +183,8 @@ static int cros_ec_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	return 0;
+ }
+ 
+-static void cros_ec_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-				  struct pwm_state *state)
++static int cros_ec_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++				 struct pwm_state *state)
+ {
+ 	struct cros_ec_pwm_device *ec_pwm = pwm_to_cros_ec_pwm(chip);
+ 	struct cros_ec_pwm *channel = pwm_get_chip_data(pwm);
+@@ -193,11 +193,12 @@ static void cros_ec_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	ret = cros_ec_pwm_get_duty(ec_pwm, pwm->hwpwm);
+ 	if (ret < 0) {
+ 		dev_err(chip->dev, "error getting initial duty: %d\n", ret);
+-		return;
++		return 0;
+ 	}
+ 
+ 	state->enabled = (ret > 0);
+ 	state->period = EC_PWM_MAX_DUTY;
++	state->polarity = PWM_POLARITY_NORMAL;
+ 
+ 	/*
+ 	 * Note that "disabled" and "duty cycle == 0" are treated the same. If
+@@ -212,6 +213,8 @@ static void cros_ec_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 		state->duty_cycle = channel->duty_cycle;
+ 	else
+ 		state->duty_cycle = ret;
++
++	return 0;
+ }
+ 
+ static struct pwm_device *
+diff --git a/drivers/pwm/pwm-dwc.c b/drivers/pwm/pwm-dwc.c
+index 7568300bb11e5..bd2308812096d 100644
+--- a/drivers/pwm/pwm-dwc.c
++++ b/drivers/pwm/pwm-dwc.c
+@@ -163,8 +163,8 @@ static int dwc_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	return 0;
+ }
+ 
+-static void dwc_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-			      struct pwm_state *state)
++static int dwc_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++			     struct pwm_state *state)
+ {
+ 	struct dwc_pwm *dwc = to_dwc_pwm(chip);
+ 	u64 duty, period;
+@@ -188,6 +188,8 @@ static void dwc_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	state->polarity = PWM_POLARITY_INVERSED;
+ 
+ 	pm_runtime_put_sync(chip->dev);
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops dwc_pwm_ops = {
+diff --git a/drivers/pwm/pwm-hibvt.c b/drivers/pwm/pwm-hibvt.c
+index 333f1b18ff4e6..1b9274c5ad872 100644
+--- a/drivers/pwm/pwm-hibvt.c
++++ b/drivers/pwm/pwm-hibvt.c
+@@ -128,8 +128,8 @@ static void hibvt_pwm_set_polarity(struct pwm_chip *chip,
+ 				PWM_POLARITY_MASK, (0x0 << PWM_POLARITY_SHIFT));
+ }
+ 
+-static void hibvt_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-				struct pwm_state *state)
++static int hibvt_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++			       struct pwm_state *state)
+ {
+ 	struct hibvt_pwm_chip *hi_pwm_chip = to_hibvt_pwm_chip(chip);
+ 	void __iomem *base;
+@@ -146,6 +146,9 @@ static void hibvt_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 
+ 	value = readl(base + PWM_CTRL_ADDR(pwm->hwpwm));
+ 	state->enabled = (PWM_ENABLE_MASK & value);
++	state->polarity = (PWM_POLARITY_MASK & value) ? PWM_POLARITY_INVERSED : PWM_POLARITY_NORMAL;
++
++	return 0;
+ }
+ 
+ static int hibvt_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+diff --git a/drivers/pwm/pwm-imx-tpm.c b/drivers/pwm/pwm-imx-tpm.c
+index e5e7b7c339a8f..ed1aad96fff04 100644
+--- a/drivers/pwm/pwm-imx-tpm.c
++++ b/drivers/pwm/pwm-imx-tpm.c
+@@ -132,9 +132,9 @@ static int pwm_imx_tpm_round_state(struct pwm_chip *chip,
+ 	return 0;
+ }
+ 
+-static void pwm_imx_tpm_get_state(struct pwm_chip *chip,
+-				  struct pwm_device *pwm,
+-				  struct pwm_state *state)
++static int pwm_imx_tpm_get_state(struct pwm_chip *chip,
++				 struct pwm_device *pwm,
++				 struct pwm_state *state)
+ {
+ 	struct imx_tpm_pwm_chip *tpm = to_imx_tpm_pwm_chip(chip);
+ 	u32 rate, val, prescale;
+@@ -164,6 +164,8 @@ static void pwm_imx_tpm_get_state(struct pwm_chip *chip,
+ 
+ 	/* get channel status */
+ 	state->enabled = FIELD_GET(PWM_IMX_TPM_CnSC_ELS, val) ? true : false;
++
++	return 0;
+ }
+ 
+ /* this function is supposed to be called with mutex hold */
+diff --git a/drivers/pwm/pwm-imx27.c b/drivers/pwm/pwm-imx27.c
+index ea91a2f81a9fc..3a22c2fddc452 100644
+--- a/drivers/pwm/pwm-imx27.c
++++ b/drivers/pwm/pwm-imx27.c
+@@ -118,8 +118,8 @@ static void pwm_imx27_clk_disable_unprepare(struct pwm_imx27_chip *imx)
+ 	clk_disable_unprepare(imx->clk_ipg);
+ }
+ 
+-static void pwm_imx27_get_state(struct pwm_chip *chip,
+-				struct pwm_device *pwm, struct pwm_state *state)
++static int pwm_imx27_get_state(struct pwm_chip *chip,
++			       struct pwm_device *pwm, struct pwm_state *state)
+ {
+ 	struct pwm_imx27_chip *imx = to_pwm_imx27_chip(chip);
+ 	u32 period, prescaler, pwm_clk, val;
+@@ -128,7 +128,7 @@ static void pwm_imx27_get_state(struct pwm_chip *chip,
+ 
+ 	ret = pwm_imx27_clk_prepare_enable(imx);
+ 	if (ret < 0)
+-		return;
++		return 0;
+ 
+ 	val = readl(imx->mmio_base + MX3_PWMCR);
+ 
+@@ -170,6 +170,8 @@ static void pwm_imx27_get_state(struct pwm_chip *chip,
+ 	state->duty_cycle = DIV_ROUND_UP_ULL(tmp, pwm_clk);
+ 
+ 	pwm_imx27_clk_disable_unprepare(imx);
++
++	return 0;
+ }
+ 
+ static void pwm_imx27_sw_reset(struct pwm_chip *chip)
+diff --git a/drivers/pwm/pwm-intel-lgm.c b/drivers/pwm/pwm-intel-lgm.c
+index b66c350740870..0cd7dd548e82f 100644
+--- a/drivers/pwm/pwm-intel-lgm.c
++++ b/drivers/pwm/pwm-intel-lgm.c
+@@ -86,8 +86,8 @@ static int lgm_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	return lgm_pwm_enable(chip, 1);
+ }
+ 
+-static void lgm_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-			      struct pwm_state *state)
++static int lgm_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++			     struct pwm_state *state)
+ {
+ 	struct lgm_pwm_chip *pc = to_lgm_pwm_chip(chip);
+ 	u32 duty, val;
+@@ -100,6 +100,8 @@ static void lgm_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	regmap_read(pc->regmap, LGM_PWM_FAN_CON0, &val);
+ 	duty = FIELD_GET(LGM_PWM_FAN_DC_MSK, val);
+ 	state->duty_cycle = DIV_ROUND_UP(duty * pc->period, LGM_PWM_MAX_DUTY_CYCLE);
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops lgm_pwm_ops = {
+diff --git a/drivers/pwm/pwm-iqs620a.c b/drivers/pwm/pwm-iqs620a.c
+index 54bd95a5cab0c..aeb19a274accf 100644
+--- a/drivers/pwm/pwm-iqs620a.c
++++ b/drivers/pwm/pwm-iqs620a.c
+@@ -104,8 +104,8 @@ static int iqs620_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	return ret;
+ }
+ 
+-static void iqs620_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-				 struct pwm_state *state)
++static int iqs620_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++				struct pwm_state *state)
+ {
+ 	struct iqs620_pwm_private *iqs620_pwm;
+ 
+@@ -126,6 +126,9 @@ static void iqs620_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	mutex_unlock(&iqs620_pwm->lock);
+ 
+ 	state->period = IQS620_PWM_PERIOD_NS;
++	state->polarity = PWM_POLARITY_NORMAL;
++
++	return 0;
+ }
+ 
+ static int iqs620_pwm_notifier(struct notifier_block *notifier,
+diff --git a/drivers/pwm/pwm-keembay.c b/drivers/pwm/pwm-keembay.c
+index 733811b057219..ac02d8bb4a0b5 100644
+--- a/drivers/pwm/pwm-keembay.c
++++ b/drivers/pwm/pwm-keembay.c
+@@ -89,8 +89,8 @@ static void keembay_pwm_disable(struct keembay_pwm *priv, int ch)
+ 				KMB_PWM_LEADIN_OFFSET(ch));
+ }
+ 
+-static void keembay_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-				  struct pwm_state *state)
++static int keembay_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++				 struct pwm_state *state)
+ {
+ 	struct keembay_pwm *priv = to_keembay_pwm_dev(chip);
+ 	unsigned long long high, low;
+@@ -113,6 +113,8 @@ static void keembay_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	state->duty_cycle = DIV_ROUND_UP_ULL(high, clk_rate);
+ 	state->period = DIV_ROUND_UP_ULL(high + low, clk_rate);
+ 	state->polarity = PWM_POLARITY_NORMAL;
++
++	return 0;
+ }
+ 
+ static int keembay_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+diff --git a/drivers/pwm/pwm-lpss.c b/drivers/pwm/pwm-lpss.c
+index accdef5dd58e7..81ac297b8ba50 100644
+--- a/drivers/pwm/pwm-lpss.c
++++ b/drivers/pwm/pwm-lpss.c
+@@ -205,8 +205,8 @@ static int pwm_lpss_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	return ret;
+ }
+ 
+-static void pwm_lpss_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-			       struct pwm_state *state)
++static int pwm_lpss_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++			      struct pwm_state *state)
+ {
+ 	struct pwm_lpss_chip *lpwm = to_lpwm(chip);
+ 	unsigned long base_unit_range;
+@@ -236,6 +236,8 @@ static void pwm_lpss_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	state->enabled = !!(ctrl & PWM_ENABLE);
+ 
+ 	pm_runtime_put(chip->dev);
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops pwm_lpss_ops = {
+diff --git a/drivers/pwm/pwm-meson.c b/drivers/pwm/pwm-meson.c
+index 57112f438c6dd..5cd7b90872c62 100644
+--- a/drivers/pwm/pwm-meson.c
++++ b/drivers/pwm/pwm-meson.c
+@@ -162,6 +162,12 @@ static int meson_pwm_calc(struct meson_pwm *meson, struct pwm_device *pwm,
+ 	duty = state->duty_cycle;
+ 	period = state->period;
+ 
++	/*
++	 * Note this is wrong. The result is an output wave that isn't really
++	 * inverted and so is wrongly identified by .get_state as normal.
++	 * Fixing this needs some care however as some machines might rely on
++	 * this.
++	 */
+ 	if (state->polarity == PWM_POLARITY_INVERSED)
+ 		duty = period - duty;
+ 
+@@ -318,8 +324,8 @@ static unsigned int meson_pwm_cnt_to_ns(struct pwm_chip *chip,
+ 	return cnt * fin_ns * (channel->pre_div + 1);
+ }
+ 
+-static void meson_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-				struct pwm_state *state)
++static int meson_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++			       struct pwm_state *state)
+ {
+ 	struct meson_pwm *meson = to_meson_pwm(chip);
+ 	struct meson_pwm_channel_data *channel_data;
+@@ -327,7 +333,7 @@ static void meson_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	u32 value, tmp;
+ 
+ 	if (!state)
+-		return;
++		return 0;
+ 
+ 	channel = &meson->channels[pwm->hwpwm];
+ 	channel_data = &meson_pwm_per_channel_data[pwm->hwpwm];
+@@ -357,6 +363,10 @@ static void meson_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 		state->period = 0;
+ 		state->duty_cycle = 0;
+ 	}
++
++	state->polarity = PWM_POLARITY_NORMAL;
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops meson_pwm_ops = {
+diff --git a/drivers/pwm/pwm-mtk-disp.c b/drivers/pwm/pwm-mtk-disp.c
+index 3fbb4bae93a4e..ccf0ccdef29df 100644
+--- a/drivers/pwm/pwm-mtk-disp.c
++++ b/drivers/pwm/pwm-mtk-disp.c
+@@ -172,9 +172,9 @@ static int mtk_disp_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	return 0;
+ }
+ 
+-static void mtk_disp_pwm_get_state(struct pwm_chip *chip,
+-				   struct pwm_device *pwm,
+-				   struct pwm_state *state)
++static int mtk_disp_pwm_get_state(struct pwm_chip *chip,
++				  struct pwm_device *pwm,
++				  struct pwm_state *state)
+ {
+ 	struct mtk_disp_pwm *mdp = to_mtk_disp_pwm(chip);
+ 	u64 rate, period, high_width;
+@@ -184,14 +184,14 @@ static void mtk_disp_pwm_get_state(struct pwm_chip *chip,
+ 	err = clk_prepare_enable(mdp->clk_main);
+ 	if (err < 0) {
+ 		dev_err(chip->dev, "Can't enable mdp->clk_main: %pe\n", ERR_PTR(err));
+-		return;
++		return 0;
+ 	}
+ 
+ 	err = clk_prepare_enable(mdp->clk_mm);
+ 	if (err < 0) {
+ 		dev_err(chip->dev, "Can't enable mdp->clk_mm: %pe\n", ERR_PTR(err));
+ 		clk_disable_unprepare(mdp->clk_main);
+-		return;
++		return 0;
+ 	}
+ 
+ 	rate = clk_get_rate(mdp->clk_main);
+@@ -212,6 +212,8 @@ static void mtk_disp_pwm_get_state(struct pwm_chip *chip,
+ 	state->polarity = PWM_POLARITY_NORMAL;
+ 	clk_disable_unprepare(mdp->clk_mm);
+ 	clk_disable_unprepare(mdp->clk_main);
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops mtk_disp_pwm_ops = {
+diff --git a/drivers/pwm/pwm-pca9685.c b/drivers/pwm/pwm-pca9685.c
+index f230c10d28bb2..41be244e7dd3d 100644
+--- a/drivers/pwm/pwm-pca9685.c
++++ b/drivers/pwm/pwm-pca9685.c
+@@ -431,8 +431,8 @@ static int pca9685_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	return ret;
+ }
+ 
+-static void pca9685_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-				  struct pwm_state *state)
++static int pca9685_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++				 struct pwm_state *state)
+ {
+ 	struct pca9685 *pca = to_pca(chip);
+ 	unsigned long long duty;
+@@ -458,12 +458,14 @@ static void pca9685_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 		 */
+ 		state->duty_cycle = 0;
+ 		state->enabled = false;
+-		return;
++		return 0;
+ 	}
+ 
+ 	state->enabled = true;
+ 	duty = pca9685_pwm_get_duty(pca, pwm->hwpwm);
+ 	state->duty_cycle = DIV_ROUND_DOWN_ULL(duty * state->period, PCA9685_COUNTER_RANGE);
++
++	return 0;
+ }
+ 
+ static int pca9685_pwm_request(struct pwm_chip *chip, struct pwm_device *pwm)
+diff --git a/drivers/pwm/pwm-raspberrypi-poe.c b/drivers/pwm/pwm-raspberrypi-poe.c
+index 6ff73029f367f..2939b71a7ba7b 100644
+--- a/drivers/pwm/pwm-raspberrypi-poe.c
++++ b/drivers/pwm/pwm-raspberrypi-poe.c
+@@ -82,9 +82,9 @@ static int raspberrypi_pwm_get_property(struct rpi_firmware *firmware,
+ 	return 0;
+ }
+ 
+-static void raspberrypi_pwm_get_state(struct pwm_chip *chip,
+-				      struct pwm_device *pwm,
+-				      struct pwm_state *state)
++static int raspberrypi_pwm_get_state(struct pwm_chip *chip,
++				     struct pwm_device *pwm,
++				     struct pwm_state *state)
+ {
+ 	struct raspberrypi_pwm *rpipwm = raspberrypi_pwm_from_chip(chip);
+ 
+@@ -93,6 +93,8 @@ static void raspberrypi_pwm_get_state(struct pwm_chip *chip,
+ 					 RPI_PWM_MAX_DUTY);
+ 	state->enabled = !!(rpipwm->duty_cycle);
+ 	state->polarity = PWM_POLARITY_NORMAL;
++
++	return 0;
+ }
+ 
+ static int raspberrypi_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+diff --git a/drivers/pwm/pwm-rockchip.c b/drivers/pwm/pwm-rockchip.c
+index a5af859217c19..3ec7d17569034 100644
+--- a/drivers/pwm/pwm-rockchip.c
++++ b/drivers/pwm/pwm-rockchip.c
+@@ -57,9 +57,9 @@ static inline struct rockchip_pwm_chip *to_rockchip_pwm_chip(struct pwm_chip *c)
+ 	return container_of(c, struct rockchip_pwm_chip, chip);
+ }
+ 
+-static void rockchip_pwm_get_state(struct pwm_chip *chip,
+-				   struct pwm_device *pwm,
+-				   struct pwm_state *state)
++static int rockchip_pwm_get_state(struct pwm_chip *chip,
++				  struct pwm_device *pwm,
++				  struct pwm_state *state)
+ {
+ 	struct rockchip_pwm_chip *pc = to_rockchip_pwm_chip(chip);
+ 	u32 enable_conf = pc->data->enable_conf;
+@@ -70,11 +70,11 @@ static void rockchip_pwm_get_state(struct pwm_chip *chip,
+ 
+ 	ret = clk_enable(pc->pclk);
+ 	if (ret)
+-		return;
++		return 0;
+ 
+ 	ret = clk_enable(pc->clk);
+ 	if (ret)
+-		return;
++		return 0;
+ 
+ 	clk_rate = clk_get_rate(pc->clk);
+ 
+@@ -96,6 +96,8 @@ static void rockchip_pwm_get_state(struct pwm_chip *chip,
+ 
+ 	clk_disable(pc->clk);
+ 	clk_disable(pc->pclk);
++
++	return 0;
+ }
+ 
+ static void rockchip_pwm_config(struct pwm_chip *chip, struct pwm_device *pwm,
+diff --git a/drivers/pwm/pwm-sifive.c b/drivers/pwm/pwm-sifive.c
+index 89d53a0f91e65..393a4b97fc19e 100644
+--- a/drivers/pwm/pwm-sifive.c
++++ b/drivers/pwm/pwm-sifive.c
+@@ -105,8 +105,8 @@ static void pwm_sifive_update_clock(struct pwm_sifive_ddata *ddata,
+ 		"New real_period = %u ns\n", ddata->real_period);
+ }
+ 
+-static void pwm_sifive_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-				 struct pwm_state *state)
++static int pwm_sifive_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++				struct pwm_state *state)
+ {
+ 	struct pwm_sifive_ddata *ddata = pwm_sifive_chip_to_ddata(chip);
+ 	u32 duty, val;
+@@ -123,6 +123,8 @@ static void pwm_sifive_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	state->duty_cycle =
+ 		(u64)duty * ddata->real_period >> PWM_SIFIVE_CMPWIDTH;
+ 	state->polarity = PWM_POLARITY_INVERSED;
++
++	return 0;
+ }
+ 
+ static int pwm_sifive_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+diff --git a/drivers/pwm/pwm-sl28cpld.c b/drivers/pwm/pwm-sl28cpld.c
+index 589aeaaa6ac86..e64900ad4ba1e 100644
+--- a/drivers/pwm/pwm-sl28cpld.c
++++ b/drivers/pwm/pwm-sl28cpld.c
+@@ -87,9 +87,9 @@ struct sl28cpld_pwm {
+ #define sl28cpld_pwm_from_chip(_chip) \
+ 	container_of(_chip, struct sl28cpld_pwm, pwm_chip)
+ 
+-static void sl28cpld_pwm_get_state(struct pwm_chip *chip,
+-				   struct pwm_device *pwm,
+-				   struct pwm_state *state)
++static int sl28cpld_pwm_get_state(struct pwm_chip *chip,
++				  struct pwm_device *pwm,
++				  struct pwm_state *state)
+ {
+ 	struct sl28cpld_pwm *priv = sl28cpld_pwm_from_chip(chip);
+ 	unsigned int reg;
+@@ -115,6 +115,8 @@ static void sl28cpld_pwm_get_state(struct pwm_chip *chip,
+ 	 * the PWM core.
+ 	 */
+ 	state->duty_cycle = min(state->duty_cycle, state->period);
++
++	return 0;
+ }
+ 
+ static int sl28cpld_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+diff --git a/drivers/pwm/pwm-sprd.c b/drivers/pwm/pwm-sprd.c
+index 7004f55bbf115..87d5cb7f67d67 100644
+--- a/drivers/pwm/pwm-sprd.c
++++ b/drivers/pwm/pwm-sprd.c
+@@ -65,8 +65,8 @@ static void sprd_pwm_write(struct sprd_pwm_chip *spc, u32 hwid,
+ 	writel_relaxed(val, spc->base + offset);
+ }
+ 
+-static void sprd_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-			       struct pwm_state *state)
++static int sprd_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++			      struct pwm_state *state)
+ {
+ 	struct sprd_pwm_chip *spc =
+ 		container_of(chip, struct sprd_pwm_chip, chip);
+@@ -83,7 +83,7 @@ static void sprd_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	if (ret) {
+ 		dev_err(spc->dev, "failed to enable pwm%u clocks\n",
+ 			pwm->hwpwm);
+-		return;
++		return 0;
+ 	}
+ 
+ 	val = sprd_pwm_read(spc, pwm->hwpwm, SPRD_PWM_ENABLE);
+@@ -109,10 +109,13 @@ static void sprd_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	duty = val & SPRD_PWM_DUTY_MSK;
+ 	tmp = (prescale + 1) * NSEC_PER_SEC * duty;
+ 	state->duty_cycle = DIV_ROUND_CLOSEST_ULL(tmp, chn->clk_rate);
++	state->polarity = PWM_POLARITY_NORMAL;
+ 
+ 	/* Disable PWM clocks if the PWM channel is not in enable state. */
+ 	if (!state->enabled)
+ 		clk_bulk_disable_unprepare(SPRD_PWM_CHN_CLKS_NUM, chn->clks);
++
++	return 0;
+ }
+ 
+ static int sprd_pwm_config(struct sprd_pwm_chip *spc, struct pwm_device *pwm,
+diff --git a/drivers/pwm/pwm-stm32-lp.c b/drivers/pwm/pwm-stm32-lp.c
+index 61a1c87cd5016..31a185c6b8da4 100644
+--- a/drivers/pwm/pwm-stm32-lp.c
++++ b/drivers/pwm/pwm-stm32-lp.c
+@@ -157,9 +157,9 @@ err:
+ 	return ret;
+ }
+ 
+-static void stm32_pwm_lp_get_state(struct pwm_chip *chip,
+-				   struct pwm_device *pwm,
+-				   struct pwm_state *state)
++static int stm32_pwm_lp_get_state(struct pwm_chip *chip,
++				  struct pwm_device *pwm,
++				  struct pwm_state *state)
+ {
+ 	struct stm32_pwm_lp *priv = to_stm32_pwm_lp(chip);
+ 	unsigned long rate = clk_get_rate(priv->clk);
+@@ -185,6 +185,8 @@ static void stm32_pwm_lp_get_state(struct pwm_chip *chip,
+ 	tmp = prd - val;
+ 	tmp = (tmp << presc) * NSEC_PER_SEC;
+ 	state->duty_cycle = DIV_ROUND_CLOSEST_ULL(tmp, rate);
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops stm32_pwm_lp_ops = {
+diff --git a/drivers/pwm/pwm-sun4i.c b/drivers/pwm/pwm-sun4i.c
+index c8445b0a33392..37d75e252d4e7 100644
+--- a/drivers/pwm/pwm-sun4i.c
++++ b/drivers/pwm/pwm-sun4i.c
+@@ -108,9 +108,9 @@ static inline void sun4i_pwm_writel(struct sun4i_pwm_chip *chip,
+ 	writel(val, chip->base + offset);
+ }
+ 
+-static void sun4i_pwm_get_state(struct pwm_chip *chip,
+-				struct pwm_device *pwm,
+-				struct pwm_state *state)
++static int sun4i_pwm_get_state(struct pwm_chip *chip,
++			       struct pwm_device *pwm,
++			       struct pwm_state *state)
+ {
+ 	struct sun4i_pwm_chip *sun4i_pwm = to_sun4i_pwm_chip(chip);
+ 	u64 clk_rate, tmp;
+@@ -132,7 +132,7 @@ static void sun4i_pwm_get_state(struct pwm_chip *chip,
+ 		state->duty_cycle = DIV_ROUND_UP_ULL(state->period, 2);
+ 		state->polarity = PWM_POLARITY_NORMAL;
+ 		state->enabled = true;
+-		return;
++		return 0;
+ 	}
+ 
+ 	if ((PWM_REG_PRESCAL(val, pwm->hwpwm) == PWM_PRESCAL_MASK) &&
+@@ -142,7 +142,7 @@ static void sun4i_pwm_get_state(struct pwm_chip *chip,
+ 		prescaler = prescaler_table[PWM_REG_PRESCAL(val, pwm->hwpwm)];
+ 
+ 	if (prescaler == 0)
+-		return;
++		return 0;
+ 
+ 	if (val & BIT_CH(PWM_ACT_STATE, pwm->hwpwm))
+ 		state->polarity = PWM_POLARITY_NORMAL;
+@@ -162,6 +162,8 @@ static void sun4i_pwm_get_state(struct pwm_chip *chip,
+ 
+ 	tmp = (u64)prescaler * NSEC_PER_SEC * PWM_REG_PRD(val);
+ 	state->period = DIV_ROUND_CLOSEST_ULL(tmp, clk_rate);
++
++	return 0;
+ }
+ 
+ static int sun4i_pwm_calculate(struct sun4i_pwm_chip *sun4i_pwm,
+diff --git a/drivers/pwm/pwm-sunplus.c b/drivers/pwm/pwm-sunplus.c
+index e776fd16512de..d6ebe9f03b354 100644
+--- a/drivers/pwm/pwm-sunplus.c
++++ b/drivers/pwm/pwm-sunplus.c
+@@ -124,8 +124,8 @@ static int sunplus_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	return 0;
+ }
+ 
+-static void sunplus_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-				  struct pwm_state *state)
++static int sunplus_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++				 struct pwm_state *state)
+ {
+ 	struct sunplus_pwm *priv = to_sunplus_pwm(chip);
+ 	u32 mode0, dd_freq, duty;
+@@ -155,6 +155,8 @@ static void sunplus_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	}
+ 
+ 	state->polarity = PWM_POLARITY_NORMAL;
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops sunplus_pwm_ops = {
+diff --git a/drivers/pwm/pwm-visconti.c b/drivers/pwm/pwm-visconti.c
+index 927c4cbb1daf0..e3fb79b3e2a7a 100644
+--- a/drivers/pwm/pwm-visconti.c
++++ b/drivers/pwm/pwm-visconti.c
+@@ -103,8 +103,8 @@ static int visconti_pwm_apply(struct pwm_chip *chip, struct pwm_device *pwm,
+ 	return 0;
+ }
+ 
+-static void visconti_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
+-				   struct pwm_state *state)
++static int visconti_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm,
++				  struct pwm_state *state)
+ {
+ 	struct visconti_pwm_chip *priv = visconti_pwm_from_chip(chip);
+ 	u32 period, duty, pwmc0, pwmc0_clk;
+@@ -122,6 +122,8 @@ static void visconti_pwm_get_state(struct pwm_chip *chip, struct pwm_device *pwm
+ 		state->polarity = PWM_POLARITY_NORMAL;
+ 
+ 	state->enabled = true;
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops visconti_pwm_ops = {
+diff --git a/drivers/pwm/pwm-xilinx.c b/drivers/pwm/pwm-xilinx.c
+index 4dab2b86c4276..f7a50fdcd9a52 100644
+--- a/drivers/pwm/pwm-xilinx.c
++++ b/drivers/pwm/pwm-xilinx.c
+@@ -169,9 +169,9 @@ static int xilinx_pwm_apply(struct pwm_chip *chip, struct pwm_device *unused,
+ 	return 0;
+ }
+ 
+-static void xilinx_pwm_get_state(struct pwm_chip *chip,
+-				 struct pwm_device *unused,
+-				 struct pwm_state *state)
++static int xilinx_pwm_get_state(struct pwm_chip *chip,
++				struct pwm_device *unused,
++				struct pwm_state *state)
+ {
+ 	struct xilinx_timer_priv *priv = xilinx_pwm_chip_to_priv(chip);
+ 	u32 tlr0, tlr1, tcsr0, tcsr1;
+@@ -191,6 +191,8 @@ static void xilinx_pwm_get_state(struct pwm_chip *chip,
+ 	 */
+ 	if (state->period == state->duty_cycle)
+ 		state->duty_cycle = 0;
++
++	return 0;
+ }
+ 
+ static const struct pwm_ops xilinx_pwm_ops = {
+diff --git a/drivers/scsi/iscsi_tcp.c b/drivers/scsi/iscsi_tcp.c
+index c3ad04ad66e0a..8009eab3b7bee 100644
+--- a/drivers/scsi/iscsi_tcp.c
++++ b/drivers/scsi/iscsi_tcp.c
+@@ -767,13 +767,12 @@ static int iscsi_sw_tcp_conn_set_param(struct iscsi_cls_conn *cls_conn,
+ 		iscsi_set_param(cls_conn, param, buf, buflen);
+ 		break;
+ 	case ISCSI_PARAM_DATADGST_EN:
+-		iscsi_set_param(cls_conn, param, buf, buflen);
+-
+ 		mutex_lock(&tcp_sw_conn->sock_lock);
+ 		if (!tcp_sw_conn->sock) {
+ 			mutex_unlock(&tcp_sw_conn->sock_lock);
+ 			return -ENOTCONN;
+ 		}
++		iscsi_set_param(cls_conn, param, buf, buflen);
+ 		tcp_sw_conn->sendpage = conn->datadgst_en ?
+ 			sock_no_sendpage : tcp_sw_conn->sock->ops->sendpage;
+ 		mutex_unlock(&tcp_sw_conn->sock_lock);
+diff --git a/drivers/scsi/qla2xxx/qla_os.c b/drivers/scsi/qla2xxx/qla_os.c
+index 7d2d872bae3c5..08dc825fbf4f6 100644
+--- a/drivers/scsi/qla2xxx/qla_os.c
++++ b/drivers/scsi/qla2xxx/qla_os.c
+@@ -3608,6 +3608,7 @@ skip_dpc:
+ probe_failed:
+ 	qla_enode_stop(base_vha);
+ 	qla_edb_stop(base_vha);
++	vfree(base_vha->scan.l);
+ 	if (base_vha->gnl.l) {
+ 		dma_free_coherent(&ha->pdev->dev, base_vha->gnl.size,
+ 				base_vha->gnl.l, base_vha->gnl.ldma);
+diff --git a/drivers/tty/serial/8250/8250_port.c b/drivers/tty/serial/8250/8250_port.c
+index 388172289627a..2a3bd6918c77e 100644
+--- a/drivers/tty/serial/8250/8250_port.c
++++ b/drivers/tty/serial/8250/8250_port.c
+@@ -1897,6 +1897,17 @@ EXPORT_SYMBOL_GPL(serial8250_modem_status);
+ static bool handle_rx_dma(struct uart_8250_port *up, unsigned int iir)
+ {
+ 	switch (iir & 0x3f) {
++	case UART_IIR_THRI:
++		/*
++		 * Postpone DMA or not decision to IIR_RDI or IIR_RX_TIMEOUT
++		 * because it's impossible to do an informed decision about
++		 * that with IIR_THRI.
++		 *
++		 * This also fixes one known DMA Rx corruption issue where
++		 * DR is asserted but DMA Rx only gets a corrupted zero byte
++		 * (too early DR?).
++		 */
++		return false;
+ 	case UART_IIR_RDI:
+ 		if (!up->dma->rx_running)
+ 			break;
+diff --git a/drivers/tty/serial/fsl_lpuart.c b/drivers/tty/serial/fsl_lpuart.c
+index cd98c04de0330..48eb5fea62fd0 100644
+--- a/drivers/tty/serial/fsl_lpuart.c
++++ b/drivers/tty/serial/fsl_lpuart.c
+@@ -838,11 +838,17 @@ static unsigned int lpuart32_tx_empty(struct uart_port *port)
+ 			struct lpuart_port, port);
+ 	unsigned long stat = lpuart32_read(port, UARTSTAT);
+ 	unsigned long sfifo = lpuart32_read(port, UARTFIFO);
++	unsigned long ctrl = lpuart32_read(port, UARTCTRL);
+ 
+ 	if (sport->dma_tx_in_progress)
+ 		return 0;
+ 
+-	if (stat & UARTSTAT_TC && sfifo & UARTFIFO_TXEMPT)
++	/*
++	 * LPUART Transmission Complete Flag may never be set while queuing a break
++	 * character, so avoid checking for transmission complete when UARTCTRL_SBK
++	 * is asserted.
++	 */
++	if ((stat & UARTSTAT_TC && sfifo & UARTFIFO_TXEMPT) || ctrl & UARTCTRL_SBK)
+ 		return TIOCSER_TEMT;
+ 
+ 	return 0;
+diff --git a/drivers/tty/serial/sh-sci.c b/drivers/tty/serial/sh-sci.c
+index 62f773286d44b..e67d3a886bf4f 100644
+--- a/drivers/tty/serial/sh-sci.c
++++ b/drivers/tty/serial/sh-sci.c
+@@ -31,6 +31,7 @@
+ #include <linux/ioport.h>
+ #include <linux/ktime.h>
+ #include <linux/major.h>
++#include <linux/minmax.h>
+ #include <linux/module.h>
+ #include <linux/mm.h>
+ #include <linux/of.h>
+@@ -2867,6 +2868,13 @@ static int sci_init_single(struct platform_device *dev,
+ 			sci_port->irqs[i] = platform_get_irq(dev, i);
+ 	}
+ 
++	/*
++	 * The fourth interrupt on SCI port is transmit end interrupt, so
++	 * shuffle the interrupts.
++	 */
++	if (p->type == PORT_SCI)
++		swap(sci_port->irqs[SCIx_BRI_IRQ], sci_port->irqs[SCIx_TEI_IRQ]);
++
+ 	/* The SCI generates several interrupts. They can be muxed together or
+ 	 * connected to different interrupt lines. In the muxed case only one
+ 	 * interrupt resource is specified as there is only one interrupt ID.
+@@ -2932,7 +2940,7 @@ static int sci_init_single(struct platform_device *dev,
+ 	port->flags		= UPF_FIXED_PORT | UPF_BOOT_AUTOCONF | p->flags;
+ 	port->fifosize		= sci_port->params->fifosize;
+ 
+-	if (port->type == PORT_SCI) {
++	if (port->type == PORT_SCI && !dev->dev.of_node) {
+ 		if (sci_port->reg_size >= 0x20)
+ 			port->regshift = 2;
+ 		else
+diff --git a/drivers/usb/cdns3/cdnsp-ep0.c b/drivers/usb/cdns3/cdnsp-ep0.c
+index d63d5d92f2554..f317d3c847810 100644
+--- a/drivers/usb/cdns3/cdnsp-ep0.c
++++ b/drivers/usb/cdns3/cdnsp-ep0.c
+@@ -414,7 +414,7 @@ static int cdnsp_ep0_std_request(struct cdnsp_device *pdev,
+ void cdnsp_setup_analyze(struct cdnsp_device *pdev)
+ {
+ 	struct usb_ctrlrequest *ctrl = &pdev->setup;
+-	int ret = 0;
++	int ret = -EINVAL;
+ 	u16 len;
+ 
+ 	trace_cdnsp_ctrl_req(ctrl);
+@@ -424,7 +424,6 @@ void cdnsp_setup_analyze(struct cdnsp_device *pdev)
+ 
+ 	if (pdev->gadget.state == USB_STATE_NOTATTACHED) {
+ 		dev_err(pdev->dev, "ERR: Setup detected in unattached state\n");
+-		ret = -EINVAL;
+ 		goto out;
+ 	}
+ 
+diff --git a/drivers/usb/dwc3/dwc3-pci.c b/drivers/usb/dwc3/dwc3-pci.c
+index a23ddbb819795..560793545362a 100644
+--- a/drivers/usb/dwc3/dwc3-pci.c
++++ b/drivers/usb/dwc3/dwc3-pci.c
+@@ -49,6 +49,7 @@
+ #define PCI_DEVICE_ID_INTEL_RPLS		0x7a61
+ #define PCI_DEVICE_ID_INTEL_MTLM		0x7eb1
+ #define PCI_DEVICE_ID_INTEL_MTLP		0x7ec1
++#define PCI_DEVICE_ID_INTEL_MTLS		0x7f6f
+ #define PCI_DEVICE_ID_INTEL_MTL			0x7e7e
+ #define PCI_DEVICE_ID_INTEL_TGL			0x9a15
+ #define PCI_DEVICE_ID_AMD_MR			0x163a
+@@ -474,6 +475,9 @@ static const struct pci_device_id dwc3_pci_id_table[] = {
+ 	{ PCI_VDEVICE(INTEL, PCI_DEVICE_ID_INTEL_MTLP),
+ 	  (kernel_ulong_t) &dwc3_pci_intel_swnode, },
+ 
++	{ PCI_VDEVICE(INTEL, PCI_DEVICE_ID_INTEL_MTLS),
++	  (kernel_ulong_t) &dwc3_pci_intel_swnode, },
++
+ 	{ PCI_VDEVICE(INTEL, PCI_DEVICE_ID_INTEL_MTL),
+ 	  (kernel_ulong_t) &dwc3_pci_intel_swnode, },
+ 
+diff --git a/drivers/usb/host/xhci-tegra.c b/drivers/usb/host/xhci-tegra.c
+index bdb776553826b..32df571bb2339 100644
+--- a/drivers/usb/host/xhci-tegra.c
++++ b/drivers/usb/host/xhci-tegra.c
+@@ -1225,6 +1225,9 @@ static void tegra_xhci_id_work(struct work_struct *work)
+ 
+ 	mutex_unlock(&tegra->lock);
+ 
++	tegra->otg_usb3_port = tegra_xusb_padctl_get_usb3_companion(tegra->padctl,
++								    tegra->otg_usb2_port);
++
+ 	if (tegra->host_mode) {
+ 		/* switch to host mode */
+ 		if (tegra->otg_usb3_port >= 0) {
+@@ -1339,9 +1342,6 @@ static int tegra_xhci_id_notify(struct notifier_block *nb,
+ 	}
+ 
+ 	tegra->otg_usb2_port = tegra_xusb_get_usb2_port(tegra, usbphy);
+-	tegra->otg_usb3_port = tegra_xusb_padctl_get_usb3_companion(
+-							tegra->padctl,
+-							tegra->otg_usb2_port);
+ 
+ 	tegra->host_mode = (usbphy->last_event == USB_EVENT_ID) ? true : false;
+ 
+diff --git a/drivers/usb/host/xhci.c b/drivers/usb/host/xhci.c
+index 2b280beb00115..c02ad4f76bb3c 100644
+--- a/drivers/usb/host/xhci.c
++++ b/drivers/usb/host/xhci.c
+@@ -9,6 +9,7 @@
+  */
+ 
+ #include <linux/pci.h>
++#include <linux/iommu.h>
+ #include <linux/iopoll.h>
+ #include <linux/irq.h>
+ #include <linux/log2.h>
+@@ -228,6 +229,7 @@ int xhci_reset(struct xhci_hcd *xhci, u64 timeout_us)
+ static void xhci_zero_64b_regs(struct xhci_hcd *xhci)
+ {
+ 	struct device *dev = xhci_to_hcd(xhci)->self.sysdev;
++	struct iommu_domain *domain;
+ 	int err, i;
+ 	u64 val;
+ 	u32 intrs;
+@@ -246,7 +248,9 @@ static void xhci_zero_64b_regs(struct xhci_hcd *xhci)
+ 	 * an iommu. Doing anything when there is no iommu is definitely
+ 	 * unsafe...
+ 	 */
+-	if (!(xhci->quirks & XHCI_ZERO_64B_REGS) || !device_iommu_mapped(dev))
++	domain = iommu_get_domain_for_dev(dev);
++	if (!(xhci->quirks & XHCI_ZERO_64B_REGS) || !domain ||
++	    domain->type == IOMMU_DOMAIN_IDENTITY)
+ 		return;
+ 
+ 	xhci_info(xhci, "Zeroing 64bit base registers, expecting fault\n");
+@@ -4406,6 +4410,7 @@ static int __maybe_unused xhci_change_max_exit_latency(struct xhci_hcd *xhci,
+ 
+ 	if (!virt_dev || max_exit_latency == virt_dev->current_mel) {
+ 		spin_unlock_irqrestore(&xhci->lock, flags);
++		xhci_free_command(xhci, command);
+ 		return 0;
+ 	}
+ 
+diff --git a/drivers/usb/serial/cp210x.c b/drivers/usb/serial/cp210x.c
+index ba5638471de49..f1d7a5a863aa4 100644
+--- a/drivers/usb/serial/cp210x.c
++++ b/drivers/usb/serial/cp210x.c
+@@ -120,6 +120,7 @@ static const struct usb_device_id id_table[] = {
+ 	{ USB_DEVICE(0x10C4, 0x826B) }, /* Cygnal Integrated Products, Inc., Fasttrax GPS demonstration module */
+ 	{ USB_DEVICE(0x10C4, 0x8281) }, /* Nanotec Plug & Drive */
+ 	{ USB_DEVICE(0x10C4, 0x8293) }, /* Telegesis ETRX2USB */
++	{ USB_DEVICE(0x10C4, 0x82AA) }, /* Silicon Labs IFS-USB-DATACABLE used with Quint UPS */
+ 	{ USB_DEVICE(0x10C4, 0x82EF) }, /* CESINEL FALCO 6105 AC Power Supply */
+ 	{ USB_DEVICE(0x10C4, 0x82F1) }, /* CESINEL MEDCAL EFD Earth Fault Detector */
+ 	{ USB_DEVICE(0x10C4, 0x82F2) }, /* CESINEL MEDCAL ST Network Analyzer */
+diff --git a/drivers/usb/serial/option.c b/drivers/usb/serial/option.c
+index a8534065e0d6d..fc12fee66141f 100644
+--- a/drivers/usb/serial/option.c
++++ b/drivers/usb/serial/option.c
+@@ -1198,6 +1198,8 @@ static const struct usb_device_id option_ids[] = {
+ 	{ USB_DEVICE_AND_INTERFACE_INFO(QUECTEL_VENDOR_ID, QUECTEL_PRODUCT_RM520N, 0xff, 0xff, 0x30) },
+ 	{ USB_DEVICE_AND_INTERFACE_INFO(QUECTEL_VENDOR_ID, QUECTEL_PRODUCT_RM520N, 0xff, 0, 0x40) },
+ 	{ USB_DEVICE_AND_INTERFACE_INFO(QUECTEL_VENDOR_ID, QUECTEL_PRODUCT_RM520N, 0xff, 0, 0) },
++	{ USB_DEVICE_AND_INTERFACE_INFO(QUECTEL_VENDOR_ID, 0x0900, 0xff, 0, 0), /* RM500U-CN */
++	  .driver_info = ZLP },
+ 	{ USB_DEVICE_AND_INTERFACE_INFO(QUECTEL_VENDOR_ID, QUECTEL_PRODUCT_EC200U, 0xff, 0, 0) },
+ 	{ USB_DEVICE_AND_INTERFACE_INFO(QUECTEL_VENDOR_ID, QUECTEL_PRODUCT_EC200S_CN, 0xff, 0, 0) },
+ 	{ USB_DEVICE_AND_INTERFACE_INFO(QUECTEL_VENDOR_ID, QUECTEL_PRODUCT_EC200T, 0xff, 0, 0) },
+@@ -1300,6 +1302,14 @@ static const struct usb_device_id option_ids[] = {
+ 	  .driver_info = NCTRL(0) | RSVD(1) },
+ 	{ USB_DEVICE_INTERFACE_CLASS(TELIT_VENDOR_ID, 0x1075, 0xff),	/* Telit FN990 (PCIe) */
+ 	  .driver_info = RSVD(0) },
++	{ USB_DEVICE_INTERFACE_CLASS(TELIT_VENDOR_ID, 0x1080, 0xff),	/* Telit FE990 (rmnet) */
++	  .driver_info = NCTRL(0) | RSVD(1) | RSVD(2) },
++	{ USB_DEVICE_INTERFACE_CLASS(TELIT_VENDOR_ID, 0x1081, 0xff),	/* Telit FE990 (MBIM) */
++	  .driver_info = NCTRL(0) | RSVD(1) },
++	{ USB_DEVICE_INTERFACE_CLASS(TELIT_VENDOR_ID, 0x1082, 0xff),	/* Telit FE990 (RNDIS) */
++	  .driver_info = NCTRL(2) | RSVD(3) },
++	{ USB_DEVICE_INTERFACE_CLASS(TELIT_VENDOR_ID, 0x1083, 0xff),	/* Telit FE990 (ECM) */
++	  .driver_info = NCTRL(0) | RSVD(1) },
+ 	{ USB_DEVICE(TELIT_VENDOR_ID, TELIT_PRODUCT_ME910),
+ 	  .driver_info = NCTRL(0) | RSVD(1) | RSVD(3) },
+ 	{ USB_DEVICE(TELIT_VENDOR_ID, TELIT_PRODUCT_ME910_DUAL_MODEM),
+diff --git a/drivers/usb/typec/altmodes/displayport.c b/drivers/usb/typec/altmodes/displayport.c
+index 3a42313d0d66e..4075c0d7e6a2c 100644
+--- a/drivers/usb/typec/altmodes/displayport.c
++++ b/drivers/usb/typec/altmodes/displayport.c
+@@ -112,8 +112,12 @@ static int dp_altmode_configure(struct dp_altmode *dp, u8 con)
+ 		if (dp->data.status & DP_STATUS_PREFER_MULTI_FUNC &&
+ 		    pin_assign & DP_PIN_ASSIGN_MULTI_FUNC_MASK)
+ 			pin_assign &= DP_PIN_ASSIGN_MULTI_FUNC_MASK;
+-		else if (pin_assign & DP_PIN_ASSIGN_DP_ONLY_MASK)
++		else if (pin_assign & DP_PIN_ASSIGN_DP_ONLY_MASK) {
+ 			pin_assign &= DP_PIN_ASSIGN_DP_ONLY_MASK;
++			/* Default to pin assign C if available */
++			if (pin_assign & BIT(DP_PIN_ASSIGN_C))
++				pin_assign = BIT(DP_PIN_ASSIGN_C);
++		}
+ 
+ 		if (!pin_assign)
+ 			return -EINVAL;
+diff --git a/fs/cifs/fs_context.c b/fs/cifs/fs_context.c
+index 45119597c7655..89e810b27a4bf 100644
+--- a/fs/cifs/fs_context.c
++++ b/fs/cifs/fs_context.c
+@@ -441,13 +441,14 @@ out:
+  * but there are some bugs that prevent rename from working if there are
+  * multiple delimiters.
+  *
+- * Returns a sanitized duplicate of @path. The caller is responsible for
+- * cleaning up the original.
++ * Returns a sanitized duplicate of @path. @gfp indicates the GFP_* flags
++ * for kstrdup.
++ * The caller is responsible for freeing the original.
+  */
+ #define IS_DELIM(c) ((c) == '/' || (c) == '\\')
+-static char *sanitize_path(char *path)
++char *cifs_sanitize_prepath(char *prepath, gfp_t gfp)
+ {
+-	char *cursor1 = path, *cursor2 = path;
++	char *cursor1 = prepath, *cursor2 = prepath;
+ 
+ 	/* skip all prepended delimiters */
+ 	while (IS_DELIM(*cursor1))
+@@ -469,7 +470,7 @@ static char *sanitize_path(char *path)
+ 		cursor2--;
+ 
+ 	*(cursor2) = '\0';
+-	return kstrdup(path, GFP_KERNEL);
++	return kstrdup(prepath, gfp);
+ }
+ 
+ /*
+@@ -531,7 +532,7 @@ smb3_parse_devname(const char *devname, struct smb3_fs_context *ctx)
+ 	if (!*pos)
+ 		return 0;
+ 
+-	ctx->prepath = sanitize_path(pos);
++	ctx->prepath = cifs_sanitize_prepath(pos, GFP_KERNEL);
+ 	if (!ctx->prepath)
+ 		return -ENOMEM;
+ 
+diff --git a/fs/cifs/fs_context.h b/fs/cifs/fs_context.h
+index a268896d05d57..26093f54d3e65 100644
+--- a/fs/cifs/fs_context.h
++++ b/fs/cifs/fs_context.h
+@@ -287,4 +287,7 @@ extern void smb3_update_mnt_flags(struct cifs_sb_info *cifs_sb);
+  */
+ #define SMB3_MAX_DCLOSETIMEO (1 << 30)
+ #define SMB3_DEF_DCLOSETIMEO (1 * HZ) /* even 1 sec enough to help eg open/write/close/open/read */
++
++extern char *cifs_sanitize_prepath(char *prepath, gfp_t gfp);
++
+ #endif
+diff --git a/fs/cifs/misc.c b/fs/cifs/misc.c
+index 832856aef4b7a..cf19e6a81ed99 100644
+--- a/fs/cifs/misc.c
++++ b/fs/cifs/misc.c
+@@ -1304,7 +1304,7 @@ int cifs_update_super_prepath(struct cifs_sb_info *cifs_sb, char *prefix)
+ 	kfree(cifs_sb->prepath);
+ 
+ 	if (prefix && *prefix) {
+-		cifs_sb->prepath = kstrdup(prefix, GFP_ATOMIC);
++		cifs_sb->prepath = cifs_sanitize_prepath(prefix, GFP_ATOMIC);
+ 		if (!cifs_sb->prepath)
+ 			return -ENOMEM;
+ 
+diff --git a/fs/ksmbd/connection.c b/fs/ksmbd/connection.c
+index 2be9d7460494b..b8f9d627f241d 100644
+--- a/fs/ksmbd/connection.c
++++ b/fs/ksmbd/connection.c
+@@ -326,10 +326,7 @@ int ksmbd_conn_handler_loop(void *p)
+ 
+ 		/* 4 for rfc1002 length field */
+ 		size = pdu_size + 4;
+-		conn->request_buf = kvmalloc(size,
+-					     GFP_KERNEL |
+-					     __GFP_NOWARN |
+-					     __GFP_NORETRY);
++		conn->request_buf = kvmalloc(size, GFP_KERNEL);
+ 		if (!conn->request_buf)
+ 			break;
+ 
+diff --git a/fs/ksmbd/server.c b/fs/ksmbd/server.c
+index a0d635304754a..651d1d01234be 100644
+--- a/fs/ksmbd/server.c
++++ b/fs/ksmbd/server.c
+@@ -289,10 +289,7 @@ static int queue_ksmbd_work(struct ksmbd_conn *conn)
+ 	work->request_buf = conn->request_buf;
+ 	conn->request_buf = NULL;
+ 
+-	if (ksmbd_init_smb_server(work)) {
+-		ksmbd_free_work_struct(work);
+-		return -EINVAL;
+-	}
++	ksmbd_init_smb_server(work);
+ 
+ 	ksmbd_conn_enqueue_request(work);
+ 	atomic_inc(&conn->r_count);
+diff --git a/fs/ksmbd/smb2pdu.c b/fs/ksmbd/smb2pdu.c
+index daaee7a89e050..4effe8df5ae92 100644
+--- a/fs/ksmbd/smb2pdu.c
++++ b/fs/ksmbd/smb2pdu.c
+@@ -235,9 +235,6 @@ int init_smb2_neg_rsp(struct ksmbd_work *work)
+ 	struct smb2_negotiate_rsp *rsp;
+ 	struct ksmbd_conn *conn = work->conn;
+ 
+-	if (conn->need_neg == false)
+-		return -EINVAL;
+-
+ 	*(__be32 *)work->response_buf =
+ 		cpu_to_be32(conn->vals->header_size);
+ 
+diff --git a/fs/ksmbd/smb_common.c b/fs/ksmbd/smb_common.c
+index 3d9e8d8a5762b..95afb6b23a91c 100644
+--- a/fs/ksmbd/smb_common.c
++++ b/fs/ksmbd/smb_common.c
+@@ -283,20 +283,121 @@ err_out:
+ 	return BAD_PROT_ID;
+ }
+ 
+-int ksmbd_init_smb_server(struct ksmbd_work *work)
++#define SMB_COM_NEGOTIATE_EX	0x0
++
++/**
++ * get_smb1_cmd_val() - get smb command value from smb header
++ * @work:	smb work containing smb header
++ *
++ * Return:      smb command value
++ */
++static u16 get_smb1_cmd_val(struct ksmbd_work *work)
+ {
+-	struct ksmbd_conn *conn = work->conn;
++	return SMB_COM_NEGOTIATE_EX;
++}
+ 
+-	if (conn->need_neg == false)
++/**
++ * init_smb1_rsp_hdr() - initialize smb negotiate response header
++ * @work:	smb work containing smb request
++ *
++ * Return:      0 on success, otherwise -EINVAL
++ */
++static int init_smb1_rsp_hdr(struct ksmbd_work *work)
++{
++	struct smb_hdr *rsp_hdr = (struct smb_hdr *)work->response_buf;
++	struct smb_hdr *rcv_hdr = (struct smb_hdr *)work->request_buf;
++
++	/*
++	 * Remove 4 byte direct TCP header.
++	 */
++	*(__be32 *)work->response_buf =
++		cpu_to_be32(sizeof(struct smb_hdr) - 4);
++
++	rsp_hdr->Command = SMB_COM_NEGOTIATE;
++	*(__le32 *)rsp_hdr->Protocol = SMB1_PROTO_NUMBER;
++	rsp_hdr->Flags = SMBFLG_RESPONSE;
++	rsp_hdr->Flags2 = SMBFLG2_UNICODE | SMBFLG2_ERR_STATUS |
++		SMBFLG2_EXT_SEC | SMBFLG2_IS_LONG_NAME;
++	rsp_hdr->Pid = rcv_hdr->Pid;
++	rsp_hdr->Mid = rcv_hdr->Mid;
++	return 0;
++}
++
++/**
++ * smb1_check_user_session() - check for valid session for a user
++ * @work:	smb work containing smb request buffer
++ *
++ * Return:      0 on success, otherwise error
++ */
++static int smb1_check_user_session(struct ksmbd_work *work)
++{
++	unsigned int cmd = work->conn->ops->get_cmd_val(work);
++
++	if (cmd == SMB_COM_NEGOTIATE_EX)
+ 		return 0;
+ 
+-	init_smb3_11_server(conn);
++	return -EINVAL;
++}
++
++/**
++ * smb1_allocate_rsp_buf() - allocate response buffer for a command
++ * @work:	smb work containing smb request
++ *
++ * Return:      0 on success, otherwise -ENOMEM
++ */
++static int smb1_allocate_rsp_buf(struct ksmbd_work *work)
++{
++	work->response_buf = kmalloc(MAX_CIFS_SMALL_BUFFER_SIZE,
++			GFP_KERNEL | __GFP_ZERO);
++	work->response_sz = MAX_CIFS_SMALL_BUFFER_SIZE;
++
++	if (!work->response_buf) {
++		pr_err("Failed to allocate %u bytes buffer\n",
++				MAX_CIFS_SMALL_BUFFER_SIZE);
++		return -ENOMEM;
++	}
+ 
+-	if (conn->ops->get_cmd_val(work) != SMB_COM_NEGOTIATE)
+-		conn->need_neg = false;
+ 	return 0;
+ }
+ 
++static struct smb_version_ops smb1_server_ops = {
++	.get_cmd_val = get_smb1_cmd_val,
++	.init_rsp_hdr = init_smb1_rsp_hdr,
++	.allocate_rsp_buf = smb1_allocate_rsp_buf,
++	.check_user_session = smb1_check_user_session,
++};
++
++static int smb1_negotiate(struct ksmbd_work *work)
++{
++	return ksmbd_smb_negotiate_common(work, SMB_COM_NEGOTIATE);
++}
++
++static struct smb_version_cmds smb1_server_cmds[1] = {
++	[SMB_COM_NEGOTIATE_EX]	= { .proc = smb1_negotiate, },
++};
++
++static void init_smb1_server(struct ksmbd_conn *conn)
++{
++	conn->ops = &smb1_server_ops;
++	conn->cmds = smb1_server_cmds;
++	conn->max_cmds = ARRAY_SIZE(smb1_server_cmds);
++}
++
++void ksmbd_init_smb_server(struct ksmbd_work *work)
++{
++	struct ksmbd_conn *conn = work->conn;
++	__le32 proto;
++
++	if (conn->need_neg == false)
++		return;
++
++	proto = *(__le32 *)((struct smb_hdr *)work->request_buf)->Protocol;
++	if (proto == SMB1_PROTO_NUMBER)
++		init_smb1_server(conn);
++	else
++		init_smb3_11_server(conn);
++}
++
+ int ksmbd_populate_dot_dotdot_entries(struct ksmbd_work *work, int info_level,
+ 				      struct ksmbd_file *dir,
+ 				      struct ksmbd_dir_info *d_info,
+@@ -444,20 +545,10 @@ static int smb_handle_negotiate(struct ksmbd_work *work)
+ 
+ 	ksmbd_debug(SMB, "Unsupported SMB1 protocol\n");
+ 
+-	/*
+-	 * Remove 4 byte direct TCP header, add 2 byte bcc and
+-	 * 2 byte DialectIndex.
+-	 */
+-	*(__be32 *)work->response_buf =
+-		cpu_to_be32(sizeof(struct smb_hdr) - 4 + 2 + 2);
++	/* Add 2 byte bcc and 2 byte DialectIndex. */
++	inc_rfc1001_len(work->response_buf, 4);
+ 	neg_rsp->hdr.Status.CifsError = STATUS_SUCCESS;
+ 
+-	neg_rsp->hdr.Command = SMB_COM_NEGOTIATE;
+-	*(__le32 *)neg_rsp->hdr.Protocol = SMB1_PROTO_NUMBER;
+-	neg_rsp->hdr.Flags = SMBFLG_RESPONSE;
+-	neg_rsp->hdr.Flags2 = SMBFLG2_UNICODE | SMBFLG2_ERR_STATUS |
+-		SMBFLG2_EXT_SEC | SMBFLG2_IS_LONG_NAME;
+-
+ 	neg_rsp->hdr.WordCount = 1;
+ 	neg_rsp->DialectIndex = cpu_to_le16(work->conn->dialect);
+ 	neg_rsp->ByteCount = 0;
+@@ -473,24 +564,13 @@ int ksmbd_smb_negotiate_common(struct ksmbd_work *work, unsigned int command)
+ 		ksmbd_negotiate_smb_dialect(work->request_buf);
+ 	ksmbd_debug(SMB, "conn->dialect 0x%x\n", conn->dialect);
+ 
+-	if (command == SMB2_NEGOTIATE_HE) {
+-		struct smb2_hdr *smb2_hdr = smb2_get_msg(work->request_buf);
+-
+-		if (smb2_hdr->ProtocolId != SMB2_PROTO_NUMBER) {
+-			ksmbd_debug(SMB, "Downgrade to SMB1 negotiation\n");
+-			command = SMB_COM_NEGOTIATE;
+-		}
+-	}
+-
+ 	if (command == SMB2_NEGOTIATE_HE) {
+ 		ret = smb2_handle_negotiate(work);
+-		init_smb2_neg_rsp(work);
+ 		return ret;
+ 	}
+ 
+ 	if (command == SMB_COM_NEGOTIATE) {
+ 		if (__smb2_negotiate(conn)) {
+-			conn->need_neg = true;
+ 			init_smb3_11_server(conn);
+ 			init_smb2_neg_rsp(work);
+ 			ksmbd_debug(SMB, "Upgrade to SMB2 negotiation\n");
+diff --git a/fs/ksmbd/smb_common.h b/fs/ksmbd/smb_common.h
+index c1f3006792ff6..78c44978a906a 100644
+--- a/fs/ksmbd/smb_common.h
++++ b/fs/ksmbd/smb_common.h
+@@ -427,7 +427,7 @@ bool ksmbd_smb_request(struct ksmbd_conn *conn);
+ 
+ int ksmbd_lookup_dialect_by_id(__le16 *cli_dialects, __le16 dialects_count);
+ 
+-int ksmbd_init_smb_server(struct ksmbd_work *work);
++void ksmbd_init_smb_server(struct ksmbd_work *work);
+ 
+ struct ksmbd_kstat;
+ int ksmbd_populate_dot_dotdot_entries(struct ksmbd_work *work,
+diff --git a/fs/namespace.c b/fs/namespace.c
+index df137ba19d375..e04a9e9e3f14f 100644
+--- a/fs/namespace.c
++++ b/fs/namespace.c
+@@ -4180,9 +4180,9 @@ out:
+ 	unlock_mount_hash();
+ 
+ 	if (kattr->propagation) {
+-		namespace_unlock();
+ 		if (err)
+ 			cleanup_group_ids(mnt, NULL);
++		namespace_unlock();
+ 	}
+ 
+ 	return err;
+diff --git a/fs/nfsd/blocklayout.c b/fs/nfsd/blocklayout.c
+index b6d01d51a7465..e7e6e78d965db 100644
+--- a/fs/nfsd/blocklayout.c
++++ b/fs/nfsd/blocklayout.c
+@@ -296,6 +296,7 @@ nfsd4_block_get_device_info_scsi(struct super_block *sb,
+ 
+ out_free_dev:
+ 	kfree(dev);
++	gdp->gd_device = NULL;
+ 	return ret;
+ }
+ 
+diff --git a/fs/nfsd/nfs4callback.c b/fs/nfsd/nfs4callback.c
+index 6253cbe5f81b4..39989c14c8a1e 100644
+--- a/fs/nfsd/nfs4callback.c
++++ b/fs/nfsd/nfs4callback.c
+@@ -874,8 +874,8 @@ static const struct cred *get_backchannel_cred(struct nfs4_client *clp, struct r
+ 		if (!kcred)
+ 			return NULL;
+ 
+-		kcred->uid = ses->se_cb_sec.uid;
+-		kcred->gid = ses->se_cb_sec.gid;
++		kcred->fsuid = ses->se_cb_sec.uid;
++		kcred->fsgid = ses->se_cb_sec.gid;
+ 		return kcred;
+ 	}
+ }
+diff --git a/fs/nfsd/nfs4xdr.c b/fs/nfsd/nfs4xdr.c
+index 8377e14b8fba9..8f5b41dc07734 100644
+--- a/fs/nfsd/nfs4xdr.c
++++ b/fs/nfsd/nfs4xdr.c
+@@ -2417,10 +2417,12 @@ nfsd4_decode_compound(struct nfsd4_compoundargs *argp)
+ 	for (i = 0; i < argp->opcnt; i++) {
+ 		op = &argp->ops[i];
+ 		op->replay = NULL;
++		op->opdesc = NULL;
+ 
+ 		if (xdr_stream_decode_u32(argp->xdr, &op->opnum) < 0)
+ 			return false;
+ 		if (nfsd4_opnum_in_range(argp, op)) {
++			op->opdesc = OPDESC(op);
+ 			op->status = nfsd4_dec_ops[op->opnum](argp, &op->u);
+ 			if (op->status != nfs_ok)
+ 				trace_nfsd_compound_decode_err(argp->rqstp,
+@@ -2431,7 +2433,7 @@ nfsd4_decode_compound(struct nfsd4_compoundargs *argp)
+ 			op->opnum = OP_ILLEGAL;
+ 			op->status = nfserr_op_illegal;
+ 		}
+-		op->opdesc = OPDESC(op);
++
+ 		/*
+ 		 * We'll try to cache the result in the DRC if any one
+ 		 * op in the compound wants to be cached:
+@@ -5351,10 +5353,8 @@ nfsd4_encode_operation(struct nfsd4_compoundres *resp, struct nfsd4_op *op)
+ 	__be32 *p;
+ 
+ 	p = xdr_reserve_space(xdr, 8);
+-	if (!p) {
+-		WARN_ON_ONCE(1);
+-		return;
+-	}
++	if (!p)
++		goto release;
+ 	*p++ = cpu_to_be32(op->opnum);
+ 	post_err_offset = xdr->buf->len;
+ 
+@@ -5369,8 +5369,6 @@ nfsd4_encode_operation(struct nfsd4_compoundres *resp, struct nfsd4_op *op)
+ 	op->status = encoder(resp, op->status, &op->u);
+ 	if (op->status)
+ 		trace_nfsd_compound_encode_err(rqstp, op->opnum, op->status);
+-	if (opdesc && opdesc->op_release)
+-		opdesc->op_release(&op->u);
+ 	xdr_commit_encode(xdr);
+ 
+ 	/* nfsd4_check_resp_size guarantees enough room for error status */
+@@ -5411,6 +5409,9 @@ nfsd4_encode_operation(struct nfsd4_compoundres *resp, struct nfsd4_op *op)
+ 	}
+ status:
+ 	*p = op->status;
++release:
++	if (opdesc && opdesc->op_release)
++		opdesc->op_release(&op->u);
+ }
+ 
+ /* 
+diff --git a/fs/nilfs2/segment.c b/fs/nilfs2/segment.c
+index 3335ef3529155..63d96a1733b2a 100644
+--- a/fs/nilfs2/segment.c
++++ b/fs/nilfs2/segment.c
+@@ -2607,11 +2607,10 @@ static int nilfs_segctor_thread(void *arg)
+ 	goto loop;
+ 
+  end_thread:
+-	spin_unlock(&sci->sc_state_lock);
+-
+ 	/* end sync. */
+ 	sci->sc_task = NULL;
+ 	wake_up(&sci->sc_wait_task); /* for nilfs_segctor_kill_thread() */
++	spin_unlock(&sci->sc_state_lock);
+ 	return 0;
+ }
+ 
+diff --git a/fs/nilfs2/super.c b/fs/nilfs2/super.c
+index 1422b8ba24ed6..77f1e5778d1c8 100644
+--- a/fs/nilfs2/super.c
++++ b/fs/nilfs2/super.c
+@@ -482,6 +482,7 @@ static void nilfs_put_super(struct super_block *sb)
+ 		up_write(&nilfs->ns_sem);
+ 	}
+ 
++	nilfs_sysfs_delete_device_group(nilfs);
+ 	iput(nilfs->ns_sufile);
+ 	iput(nilfs->ns_cpfile);
+ 	iput(nilfs->ns_dat);
+@@ -1105,6 +1106,7 @@ nilfs_fill_super(struct super_block *sb, void *data, int silent)
+ 	nilfs_put_root(fsroot);
+ 
+  failed_unload:
++	nilfs_sysfs_delete_device_group(nilfs);
+ 	iput(nilfs->ns_sufile);
+ 	iput(nilfs->ns_cpfile);
+ 	iput(nilfs->ns_dat);
+diff --git a/fs/nilfs2/the_nilfs.c b/fs/nilfs2/the_nilfs.c
+index 3a4c9c150cbf5..2894152a6b25c 100644
+--- a/fs/nilfs2/the_nilfs.c
++++ b/fs/nilfs2/the_nilfs.c
+@@ -87,7 +87,6 @@ void destroy_nilfs(struct the_nilfs *nilfs)
+ {
+ 	might_sleep();
+ 	if (nilfs_init(nilfs)) {
+-		nilfs_sysfs_delete_device_group(nilfs);
+ 		brelse(nilfs->ns_sbh[0]);
+ 		brelse(nilfs->ns_sbh[1]);
+ 	}
+@@ -305,6 +304,10 @@ int load_nilfs(struct the_nilfs *nilfs, struct super_block *sb)
+ 		goto failed;
+ 	}
+ 
++	err = nilfs_sysfs_create_device_group(sb);
++	if (unlikely(err))
++		goto sysfs_error;
++
+ 	if (valid_fs)
+ 		goto skip_recovery;
+ 
+@@ -366,6 +369,9 @@ int load_nilfs(struct the_nilfs *nilfs, struct super_block *sb)
+ 	goto failed;
+ 
+  failed_unload:
++	nilfs_sysfs_delete_device_group(nilfs);
++
++ sysfs_error:
+ 	iput(nilfs->ns_cpfile);
+ 	iput(nilfs->ns_sufile);
+ 	iput(nilfs->ns_dat);
+@@ -697,10 +703,6 @@ int init_nilfs(struct the_nilfs *nilfs, struct super_block *sb, char *data)
+ 	if (err)
+ 		goto failed_sbh;
+ 
+-	err = nilfs_sysfs_create_device_group(sb);
+-	if (err)
+-		goto failed_sbh;
+-
+ 	set_nilfs_init(nilfs);
+ 	err = 0;
+  out:
+diff --git a/include/acpi/video.h b/include/acpi/video.h
+index 8ed9bec03e534..ff5a8da5d8832 100644
+--- a/include/acpi/video.h
++++ b/include/acpi/video.h
+@@ -59,8 +59,6 @@ extern void acpi_video_unregister(void);
+ extern void acpi_video_register_backlight(void);
+ extern int acpi_video_get_edid(struct acpi_device *device, int type,
+ 			       int device_id, void **edid);
+-extern enum acpi_backlight_type acpi_video_get_backlight_type(void);
+-extern bool acpi_video_backlight_use_native(void);
+ /*
+  * Note: The value returned by acpi_video_handles_brightness_key_presses()
+  * may change over time and should not be cached.
+@@ -69,6 +67,19 @@ extern bool acpi_video_handles_brightness_key_presses(void);
+ extern int acpi_video_get_levels(struct acpi_device *device,
+ 				 struct acpi_video_device_brightness **dev_br,
+ 				 int *pmax_level);
++
++extern enum acpi_backlight_type __acpi_video_get_backlight_type(bool native,
++								bool *auto_detect);
++
++static inline enum acpi_backlight_type acpi_video_get_backlight_type(void)
++{
++	return __acpi_video_get_backlight_type(false, NULL);
++}
++
++static inline bool acpi_video_backlight_use_native(void)
++{
++	return __acpi_video_get_backlight_type(true, NULL) == acpi_backlight_native;
++}
+ #else
+ static inline void acpi_video_report_nolcd(void) { return; };
+ static inline int acpi_video_register(void) { return -ENODEV; }
+diff --git a/include/drm/display/drm_dp_mst_helper.h b/include/drm/display/drm_dp_mst_helper.h
+index 6c0c3dc3d3aca..32c764fb9cb56 100644
+--- a/include/drm/display/drm_dp_mst_helper.h
++++ b/include/drm/display/drm_dp_mst_helper.h
+@@ -841,7 +841,8 @@ int drm_dp_add_payload_part2(struct drm_dp_mst_topology_mgr *mgr,
+ 			     struct drm_dp_mst_atomic_payload *payload);
+ void drm_dp_remove_payload(struct drm_dp_mst_topology_mgr *mgr,
+ 			   struct drm_dp_mst_topology_state *mst_state,
+-			   struct drm_dp_mst_atomic_payload *payload);
++			   const struct drm_dp_mst_atomic_payload *old_payload,
++			   struct drm_dp_mst_atomic_payload *new_payload);
+ 
+ int drm_dp_check_act_status(struct drm_dp_mst_topology_mgr *mgr);
+ 
+diff --git a/include/kvm/arm_pmu.h b/include/kvm/arm_pmu.h
+index c0b868ce6a8f2..96b192139a23a 100644
+--- a/include/kvm/arm_pmu.h
++++ b/include/kvm/arm_pmu.h
+@@ -11,7 +11,6 @@
+ #include <asm/perf_event.h>
+ 
+ #define ARMV8_PMU_CYCLE_IDX		(ARMV8_PMU_MAX_COUNTERS - 1)
+-#define ARMV8_PMU_MAX_COUNTER_PAIRS	((ARMV8_PMU_MAX_COUNTERS + 1) >> 1)
+ 
+ #ifdef CONFIG_HW_PERF_EVENTS
+ 
+@@ -29,7 +28,6 @@ struct kvm_pmu {
+ 	struct irq_work overflow_work;
+ 	struct kvm_pmu_events events;
+ 	struct kvm_pmc pmc[ARMV8_PMU_MAX_COUNTERS];
+-	DECLARE_BITMAP(chained, ARMV8_PMU_MAX_COUNTER_PAIRS);
+ 	int irq_num;
+ 	bool created;
+ 	bool irq_level;
+diff --git a/include/linux/device-mapper.h b/include/linux/device-mapper.h
+index 04c6acf7faaa5..201dd1ab7f1c6 100644
+--- a/include/linux/device-mapper.h
++++ b/include/linux/device-mapper.h
+@@ -87,10 +87,10 @@ typedef int (*dm_preresume_fn) (struct dm_target *ti);
+ typedef void (*dm_resume_fn) (struct dm_target *ti);
+ 
+ typedef void (*dm_status_fn) (struct dm_target *ti, status_type_t status_type,
+-			      unsigned status_flags, char *result, unsigned maxlen);
++			      unsigned int status_flags, char *result, unsigned int maxlen);
+ 
+-typedef int (*dm_message_fn) (struct dm_target *ti, unsigned argc, char **argv,
+-			      char *result, unsigned maxlen);
++typedef int (*dm_message_fn) (struct dm_target *ti, unsigned int argc, char **argv,
++			      char *result, unsigned int maxlen);
+ 
+ typedef int (*dm_prepare_ioctl_fn) (struct dm_target *ti, struct block_device **bdev);
+ 
+@@ -187,7 +187,7 @@ struct target_type {
+ 	uint64_t features;
+ 	const char *name;
+ 	struct module *module;
+-	unsigned version[3];
++	unsigned int version[3];
+ 	dm_ctr_fn ctr;
+ 	dm_dtr_fn dtr;
+ 	dm_map_fn map;
+@@ -313,31 +313,31 @@ struct dm_target {
+ 	 * It is a responsibility of the target driver to remap these bios
+ 	 * to the real underlying devices.
+ 	 */
+-	unsigned num_flush_bios;
++	unsigned int num_flush_bios;
+ 
+ 	/*
+ 	 * The number of discard bios that will be submitted to the target.
+ 	 * The bio number can be accessed with dm_bio_get_target_bio_nr.
+ 	 */
+-	unsigned num_discard_bios;
++	unsigned int num_discard_bios;
+ 
+ 	/*
+ 	 * The number of secure erase bios that will be submitted to the target.
+ 	 * The bio number can be accessed with dm_bio_get_target_bio_nr.
+ 	 */
+-	unsigned num_secure_erase_bios;
++	unsigned int num_secure_erase_bios;
+ 
+ 	/*
+ 	 * The number of WRITE ZEROES bios that will be submitted to the target.
+ 	 * The bio number can be accessed with dm_bio_get_target_bio_nr.
+ 	 */
+-	unsigned num_write_zeroes_bios;
++	unsigned int num_write_zeroes_bios;
+ 
+ 	/*
+ 	 * The minimum number of extra bytes allocated in each io for the
+ 	 * target to use.
+ 	 */
+-	unsigned per_io_data_size;
++	unsigned int per_io_data_size;
+ 
+ 	/* target specific data */
+ 	void *private;
+@@ -383,7 +383,7 @@ struct dm_target {
+ 
+ void *dm_per_bio_data(struct bio *bio, size_t data_size);
+ struct bio *dm_bio_from_per_bio_data(void *data, size_t data_size);
+-unsigned dm_bio_get_target_bio_nr(const struct bio *bio);
++unsigned int dm_bio_get_target_bio_nr(const struct bio *bio);
+ 
+ u64 dm_start_time_ns_from_clone(struct bio *bio);
+ 
+@@ -394,7 +394,7 @@ void dm_unregister_target(struct target_type *t);
+  * Target argument parsing.
+  */
+ struct dm_arg_set {
+-	unsigned argc;
++	unsigned int argc;
+ 	char **argv;
+ };
+ 
+@@ -403,8 +403,8 @@ struct dm_arg_set {
+  * the error message to use if the number is found to be outside that range.
+  */
+ struct dm_arg {
+-	unsigned min;
+-	unsigned max;
++	unsigned int min;
++	unsigned int max;
+ 	char *error;
+ };
+ 
+@@ -413,7 +413,7 @@ struct dm_arg {
+  * returning -EINVAL and setting *error.
+  */
+ int dm_read_arg(const struct dm_arg *arg, struct dm_arg_set *arg_set,
+-		unsigned *value, char **error);
++		unsigned int *value, char **error);
+ 
+ /*
+  * Process the next argument as the start of a group containing between
+@@ -421,7 +421,7 @@ int dm_read_arg(const struct dm_arg *arg, struct dm_arg_set *arg_set,
+  * *num_args or, if invalid, return -EINVAL and set *error.
+  */
+ int dm_read_arg_group(const struct dm_arg *arg, struct dm_arg_set *arg_set,
+-		      unsigned *num_args, char **error);
++		      unsigned int *num_args, char **error);
+ 
+ /*
+  * Return the current argument and shift to the next.
+@@ -431,7 +431,7 @@ const char *dm_shift_arg(struct dm_arg_set *as);
+ /*
+  * Move through num_args arguments.
+  */
+-void dm_consume_args(struct dm_arg_set *as, unsigned num_args);
++void dm_consume_args(struct dm_arg_set *as, unsigned int num_args);
+ 
+ /*-----------------------------------------------------------------
+  * Functions for creating and manipulating mapped devices.
+@@ -461,7 +461,7 @@ void *dm_get_mdptr(struct mapped_device *md);
+ /*
+  * A device can still be used while suspended, but I/O is deferred.
+  */
+-int dm_suspend(struct mapped_device *md, unsigned suspend_flags);
++int dm_suspend(struct mapped_device *md, unsigned int suspend_flags);
+ int dm_resume(struct mapped_device *md);
+ 
+ /*
+@@ -481,7 +481,7 @@ struct gendisk *dm_disk(struct mapped_device *md);
+ int dm_suspended(struct dm_target *ti);
+ int dm_post_suspending(struct dm_target *ti);
+ int dm_noflush_suspending(struct dm_target *ti);
+-void dm_accept_partial_bio(struct bio *bio, unsigned n_sectors);
++void dm_accept_partial_bio(struct bio *bio, unsigned int n_sectors);
+ void dm_submit_bio_remap(struct bio *clone, struct bio *tgt_clone);
+ union map_info *dm_get_rq_mapinfo(struct request *rq);
+ 
+@@ -525,7 +525,7 @@ int dm_set_geometry(struct mapped_device *md, struct hd_geometry *geo);
+  * First create an empty table.
+  */
+ int dm_table_create(struct dm_table **result, fmode_t mode,
+-		    unsigned num_targets, struct mapped_device *md);
++		    unsigned int num_targets, struct mapped_device *md);
+ 
+ /*
+  * Then call this once for each target.
+diff --git a/include/linux/dm-bufio.h b/include/linux/dm-bufio.h
+index 15d9e15ca830d..1262d92ab88fc 100644
+--- a/include/linux/dm-bufio.h
++++ b/include/linux/dm-bufio.h
+@@ -26,8 +26,8 @@ struct dm_buffer;
+  * Create a buffered IO cache on a given device
+  */
+ struct dm_bufio_client *
+-dm_bufio_client_create(struct block_device *bdev, unsigned block_size,
+-		       unsigned reserved_buffers, unsigned aux_size,
++dm_bufio_client_create(struct block_device *bdev, unsigned int block_size,
++		       unsigned int reserved_buffers, unsigned int aux_size,
+ 		       void (*alloc_callback)(struct dm_buffer *),
+ 		       void (*write_callback)(struct dm_buffer *),
+ 		       unsigned int flags);
+@@ -81,7 +81,7 @@ void *dm_bufio_new(struct dm_bufio_client *c, sector_t block,
+  * I/O to finish.
+  */
+ void dm_bufio_prefetch(struct dm_bufio_client *c,
+-		       sector_t block, unsigned n_blocks);
++		       sector_t block, unsigned int n_blocks);
+ 
+ /*
+  * Release a reference obtained with dm_bufio_{read,get,new}. The data
+@@ -106,7 +106,7 @@ void dm_bufio_mark_buffer_dirty(struct dm_buffer *b);
+  * write the specified part of the buffer or it may write a larger superset.
+  */
+ void dm_bufio_mark_partial_buffer_dirty(struct dm_buffer *b,
+-					unsigned start, unsigned end);
++					unsigned int start, unsigned int end);
+ 
+ /*
+  * Initiate writing of dirty buffers, without waiting for completion.
+@@ -152,9 +152,9 @@ void dm_bufio_forget_buffers(struct dm_bufio_client *c, sector_t block, sector_t
+ /*
+  * Set the minimum number of buffers before cleanup happens.
+  */
+-void dm_bufio_set_minimum_buffers(struct dm_bufio_client *c, unsigned n);
++void dm_bufio_set_minimum_buffers(struct dm_bufio_client *c, unsigned int n);
+ 
+-unsigned dm_bufio_get_block_size(struct dm_bufio_client *c);
++unsigned int dm_bufio_get_block_size(struct dm_bufio_client *c);
+ sector_t dm_bufio_get_device_size(struct dm_bufio_client *c);
+ struct dm_io_client *dm_bufio_get_dm_io_client(struct dm_bufio_client *c);
+ sector_t dm_bufio_get_block_number(struct dm_buffer *b);
+diff --git a/include/linux/dm-dirty-log.h b/include/linux/dm-dirty-log.h
+index 7084503c3405f..843c857f07b0d 100644
+--- a/include/linux/dm-dirty-log.h
++++ b/include/linux/dm-dirty-log.h
+@@ -33,7 +33,7 @@ struct dm_dirty_log_type {
+ 	struct list_head list;
+ 
+ 	int (*ctr)(struct dm_dirty_log *log, struct dm_target *ti,
+-		   unsigned argc, char **argv);
++		   unsigned int argc, char **argv);
+ 	void (*dtr)(struct dm_dirty_log *log);
+ 
+ 	/*
+@@ -116,7 +116,7 @@ struct dm_dirty_log_type {
+ 	 * Support function for mirror status requests.
+ 	 */
+ 	int (*status)(struct dm_dirty_log *log, status_type_t status_type,
+-		      char *result, unsigned maxlen);
++		      char *result, unsigned int maxlen);
+ 
+ 	/*
+ 	 * is_remote_recovering is necessary for cluster mirroring. It provides
+@@ -139,7 +139,7 @@ int dm_dirty_log_type_unregister(struct dm_dirty_log_type *type);
+ struct dm_dirty_log *dm_dirty_log_create(const char *type_name,
+ 			struct dm_target *ti,
+ 			int (*flush_callback_fn)(struct dm_target *ti),
+-			unsigned argc, char **argv);
++			unsigned int argc, char **argv);
+ void dm_dirty_log_destroy(struct dm_dirty_log *log);
+ 
+ #endif	/* __KERNEL__ */
+diff --git a/include/linux/dm-io.h b/include/linux/dm-io.h
+index 8e1c4ab5df043..92e7abfe04f92 100644
+--- a/include/linux/dm-io.h
++++ b/include/linux/dm-io.h
+@@ -26,7 +26,7 @@ struct page_list {
+ 	struct page *page;
+ };
+ 
+-typedef void (*io_notify_fn)(unsigned long error, void *context);
++typedef void (*io_notify_fn)(unsigned int long error, void *context);
+ 
+ enum dm_io_mem_type {
+ 	DM_IO_PAGE_LIST,/* Page list */
+@@ -38,7 +38,7 @@ enum dm_io_mem_type {
+ struct dm_io_memory {
+ 	enum dm_io_mem_type type;
+ 
+-	unsigned offset;
++	unsigned int offset;
+ 
+ 	union {
+ 		struct page_list *pl;
+@@ -78,8 +78,8 @@ void dm_io_client_destroy(struct dm_io_client *client);
+  * Each bit in the optional 'sync_error_bits' bitset indicates whether an
+  * error occurred doing io to the corresponding region.
+  */
+-int dm_io(struct dm_io_request *io_req, unsigned num_regions,
+-	  struct dm_io_region *region, unsigned long *sync_error_bits);
++int dm_io(struct dm_io_request *io_req, unsigned int num_regions,
++	  struct dm_io_region *region, unsigned int long *sync_error_bits);
+ 
+ #endif	/* __KERNEL__ */
+ #endif	/* _LINUX_DM_IO_H */
+diff --git a/include/linux/dm-kcopyd.h b/include/linux/dm-kcopyd.h
+index c1707ee5b5408..68c412b31b788 100644
+--- a/include/linux/dm-kcopyd.h
++++ b/include/linux/dm-kcopyd.h
+@@ -23,11 +23,11 @@
+ #define DM_KCOPYD_WRITE_SEQ    2
+ 
+ struct dm_kcopyd_throttle {
+-	unsigned throttle;
+-	unsigned num_io_jobs;
+-	unsigned io_period;
+-	unsigned total_period;
+-	unsigned last_jiffies;
++	unsigned int throttle;
++	unsigned int num_io_jobs;
++	unsigned int io_period;
++	unsigned int total_period;
++	unsigned int last_jiffies;
+ };
+ 
+ /*
+@@ -60,12 +60,12 @@ void dm_kcopyd_client_flush(struct dm_kcopyd_client *kc);
+  * read_err is a boolean,
+  * write_err is a bitset, with 1 bit for each destination region
+  */
+-typedef void (*dm_kcopyd_notify_fn)(int read_err, unsigned long write_err,
++typedef void (*dm_kcopyd_notify_fn)(int read_err, unsigned int long write_err,
+ 				    void *context);
+ 
+ void dm_kcopyd_copy(struct dm_kcopyd_client *kc, struct dm_io_region *from,
+-		    unsigned num_dests, struct dm_io_region *dests,
+-		    unsigned flags, dm_kcopyd_notify_fn fn, void *context);
++		    unsigned int num_dests, struct dm_io_region *dests,
++		    unsigned int flags, dm_kcopyd_notify_fn fn, void *context);
+ 
+ /*
+  * Prepare a callback and submit it via the kcopyd thread.
+@@ -80,11 +80,11 @@ void dm_kcopyd_copy(struct dm_kcopyd_client *kc, struct dm_io_region *from,
+  */
+ void *dm_kcopyd_prepare_callback(struct dm_kcopyd_client *kc,
+ 				 dm_kcopyd_notify_fn fn, void *context);
+-void dm_kcopyd_do_callback(void *job, int read_err, unsigned long write_err);
++void dm_kcopyd_do_callback(void *job, int read_err, unsigned int long write_err);
+ 
+ void dm_kcopyd_zero(struct dm_kcopyd_client *kc,
+-		    unsigned num_dests, struct dm_io_region *dests,
+-		    unsigned flags, dm_kcopyd_notify_fn fn, void *context);
++		    unsigned int num_dests, struct dm_io_region *dests,
++		    unsigned int flags, dm_kcopyd_notify_fn fn, void *context);
+ 
+ #endif	/* __KERNEL__ */
+ #endif	/* _LINUX_DM_KCOPYD_H */
+diff --git a/include/linux/dm-region-hash.h b/include/linux/dm-region-hash.h
+index 9e2a7a401df50..e8691539e1d77 100644
+--- a/include/linux/dm-region-hash.h
++++ b/include/linux/dm-region-hash.h
+@@ -37,7 +37,7 @@ struct dm_region_hash *dm_region_hash_create(
+ 						     struct bio_list *bios),
+ 		void (*wakeup_workers)(void *context),
+ 		void (*wakeup_all_recovery_waiters)(void *context),
+-		sector_t target_begin, unsigned max_recovery,
++		sector_t target_begin, unsigned int max_recovery,
+ 		struct dm_dirty_log *log, uint32_t region_size,
+ 		region_t nr_regions);
+ void dm_region_hash_destroy(struct dm_region_hash *rh);
+diff --git a/include/linux/ftrace.h b/include/linux/ftrace.h
+index 62557d4bffc2b..8128059db5ed7 100644
+--- a/include/linux/ftrace.h
++++ b/include/linux/ftrace.h
+@@ -943,7 +943,7 @@ static inline void __ftrace_enabled_restore(int enabled)
+ #define CALLER_ADDR5 ((unsigned long)ftrace_return_address(5))
+ #define CALLER_ADDR6 ((unsigned long)ftrace_return_address(6))
+ 
+-static inline unsigned long get_lock_parent_ip(void)
++static __always_inline unsigned long get_lock_parent_ip(void)
+ {
+ 	unsigned long addr = CALLER_ADDR0;
+ 
+diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h
+index 500e536796ca4..247aedb18d5c3 100644
+--- a/include/linux/mm_types.h
++++ b/include/linux/mm_types.h
+@@ -725,7 +725,8 @@ struct mm_struct {
+ 	unsigned long cpu_bitmap[];
+ };
+ 
+-#define MM_MT_FLAGS	(MT_FLAGS_ALLOC_RANGE | MT_FLAGS_LOCK_EXTERN)
++#define MM_MT_FLAGS	(MT_FLAGS_ALLOC_RANGE | MT_FLAGS_LOCK_EXTERN | \
++			 MT_FLAGS_USE_RCU)
+ extern struct mm_struct init_mm;
+ 
+ /* Pointer magic because the dynamic array size confuses some compilers. */
+diff --git a/include/linux/pci-doe.h b/include/linux/pci-doe.h
+index ed9b4df792b88..43765eaf2342c 100644
+--- a/include/linux/pci-doe.h
++++ b/include/linux/pci-doe.h
+@@ -34,6 +34,10 @@ struct pci_doe_mb;
+  * @work: Used internally by the mailbox
+  * @doe_mb: Used internally by the mailbox
+  *
++ * Payloads are treated as opaque byte streams which are transmitted verbatim,
++ * without byte-swapping.  If payloads contain little-endian register values,
++ * the caller is responsible for conversion with cpu_to_le32() / le32_to_cpu().
++ *
+  * The payload sizes and rv are specified in bytes with the following
+  * restrictions concerning the protocol.
+  *
+@@ -45,9 +49,9 @@ struct pci_doe_mb;
+  */
+ struct pci_doe_task {
+ 	struct pci_doe_protocol prot;
+-	u32 *request_pl;
++	__le32 *request_pl;
+ 	size_t request_pl_sz;
+-	u32 *response_pl;
++	__le32 *response_pl;
+ 	size_t response_pl_sz;
+ 	int rv;
+ 	void (*complete)(struct pci_doe_task *task);
+diff --git a/include/linux/phylink.h b/include/linux/phylink.h
+index 3f01ac8017e06..6d62c9ea2e040 100644
+--- a/include/linux/phylink.h
++++ b/include/linux/phylink.h
+@@ -566,6 +566,7 @@ struct phylink *phylink_create(struct phylink_config *, struct fwnode_handle *,
+ 			       phy_interface_t iface,
+ 			       const struct phylink_mac_ops *mac_ops);
+ void phylink_destroy(struct phylink *);
++bool phylink_expects_phy(struct phylink *pl);
+ 
+ int phylink_connect_phy(struct phylink *, struct phy_device *);
+ int phylink_of_phy_connect(struct phylink *, struct device_node *, u32 flags);
+diff --git a/include/linux/pwm.h b/include/linux/pwm.h
+index d70c6e5a839d6..4de09163c968a 100644
+--- a/include/linux/pwm.h
++++ b/include/linux/pwm.h
+@@ -276,8 +276,8 @@ struct pwm_ops {
+ 		       struct pwm_capture *result, unsigned long timeout);
+ 	int (*apply)(struct pwm_chip *chip, struct pwm_device *pwm,
+ 		     const struct pwm_state *state);
+-	void (*get_state)(struct pwm_chip *chip, struct pwm_device *pwm,
+-			  struct pwm_state *state);
++	int (*get_state)(struct pwm_chip *chip, struct pwm_device *pwm,
++			 struct pwm_state *state);
+ 	struct module *owner;
+ };
+ 
+diff --git a/include/net/raw.h b/include/net/raw.h
+index 5e665934ebc7c..3af5289fdead9 100644
+--- a/include/net/raw.h
++++ b/include/net/raw.h
+@@ -15,6 +15,8 @@
+ 
+ #include <net/inet_sock.h>
+ #include <net/protocol.h>
++#include <net/netns/hash.h>
++#include <linux/hash.h>
+ #include <linux/icmp.h>
+ 
+ extern struct proto raw_prot;
+@@ -29,20 +31,27 @@ int raw_local_deliver(struct sk_buff *, int);
+ 
+ int raw_rcv(struct sock *, struct sk_buff *);
+ 
+-#define RAW_HTABLE_SIZE	MAX_INET_PROTOS
++#define RAW_HTABLE_LOG	8
++#define RAW_HTABLE_SIZE	(1U << RAW_HTABLE_LOG)
+ 
+ struct raw_hashinfo {
+ 	spinlock_t lock;
+-	struct hlist_nulls_head ht[RAW_HTABLE_SIZE];
++
++	struct hlist_head ht[RAW_HTABLE_SIZE] ____cacheline_aligned;
+ };
+ 
++static inline u32 raw_hashfunc(const struct net *net, u32 proto)
++{
++	return hash_32(net_hash_mix(net) ^ proto, RAW_HTABLE_LOG);
++}
++
+ static inline void raw_hashinfo_init(struct raw_hashinfo *hashinfo)
+ {
+ 	int i;
+ 
+ 	spin_lock_init(&hashinfo->lock);
+ 	for (i = 0; i < RAW_HTABLE_SIZE; i++)
+-		INIT_HLIST_NULLS_HEAD(&hashinfo->ht[i], i);
++		INIT_HLIST_HEAD(&hashinfo->ht[i]);
+ }
+ 
+ #ifdef CONFIG_PROC_FS
+diff --git a/io_uring/io_uring.c b/io_uring/io_uring.c
+index ce4969d3e20de..cc35aba1e4957 100644
+--- a/io_uring/io_uring.c
++++ b/io_uring/io_uring.c
+@@ -2585,8 +2585,8 @@ static __cold void io_ring_ctx_free(struct io_ring_ctx *ctx)
+ 	io_eventfd_unregister(ctx);
+ 	io_alloc_cache_free(&ctx->apoll_cache, io_apoll_cache_free);
+ 	io_alloc_cache_free(&ctx->netmsg_cache, io_netmsg_cache_free);
+-	mutex_unlock(&ctx->uring_lock);
+ 	io_destroy_buffers(ctx);
++	mutex_unlock(&ctx->uring_lock);
+ 	if (ctx->sq_creds)
+ 		put_cred(ctx->sq_creds);
+ 	if (ctx->submitter_task)
+diff --git a/io_uring/kbuf.c b/io_uring/kbuf.c
+index 746b137b96e9b..acc37e5a6d4e1 100644
+--- a/io_uring/kbuf.c
++++ b/io_uring/kbuf.c
+@@ -228,17 +228,18 @@ static int __io_remove_buffers(struct io_ring_ctx *ctx,
+ 		return i;
+ 	}
+ 
+-	/* the head kbuf is the list itself */
++	/* protects io_buffers_cache */
++	lockdep_assert_held(&ctx->uring_lock);
++
+ 	while (!list_empty(&bl->buf_list)) {
+ 		struct io_buffer *nxt;
+ 
+ 		nxt = list_first_entry(&bl->buf_list, struct io_buffer, list);
+-		list_del(&nxt->list);
++		list_move(&nxt->list, &ctx->io_buffers_cache);
+ 		if (++i == nbufs)
+ 			return i;
+ 		cond_resched();
+ 	}
+-	i++;
+ 
+ 	return i;
+ }
+diff --git a/kernel/events/core.c b/kernel/events/core.c
+index 2aa286b4151b3..7699b99706ad4 100644
+--- a/kernel/events/core.c
++++ b/kernel/events/core.c
+@@ -12040,7 +12040,7 @@ perf_event_set_output(struct perf_event *event, struct perf_event *output_event)
+ 	/*
+ 	 * If its not a per-cpu rb, it must be the same task.
+ 	 */
+-	if (output_event->cpu == -1 && output_event->ctx != event->ctx)
++	if (output_event->cpu == -1 && output_event->hw.target != event->hw.target)
+ 		goto out;
+ 
+ 	/*
+diff --git a/kernel/fork.c b/kernel/fork.c
+index a6d243a50be3e..ec913b13c5edb 100644
+--- a/kernel/fork.c
++++ b/kernel/fork.c
+@@ -617,6 +617,7 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm,
+ 	if (retval)
+ 		goto out;
+ 
++	mt_clear_in_rcu(mas.tree);
+ 	mas_for_each(&old_mas, mpnt, ULONG_MAX) {
+ 		struct file *file;
+ 
+@@ -703,6 +704,8 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm,
+ 	retval = arch_dup_mmap(oldmm, mm);
+ loop_out:
+ 	mas_destroy(&mas);
++	if (!retval)
++		mt_set_in_rcu(mas.tree);
+ out:
+ 	mmap_write_unlock(mm);
+ 	flush_tlb_mm(oldmm);
+diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c
+index 59062fedeaf7c..57db50c2dce80 100644
+--- a/kernel/trace/ftrace.c
++++ b/kernel/trace/ftrace.c
+@@ -5557,12 +5557,15 @@ int modify_ftrace_direct(unsigned long ip,
+ 		ret = 0;
+ 	}
+ 
+-	if (unlikely(ret && new_direct)) {
+-		direct->count++;
+-		list_del_rcu(&new_direct->next);
+-		synchronize_rcu_tasks();
+-		kfree(new_direct);
+-		ftrace_direct_func_count--;
++	if (ret) {
++		direct->addr = old_addr;
++		if (unlikely(new_direct)) {
++			direct->count++;
++			list_del_rcu(&new_direct->next);
++			synchronize_rcu_tasks();
++			kfree(new_direct);
++			ftrace_direct_func_count--;
++		}
+ 	}
+ 
+  out_unlock:
+diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c
+index 361bd8beafdff..9d8538531a545 100644
+--- a/kernel/trace/ring_buffer.c
++++ b/kernel/trace/ring_buffer.c
+@@ -3084,6 +3084,10 @@ rb_set_commit_to_write(struct ring_buffer_per_cpu *cpu_buffer)
+ 		if (RB_WARN_ON(cpu_buffer,
+ 			       rb_is_reader_page(cpu_buffer->tail_page)))
+ 			return;
++		/*
++		 * No need for a memory barrier here, as the update
++		 * of the tail_page did it for this page.
++		 */
+ 		local_set(&cpu_buffer->commit_page->page->commit,
+ 			  rb_page_write(cpu_buffer->commit_page));
+ 		rb_inc_page(&cpu_buffer->commit_page);
+@@ -3093,6 +3097,8 @@ rb_set_commit_to_write(struct ring_buffer_per_cpu *cpu_buffer)
+ 	while (rb_commit_index(cpu_buffer) !=
+ 	       rb_page_write(cpu_buffer->commit_page)) {
+ 
++		/* Make sure the readers see the content of what is committed. */
++		smp_wmb();
+ 		local_set(&cpu_buffer->commit_page->page->commit,
+ 			  rb_page_write(cpu_buffer->commit_page));
+ 		RB_WARN_ON(cpu_buffer,
+@@ -4672,7 +4678,12 @@ rb_get_reader_page(struct ring_buffer_per_cpu *cpu_buffer)
+ 
+ 	/*
+ 	 * Make sure we see any padding after the write update
+-	 * (see rb_reset_tail())
++	 * (see rb_reset_tail()).
++	 *
++	 * In addition, a writer may be writing on the reader page
++	 * if the page has not been fully filled, so the read barrier
++	 * is also needed to make sure we see the content of what is
++	 * committed by the writer (see rb_set_commit_to_write()).
+ 	 */
+ 	smp_rmb();
+ 
+diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c
+index 78d69b9488e45..78855e74e355f 100644
+--- a/kernel/trace/trace.c
++++ b/kernel/trace/trace.c
+@@ -9470,6 +9470,7 @@ static int __remove_instance(struct trace_array *tr)
+ 	tracefs_remove(tr->dir);
+ 	free_percpu(tr->last_func_repeats);
+ 	free_trace_buffers(tr);
++	clear_tracing_err_log(tr);
+ 
+ 	for (i = 0; i < tr->nr_topts; i++) {
+ 		kfree(tr->topts[i].topts);
+diff --git a/kernel/trace/trace_events_synth.c b/kernel/trace/trace_events_synth.c
+index 67592eed0be8d..89083ae1aebe3 100644
+--- a/kernel/trace/trace_events_synth.c
++++ b/kernel/trace/trace_events_synth.c
+@@ -44,14 +44,21 @@ enum { ERRORS };
+ 
+ static const char *err_text[] = { ERRORS };
+ 
++static DEFINE_MUTEX(lastcmd_mutex);
+ static char *last_cmd;
+ 
+ static int errpos(const char *str)
+ {
++	int ret = 0;
++
++	mutex_lock(&lastcmd_mutex);
+ 	if (!str || !last_cmd)
+-		return 0;
++		goto out;
+ 
+-	return err_pos(last_cmd, str);
++	ret = err_pos(last_cmd, str);
++ out:
++	mutex_unlock(&lastcmd_mutex);
++	return ret;
+ }
+ 
+ static void last_cmd_set(const char *str)
+@@ -59,18 +66,22 @@ static void last_cmd_set(const char *str)
+ 	if (!str)
+ 		return;
+ 
++	mutex_lock(&lastcmd_mutex);
+ 	kfree(last_cmd);
+-
+ 	last_cmd = kstrdup(str, GFP_KERNEL);
++	mutex_unlock(&lastcmd_mutex);
+ }
+ 
+ static void synth_err(u8 err_type, u16 err_pos)
+ {
++	mutex_lock(&lastcmd_mutex);
+ 	if (!last_cmd)
+-		return;
++		goto out;
+ 
+ 	tracing_log_err(NULL, "synthetic_events", last_cmd, err_text,
+ 			err_type, err_pos);
++ out:
++	mutex_unlock(&lastcmd_mutex);
+ }
+ 
+ static int create_synth_event(const char *raw_command);
+diff --git a/kernel/trace/trace_osnoise.c b/kernel/trace/trace_osnoise.c
+index 1c07efcb3d466..689361097bb0c 100644
+--- a/kernel/trace/trace_osnoise.c
++++ b/kernel/trace/trace_osnoise.c
+@@ -1270,7 +1270,7 @@ static void notify_new_max_latency(u64 latency)
+ 	rcu_read_lock();
+ 	list_for_each_entry_rcu(inst, &osnoise_instances, list) {
+ 		tr = inst->tr;
+-		if (tr->max_latency < latency) {
++		if (tracer_tracing_is_on(tr) && tr->max_latency < latency) {
+ 			tr->max_latency = latency;
+ 			latency_fsnotify(tr);
+ 		}
+@@ -1681,6 +1681,8 @@ static int timerlat_main(void *data)
+ 
+ 		trace_timerlat_sample(&s);
+ 
++		notify_new_max_latency(diff);
++
+ 		timerlat_dump_stack(time_to_us(diff));
+ 
+ 		tlat->tracing_thread = false;
+diff --git a/lib/maple_tree.c b/lib/maple_tree.c
+index 96be96a1f35dc..84530fb73bd9e 100644
+--- a/lib/maple_tree.c
++++ b/lib/maple_tree.c
+@@ -149,13 +149,12 @@ struct maple_subtree_state {
+ /* Functions */
+ static inline struct maple_node *mt_alloc_one(gfp_t gfp)
+ {
+-	return kmem_cache_alloc(maple_node_cache, gfp | __GFP_ZERO);
++	return kmem_cache_alloc(maple_node_cache, gfp);
+ }
+ 
+ static inline int mt_alloc_bulk(gfp_t gfp, size_t size, void **nodes)
+ {
+-	return kmem_cache_alloc_bulk(maple_node_cache, gfp | __GFP_ZERO, size,
+-				     nodes);
++	return kmem_cache_alloc_bulk(maple_node_cache, gfp, size, nodes);
+ }
+ 
+ static inline void mt_free_bulk(size_t size, void __rcu **nodes)
+@@ -530,11 +529,14 @@ static inline struct maple_node *mte_parent(const struct maple_enode *enode)
+  */
+ static inline bool ma_dead_node(const struct maple_node *node)
+ {
+-	struct maple_node *parent = (void *)((unsigned long)
+-					     node->parent & ~MAPLE_NODE_MASK);
++	struct maple_node *parent;
+ 
++	/* Do not reorder reads from the node prior to the parent check */
++	smp_rmb();
++	parent = (void *)((unsigned long) node->parent & ~MAPLE_NODE_MASK);
+ 	return (parent == node);
+ }
++
+ /*
+  * mte_dead_node() - check if the @enode is dead.
+  * @enode: The encoded maple node
+@@ -546,6 +548,8 @@ static inline bool mte_dead_node(const struct maple_enode *enode)
+ 	struct maple_node *parent, *node;
+ 
+ 	node = mte_to_node(enode);
++	/* Do not reorder reads from the node prior to the parent check */
++	smp_rmb();
+ 	parent = mte_parent(enode);
+ 	return (parent == node);
+ }
+@@ -616,6 +620,8 @@ static inline unsigned int mas_alloc_req(const struct ma_state *mas)
+  * @node - the maple node
+  * @type - the node type
+  *
++ * In the event of a dead node, this array may be %NULL
++ *
+  * Return: A pointer to the maple node pivots
+  */
+ static inline unsigned long *ma_pivots(struct maple_node *node,
+@@ -808,6 +814,11 @@ static inline void *mt_slot(const struct maple_tree *mt,
+ 	return rcu_dereference_check(slots[offset], mt_locked(mt));
+ }
+ 
++static inline void *mt_slot_locked(struct maple_tree *mt, void __rcu **slots,
++				   unsigned char offset)
++{
++	return rcu_dereference_protected(slots[offset], mt_locked(mt));
++}
+ /*
+  * mas_slot_locked() - Get the slot value when holding the maple tree lock.
+  * @mas: The maple state
+@@ -819,7 +830,7 @@ static inline void *mt_slot(const struct maple_tree *mt,
+ static inline void *mas_slot_locked(struct ma_state *mas, void __rcu **slots,
+ 				       unsigned char offset)
+ {
+-	return rcu_dereference_protected(slots[offset], mt_locked(mas->tree));
++	return mt_slot_locked(mas->tree, slots, offset);
+ }
+ 
+ /*
+@@ -890,6 +901,45 @@ static inline void ma_set_meta(struct maple_node *mn, enum maple_type mt,
+ 	meta->end = end;
+ }
+ 
++/*
++ * mt_clear_meta() - clear the metadata information of a node, if it exists
++ * @mt: The maple tree
++ * @mn: The maple node
++ * @type: The maple node type
++ * @offset: The offset of the highest sub-gap in this node.
++ * @end: The end of the data in this node.
++ */
++static inline void mt_clear_meta(struct maple_tree *mt, struct maple_node *mn,
++				  enum maple_type type)
++{
++	struct maple_metadata *meta;
++	unsigned long *pivots;
++	void __rcu **slots;
++	void *next;
++
++	switch (type) {
++	case maple_range_64:
++		pivots = mn->mr64.pivot;
++		if (unlikely(pivots[MAPLE_RANGE64_SLOTS - 2])) {
++			slots = mn->mr64.slot;
++			next = mt_slot_locked(mt, slots,
++					      MAPLE_RANGE64_SLOTS - 1);
++			if (unlikely((mte_to_node(next) &&
++				      mte_node_type(next))))
++				return; /* no metadata, could be node */
++		}
++		fallthrough;
++	case maple_arange_64:
++		meta = ma_meta(mn, type);
++		break;
++	default:
++		return;
++	}
++
++	meta->gap = 0;
++	meta->end = 0;
++}
++
+ /*
+  * ma_meta_end() - Get the data end of a node from the metadata
+  * @mn: The maple node
+@@ -1087,8 +1137,11 @@ static int mas_ascend(struct ma_state *mas)
+ 		a_type = mas_parent_enum(mas, p_enode);
+ 		a_node = mte_parent(p_enode);
+ 		a_slot = mte_parent_slot(p_enode);
+-		pivots = ma_pivots(a_node, a_type);
+ 		a_enode = mt_mk_node(a_node, a_type);
++		pivots = ma_pivots(a_node, a_type);
++
++		if (unlikely(ma_dead_node(a_node)))
++			return 1;
+ 
+ 		if (!set_min && a_slot) {
+ 			set_min = true;
+@@ -1123,9 +1176,10 @@ static inline struct maple_node *mas_pop_node(struct ma_state *mas)
+ {
+ 	struct maple_alloc *ret, *node = mas->alloc;
+ 	unsigned long total = mas_allocated(mas);
++	unsigned int req = mas_alloc_req(mas);
+ 
+ 	/* nothing or a request pending. */
+-	if (unlikely(!total))
++	if (WARN_ON(!total))
+ 		return NULL;
+ 
+ 	if (total == 1) {
+@@ -1135,27 +1189,25 @@ static inline struct maple_node *mas_pop_node(struct ma_state *mas)
+ 		goto single_node;
+ 	}
+ 
+-	if (!node->node_count) {
++	if (node->node_count == 1) {
+ 		/* Single allocation in this node. */
+ 		mas->alloc = node->slot[0];
+-		node->slot[0] = NULL;
+ 		mas->alloc->total = node->total - 1;
+ 		ret = node;
+ 		goto new_head;
+ 	}
+-
+ 	node->total--;
+-	ret = node->slot[node->node_count];
+-	node->slot[node->node_count--] = NULL;
++	ret = node->slot[--node->node_count];
++	node->slot[node->node_count] = NULL;
+ 
+ single_node:
+ new_head:
+-	ret->total = 0;
+-	ret->node_count = 0;
+-	if (ret->request_count) {
+-		mas_set_alloc_req(mas, ret->request_count + 1);
+-		ret->request_count = 0;
++	if (req) {
++		req++;
++		mas_set_alloc_req(mas, req);
+ 	}
++
++	memset(ret, 0, sizeof(*ret));
+ 	return (struct maple_node *)ret;
+ }
+ 
+@@ -1174,21 +1226,20 @@ static inline void mas_push_node(struct ma_state *mas, struct maple_node *used)
+ 	unsigned long count;
+ 	unsigned int requested = mas_alloc_req(mas);
+ 
+-	memset(reuse, 0, sizeof(*reuse));
+ 	count = mas_allocated(mas);
+ 
+-	if (count && (head->node_count < MAPLE_ALLOC_SLOTS - 1)) {
+-		if (head->slot[0])
+-			head->node_count++;
+-		head->slot[head->node_count] = reuse;
++	reuse->request_count = 0;
++	reuse->node_count = 0;
++	if (count && (head->node_count < MAPLE_ALLOC_SLOTS)) {
++		head->slot[head->node_count++] = reuse;
+ 		head->total++;
+ 		goto done;
+ 	}
+ 
+ 	reuse->total = 1;
+ 	if ((head) && !((unsigned long)head & 0x1)) {
+-		head->request_count = 0;
+ 		reuse->slot[0] = head;
++		reuse->node_count = 1;
+ 		reuse->total += head->total;
+ 	}
+ 
+@@ -1207,7 +1258,6 @@ static inline void mas_alloc_nodes(struct ma_state *mas, gfp_t gfp)
+ {
+ 	struct maple_alloc *node;
+ 	unsigned long allocated = mas_allocated(mas);
+-	unsigned long success = allocated;
+ 	unsigned int requested = mas_alloc_req(mas);
+ 	unsigned int count;
+ 	void **slots = NULL;
+@@ -1223,24 +1273,29 @@ static inline void mas_alloc_nodes(struct ma_state *mas, gfp_t gfp)
+ 		WARN_ON(!allocated);
+ 	}
+ 
+-	if (!allocated || mas->alloc->node_count == MAPLE_ALLOC_SLOTS - 1) {
++	if (!allocated || mas->alloc->node_count == MAPLE_ALLOC_SLOTS) {
+ 		node = (struct maple_alloc *)mt_alloc_one(gfp);
+ 		if (!node)
+ 			goto nomem_one;
+ 
+-		if (allocated)
++		if (allocated) {
+ 			node->slot[0] = mas->alloc;
++			node->node_count = 1;
++		} else {
++			node->node_count = 0;
++		}
+ 
+-		success++;
+ 		mas->alloc = node;
++		node->total = ++allocated;
+ 		requested--;
+ 	}
+ 
+ 	node = mas->alloc;
++	node->request_count = 0;
+ 	while (requested) {
+ 		max_req = MAPLE_ALLOC_SLOTS;
+-		if (node->slot[0]) {
+-			unsigned int offset = node->node_count + 1;
++		if (node->node_count) {
++			unsigned int offset = node->node_count;
+ 
+ 			slots = (void **)&node->slot[offset];
+ 			max_req -= offset;
+@@ -1254,15 +1309,13 @@ static inline void mas_alloc_nodes(struct ma_state *mas, gfp_t gfp)
+ 			goto nomem_bulk;
+ 
+ 		node->node_count += count;
+-		/* zero indexed. */
+-		if (slots == (void **)&node->slot)
+-			node->node_count--;
+-
+-		success += count;
++		allocated += count;
+ 		node = node->slot[0];
++		node->node_count = 0;
++		node->request_count = 0;
+ 		requested -= count;
+ 	}
+-	mas->alloc->total = success;
++	mas->alloc->total = allocated;
+ 	return;
+ 
+ nomem_bulk:
+@@ -1271,7 +1324,7 @@ nomem_bulk:
+ nomem_one:
+ 	mas_set_alloc_req(mas, requested);
+ 	if (mas->alloc && !(((unsigned long)mas->alloc & 0x1)))
+-		mas->alloc->total = success;
++		mas->alloc->total = allocated;
+ 	mas_set_err(mas, -ENOMEM);
+ 	return;
+ 
+@@ -1329,7 +1382,7 @@ static void mas_node_count(struct ma_state *mas, int count)
+  * mas_start() - Sets up maple state for operations.
+  * @mas: The maple state.
+  *
+- * If mas->node == MAS_START, then set the min, max, depth, and offset to
++ * If mas->node == MAS_START, then set the min, max and depth to
+  * defaults.
+  *
+  * Return:
+@@ -1343,22 +1396,26 @@ static inline struct maple_enode *mas_start(struct ma_state *mas)
+ 	if (likely(mas_is_start(mas))) {
+ 		struct maple_enode *root;
+ 
+-		mas->node = MAS_NONE;
+ 		mas->min = 0;
+ 		mas->max = ULONG_MAX;
+ 		mas->depth = 0;
+-		mas->offset = 0;
+ 
++retry:
+ 		root = mas_root(mas);
+ 		/* Tree with nodes */
+ 		if (likely(xa_is_node(root))) {
+ 			mas->depth = 1;
+ 			mas->node = mte_safe_root(root);
++			mas->offset = 0;
++			if (mte_dead_node(mas->node))
++				goto retry;
++
+ 			return NULL;
+ 		}
+ 
+ 		/* empty tree */
+ 		if (unlikely(!root)) {
++			mas->node = MAS_NONE;
+ 			mas->offset = MAPLE_NODE_SLOTS;
+ 			return NULL;
+ 		}
+@@ -1394,6 +1451,9 @@ static inline unsigned char ma_data_end(struct maple_node *node,
+ {
+ 	unsigned char offset;
+ 
++	if (!pivots)
++		return 0;
++
+ 	if (type == maple_arange_64)
+ 		return ma_meta_end(node, type);
+ 
+@@ -1429,6 +1489,9 @@ static inline unsigned char mas_data_end(struct ma_state *mas)
+ 		return ma_meta_end(node, type);
+ 
+ 	pivots = ma_pivots(node, type);
++	if (unlikely(ma_dead_node(node)))
++		return 0;
++
+ 	offset = mt_pivots[type] - 1;
+ 	if (likely(!pivots[offset]))
+ 		return ma_meta_end(node, type);
+@@ -3654,10 +3717,9 @@ static inline int mas_root_expand(struct ma_state *mas, void *entry)
+ 		slot++;
+ 	mas->depth = 1;
+ 	mas_set_height(mas);
+-
++	ma_set_meta(node, maple_leaf_64, 0, slot);
+ 	/* swap the new root into the tree */
+ 	rcu_assign_pointer(mas->tree->ma_root, mte_mk_root(mas->node));
+-	ma_set_meta(node, maple_leaf_64, 0, slot);
+ 	return slot;
+ }
+ 
+@@ -3870,18 +3932,13 @@ static inline void *mtree_lookup_walk(struct ma_state *mas)
+ 		end = ma_data_end(node, type, pivots, max);
+ 		if (unlikely(ma_dead_node(node)))
+ 			goto dead_node;
+-
+-		if (pivots[offset] >= mas->index)
+-			goto next;
+-
+ 		do {
+-			offset++;
+-		} while ((offset < end) && (pivots[offset] < mas->index));
+-
+-		if (likely(offset > end))
+-			max = pivots[offset];
++			if (pivots[offset] >= mas->index) {
++				max = pivots[offset];
++				break;
++			}
++		} while (++offset < end);
+ 
+-next:
+ 		slots = ma_slots(node, type);
+ 		next = mt_slot(mas->tree, slots, offset);
+ 		if (unlikely(ma_dead_node(node)))
+@@ -4500,6 +4557,9 @@ static inline int mas_prev_node(struct ma_state *mas, unsigned long min)
+ 	node = mas_mn(mas);
+ 	slots = ma_slots(node, mt);
+ 	pivots = ma_pivots(node, mt);
++	if (unlikely(ma_dead_node(node)))
++		return 1;
++
+ 	mas->max = pivots[offset];
+ 	if (offset)
+ 		mas->min = pivots[offset - 1] + 1;
+@@ -4521,6 +4581,9 @@ static inline int mas_prev_node(struct ma_state *mas, unsigned long min)
+ 		slots = ma_slots(node, mt);
+ 		pivots = ma_pivots(node, mt);
+ 		offset = ma_data_end(node, mt, pivots, mas->max);
++		if (unlikely(ma_dead_node(node)))
++			return 1;
++
+ 		if (offset)
+ 			mas->min = pivots[offset - 1] + 1;
+ 
+@@ -4569,6 +4632,7 @@ static inline int mas_next_node(struct ma_state *mas, struct maple_node *node,
+ 	struct maple_enode *enode;
+ 	int level = 0;
+ 	unsigned char offset;
++	unsigned char node_end;
+ 	enum maple_type mt;
+ 	void __rcu **slots;
+ 
+@@ -4592,7 +4656,11 @@ static inline int mas_next_node(struct ma_state *mas, struct maple_node *node,
+ 		node = mas_mn(mas);
+ 		mt = mte_node_type(mas->node);
+ 		pivots = ma_pivots(node, mt);
+-	} while (unlikely(offset == ma_data_end(node, mt, pivots, mas->max)));
++		node_end = ma_data_end(node, mt, pivots, mas->max);
++		if (unlikely(ma_dead_node(node)))
++			return 1;
++
++	} while (unlikely(offset == node_end));
+ 
+ 	slots = ma_slots(node, mt);
+ 	pivot = mas_safe_pivot(mas, pivots, ++offset, mt);
+@@ -4608,6 +4676,9 @@ static inline int mas_next_node(struct ma_state *mas, struct maple_node *node,
+ 		mt = mte_node_type(mas->node);
+ 		slots = ma_slots(node, mt);
+ 		pivots = ma_pivots(node, mt);
++		if (unlikely(ma_dead_node(node)))
++			return 1;
++
+ 		offset = 0;
+ 		pivot = pivots[0];
+ 	}
+@@ -4654,16 +4725,19 @@ static inline void *mas_next_nentry(struct ma_state *mas,
+ 		return NULL;
+ 	}
+ 
+-	pivots = ma_pivots(node, type);
+ 	slots = ma_slots(node, type);
++	pivots = ma_pivots(node, type);
++	count = ma_data_end(node, type, pivots, mas->max);
++	if (unlikely(ma_dead_node(node)))
++		return NULL;
++
+ 	mas->index = mas_safe_min(mas, pivots, mas->offset);
+-	if (ma_dead_node(node))
++	if (unlikely(ma_dead_node(node)))
+ 		return NULL;
+ 
+ 	if (mas->index > max)
+ 		return NULL;
+ 
+-	count = ma_data_end(node, type, pivots, mas->max);
+ 	if (mas->offset > count)
+ 		return NULL;
+ 
+@@ -4738,6 +4812,11 @@ static inline void *mas_next_entry(struct ma_state *mas, unsigned long limit)
+ 	unsigned long last;
+ 	enum maple_type mt;
+ 
++	if (mas->index > limit) {
++		mas->index = mas->last = limit;
++		mas_pause(mas);
++		return NULL;
++	}
+ 	last = mas->last;
+ retry:
+ 	offset = mas->offset;
+@@ -4811,6 +4890,11 @@ retry:
+ 
+ 	slots = ma_slots(mn, mt);
+ 	pivots = ma_pivots(mn, mt);
++	if (unlikely(ma_dead_node(mn))) {
++		mas_rewalk(mas, index);
++		goto retry;
++	}
++
+ 	if (offset == mt_pivots[mt])
+ 		pivot = mas->max;
+ 	else
+@@ -4844,6 +4928,11 @@ static inline void *mas_prev_entry(struct ma_state *mas, unsigned long min)
+ {
+ 	void *entry;
+ 
++	if (mas->index < min) {
++		mas->index = mas->last = min;
++		mas->node = MAS_NONE;
++		return NULL;
++	}
+ retry:
+ 	while (likely(!mas_is_none(mas))) {
+ 		entry = mas_prev_nentry(mas, min, mas->index);
+@@ -5389,24 +5478,26 @@ no_gap:
+ }
+ 
+ /*
+- * mas_dead_leaves() - Mark all leaves of a node as dead.
++ * mte_dead_leaves() - Mark all leaves of a node as dead.
+  * @mas: The maple state
+  * @slots: Pointer to the slot array
++ * @type: The maple node type
+  *
+  * Must hold the write lock.
+  *
+  * Return: The number of leaves marked as dead.
+  */
+ static inline
+-unsigned char mas_dead_leaves(struct ma_state *mas, void __rcu **slots)
++unsigned char mte_dead_leaves(struct maple_enode *enode, struct maple_tree *mt,
++			      void __rcu **slots)
+ {
+ 	struct maple_node *node;
+ 	enum maple_type type;
+ 	void *entry;
+ 	int offset;
+ 
+-	for (offset = 0; offset < mt_slot_count(mas->node); offset++) {
+-		entry = mas_slot_locked(mas, slots, offset);
++	for (offset = 0; offset < mt_slot_count(enode); offset++) {
++		entry = mt_slot(mt, slots, offset);
+ 		type = mte_node_type(entry);
+ 		node = mte_to_node(entry);
+ 		/* Use both node and type to catch LE & BE metadata */
+@@ -5414,7 +5505,6 @@ unsigned char mas_dead_leaves(struct ma_state *mas, void __rcu **slots)
+ 			break;
+ 
+ 		mte_set_node_dead(entry);
+-		smp_wmb(); /* Needed for RCU */
+ 		node->type = type;
+ 		rcu_assign_pointer(slots[offset], node);
+ 	}
+@@ -5422,151 +5512,160 @@ unsigned char mas_dead_leaves(struct ma_state *mas, void __rcu **slots)
+ 	return offset;
+ }
+ 
+-static void __rcu **mas_dead_walk(struct ma_state *mas, unsigned char offset)
++/**
++ * mte_dead_walk() - Walk down a dead tree to just before the leaves
++ * @enode: The maple encoded node
++ * @offset: The starting offset
++ *
++ * Note: This can only be used from the RCU callback context.
++ */
++static void __rcu **mte_dead_walk(struct maple_enode **enode, unsigned char offset)
+ {
+ 	struct maple_node *node, *next;
+ 	void __rcu **slots = NULL;
+ 
+-	next = mas_mn(mas);
++	next = mte_to_node(*enode);
+ 	do {
+-		mas->node = ma_enode_ptr(next);
+-		node = mas_mn(mas);
++		*enode = ma_enode_ptr(next);
++		node = mte_to_node(*enode);
+ 		slots = ma_slots(node, node->type);
+-		next = mas_slot_locked(mas, slots, offset);
++		next = rcu_dereference_protected(slots[offset],
++					lock_is_held(&rcu_callback_map));
+ 		offset = 0;
+ 	} while (!ma_is_leaf(next->type));
+ 
+ 	return slots;
+ }
+ 
++/**
++ * mt_free_walk() - Walk & free a tree in the RCU callback context
++ * @head: The RCU head that's within the node.
++ *
++ * Note: This can only be used from the RCU callback context.
++ */
+ static void mt_free_walk(struct rcu_head *head)
+ {
+ 	void __rcu **slots;
+ 	struct maple_node *node, *start;
+-	struct maple_tree mt;
++	struct maple_enode *enode;
+ 	unsigned char offset;
+ 	enum maple_type type;
+-	MA_STATE(mas, &mt, 0, 0);
+ 
+ 	node = container_of(head, struct maple_node, rcu);
+ 
+ 	if (ma_is_leaf(node->type))
+ 		goto free_leaf;
+ 
+-	mt_init_flags(&mt, node->ma_flags);
+-	mas_lock(&mas);
+ 	start = node;
+-	mas.node = mt_mk_node(node, node->type);
+-	slots = mas_dead_walk(&mas, 0);
+-	node = mas_mn(&mas);
++	enode = mt_mk_node(node, node->type);
++	slots = mte_dead_walk(&enode, 0);
++	node = mte_to_node(enode);
+ 	do {
+ 		mt_free_bulk(node->slot_len, slots);
+ 		offset = node->parent_slot + 1;
+-		mas.node = node->piv_parent;
+-		if (mas_mn(&mas) == node)
+-			goto start_slots_free;
+-
+-		type = mte_node_type(mas.node);
+-		slots = ma_slots(mte_to_node(mas.node), type);
+-		if ((offset < mt_slots[type]) && (slots[offset]))
+-			slots = mas_dead_walk(&mas, offset);
+-
+-		node = mas_mn(&mas);
++		enode = node->piv_parent;
++		if (mte_to_node(enode) == node)
++			goto free_leaf;
++
++		type = mte_node_type(enode);
++		slots = ma_slots(mte_to_node(enode), type);
++		if ((offset < mt_slots[type]) &&
++		    rcu_dereference_protected(slots[offset],
++					      lock_is_held(&rcu_callback_map)))
++			slots = mte_dead_walk(&enode, offset);
++		node = mte_to_node(enode);
+ 	} while ((node != start) || (node->slot_len < offset));
+ 
+ 	slots = ma_slots(node, node->type);
+ 	mt_free_bulk(node->slot_len, slots);
+ 
+-start_slots_free:
+-	mas_unlock(&mas);
+ free_leaf:
+ 	mt_free_rcu(&node->rcu);
+ }
+ 
+-static inline void __rcu **mas_destroy_descend(struct ma_state *mas,
+-			struct maple_enode *prev, unsigned char offset)
++static inline void __rcu **mte_destroy_descend(struct maple_enode **enode,
++	struct maple_tree *mt, struct maple_enode *prev, unsigned char offset)
+ {
+ 	struct maple_node *node;
+-	struct maple_enode *next = mas->node;
++	struct maple_enode *next = *enode;
+ 	void __rcu **slots = NULL;
++	enum maple_type type;
++	unsigned char next_offset = 0;
+ 
+ 	do {
+-		mas->node = next;
+-		node = mas_mn(mas);
+-		slots = ma_slots(node, mte_node_type(mas->node));
+-		next = mas_slot_locked(mas, slots, 0);
++		*enode = next;
++		node = mte_to_node(*enode);
++		type = mte_node_type(*enode);
++		slots = ma_slots(node, type);
++		next = mt_slot_locked(mt, slots, next_offset);
+ 		if ((mte_dead_node(next)))
+-			next = mas_slot_locked(mas, slots, 1);
++			next = mt_slot_locked(mt, slots, ++next_offset);
+ 
+-		mte_set_node_dead(mas->node);
+-		node->type = mte_node_type(mas->node);
++		mte_set_node_dead(*enode);
++		node->type = type;
+ 		node->piv_parent = prev;
+ 		node->parent_slot = offset;
+-		offset = 0;
+-		prev = mas->node;
++		offset = next_offset;
++		next_offset = 0;
++		prev = *enode;
+ 	} while (!mte_is_leaf(next));
+ 
+ 	return slots;
+ }
+ 
+-static void mt_destroy_walk(struct maple_enode *enode, unsigned char ma_flags,
++static void mt_destroy_walk(struct maple_enode *enode, struct maple_tree *mt,
+ 			    bool free)
+ {
+ 	void __rcu **slots;
+ 	struct maple_node *node = mte_to_node(enode);
+ 	struct maple_enode *start;
+-	struct maple_tree mt;
+-
+-	MA_STATE(mas, &mt, 0, 0);
+ 
+-	if (mte_is_leaf(enode))
++	if (mte_is_leaf(enode)) {
++		node->type = mte_node_type(enode);
+ 		goto free_leaf;
++	}
+ 
+-	mt_init_flags(&mt, ma_flags);
+-	mas_lock(&mas);
+-
+-	mas.node = start = enode;
+-	slots = mas_destroy_descend(&mas, start, 0);
+-	node = mas_mn(&mas);
++	start = enode;
++	slots = mte_destroy_descend(&enode, mt, start, 0);
++	node = mte_to_node(enode); // Updated in the above call.
+ 	do {
+ 		enum maple_type type;
+ 		unsigned char offset;
+ 		struct maple_enode *parent, *tmp;
+ 
+-		node->slot_len = mas_dead_leaves(&mas, slots);
++		node->slot_len = mte_dead_leaves(enode, mt, slots);
+ 		if (free)
+ 			mt_free_bulk(node->slot_len, slots);
+ 		offset = node->parent_slot + 1;
+-		mas.node = node->piv_parent;
+-		if (mas_mn(&mas) == node)
+-			goto start_slots_free;
++		enode = node->piv_parent;
++		if (mte_to_node(enode) == node)
++			goto free_leaf;
+ 
+-		type = mte_node_type(mas.node);
+-		slots = ma_slots(mte_to_node(mas.node), type);
++		type = mte_node_type(enode);
++		slots = ma_slots(mte_to_node(enode), type);
+ 		if (offset >= mt_slots[type])
+ 			goto next;
+ 
+-		tmp = mas_slot_locked(&mas, slots, offset);
++		tmp = mt_slot_locked(mt, slots, offset);
+ 		if (mte_node_type(tmp) && mte_to_node(tmp)) {
+-			parent = mas.node;
+-			mas.node = tmp;
+-			slots = mas_destroy_descend(&mas, parent, offset);
++			parent = enode;
++			enode = tmp;
++			slots = mte_destroy_descend(&enode, mt, parent, offset);
+ 		}
+ next:
+-		node = mas_mn(&mas);
+-	} while (start != mas.node);
++		node = mte_to_node(enode);
++	} while (start != enode);
+ 
+-	node = mas_mn(&mas);
+-	node->slot_len = mas_dead_leaves(&mas, slots);
++	node = mte_to_node(enode);
++	node->slot_len = mte_dead_leaves(enode, mt, slots);
+ 	if (free)
+ 		mt_free_bulk(node->slot_len, slots);
+ 
+-start_slots_free:
+-	mas_unlock(&mas);
+-
+ free_leaf:
+ 	if (free)
+ 		mt_free_rcu(&node->rcu);
++	else
++		mt_clear_meta(mt, node, node->type);
+ }
+ 
+ /*
+@@ -5582,15 +5681,18 @@ static inline void mte_destroy_walk(struct maple_enode *enode,
+ 	struct maple_node *node = mte_to_node(enode);
+ 
+ 	if (mt_in_rcu(mt)) {
+-		mt_destroy_walk(enode, mt->ma_flags, false);
++		mt_destroy_walk(enode, mt, false);
+ 		call_rcu(&node->rcu, mt_free_walk);
+ 	} else {
+-		mt_destroy_walk(enode, mt->ma_flags, true);
++		mt_destroy_walk(enode, mt, true);
+ 	}
+ }
+ 
+ static void mas_wr_store_setup(struct ma_wr_state *wr_mas)
+ {
++	if (unlikely(mas_is_paused(wr_mas->mas)))
++		mas_reset(wr_mas->mas);
++
+ 	if (!mas_is_start(wr_mas->mas)) {
+ 		if (mas_is_none(wr_mas->mas)) {
+ 			mas_reset(wr_mas->mas);
+@@ -5726,6 +5828,7 @@ int mas_preallocate(struct ma_state *mas, void *entry, gfp_t gfp)
+ void mas_destroy(struct ma_state *mas)
+ {
+ 	struct maple_alloc *node;
++	unsigned long total;
+ 
+ 	/*
+ 	 * When using mas_for_each() to insert an expected number of elements,
+@@ -5748,14 +5851,20 @@ void mas_destroy(struct ma_state *mas)
+ 	}
+ 	mas->mas_flags &= ~(MA_STATE_BULK|MA_STATE_PREALLOC);
+ 
+-	while (mas->alloc && !((unsigned long)mas->alloc & 0x1)) {
++	total = mas_allocated(mas);
++	while (total) {
+ 		node = mas->alloc;
+ 		mas->alloc = node->slot[0];
+-		if (node->node_count > 0)
+-			mt_free_bulk(node->node_count,
+-				     (void __rcu **)&node->slot[1]);
++		if (node->node_count > 1) {
++			size_t count = node->node_count - 1;
++
++			mt_free_bulk(count, (void __rcu **)&node->slot[1]);
++			total -= count;
++		}
+ 		kmem_cache_free(maple_node_cache, node);
++		total--;
+ 	}
++
+ 	mas->alloc = NULL;
+ }
+ EXPORT_SYMBOL_GPL(mas_destroy);
+@@ -5893,6 +6002,7 @@ void *mas_prev(struct ma_state *mas, unsigned long min)
+ 	if (!mas->index) {
+ 		/* Nothing comes before 0 */
+ 		mas->last = 0;
++		mas->node = MAS_NONE;
+ 		return NULL;
+ 	}
+ 
+@@ -5983,6 +6093,9 @@ void *mas_find(struct ma_state *mas, unsigned long max)
+ 		mas->index = ++mas->last;
+ 	}
+ 
++	if (unlikely(mas_is_none(mas)))
++		mas->node = MAS_START;
++
+ 	if (unlikely(mas_is_start(mas))) {
+ 		/* First run or continue */
+ 		void *entry;
+@@ -6594,11 +6707,11 @@ static inline void *mas_first_entry(struct ma_state *mas, struct maple_node *mn,
+ 	while (likely(!ma_is_leaf(mt))) {
+ 		MT_BUG_ON(mas->tree, mte_dead_node(mas->node));
+ 		slots = ma_slots(mn, mt);
+-		pivots = ma_pivots(mn, mt);
+-		max = pivots[0];
+ 		entry = mas_slot(mas, slots, 0);
++		pivots = ma_pivots(mn, mt);
+ 		if (unlikely(ma_dead_node(mn)))
+ 			return NULL;
++		max = pivots[0];
+ 		mas->node = entry;
+ 		mn = mas_mn(mas);
+ 		mt = mte_node_type(mas->node);
+@@ -6618,13 +6731,13 @@ static inline void *mas_first_entry(struct ma_state *mas, struct maple_node *mn,
+ 	if (likely(entry))
+ 		return entry;
+ 
+-	pivots = ma_pivots(mn, mt);
+-	mas->index = pivots[0] + 1;
+ 	mas->offset = 1;
+ 	entry = mas_slot(mas, slots, 1);
++	pivots = ma_pivots(mn, mt);
+ 	if (unlikely(ma_dead_node(mn)))
+ 		return NULL;
+ 
++	mas->index = pivots[0] + 1;
+ 	if (mas->index > limit)
+ 		goto none;
+ 
+diff --git a/mm/hugetlb.c b/mm/hugetlb.c
+index 52475c4262e45..d3ffa0fd49e57 100644
+--- a/mm/hugetlb.c
++++ b/mm/hugetlb.c
+@@ -5469,7 +5469,7 @@ static vm_fault_t hugetlb_wp(struct mm_struct *mm, struct vm_area_struct *vma,
+ 		       struct page *pagecache_page, spinlock_t *ptl)
+ {
+ 	const bool unshare = flags & FAULT_FLAG_UNSHARE;
+-	pte_t pte;
++	pte_t pte = huge_ptep_get(ptep);
+ 	struct hstate *h = hstate_vma(vma);
+ 	struct page *old_page, *new_page;
+ 	int outside_reserve = 0;
+@@ -5480,6 +5480,17 @@ static vm_fault_t hugetlb_wp(struct mm_struct *mm, struct vm_area_struct *vma,
+ 	VM_BUG_ON(unshare && (flags & FOLL_WRITE));
+ 	VM_BUG_ON(!unshare && !(flags & FOLL_WRITE));
+ 
++	/*
++	 * Never handle CoW for uffd-wp protected pages.  It should be only
++	 * handled when the uffd-wp protection is removed.
++	 *
++	 * Note that only the CoW optimization path (in hugetlb_no_page())
++	 * can trigger this, because hugetlb_fault() will always resolve
++	 * uffd-wp bit first.
++	 */
++	if (!unshare && huge_pte_uffd_wp(pte))
++		return 0;
++
+ 	/*
+ 	 * hugetlb does not support FOLL_FORCE-style write faults that keep the
+ 	 * PTE mapped R/O such as maybe_mkwrite() would do.
+@@ -5495,7 +5506,6 @@ static vm_fault_t hugetlb_wp(struct mm_struct *mm, struct vm_area_struct *vma,
+ 		return 0;
+ 	}
+ 
+-	pte = huge_ptep_get(ptep);
+ 	old_page = pte_page(pte);
+ 
+ 	delayacct_wpcopy_start();
+diff --git a/mm/kfence/core.c b/mm/kfence/core.c
+index c3d04753806a2..a477b7fb8aa33 100644
+--- a/mm/kfence/core.c
++++ b/mm/kfence/core.c
+@@ -557,15 +557,11 @@ static unsigned long kfence_init_pool(void)
+ 	 * enters __slab_free() slow-path.
+ 	 */
+ 	for (i = 0; i < KFENCE_POOL_SIZE / PAGE_SIZE; i++) {
+-		struct slab *slab = page_slab(&pages[i]);
++		struct slab *slab = page_slab(nth_page(pages, i));
+ 
+ 		if (!i || (i % 2))
+ 			continue;
+ 
+-		/* Verify we do not have a compound head page. */
+-		if (WARN_ON(compound_head(&pages[i]) != &pages[i]))
+-			return addr;
+-
+ 		__folio_set_slab(slab_folio(slab));
+ #ifdef CONFIG_MEMCG
+ 		slab->memcg_data = (unsigned long)&kfence_metadata[i / 2 - 1].objcg |
+@@ -598,12 +594,26 @@ static unsigned long kfence_init_pool(void)
+ 
+ 		/* Protect the right redzone. */
+ 		if (unlikely(!kfence_protect(addr + PAGE_SIZE)))
+-			return addr;
++			goto reset_slab;
+ 
+ 		addr += 2 * PAGE_SIZE;
+ 	}
+ 
+ 	return 0;
++
++reset_slab:
++	for (i = 0; i < KFENCE_POOL_SIZE / PAGE_SIZE; i++) {
++		struct slab *slab = page_slab(nth_page(pages, i));
++
++		if (!i || (i % 2))
++			continue;
++#ifdef CONFIG_MEMCG
++		slab->memcg_data = 0;
++#endif
++		__folio_clear_slab(slab_folio(slab));
++	}
++
++	return addr;
+ }
+ 
+ static bool __init kfence_init_pool_early(void)
+@@ -633,16 +643,6 @@ static bool __init kfence_init_pool_early(void)
+ 	 * fails for the first page, and therefore expect addr==__kfence_pool in
+ 	 * most failure cases.
+ 	 */
+-	for (char *p = (char *)addr; p < __kfence_pool + KFENCE_POOL_SIZE; p += PAGE_SIZE) {
+-		struct slab *slab = virt_to_slab(p);
+-
+-		if (!slab)
+-			continue;
+-#ifdef CONFIG_MEMCG
+-		slab->memcg_data = 0;
+-#endif
+-		__folio_clear_slab(slab_folio(slab));
+-	}
+ 	memblock_free_late(__pa(addr), KFENCE_POOL_SIZE - (addr - (unsigned long)__kfence_pool));
+ 	__kfence_pool = NULL;
+ 	return false;
+diff --git a/mm/memory.c b/mm/memory.c
+index f6f93e5b6b023..747b7ea30f890 100644
+--- a/mm/memory.c
++++ b/mm/memory.c
+@@ -3619,8 +3619,21 @@ static vm_fault_t remove_device_exclusive_entry(struct vm_fault *vmf)
+ 	struct vm_area_struct *vma = vmf->vma;
+ 	struct mmu_notifier_range range;
+ 
+-	if (!folio_lock_or_retry(folio, vma->vm_mm, vmf->flags))
++	/*
++	 * We need a reference to lock the folio because we don't hold
++	 * the PTL so a racing thread can remove the device-exclusive
++	 * entry and unmap it. If the folio is free the entry must
++	 * have been removed already. If it happens to have already
++	 * been re-allocated after being freed all we do is lock and
++	 * unlock it.
++	 */
++	if (!folio_try_get(folio))
++		return 0;
++
++	if (!folio_lock_or_retry(folio, vma->vm_mm, vmf->flags)) {
++		folio_put(folio);
+ 		return VM_FAULT_RETRY;
++	}
+ 	mmu_notifier_range_init_owner(&range, MMU_NOTIFY_EXCLUSIVE, 0, vma,
+ 				vma->vm_mm, vmf->address & PAGE_MASK,
+ 				(vmf->address & PAGE_MASK) + PAGE_SIZE, NULL);
+@@ -3633,6 +3646,7 @@ static vm_fault_t remove_device_exclusive_entry(struct vm_fault *vmf)
+ 
+ 	pte_unmap_unlock(vmf->pte, vmf->ptl);
+ 	folio_unlock(folio);
++	folio_put(folio);
+ 
+ 	mmu_notifier_invalidate_range_end(&range);
+ 	return 0;
+diff --git a/mm/mmap.c b/mm/mmap.c
+index 1777148868494..fe1db604dc49e 100644
+--- a/mm/mmap.c
++++ b/mm/mmap.c
+@@ -2308,7 +2308,7 @@ do_mas_align_munmap(struct ma_state *mas, struct vm_area_struct *vma,
+ 	int count = 0;
+ 	int error = -ENOMEM;
+ 	MA_STATE(mas_detach, &mt_detach, 0, 0);
+-	mt_init_flags(&mt_detach, MT_FLAGS_LOCK_EXTERN);
++	mt_init_flags(&mt_detach, mas->tree->ma_flags & MT_FLAGS_LOCK_MASK);
+ 	mt_set_external_lock(&mt_detach, &mm->mmap_lock);
+ 
+ 	if (mas_preallocate(mas, vma, GFP_KERNEL))
+@@ -3095,6 +3095,7 @@ void exit_mmap(struct mm_struct *mm)
+ 	 */
+ 	set_bit(MMF_OOM_SKIP, &mm->flags);
+ 	mmap_write_lock(mm);
++	mt_clear_in_rcu(&mm->mm_mt);
+ 	free_pgtables(&tlb, &mm->mm_mt, vma, FIRST_USER_ADDRESS,
+ 		      USER_PGTABLES_CEILING);
+ 	tlb_finish_mmu(&tlb);
+diff --git a/mm/swapfile.c b/mm/swapfile.c
+index b72908df52ac9..71db6d8a1ea30 100644
+--- a/mm/swapfile.c
++++ b/mm/swapfile.c
+@@ -679,6 +679,7 @@ static void __del_from_avail_list(struct swap_info_struct *p)
+ {
+ 	int nid;
+ 
++	assert_spin_locked(&p->lock);
+ 	for_each_node(nid)
+ 		plist_del(&p->avail_lists[nid], &swap_avail_heads[nid]);
+ }
+@@ -2428,8 +2429,8 @@ SYSCALL_DEFINE1(swapoff, const char __user *, specialfile)
+ 		spin_unlock(&swap_lock);
+ 		goto out_dput;
+ 	}
+-	del_from_avail_list(p);
+ 	spin_lock(&p->lock);
++	del_from_avail_list(p);
+ 	if (p->prio < 0) {
+ 		struct swap_info_struct *si = p;
+ 		int nid;
+diff --git a/mm/vmalloc.c b/mm/vmalloc.c
+index ccaa461998f3c..d606e53c650e5 100644
+--- a/mm/vmalloc.c
++++ b/mm/vmalloc.c
+@@ -3056,9 +3056,11 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
+ 	 * allocation request, free them via __vfree() if any.
+ 	 */
+ 	if (area->nr_pages != nr_small_pages) {
+-		warn_alloc(gfp_mask, NULL,
+-			"vmalloc error: size %lu, page order %u, failed to allocate pages",
+-			area->nr_pages * PAGE_SIZE, page_order);
++		/* vm_area_alloc_pages() can also fail due to a fatal signal */
++		if (!fatal_signal_pending(current))
++			warn_alloc(gfp_mask, NULL,
++				"vmalloc error: size %lu, page order %u, failed to allocate pages",
++				area->nr_pages * PAGE_SIZE, page_order);
+ 		goto fail;
+ 	}
+ 
+diff --git a/net/can/isotp.c b/net/can/isotp.c
+index 9bc344851704e..5761d4ab839dd 100644
+--- a/net/can/isotp.c
++++ b/net/can/isotp.c
+@@ -119,7 +119,8 @@ enum {
+ 	ISOTP_WAIT_FIRST_FC,
+ 	ISOTP_WAIT_FC,
+ 	ISOTP_WAIT_DATA,
+-	ISOTP_SENDING
++	ISOTP_SENDING,
++	ISOTP_SHUTDOWN,
+ };
+ 
+ struct tpcon {
+@@ -880,8 +881,8 @@ static enum hrtimer_restart isotp_tx_timer_handler(struct hrtimer *hrtimer)
+ 					     txtimer);
+ 	struct sock *sk = &so->sk;
+ 
+-	/* don't handle timeouts in IDLE state */
+-	if (so->tx.state == ISOTP_IDLE)
++	/* don't handle timeouts in IDLE or SHUTDOWN state */
++	if (so->tx.state == ISOTP_IDLE || so->tx.state == ISOTP_SHUTDOWN)
+ 		return HRTIMER_NORESTART;
+ 
+ 	/* we did not get any flow control or echo frame in time */
+@@ -918,7 +919,6 @@ static int isotp_sendmsg(struct socket *sock, struct msghdr *msg, size_t size)
+ {
+ 	struct sock *sk = sock->sk;
+ 	struct isotp_sock *so = isotp_sk(sk);
+-	u32 old_state = so->tx.state;
+ 	struct sk_buff *skb;
+ 	struct net_device *dev;
+ 	struct canfd_frame *cf;
+@@ -928,23 +928,24 @@ static int isotp_sendmsg(struct socket *sock, struct msghdr *msg, size_t size)
+ 	int off;
+ 	int err;
+ 
+-	if (!so->bound)
++	if (!so->bound || so->tx.state == ISOTP_SHUTDOWN)
+ 		return -EADDRNOTAVAIL;
+ 
++wait_free_buffer:
+ 	/* we do not support multiple buffers - for now */
+-	if (cmpxchg(&so->tx.state, ISOTP_IDLE, ISOTP_SENDING) != ISOTP_IDLE ||
+-	    wq_has_sleeper(&so->wait)) {
+-		if (msg->msg_flags & MSG_DONTWAIT) {
+-			err = -EAGAIN;
+-			goto err_out;
+-		}
++	if (wq_has_sleeper(&so->wait) && (msg->msg_flags & MSG_DONTWAIT))
++		return -EAGAIN;
+ 
+-		/* wait for complete transmission of current pdu */
+-		err = wait_event_interruptible(so->wait, so->tx.state == ISOTP_IDLE);
+-		if (err)
+-			goto err_out;
++	/* wait for complete transmission of current pdu */
++	err = wait_event_interruptible(so->wait, so->tx.state == ISOTP_IDLE);
++	if (err)
++		goto err_event_drop;
+ 
+-		so->tx.state = ISOTP_SENDING;
++	if (cmpxchg(&so->tx.state, ISOTP_IDLE, ISOTP_SENDING) != ISOTP_IDLE) {
++		if (so->tx.state == ISOTP_SHUTDOWN)
++			return -EADDRNOTAVAIL;
++
++		goto wait_free_buffer;
+ 	}
+ 
+ 	if (!size || size > MAX_MSG_LENGTH) {
+@@ -1074,7 +1075,9 @@ static int isotp_sendmsg(struct socket *sock, struct msghdr *msg, size_t size)
+ 
+ 	if (wait_tx_done) {
+ 		/* wait for complete transmission of current pdu */
+-		wait_event_interruptible(so->wait, so->tx.state == ISOTP_IDLE);
++		err = wait_event_interruptible(so->wait, so->tx.state == ISOTP_IDLE);
++		if (err)
++			goto err_event_drop;
+ 
+ 		if (sk->sk_err)
+ 			return -sk->sk_err;
+@@ -1082,13 +1085,15 @@ static int isotp_sendmsg(struct socket *sock, struct msghdr *msg, size_t size)
+ 
+ 	return size;
+ 
++err_event_drop:
++	/* got signal: force tx state machine to be idle */
++	so->tx.state = ISOTP_IDLE;
++	hrtimer_cancel(&so->txfrtimer);
++	hrtimer_cancel(&so->txtimer);
+ err_out_drop:
+ 	/* drop this PDU and unlock a potential wait queue */
+-	old_state = ISOTP_IDLE;
+-err_out:
+-	so->tx.state = old_state;
+-	if (so->tx.state == ISOTP_IDLE)
+-		wake_up_interruptible(&so->wait);
++	so->tx.state = ISOTP_IDLE;
++	wake_up_interruptible(&so->wait);
+ 
+ 	return err;
+ }
+@@ -1120,7 +1125,7 @@ static int isotp_recvmsg(struct socket *sock, struct msghdr *msg, size_t size,
+ 	if (ret < 0)
+ 		goto out_err;
+ 
+-	sock_recv_timestamp(msg, sk, skb);
++	sock_recv_cmsgs(msg, sk, skb);
+ 
+ 	if (msg->msg_name) {
+ 		__sockaddr_check_size(ISOTP_MIN_NAMELEN);
+@@ -1150,10 +1155,12 @@ static int isotp_release(struct socket *sock)
+ 	net = sock_net(sk);
+ 
+ 	/* wait for complete transmission of current pdu */
+-	wait_event_interruptible(so->wait, so->tx.state == ISOTP_IDLE);
++	while (wait_event_interruptible(so->wait, so->tx.state == ISOTP_IDLE) == 0 &&
++	       cmpxchg(&so->tx.state, ISOTP_IDLE, ISOTP_SHUTDOWN) != ISOTP_IDLE)
++		;
+ 
+ 	/* force state machines to be idle also when a signal occurred */
+-	so->tx.state = ISOTP_IDLE;
++	so->tx.state = ISOTP_SHUTDOWN;
+ 	so->rx.state = ISOTP_IDLE;
+ 
+ 	spin_lock(&isotp_notifier_lock);
+@@ -1608,6 +1615,21 @@ static int isotp_init(struct sock *sk)
+ 	return 0;
+ }
+ 
++static __poll_t isotp_poll(struct file *file, struct socket *sock, poll_table *wait)
++{
++	struct sock *sk = sock->sk;
++	struct isotp_sock *so = isotp_sk(sk);
++
++	__poll_t mask = datagram_poll(file, sock, wait);
++	poll_wait(file, &so->wait, wait);
++
++	/* Check for false positives due to TX state */
++	if ((mask & EPOLLWRNORM) && (so->tx.state != ISOTP_IDLE))
++		mask &= ~(EPOLLOUT | EPOLLWRNORM);
++
++	return mask;
++}
++
+ static int isotp_sock_no_ioctlcmd(struct socket *sock, unsigned int cmd,
+ 				  unsigned long arg)
+ {
+@@ -1623,7 +1645,7 @@ static const struct proto_ops isotp_ops = {
+ 	.socketpair = sock_no_socketpair,
+ 	.accept = sock_no_accept,
+ 	.getname = isotp_getname,
+-	.poll = datagram_poll,
++	.poll = isotp_poll,
+ 	.ioctl = isotp_sock_no_ioctlcmd,
+ 	.gettstamp = sock_gettstamp,
+ 	.listen = sock_no_listen,
+diff --git a/net/can/j1939/transport.c b/net/can/j1939/transport.c
+index 848b60c9ef79e..bd8ec24338324 100644
+--- a/net/can/j1939/transport.c
++++ b/net/can/j1939/transport.c
+@@ -604,7 +604,10 @@ sk_buff *j1939_tp_tx_dat_new(struct j1939_priv *priv,
+ 	/* reserve CAN header */
+ 	skb_reserve(skb, offsetof(struct can_frame, data));
+ 
+-	memcpy(skb->cb, re_skcb, sizeof(skb->cb));
++	/* skb->cb must be large enough to hold a j1939_sk_buff_cb structure */
++	BUILD_BUG_ON(sizeof(skb->cb) < sizeof(*re_skcb));
++
++	memcpy(skb->cb, re_skcb, sizeof(*re_skcb));
+ 	skcb = j1939_skb_to_cb(skb);
+ 	if (swap_src_dst)
+ 		j1939_skbcb_swap(skcb);
+diff --git a/net/core/netpoll.c b/net/core/netpoll.c
+index 9be762e1d0428..4ac8d0ad9f6fc 100644
+--- a/net/core/netpoll.c
++++ b/net/core/netpoll.c
+@@ -137,6 +137,20 @@ static void queue_process(struct work_struct *work)
+ 	}
+ }
+ 
++static int netif_local_xmit_active(struct net_device *dev)
++{
++	int i;
++
++	for (i = 0; i < dev->num_tx_queues; i++) {
++		struct netdev_queue *txq = netdev_get_tx_queue(dev, i);
++
++		if (READ_ONCE(txq->xmit_lock_owner) == smp_processor_id())
++			return 1;
++	}
++
++	return 0;
++}
++
+ static void poll_one_napi(struct napi_struct *napi)
+ {
+ 	int work;
+@@ -183,7 +197,10 @@ void netpoll_poll_dev(struct net_device *dev)
+ 	if (!ni || down_trylock(&ni->dev_lock))
+ 		return;
+ 
+-	if (!netif_running(dev)) {
++	/* Some drivers will take the same locks in poll and xmit,
++	 * we can't poll if local CPU is already in xmit.
++	 */
++	if (!netif_running(dev) || netif_local_xmit_active(dev)) {
+ 		up(&ni->dev_lock);
+ 		return;
+ 	}
+diff --git a/net/ethtool/linkmodes.c b/net/ethtool/linkmodes.c
+index 126e06c713a3a..2d91f2a8c7626 100644
+--- a/net/ethtool/linkmodes.c
++++ b/net/ethtool/linkmodes.c
+@@ -282,11 +282,12 @@ static int ethnl_update_linkmodes(struct genl_info *info, struct nlattr **tb,
+ 					    "lanes configuration not supported by device");
+ 			return -EOPNOTSUPP;
+ 		}
+-	} else if (!lsettings->autoneg) {
+-		/* If autoneg is off and lanes parameter is not passed from user,
+-		 * set the lanes parameter to 0.
++	} else if (!lsettings->autoneg && ksettings->lanes) {
++		/* If autoneg is off and lanes parameter is not passed from user but
++		 * it was defined previously then set the lanes parameter to 0.
+ 		 */
+ 		ksettings->lanes = 0;
++		*mod = true;
+ 	}
+ 
+ 	ret = ethnl_update_bitset(ksettings->link_modes.advertising,
+diff --git a/net/ipv4/icmp.c b/net/ipv4/icmp.c
+index d5d745c3e345b..2b09ef70752f9 100644
+--- a/net/ipv4/icmp.c
++++ b/net/ipv4/icmp.c
+@@ -746,6 +746,11 @@ void __icmp_send(struct sk_buff *skb_in, int type, int code, __be32 info,
+ 		room = 576;
+ 	room -= sizeof(struct iphdr) + icmp_param.replyopts.opt.opt.optlen;
+ 	room -= sizeof(struct icmphdr);
++	/* Guard against tiny mtu. We need to include at least one
++	 * IP network header for this message to make any sense.
++	 */
++	if (room <= (int)sizeof(struct iphdr))
++		goto ende;
+ 
+ 	icmp_param.data_len = skb_in->len - icmp_param.offset;
+ 	if (icmp_param.data_len > room)
+diff --git a/net/ipv4/ping.c b/net/ipv4/ping.c
+index 409ec2a1f95b0..5178a3f3cb537 100644
+--- a/net/ipv4/ping.c
++++ b/net/ipv4/ping.c
+@@ -1089,13 +1089,13 @@ static struct sock *ping_get_idx(struct seq_file *seq, loff_t pos)
+ }
+ 
+ void *ping_seq_start(struct seq_file *seq, loff_t *pos, sa_family_t family)
+-	__acquires(RCU)
++	__acquires(ping_table.lock)
+ {
+ 	struct ping_iter_state *state = seq->private;
+ 	state->bucket = 0;
+ 	state->family = family;
+ 
+-	rcu_read_lock();
++	spin_lock(&ping_table.lock);
+ 
+ 	return *pos ? ping_get_idx(seq, *pos-1) : SEQ_START_TOKEN;
+ }
+@@ -1121,9 +1121,9 @@ void *ping_seq_next(struct seq_file *seq, void *v, loff_t *pos)
+ EXPORT_SYMBOL_GPL(ping_seq_next);
+ 
+ void ping_seq_stop(struct seq_file *seq, void *v)
+-	__releases(RCU)
++	__releases(ping_table.lock)
+ {
+-	rcu_read_unlock();
++	spin_unlock(&ping_table.lock);
+ }
+ EXPORT_SYMBOL_GPL(ping_seq_stop);
+ 
+diff --git a/net/ipv4/raw.c b/net/ipv4/raw.c
+index 006c1f0ed8b47..af03aa8a8e513 100644
+--- a/net/ipv4/raw.c
++++ b/net/ipv4/raw.c
+@@ -91,12 +91,12 @@ EXPORT_SYMBOL_GPL(raw_v4_hashinfo);
+ int raw_hash_sk(struct sock *sk)
+ {
+ 	struct raw_hashinfo *h = sk->sk_prot->h.raw_hash;
+-	struct hlist_nulls_head *hlist;
++	struct hlist_head *hlist;
+ 
+-	hlist = &h->ht[inet_sk(sk)->inet_num & (RAW_HTABLE_SIZE - 1)];
++	hlist = &h->ht[raw_hashfunc(sock_net(sk), inet_sk(sk)->inet_num)];
+ 
+ 	spin_lock(&h->lock);
+-	__sk_nulls_add_node_rcu(sk, hlist);
++	sk_add_node_rcu(sk, hlist);
+ 	sock_set_flag(sk, SOCK_RCU_FREE);
+ 	spin_unlock(&h->lock);
+ 	sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1);
+@@ -110,7 +110,7 @@ void raw_unhash_sk(struct sock *sk)
+ 	struct raw_hashinfo *h = sk->sk_prot->h.raw_hash;
+ 
+ 	spin_lock(&h->lock);
+-	if (__sk_nulls_del_node_init_rcu(sk))
++	if (sk_del_node_init_rcu(sk))
+ 		sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1);
+ 	spin_unlock(&h->lock);
+ }
+@@ -160,19 +160,18 @@ static int icmp_filter(const struct sock *sk, const struct sk_buff *skb)
+  * RFC 1122: SHOULD pass TOS value up to the transport layer.
+  * -> It does. And not only TOS, but all IP header.
+  */
+-static int raw_v4_input(struct sk_buff *skb, const struct iphdr *iph, int hash)
++static int raw_v4_input(struct net *net, struct sk_buff *skb,
++			const struct iphdr *iph, int hash)
+ {
+-	struct net *net = dev_net(skb->dev);
+-	struct hlist_nulls_head *hlist;
+-	struct hlist_nulls_node *hnode;
+ 	int sdif = inet_sdif(skb);
++	struct hlist_head *hlist;
+ 	int dif = inet_iif(skb);
+ 	int delivered = 0;
+ 	struct sock *sk;
+ 
+ 	hlist = &raw_v4_hashinfo.ht[hash];
+ 	rcu_read_lock();
+-	sk_nulls_for_each(sk, hnode, hlist) {
++	sk_for_each_rcu(sk, hlist) {
+ 		if (!raw_v4_match(net, sk, iph->protocol,
+ 				  iph->saddr, iph->daddr, dif, sdif))
+ 			continue;
+@@ -193,9 +192,10 @@ static int raw_v4_input(struct sk_buff *skb, const struct iphdr *iph, int hash)
+ 
+ int raw_local_deliver(struct sk_buff *skb, int protocol)
+ {
+-	int hash = protocol & (RAW_HTABLE_SIZE - 1);
++	struct net *net = dev_net(skb->dev);
+ 
+-	return raw_v4_input(skb, ip_hdr(skb), hash);
++	return raw_v4_input(net, skb, ip_hdr(skb),
++			    raw_hashfunc(net, protocol));
+ }
+ 
+ static void raw_err(struct sock *sk, struct sk_buff *skb, u32 info)
+@@ -263,19 +263,18 @@ static void raw_err(struct sock *sk, struct sk_buff *skb, u32 info)
+ void raw_icmp_error(struct sk_buff *skb, int protocol, u32 info)
+ {
+ 	struct net *net = dev_net(skb->dev);
+-	struct hlist_nulls_head *hlist;
+-	struct hlist_nulls_node *hnode;
+ 	int dif = skb->dev->ifindex;
+ 	int sdif = inet_sdif(skb);
++	struct hlist_head *hlist;
+ 	const struct iphdr *iph;
+ 	struct sock *sk;
+ 	int hash;
+ 
+-	hash = protocol & (RAW_HTABLE_SIZE - 1);
++	hash = raw_hashfunc(net, protocol);
+ 	hlist = &raw_v4_hashinfo.ht[hash];
+ 
+ 	rcu_read_lock();
+-	sk_nulls_for_each(sk, hnode, hlist) {
++	sk_for_each_rcu(sk, hlist) {
+ 		iph = (const struct iphdr *)skb->data;
+ 		if (!raw_v4_match(net, sk, iph->protocol,
+ 				  iph->daddr, iph->saddr, dif, sdif))
+@@ -947,14 +946,13 @@ static struct sock *raw_get_first(struct seq_file *seq, int bucket)
+ {
+ 	struct raw_hashinfo *h = pde_data(file_inode(seq->file));
+ 	struct raw_iter_state *state = raw_seq_private(seq);
+-	struct hlist_nulls_head *hlist;
+-	struct hlist_nulls_node *hnode;
++	struct hlist_head *hlist;
+ 	struct sock *sk;
+ 
+ 	for (state->bucket = bucket; state->bucket < RAW_HTABLE_SIZE;
+ 			++state->bucket) {
+ 		hlist = &h->ht[state->bucket];
+-		sk_nulls_for_each(sk, hnode, hlist) {
++		sk_for_each(sk, hlist) {
+ 			if (sock_net(sk) == seq_file_net(seq))
+ 				return sk;
+ 		}
+@@ -967,7 +965,7 @@ static struct sock *raw_get_next(struct seq_file *seq, struct sock *sk)
+ 	struct raw_iter_state *state = raw_seq_private(seq);
+ 
+ 	do {
+-		sk = sk_nulls_next(sk);
++		sk = sk_next(sk);
+ 	} while (sk && sock_net(sk) != seq_file_net(seq));
+ 
+ 	if (!sk)
+@@ -986,9 +984,12 @@ static struct sock *raw_get_idx(struct seq_file *seq, loff_t pos)
+ }
+ 
+ void *raw_seq_start(struct seq_file *seq, loff_t *pos)
+-	__acquires(RCU)
++	__acquires(&h->lock)
+ {
+-	rcu_read_lock();
++	struct raw_hashinfo *h = pde_data(file_inode(seq->file));
++
++	spin_lock(&h->lock);
++
+ 	return *pos ? raw_get_idx(seq, *pos - 1) : SEQ_START_TOKEN;
+ }
+ EXPORT_SYMBOL_GPL(raw_seq_start);
+@@ -1007,9 +1008,11 @@ void *raw_seq_next(struct seq_file *seq, void *v, loff_t *pos)
+ EXPORT_SYMBOL_GPL(raw_seq_next);
+ 
+ void raw_seq_stop(struct seq_file *seq, void *v)
+-	__releases(RCU)
++	__releases(&h->lock)
+ {
+-	rcu_read_unlock();
++	struct raw_hashinfo *h = pde_data(file_inode(seq->file));
++
++	spin_unlock(&h->lock);
+ }
+ EXPORT_SYMBOL_GPL(raw_seq_stop);
+ 
+diff --git a/net/ipv4/raw_diag.c b/net/ipv4/raw_diag.c
+index 999321834b94a..da3591a66a169 100644
+--- a/net/ipv4/raw_diag.c
++++ b/net/ipv4/raw_diag.c
+@@ -57,8 +57,7 @@ static bool raw_lookup(struct net *net, struct sock *sk,
+ static struct sock *raw_sock_get(struct net *net, const struct inet_diag_req_v2 *r)
+ {
+ 	struct raw_hashinfo *hashinfo = raw_get_hashinfo(r);
+-	struct hlist_nulls_head *hlist;
+-	struct hlist_nulls_node *hnode;
++	struct hlist_head *hlist;
+ 	struct sock *sk;
+ 	int slot;
+ 
+@@ -68,7 +67,7 @@ static struct sock *raw_sock_get(struct net *net, const struct inet_diag_req_v2
+ 	rcu_read_lock();
+ 	for (slot = 0; slot < RAW_HTABLE_SIZE; slot++) {
+ 		hlist = &hashinfo->ht[slot];
+-		sk_nulls_for_each(sk, hnode, hlist) {
++		sk_for_each_rcu(sk, hlist) {
+ 			if (raw_lookup(net, sk, r)) {
+ 				/*
+ 				 * Grab it and keep until we fill
+@@ -142,9 +141,8 @@ static void raw_diag_dump(struct sk_buff *skb, struct netlink_callback *cb,
+ 	struct raw_hashinfo *hashinfo = raw_get_hashinfo(r);
+ 	struct net *net = sock_net(skb->sk);
+ 	struct inet_diag_dump_data *cb_data;
+-	struct hlist_nulls_head *hlist;
+-	struct hlist_nulls_node *hnode;
+ 	int num, s_num, slot, s_slot;
++	struct hlist_head *hlist;
+ 	struct sock *sk = NULL;
+ 	struct nlattr *bc;
+ 
+@@ -161,7 +159,7 @@ static void raw_diag_dump(struct sk_buff *skb, struct netlink_callback *cb,
+ 		num = 0;
+ 
+ 		hlist = &hashinfo->ht[slot];
+-		sk_nulls_for_each(sk, hnode, hlist) {
++		sk_for_each_rcu(sk, hlist) {
+ 			struct inet_sock *inet = inet_sk(sk);
+ 
+ 			if (!net_eq(sock_net(sk), net))
+diff --git a/net/ipv6/ip6_output.c b/net/ipv6/ip6_output.c
+index c314fdde0097c..95a55c6630add 100644
+--- a/net/ipv6/ip6_output.c
++++ b/net/ipv6/ip6_output.c
+@@ -1965,8 +1965,13 @@ struct sk_buff *__ip6_make_skb(struct sock *sk,
+ 	IP6_UPD_PO_STATS(net, rt->rt6i_idev, IPSTATS_MIB_OUT, skb->len);
+ 	if (proto == IPPROTO_ICMPV6) {
+ 		struct inet6_dev *idev = ip6_dst_idev(skb_dst(skb));
++		u8 icmp6_type;
+ 
+-		ICMP6MSGOUT_INC_STATS(net, idev, icmp6_hdr(skb)->icmp6_type);
++		if (sk->sk_socket->type == SOCK_RAW && !inet_sk(sk)->hdrincl)
++			icmp6_type = fl6->fl6_icmp_type;
++		else
++			icmp6_type = icmp6_hdr(skb)->icmp6_type;
++		ICMP6MSGOUT_INC_STATS(net, idev, icmp6_type);
+ 		ICMP6_INC_STATS(net, idev, ICMP6_MIB_OUTMSGS);
+ 	}
+ 
+diff --git a/net/ipv6/raw.c b/net/ipv6/raw.c
+index 8ffeac7456567..9ee1506e23ab1 100644
+--- a/net/ipv6/raw.c
++++ b/net/ipv6/raw.c
+@@ -141,10 +141,9 @@ EXPORT_SYMBOL(rawv6_mh_filter_unregister);
+ static bool ipv6_raw_deliver(struct sk_buff *skb, int nexthdr)
+ {
+ 	struct net *net = dev_net(skb->dev);
+-	struct hlist_nulls_head *hlist;
+-	struct hlist_nulls_node *hnode;
+ 	const struct in6_addr *saddr;
+ 	const struct in6_addr *daddr;
++	struct hlist_head *hlist;
+ 	struct sock *sk;
+ 	bool delivered = false;
+ 	__u8 hash;
+@@ -152,10 +151,10 @@ static bool ipv6_raw_deliver(struct sk_buff *skb, int nexthdr)
+ 	saddr = &ipv6_hdr(skb)->saddr;
+ 	daddr = saddr + 1;
+ 
+-	hash = nexthdr & (RAW_HTABLE_SIZE - 1);
++	hash = raw_hashfunc(net, nexthdr);
+ 	hlist = &raw_v6_hashinfo.ht[hash];
+ 	rcu_read_lock();
+-	sk_nulls_for_each(sk, hnode, hlist) {
++	sk_for_each_rcu(sk, hlist) {
+ 		int filtered;
+ 
+ 		if (!raw_v6_match(net, sk, nexthdr, daddr, saddr,
+@@ -333,15 +332,14 @@ void raw6_icmp_error(struct sk_buff *skb, int nexthdr,
+ 		u8 type, u8 code, int inner_offset, __be32 info)
+ {
+ 	struct net *net = dev_net(skb->dev);
+-	struct hlist_nulls_head *hlist;
+-	struct hlist_nulls_node *hnode;
++	struct hlist_head *hlist;
+ 	struct sock *sk;
+ 	int hash;
+ 
+-	hash = nexthdr & (RAW_HTABLE_SIZE - 1);
++	hash = raw_hashfunc(net, nexthdr);
+ 	hlist = &raw_v6_hashinfo.ht[hash];
+ 	rcu_read_lock();
+-	sk_nulls_for_each(sk, hnode, hlist) {
++	sk_for_each_rcu(sk, hlist) {
+ 		/* Note: ipv6_hdr(skb) != skb->data */
+ 		const struct ipv6hdr *ip6h = (const struct ipv6hdr *)skb->data;
+ 
+diff --git a/net/l2tp/l2tp_ip.c b/net/l2tp/l2tp_ip.c
+index 4db5a554bdbd9..41a74fc84ca13 100644
+--- a/net/l2tp/l2tp_ip.c
++++ b/net/l2tp/l2tp_ip.c
+@@ -677,8 +677,8 @@ MODULE_AUTHOR("James Chapman <jchapman@katalix.com>");
+ MODULE_DESCRIPTION("L2TP over IP");
+ MODULE_VERSION("1.0");
+ 
+-/* Use the value of SOCK_DGRAM (2) directory, because __stringify doesn't like
+- * enums
++/* Use the values of SOCK_DGRAM (2) as type and IPPROTO_L2TP (115) as protocol,
++ * because __stringify doesn't like enums
+  */
+-MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_INET, 2, IPPROTO_L2TP);
+-MODULE_ALIAS_NET_PF_PROTO(PF_INET, IPPROTO_L2TP);
++MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_INET, 115, 2);
++MODULE_ALIAS_NET_PF_PROTO(PF_INET, 115);
+diff --git a/net/l2tp/l2tp_ip6.c b/net/l2tp/l2tp_ip6.c
+index 9dbd801ddb98c..9db7f4f5a4414 100644
+--- a/net/l2tp/l2tp_ip6.c
++++ b/net/l2tp/l2tp_ip6.c
+@@ -808,8 +808,8 @@ MODULE_AUTHOR("Chris Elston <celston@katalix.com>");
+ MODULE_DESCRIPTION("L2TP IP encapsulation for IPv6");
+ MODULE_VERSION("1.0");
+ 
+-/* Use the value of SOCK_DGRAM (2) directory, because __stringify doesn't like
+- * enums
++/* Use the values of SOCK_DGRAM (2) as type and IPPROTO_L2TP (115) as protocol,
++ * because __stringify doesn't like enums
+  */
+-MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_INET6, 2, IPPROTO_L2TP);
+-MODULE_ALIAS_NET_PF_PROTO(PF_INET6, IPPROTO_L2TP);
++MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_INET6, 115, 2);
++MODULE_ALIAS_NET_PF_PROTO(PF_INET6, 115);
+diff --git a/net/mac80211/sta_info.c b/net/mac80211/sta_info.c
+index 3603cbc167570..30efa26f977f6 100644
+--- a/net/mac80211/sta_info.c
++++ b/net/mac80211/sta_info.c
+@@ -1242,7 +1242,8 @@ static int __must_check __sta_info_destroy_part1(struct sta_info *sta)
+ 	list_del_rcu(&sta->list);
+ 	sta->removed = true;
+ 
+-	drv_sta_pre_rcu_remove(local, sta->sdata, sta);
++	if (sta->uploaded)
++		drv_sta_pre_rcu_remove(local, sta->sdata, sta);
+ 
+ 	if (sdata->vif.type == NL80211_IFTYPE_AP_VLAN &&
+ 	    rcu_access_pointer(sdata->u.vlan.sta) == sta)
+diff --git a/net/mac80211/util.c b/net/mac80211/util.c
+index ed53c51bbc321..0785d9393e718 100644
+--- a/net/mac80211/util.c
++++ b/net/mac80211/util.c
+@@ -4785,7 +4785,7 @@ u8 ieee80211_ie_len_eht_cap(struct ieee80211_sub_if_data *sdata, u8 iftype)
+ 				       &eht_cap->eht_cap_elem,
+ 				       is_ap);
+ 	return 2 + 1 +
+-	       sizeof(he_cap->he_cap_elem) + n +
++	       sizeof(eht_cap->eht_cap_elem) + n +
+ 	       ieee80211_eht_ppe_size(eht_cap->eht_ppe_thres[0],
+ 				      eht_cap->eht_cap_elem.phy_cap_info);
+ 	return 0;
+diff --git a/net/netlink/af_netlink.c b/net/netlink/af_netlink.c
+index e506712967918..99622c64081c4 100644
+--- a/net/netlink/af_netlink.c
++++ b/net/netlink/af_netlink.c
+@@ -1941,7 +1941,7 @@ static int netlink_recvmsg(struct socket *sock, struct msghdr *msg, size_t len,
+ 	struct scm_cookie scm;
+ 	struct sock *sk = sock->sk;
+ 	struct netlink_sock *nlk = nlk_sk(sk);
+-	size_t copied;
++	size_t copied, max_recvmsg_len;
+ 	struct sk_buff *skb, *data_skb;
+ 	int err, ret;
+ 
+@@ -1974,9 +1974,10 @@ static int netlink_recvmsg(struct socket *sock, struct msghdr *msg, size_t len,
+ #endif
+ 
+ 	/* Record the max length of recvmsg() calls for future allocations */
+-	nlk->max_recvmsg_len = max(nlk->max_recvmsg_len, len);
+-	nlk->max_recvmsg_len = min_t(size_t, nlk->max_recvmsg_len,
+-				     SKB_WITH_OVERHEAD(32768));
++	max_recvmsg_len = max(READ_ONCE(nlk->max_recvmsg_len), len);
++	max_recvmsg_len = min_t(size_t, max_recvmsg_len,
++				SKB_WITH_OVERHEAD(32768));
++	WRITE_ONCE(nlk->max_recvmsg_len, max_recvmsg_len);
+ 
+ 	copied = data_skb->len;
+ 	if (len < copied) {
+@@ -2225,6 +2226,7 @@ static int netlink_dump(struct sock *sk)
+ 	struct netlink_ext_ack extack = {};
+ 	struct netlink_callback *cb;
+ 	struct sk_buff *skb = NULL;
++	size_t max_recvmsg_len;
+ 	struct module *module;
+ 	int err = -ENOBUFS;
+ 	int alloc_min_size;
+@@ -2247,8 +2249,9 @@ static int netlink_dump(struct sock *sk)
+ 	cb = &nlk->cb;
+ 	alloc_min_size = max_t(int, cb->min_dump_alloc, NLMSG_GOODSIZE);
+ 
+-	if (alloc_min_size < nlk->max_recvmsg_len) {
+-		alloc_size = nlk->max_recvmsg_len;
++	max_recvmsg_len = READ_ONCE(nlk->max_recvmsg_len);
++	if (alloc_min_size < max_recvmsg_len) {
++		alloc_size = max_recvmsg_len;
+ 		skb = alloc_skb(alloc_size,
+ 				(GFP_KERNEL & ~__GFP_DIRECT_RECLAIM) |
+ 				__GFP_NOWARN | __GFP_NORETRY);
+diff --git a/net/qrtr/af_qrtr.c b/net/qrtr/af_qrtr.c
+index 5c2fb992803b7..3a70255c8d02f 100644
+--- a/net/qrtr/af_qrtr.c
++++ b/net/qrtr/af_qrtr.c
+@@ -393,10 +393,12 @@ static struct qrtr_node *qrtr_node_lookup(unsigned int nid)
+ 	struct qrtr_node *node;
+ 	unsigned long flags;
+ 
++	mutex_lock(&qrtr_node_lock);
+ 	spin_lock_irqsave(&qrtr_nodes_lock, flags);
+ 	node = radix_tree_lookup(&qrtr_nodes, nid);
+ 	node = qrtr_node_acquire(node);
+ 	spin_unlock_irqrestore(&qrtr_nodes_lock, flags);
++	mutex_unlock(&qrtr_node_lock);
+ 
+ 	return node;
+ }
+diff --git a/net/qrtr/ns.c b/net/qrtr/ns.c
+index e595079c2cafe..3e40a1ba48f79 100644
+--- a/net/qrtr/ns.c
++++ b/net/qrtr/ns.c
+@@ -273,7 +273,7 @@ err:
+ 	return NULL;
+ }
+ 
+-static int server_del(struct qrtr_node *node, unsigned int port)
++static int server_del(struct qrtr_node *node, unsigned int port, bool bcast)
+ {
+ 	struct qrtr_lookup *lookup;
+ 	struct qrtr_server *srv;
+@@ -286,7 +286,7 @@ static int server_del(struct qrtr_node *node, unsigned int port)
+ 	radix_tree_delete(&node->servers, port);
+ 
+ 	/* Broadcast the removal of local servers */
+-	if (srv->node == qrtr_ns.local_node)
++	if (srv->node == qrtr_ns.local_node && bcast)
+ 		service_announce_del(&qrtr_ns.bcast_sq, srv);
+ 
+ 	/* Announce the service's disappearance to observers */
+@@ -372,7 +372,7 @@ static int ctrl_cmd_bye(struct sockaddr_qrtr *from)
+ 		}
+ 		slot = radix_tree_iter_resume(slot, &iter);
+ 		rcu_read_unlock();
+-		server_del(node, srv->port);
++		server_del(node, srv->port, true);
+ 		rcu_read_lock();
+ 	}
+ 	rcu_read_unlock();
+@@ -458,10 +458,13 @@ static int ctrl_cmd_del_client(struct sockaddr_qrtr *from,
+ 		kfree(lookup);
+ 	}
+ 
+-	/* Remove the server belonging to this port */
++	/* Remove the server belonging to this port but don't broadcast
++	 * DEL_SERVER. Neighbours would've already removed the server belonging
++	 * to this port due to the DEL_CLIENT broadcast from qrtr_port_remove().
++	 */
+ 	node = node_get(node_id);
+ 	if (node)
+-		server_del(node, port);
++		server_del(node, port, false);
+ 
+ 	/* Advertise the removal of this client to all local servers */
+ 	local_node = node_get(qrtr_ns.local_node);
+@@ -566,7 +569,7 @@ static int ctrl_cmd_del_server(struct sockaddr_qrtr *from,
+ 	if (!node)
+ 		return -ENOENT;
+ 
+-	return server_del(node, port);
++	return server_del(node, port, true);
+ }
+ 
+ static int ctrl_cmd_new_lookup(struct sockaddr_qrtr *from,
+diff --git a/net/sctp/socket.c b/net/sctp/socket.c
+index 83628c347744b..507b2ad5ef7c7 100644
+--- a/net/sctp/socket.c
++++ b/net/sctp/socket.c
+@@ -1829,6 +1829,10 @@ static int sctp_sendmsg_to_asoc(struct sctp_association *asoc,
+ 		err = sctp_wait_for_sndbuf(asoc, &timeo, msg_len);
+ 		if (err)
+ 			goto err;
++		if (unlikely(sinfo->sinfo_stream >= asoc->stream.outcnt)) {
++			err = -EINVAL;
++			goto err;
++		}
+ 	}
+ 
+ 	if (sctp_state(asoc, CLOSED)) {
+diff --git a/net/sunrpc/svcauth_unix.c b/net/sunrpc/svcauth_unix.c
+index b1efc34db6ed8..609ade4fb49ed 100644
+--- a/net/sunrpc/svcauth_unix.c
++++ b/net/sunrpc/svcauth_unix.c
+@@ -416,14 +416,23 @@ static int unix_gid_hash(kuid_t uid)
+ 	return hash_long(from_kuid(&init_user_ns, uid), GID_HASHBITS);
+ }
+ 
+-static void unix_gid_put(struct kref *kref)
++static void unix_gid_free(struct rcu_head *rcu)
+ {
+-	struct cache_head *item = container_of(kref, struct cache_head, ref);
+-	struct unix_gid *ug = container_of(item, struct unix_gid, h);
++	struct unix_gid *ug = container_of(rcu, struct unix_gid, rcu);
++	struct cache_head *item = &ug->h;
++
+ 	if (test_bit(CACHE_VALID, &item->flags) &&
+ 	    !test_bit(CACHE_NEGATIVE, &item->flags))
+ 		put_group_info(ug->gi);
+-	kfree_rcu(ug, rcu);
++	kfree(ug);
++}
++
++static void unix_gid_put(struct kref *kref)
++{
++	struct cache_head *item = container_of(kref, struct cache_head, ref);
++	struct unix_gid *ug = container_of(item, struct unix_gid, h);
++
++	call_rcu(&ug->rcu, unix_gid_free);
+ }
+ 
+ static int unix_gid_match(struct cache_head *corig, struct cache_head *cnew)
+diff --git a/sound/pci/hda/patch_hdmi.c b/sound/pci/hda/patch_hdmi.c
+index 9ea633fe93393..4ffa3a59f419f 100644
+--- a/sound/pci/hda/patch_hdmi.c
++++ b/sound/pci/hda/patch_hdmi.c
+@@ -81,6 +81,7 @@ struct hdmi_spec_per_pin {
+ 	struct delayed_work work;
+ 	struct hdmi_pcm *pcm; /* pointer to spec->pcm_rec[n] dynamically*/
+ 	int pcm_idx; /* which pcm is attached. -1 means no pcm is attached */
++	int prev_pcm_idx; /* previously assigned pcm index */
+ 	int repoll_count;
+ 	bool setup; /* the stream has been set up by prepare callback */
+ 	bool silent_stream;
+@@ -1380,9 +1381,17 @@ static void hdmi_attach_hda_pcm(struct hdmi_spec *spec,
+ 	/* pcm already be attached to the pin */
+ 	if (per_pin->pcm)
+ 		return;
++	/* try the previously used slot at first */
++	idx = per_pin->prev_pcm_idx;
++	if (idx >= 0) {
++		if (!test_bit(idx, &spec->pcm_bitmap))
++			goto found;
++		per_pin->prev_pcm_idx = -1; /* no longer valid, clear it */
++	}
+ 	idx = hdmi_find_pcm_slot(spec, per_pin);
+ 	if (idx == -EBUSY)
+ 		return;
++ found:
+ 	per_pin->pcm_idx = idx;
+ 	per_pin->pcm = get_hdmi_pcm(spec, idx);
+ 	set_bit(idx, &spec->pcm_bitmap);
+@@ -1398,6 +1407,7 @@ static void hdmi_detach_hda_pcm(struct hdmi_spec *spec,
+ 		return;
+ 	idx = per_pin->pcm_idx;
+ 	per_pin->pcm_idx = -1;
++	per_pin->prev_pcm_idx = idx; /* remember the previous index */
+ 	per_pin->pcm = NULL;
+ 	if (idx >= 0 && idx < spec->pcm_used)
+ 		clear_bit(idx, &spec->pcm_bitmap);
+@@ -1924,6 +1934,7 @@ static int hdmi_add_pin(struct hda_codec *codec, hda_nid_t pin_nid)
+ 
+ 		per_pin->pcm = NULL;
+ 		per_pin->pcm_idx = -1;
++		per_pin->prev_pcm_idx = -1;
+ 		per_pin->pin_nid = pin_nid;
+ 		per_pin->pin_nid_idx = spec->num_nids;
+ 		per_pin->dev_id = i;
+diff --git a/sound/pci/hda/patch_realtek.c b/sound/pci/hda/patch_realtek.c
+index 070150bbd3559..50b8573b52066 100644
+--- a/sound/pci/hda/patch_realtek.c
++++ b/sound/pci/hda/patch_realtek.c
+@@ -2624,6 +2624,7 @@ static const struct snd_pci_quirk alc882_fixup_tbl[] = {
+ 	SND_PCI_QUIRK(0x1462, 0xda57, "MSI Z270-Gaming", ALC1220_FIXUP_GB_DUAL_CODECS),
+ 	SND_PCI_QUIRK_VENDOR(0x1462, "MSI", ALC882_FIXUP_GPIO3),
+ 	SND_PCI_QUIRK(0x147b, 0x107a, "Abit AW9D-MAX", ALC882_FIXUP_ABIT_AW9D_MAX),
++	SND_PCI_QUIRK(0x1558, 0x3702, "Clevo X370SN[VW]", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
+ 	SND_PCI_QUIRK(0x1558, 0x50d3, "Clevo PC50[ER][CDF]", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
+ 	SND_PCI_QUIRK(0x1558, 0x65d1, "Clevo PB51[ER][CDF]", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
+ 	SND_PCI_QUIRK(0x1558, 0x65d2, "Clevo PB51R[CDF]", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
+@@ -9442,6 +9443,7 @@ static const struct snd_pci_quirk alc269_fixup_tbl[] = {
+ 	SND_PCI_QUIRK(0x103c, 0x8b47, "HP", ALC245_FIXUP_CS35L41_SPI_2_HP_GPIO_LED),
+ 	SND_PCI_QUIRK(0x103c, 0x8b5d, "HP", ALC236_FIXUP_HP_MUTE_LED_MICMUTE_VREF),
+ 	SND_PCI_QUIRK(0x103c, 0x8b5e, "HP", ALC236_FIXUP_HP_MUTE_LED_MICMUTE_VREF),
++	SND_PCI_QUIRK(0x103c, 0x8b66, "HP", ALC236_FIXUP_HP_MUTE_LED_MICMUTE_VREF),
+ 	SND_PCI_QUIRK(0x103c, 0x8b7a, "HP", ALC236_FIXUP_HP_GPIO_LED),
+ 	SND_PCI_QUIRK(0x103c, 0x8b7d, "HP", ALC236_FIXUP_HP_GPIO_LED),
+ 	SND_PCI_QUIRK(0x103c, 0x8b87, "HP", ALC236_FIXUP_HP_GPIO_LED),
+diff --git a/sound/soc/codecs/hdac_hdmi.c b/sound/soc/codecs/hdac_hdmi.c
+index cb23650ad5223..d8e83150ea28e 100644
+--- a/sound/soc/codecs/hdac_hdmi.c
++++ b/sound/soc/codecs/hdac_hdmi.c
+@@ -436,23 +436,28 @@ static int hdac_hdmi_setup_audio_infoframe(struct hdac_device *hdev,
+ 	return 0;
+ }
+ 
+-static int hdac_hdmi_set_tdm_slot(struct snd_soc_dai *dai,
+-		unsigned int tx_mask, unsigned int rx_mask,
+-		int slots, int slot_width)
++static int hdac_hdmi_set_stream(struct snd_soc_dai *dai,
++				void *stream, int direction)
+ {
+ 	struct hdac_hdmi_priv *hdmi = snd_soc_dai_get_drvdata(dai);
+ 	struct hdac_device *hdev = hdmi->hdev;
+ 	struct hdac_hdmi_dai_port_map *dai_map;
+ 	struct hdac_hdmi_pcm *pcm;
++	struct hdac_stream *hstream;
+ 
+-	dev_dbg(&hdev->dev, "%s: strm_tag: %d\n", __func__, tx_mask);
++	if (!stream)
++		return -EINVAL;
++
++	hstream = (struct hdac_stream *)stream;
++
++	dev_dbg(&hdev->dev, "%s: strm_tag: %d\n", __func__, hstream->stream_tag);
+ 
+ 	dai_map = &hdmi->dai_map[dai->id];
+ 
+ 	pcm = hdac_hdmi_get_pcm_from_cvt(hdmi, dai_map->cvt);
+ 
+ 	if (pcm)
+-		pcm->stream_tag = (tx_mask << 4);
++		pcm->stream_tag = (hstream->stream_tag << 4);
+ 
+ 	return 0;
+ }
+@@ -1544,7 +1549,7 @@ static const struct snd_soc_dai_ops hdmi_dai_ops = {
+ 	.startup = hdac_hdmi_pcm_open,
+ 	.shutdown = hdac_hdmi_pcm_close,
+ 	.hw_params = hdac_hdmi_set_hw_params,
+-	.set_tdm_slot = hdac_hdmi_set_tdm_slot,
++	.set_stream = hdac_hdmi_set_stream,
+ };
+ 
+ /*
+diff --git a/sound/soc/codecs/lpass-rx-macro.c b/sound/soc/codecs/lpass-rx-macro.c
+index 8621cfabcf5b6..1639f3b66facb 100644
+--- a/sound/soc/codecs/lpass-rx-macro.c
++++ b/sound/soc/codecs/lpass-rx-macro.c
+@@ -3667,9 +3667,9 @@ static int __maybe_unused rx_macro_runtime_suspend(struct device *dev)
+ 	regcache_cache_only(rx->regmap, true);
+ 	regcache_mark_dirty(rx->regmap);
+ 
+-	clk_disable_unprepare(rx->mclk);
+-	clk_disable_unprepare(rx->npl);
+ 	clk_disable_unprepare(rx->fsgen);
++	clk_disable_unprepare(rx->npl);
++	clk_disable_unprepare(rx->mclk);
+ 
+ 	return 0;
+ }
+diff --git a/sound/soc/codecs/lpass-tx-macro.c b/sound/soc/codecs/lpass-tx-macro.c
+index e5611f655beda..d1d9d8d2df2d2 100644
+--- a/sound/soc/codecs/lpass-tx-macro.c
++++ b/sound/soc/codecs/lpass-tx-macro.c
+@@ -1946,9 +1946,9 @@ static int __maybe_unused tx_macro_runtime_suspend(struct device *dev)
+ 	regcache_cache_only(tx->regmap, true);
+ 	regcache_mark_dirty(tx->regmap);
+ 
+-	clk_disable_unprepare(tx->mclk);
+-	clk_disable_unprepare(tx->npl);
+ 	clk_disable_unprepare(tx->fsgen);
++	clk_disable_unprepare(tx->npl);
++	clk_disable_unprepare(tx->mclk);
+ 
+ 	return 0;
+ }
+diff --git a/sound/soc/codecs/lpass-wsa-macro.c b/sound/soc/codecs/lpass-wsa-macro.c
+index c012033fb69ed..8ed48c86ccb33 100644
+--- a/sound/soc/codecs/lpass-wsa-macro.c
++++ b/sound/soc/codecs/lpass-wsa-macro.c
+@@ -2502,9 +2502,9 @@ static int __maybe_unused wsa_macro_runtime_suspend(struct device *dev)
+ 	regcache_cache_only(wsa->regmap, true);
+ 	regcache_mark_dirty(wsa->regmap);
+ 
+-	clk_disable_unprepare(wsa->mclk);
+-	clk_disable_unprepare(wsa->npl);
+ 	clk_disable_unprepare(wsa->fsgen);
++	clk_disable_unprepare(wsa->npl);
++	clk_disable_unprepare(wsa->mclk);
+ 
+ 	return 0;
+ }
+diff --git a/sound/soc/sof/ipc4.c b/sound/soc/sof/ipc4.c
+index 6eaa18e27e5af..c08f3960ddd96 100644
+--- a/sound/soc/sof/ipc4.c
++++ b/sound/soc/sof/ipc4.c
+@@ -392,6 +392,9 @@ static int sof_ipc4_tx_msg(struct snd_sof_dev *sdev, void *msg_data, size_t msg_
+ static int sof_ipc4_set_get_data(struct snd_sof_dev *sdev, void *data,
+ 				 size_t payload_bytes, bool set)
+ {
++	const struct sof_dsp_power_state target_state = {
++			.state = SOF_DSP_PM_D0,
++	};
+ 	size_t payload_limit = sdev->ipc->max_payload_size;
+ 	struct sof_ipc4_msg *ipc4_msg = data;
+ 	struct sof_ipc4_msg tx = {{ 0 }};
+@@ -422,6 +425,11 @@ static int sof_ipc4_set_get_data(struct snd_sof_dev *sdev, void *data,
+ 
+ 	tx.extension |= SOF_IPC4_MOD_EXT_MSG_FIRST_BLOCK(1);
+ 
++	/* ensure the DSP is in D0i0 before sending IPC */
++	ret = snd_sof_dsp_set_power_state(sdev, &target_state);
++	if (ret < 0)
++		return ret;
++
+ 	/* Serialise IPC TX */
+ 	mutex_lock(&sdev->ipc->tx_mutex);
+ 
+diff --git a/tools/lib/bpf/btf_dump.c b/tools/lib/bpf/btf_dump.c
+index 56102711f395a..80a15a6802094 100644
+--- a/tools/lib/bpf/btf_dump.c
++++ b/tools/lib/bpf/btf_dump.c
+@@ -1011,7 +1011,11 @@ static void btf_dump_emit_struct_def(struct btf_dump *d,
+ 	if (is_struct)
+ 		btf_dump_emit_bit_padding(d, off, t->size * 8, align, false, lvl + 1);
+ 
+-	if (vlen)
++	/*
++	 * Keep `struct empty {}` on a single line,
++	 * only print newline when there are regular or padding fields.
++	 */
++	if (vlen || t->size)
+ 		btf_dump_printf(d, "\n");
+ 	btf_dump_printf(d, "%s}", pfx(lvl));
+ 	if (packed)
+diff --git a/tools/testing/radix-tree/maple.c b/tools/testing/radix-tree/maple.c
+index 2e91973fbaa66..aceb6011315c2 100644
+--- a/tools/testing/radix-tree/maple.c
++++ b/tools/testing/radix-tree/maple.c
+@@ -172,11 +172,11 @@ static noinline void check_new_node(struct maple_tree *mt)
+ 
+ 		if (!MAPLE_32BIT) {
+ 			if (i >= 35)
+-				e = i - 35;
++				e = i - 34;
+ 			else if (i >= 5)
+-				e = i - 5;
++				e = i - 4;
+ 			else if (i >= 2)
+-				e = i - 2;
++				e = i - 1;
+ 		} else {
+ 			if (i >= 4)
+ 				e = i - 4;
+@@ -304,17 +304,17 @@ static noinline void check_new_node(struct maple_tree *mt)
+ 	MT_BUG_ON(mt, mas.node != MA_ERROR(-ENOMEM));
+ 	MT_BUG_ON(mt, !mas_nomem(&mas, GFP_KERNEL));
+ 	MT_BUG_ON(mt, mas_allocated(&mas) != MAPLE_ALLOC_SLOTS + 1);
+-	MT_BUG_ON(mt, mas.alloc->node_count != MAPLE_ALLOC_SLOTS - 1);
++	MT_BUG_ON(mt, mas.alloc->node_count != MAPLE_ALLOC_SLOTS);
+ 
+ 	mn = mas_pop_node(&mas); /* get the next node. */
+ 	MT_BUG_ON(mt, mn == NULL);
+ 	MT_BUG_ON(mt, not_empty(mn));
+ 	MT_BUG_ON(mt, mas_allocated(&mas) != MAPLE_ALLOC_SLOTS);
+-	MT_BUG_ON(mt, mas.alloc->node_count != MAPLE_ALLOC_SLOTS - 2);
++	MT_BUG_ON(mt, mas.alloc->node_count != MAPLE_ALLOC_SLOTS - 1);
+ 
+ 	mas_push_node(&mas, mn);
+ 	MT_BUG_ON(mt, mas_allocated(&mas) != MAPLE_ALLOC_SLOTS + 1);
+-	MT_BUG_ON(mt, mas.alloc->node_count != MAPLE_ALLOC_SLOTS - 1);
++	MT_BUG_ON(mt, mas.alloc->node_count != MAPLE_ALLOC_SLOTS);
+ 
+ 	/* Check the limit of pop/push/pop */
+ 	mas_node_count(&mas, MAPLE_ALLOC_SLOTS + 2); /* Request */
+@@ -322,14 +322,14 @@ static noinline void check_new_node(struct maple_tree *mt)
+ 	MT_BUG_ON(mt, mas.node != MA_ERROR(-ENOMEM));
+ 	MT_BUG_ON(mt, !mas_nomem(&mas, GFP_KERNEL));
+ 	MT_BUG_ON(mt, mas_alloc_req(&mas));
+-	MT_BUG_ON(mt, mas.alloc->node_count);
++	MT_BUG_ON(mt, mas.alloc->node_count != 1);
+ 	MT_BUG_ON(mt, mas_allocated(&mas) != MAPLE_ALLOC_SLOTS + 2);
+ 	mn = mas_pop_node(&mas);
+ 	MT_BUG_ON(mt, not_empty(mn));
+ 	MT_BUG_ON(mt, mas_allocated(&mas) != MAPLE_ALLOC_SLOTS + 1);
+-	MT_BUG_ON(mt, mas.alloc->node_count  != MAPLE_ALLOC_SLOTS - 1);
++	MT_BUG_ON(mt, mas.alloc->node_count  != MAPLE_ALLOC_SLOTS);
+ 	mas_push_node(&mas, mn);
+-	MT_BUG_ON(mt, mas.alloc->node_count);
++	MT_BUG_ON(mt, mas.alloc->node_count != 1);
+ 	MT_BUG_ON(mt, mas_allocated(&mas) != MAPLE_ALLOC_SLOTS + 2);
+ 	mn = mas_pop_node(&mas);
+ 	MT_BUG_ON(mt, not_empty(mn));


             reply	other threads:[~2023-04-13 16:09 UTC|newest]

Thread overview: 188+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-04-13 16:09 Mike Pagano [this message]
  -- strict thread matches above, loose matches on Subject: below --
2025-03-29 10:49 [gentoo-commits] proj/linux-patches:6.1 commit in: / Mike Pagano
2025-03-13 12:56 Mike Pagano
2025-03-07 16:38 Mike Pagano
2025-02-21 13:32 Mike Pagano
2025-02-01 23:08 Mike Pagano
2025-01-30 12:56 Mike Pagano
2025-01-23 17:04 Mike Pagano
2025-01-19 10:58 Mike Pagano
2025-01-17 13:19 Mike Pagano
2025-01-09 13:54 Mike Pagano
2025-01-02 12:35 Mike Pagano
2024-12-27 14:09 Mike Pagano
2024-12-19 18:08 Mike Pagano
2024-12-14 23:49 Mike Pagano
2024-12-12 19:42 Mike Pagano
2024-11-22 17:48 Mike Pagano
2024-11-17 18:17 Mike Pagano
2024-11-14 14:55 Mike Pagano
2024-11-08 16:31 Mike Pagano
2024-11-04 20:52 Mike Pagano
2024-11-03 13:58 Mike Pagano
2024-11-01 11:33 Mike Pagano
2024-11-01 11:28 Mike Pagano
2024-10-25 11:46 Mike Pagano
2024-10-22 16:58 Mike Pagano
2024-10-17 14:24 Mike Pagano
2024-10-17 14:06 Mike Pagano
2024-09-30 16:04 Mike Pagano
2024-09-18 18:04 Mike Pagano
2024-09-12 12:35 Mike Pagano
2024-09-08 11:06 Mike Pagano
2024-09-04 13:52 Mike Pagano
2024-08-29 16:49 Mike Pagano
2024-08-19 10:43 Mike Pagano
2024-08-14 15:06 Mike Pagano
2024-08-14 14:11 Mike Pagano
2024-08-11 13:32 Mike Pagano
2024-08-11 13:29 Mike Pagano
2024-08-10 15:45 Mike Pagano
2024-08-03 15:28 Mike Pagano
2024-07-27 13:47 Mike Pagano
2024-07-25 12:15 Mike Pagano
2024-07-25 12:09 Mike Pagano
2024-07-18 12:15 Mike Pagano
2024-07-15 11:16 Mike Pagano
2024-07-11 11:49 Mike Pagano
2024-07-05 11:07 Mike Pagano
2024-06-27 13:10 Mike Pagano
2024-06-27 12:33 Mike Pagano
2024-06-21 14:07 Mike Pagano
2024-06-16 14:33 Mike Pagano
2024-06-12 10:16 Mike Pagano
2024-05-25 15:16 Mike Pagano
2024-05-17 11:36 Mike Pagano
2024-05-05 18:10 Mike Pagano
2024-05-02 15:01 Mike Pagano
2024-04-29 11:30 Mike Pagano
2024-04-29 11:27 Mike Pagano
2024-04-27 22:45 Mike Pagano
2024-04-27 17:06 Mike Pagano
2024-04-18  3:05 Alice Ferrazzi
2024-04-13 13:07 Mike Pagano
2024-04-10 15:10 Mike Pagano
2024-04-03 13:54 Mike Pagano
2024-03-27 11:24 Mike Pagano
2024-03-15 22:00 Mike Pagano
2024-03-06 18:07 Mike Pagano
2024-03-01 13:07 Mike Pagano
2024-02-23 13:19 Mike Pagano
2024-02-23 12:37 Mike Pagano
2024-02-16 19:00 Mike Pagano
2024-02-05 21:01 Mike Pagano
2024-02-01  1:23 Mike Pagano
2024-01-26  0:09 Mike Pagano
2024-01-20 11:45 Mike Pagano
2024-01-15 18:47 Mike Pagano
2024-01-10 17:16 Mike Pagano
2024-01-05 14:54 Mike Pagano
2024-01-05 14:50 Mike Pagano
2024-01-04 16:10 Mike Pagano
2024-01-01 13:46 Mike Pagano
2023-12-20 16:56 Mike Pagano
2023-12-13 18:27 Mike Pagano
2023-12-11 14:20 Mike Pagano
2023-12-08 10:55 Mike Pagano
2023-12-03 11:16 Mike Pagano
2023-12-01 10:36 Mike Pagano
2023-11-28 17:51 Mike Pagano
2023-11-20 11:23 Mike Pagano
2023-11-08 14:02 Mike Pagano
2023-11-02 11:10 Mike Pagano
2023-10-25 11:36 Mike Pagano
2023-10-22 22:53 Mike Pagano
2023-10-19 22:30 Mike Pagano
2023-10-18 20:04 Mike Pagano
2023-10-15 17:40 Mike Pagano
2023-10-10 22:56 Mike Pagano
2023-10-06 13:18 Mike Pagano
2023-10-05 14:23 Mike Pagano
2023-09-23 11:03 Mike Pagano
2023-09-23 10:16 Mike Pagano
2023-09-19 13:20 Mike Pagano
2023-09-15 18:04 Mike Pagano
2023-09-13 11:19 Mike Pagano
2023-09-13 11:05 Mike Pagano
2023-09-06 22:16 Mike Pagano
2023-09-02  9:56 Mike Pagano
2023-08-30 14:42 Mike Pagano
2023-08-27 21:41 Mike Pagano
2023-08-26 15:19 Mike Pagano
2023-08-26 15:00 Mike Pagano
2023-08-23 18:08 Mike Pagano
2023-08-16 18:32 Mike Pagano
2023-08-16 18:32 Mike Pagano
2023-08-11 11:55 Mike Pagano
2023-08-08 18:40 Mike Pagano
2023-08-03 11:54 Mike Pagano
2023-08-03 11:48 Mike Pagano
2023-07-27 11:48 Mike Pagano
2023-07-24 20:27 Mike Pagano
2023-07-23 15:14 Mike Pagano
2023-07-19 17:05 Mike Pagano
2023-07-05 20:34 Mike Pagano
2023-07-05 20:28 Mike Pagano
2023-07-04 13:15 Mike Pagano
2023-07-01 18:27 Mike Pagano
2023-06-28 10:26 Mike Pagano
2023-06-21 14:54 Alice Ferrazzi
2023-06-14 10:17 Mike Pagano
2023-06-09 12:02 Mike Pagano
2023-06-09 11:29 Mike Pagano
2023-06-05 11:48 Mike Pagano
2023-06-02 15:07 Mike Pagano
2023-05-30 16:51 Mike Pagano
2023-05-24 17:05 Mike Pagano
2023-05-17 10:57 Mike Pagano
2023-05-11 16:08 Mike Pagano
2023-05-11 14:49 Mike Pagano
2023-05-10 17:54 Mike Pagano
2023-05-10 16:18 Mike Pagano
2023-04-30 23:50 Alice Ferrazzi
2023-04-26 13:19 Mike Pagano
2023-04-20 11:16 Alice Ferrazzi
2023-04-06 10:41 Alice Ferrazzi
2023-03-30 20:52 Mike Pagano
2023-03-30 11:21 Alice Ferrazzi
2023-03-22 14:15 Alice Ferrazzi
2023-03-21 13:32 Mike Pagano
2023-03-17 10:43 Mike Pagano
2023-03-13 11:30 Alice Ferrazzi
2023-03-11 14:09 Mike Pagano
2023-03-11 11:19 Mike Pagano
2023-03-10 12:57 Mike Pagano
2023-03-10 12:47 Mike Pagano
2023-03-06 17:30 Mike Pagano
2023-03-03 13:01 Mike Pagano
2023-03-03 12:28 Mike Pagano
2023-02-27 16:59 Mike Pagano
2023-02-26 18:24 Mike Pagano
2023-02-26 18:16 Mike Pagano
2023-02-25 11:02 Alice Ferrazzi
2023-02-24  3:03 Alice Ferrazzi
2023-02-22 13:46 Alice Ferrazzi
2023-02-14 18:35 Mike Pagano
2023-02-13 13:38 Mike Pagano
2023-02-09 12:52 Mike Pagano
2023-02-09 12:49 Mike Pagano
2023-02-09 12:47 Mike Pagano
2023-02-09 12:40 Mike Pagano
2023-02-09 12:34 Mike Pagano
2023-02-06 12:46 Mike Pagano
2023-02-02 19:02 Mike Pagano
2023-02-01  8:05 Alice Ferrazzi
2023-01-24  7:19 Alice Ferrazzi
2023-01-22 14:59 Mike Pagano
2023-01-18 11:29 Mike Pagano
2023-01-14 13:48 Mike Pagano
2023-01-12 15:25 Mike Pagano
2023-01-12 12:16 Mike Pagano
2023-01-07 11:10 Mike Pagano
2023-01-04 11:37 Mike Pagano
2022-12-31 15:28 Mike Pagano
2022-12-21 19:05 Alice Ferrazzi
2022-12-16 20:25 Mike Pagano
2022-12-16 19:44 Mike Pagano
2022-12-11 23:32 Mike Pagano
2022-12-11 14:28 Mike Pagano

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1681402159.ca7e265c42f5d0ef307cd4e76877fd77b86dae6c.mpagano@gentoo \
    --to=mpagano@gentoo.org \
    --cc=gentoo-commits@lists.gentoo.org \
    --cc=gentoo-dev@lists.gentoo.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox