summaryrefslogtreecommitdiff
path: root/mm
diff options
context:
space:
mode:
authorWill Deacon <will@kernel.org>2019-10-30 17:15:01 +0000
committerWill Deacon <will@kernel.org>2020-07-21 10:50:36 +0100
commitbb7cdd38185a4f9fa32e62db115c2c6dceb2b621 (patch)
treee595c58c6d7cdb11ba37afcf5e1b35dc3708b0ad /mm
parent71c0b9a65cefa8c34eab83d337a1e3ec61fb7cc2 (diff)
alpha: Replace smp_read_barrier_depends() usage with smp_[r]mb()
In preparation for removing smp_read_barrier_depends() altogether, move the Alpha code over to using smp_rmb() and smp_mb() directly. Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org> Acked-by: Paul E. McKenney <paulmck@kernel.org> Signed-off-by: Will Deacon <will@kernel.org>
Diffstat (limited to 'mm')
-rw-r--r--mm/memory.c2
1 files changed, 1 insertions, 1 deletions
diff --git a/mm/memory.c b/mm/memory.c
index 87ec87cdc1ff..e1f2c730d8bb 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -437,7 +437,7 @@ int __pte_alloc(struct mm_struct *mm, pmd_t *pmd)
* of a chain of data-dependent loads, meaning most CPUs (alpha
* being the notable exception) will already guarantee loads are
* seen in-order. See the alpha page table accessors for the
- * smp_read_barrier_depends() barriers in page table walking code.
+ * smp_rmb() barriers in page table walking code.
*/
smp_wmb(); /* Could be smp_wmb__xxx(before|after)_spin_lock */