Skip to content

Commit

Permalink
[Xtensa] Fix volatile load/store implementation
Browse files Browse the repository at this point in the history
  • Loading branch information
andreisfr committed Oct 2, 2024
1 parent c629fed commit 6bc11f4
Show file tree
Hide file tree
Showing 2 changed files with 160 additions and 30 deletions.
9 changes: 6 additions & 3 deletions llvm/lib/Target/Xtensa/XtensaISelLowering.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,7 @@

#include "XtensaISelLowering.h"
#include "XtensaConstantPoolValue.h"
#include "XtensaInstrInfo.h"
#include "XtensaSubtarget.h"
#include "XtensaTargetMachine.h"
#include "llvm/CodeGen/CallingConvLower.h"
Expand Down Expand Up @@ -1104,7 +1105,7 @@ XtensaTargetLowering::emitSelectCC(MachineInstr &MI,
MachineBasicBlock *XtensaTargetLowering::EmitInstrWithCustomInserter(
MachineInstr &MI, MachineBasicBlock *MBB) const {
DebugLoc DL = MI.getDebugLoc();
const TargetInstrInfo &TII = *Subtarget.getInstrInfo();
const XtensaInstrInfo &TII = *Subtarget.getInstrInfo();

switch (MI.getOpcode()) {
case Xtensa::SELECT:
Expand All @@ -1116,8 +1117,10 @@ MachineBasicBlock *XtensaTargetLowering::EmitInstrWithCustomInserter(
case Xtensa::L16SI:
case Xtensa::L16UI:
case Xtensa::L32I: {
const MachineMemOperand &MMO = **MI.memoperands_begin();
if (MMO.isVolatile()) {
// Insert memory wait instruction "memw" before volatile load/store as it is
// implemented in gcc. If memoperands is empty then assume that it aslo
// maybe volatile load/store and insert "memw".
if (MI.memoperands_empty() || (*MI.memoperands_begin())->isVolatile()) {
BuildMI(*MBB, MI, DL, TII.get(Xtensa::MEMW));
}
return MBB;
Expand Down
181 changes: 154 additions & 27 deletions llvm/test/CodeGen/Xtensa/volatile.ll
Original file line number Diff line number Diff line change
@@ -1,3 +1,4 @@
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5
; RUN: llc -mtriple=xtensa -verify-machineinstrs < %s \
; RUN: | FileCheck %s

Expand All @@ -7,36 +8,36 @@
@y_i16 = common dso_local global i16 0, align 8
@x_i32 = common dso_local global i32 0, align 8
@y_i32 = common dso_local global i32 0, align 8
@x_i64 = common dso_local global i64 0, align 8
@y_i64 = common dso_local global i64 0, align 8
@x_float = common dso_local global float 0.0, align 8
@y_float = common dso_local global float 0.0, align 8
@x_double = common dso_local global double 0.0, align 8
@y_double = common dso_local global double 0.0, align 8
@x_vec = common dso_local global <4 x i32> <i32 0, i32 0, i32 0, i32 0>, align 8
@y_vec = common dso_local global <4 x i32> <i32 0, i32 0, i32 0, i32 0>, align 8

define void @test() {
; CHECK: .literal_position
; CHECK-NEXT: .literal .LCPI0_0, x_i8
; CHECK-NEXT: .literal .LCPI0_1, y_i8
; CHECK-NEXT: .literal .LCPI0_2, x_i16
; CHECK-NEXT: .literal .LCPI0_3, y_i16
; CHECK-NEXT: .literal .LCPI0_4, x_i32
; CHECK-NEXT: .literal .LCPI0_5, y_i32
; CHECK-LABEL: test:
; CHECK: # %bb.0:
; CHECK-NEXT: l32r a8, .LCPI0_0
; CHECK-NEXT: memw
; CHECK-NEXT: l8ui a8, a8, 0
; CHECK-NEXT: l32r a9, .LCPI0_1
; CHECK-NEXT: memw
; CHECK-NEXT: s8i a8, a9, 0
; CHECK-NEXT: l32r a8, .LCPI0_2
; CHECK-NEXT: memw
; CHECK-NEXT: l16ui a8, a8, 0
; CHECK-NEXT: l32r a9, .LCPI0_3
; CHECK-NEXT: memw
; CHECK-NEXT: s16i a8, a9, 0
; CHECK-NEXT: l32r a8, .LCPI0_4
; CHECK-NEXT: memw
; CHECK-NEXT: l32i a8, a8, 0
; CHECK-NEXT: l32r a9, .LCPI0_5
; CHECK-NEXT: memw
; CHECK-NEXT: s32i a8, a9, 0
; CHECK-NEXT: ret
; CHECK: l32r a8, .LCPI0_0
; CHECK-NEXT: memw
; CHECK-NEXT: l8ui a8, a8, 0
; CHECK-NEXT: l32r a9, .LCPI0_1
; CHECK-NEXT: memw
; CHECK-NEXT: s8i a8, a9, 0
; CHECK-NEXT: l32r a8, .LCPI0_2
; CHECK-NEXT: memw
; CHECK-NEXT: l16ui a8, a8, 0
; CHECK-NEXT: l32r a9, .LCPI0_3
; CHECK-NEXT: memw
; CHECK-NEXT: s16i a8, a9, 0
; CHECK-NEXT: l32r a8, .LCPI0_4
; CHECK-NEXT: memw
; CHECK-NEXT: l32i a8, a8, 0
; CHECK-NEXT: l32r a9, .LCPI0_5
; CHECK-NEXT: memw
; CHECK-NEXT: s32i a8, a9, 0
; CHECK-NEXT: ret

entry:
%0 = load volatile i8, ptr @x_i8, align 4
Expand All @@ -47,3 +48,129 @@ entry:
store volatile i32 %2, ptr @y_i32, align 4
ret void
}


define void @test_i8() {
; CHECK-LABEL: test_i8:
; CHECK: l32r a8, .LCPI1_0
; CHECK-NEXT: memw
; CHECK-NEXT: l8ui a8, a8, 0
; CHECK-NEXT: l32r a9, .LCPI1_1
; CHECK-NEXT: memw
; CHECK-NEXT: s8i a8, a9, 0
; CHECK-NEXT: ret
entry:
%a = load volatile i8, ptr @x_i8, align 4
store volatile i8 %a, ptr @y_i8, align 4
ret void
}

define void @test_i16() {
; CHECK-LABEL: test_i16:
; CHECK: l32r a8, .LCPI2_0
; CHECK-NEXT: memw
; CHECK-NEXT: l16ui a8, a8, 0
; CHECK-NEXT: l32r a9, .LCPI2_1
; CHECK-NEXT: memw
; CHECK-NEXT: s16i a8, a9, 0
; CHECK-NEXT: ret
entry:
%a = load volatile i16, ptr @x_i16, align 4
store volatile i16 %a, ptr @y_i16, align 4
ret void
}

define void @test_i32() {
; CHECK-LABEL: test_i32:
; CHECK: l32r a8, .LCPI3_0
; CHECK-NEXT: memw
; CHECK-NEXT: l32i a8, a8, 0
; CHECK-NEXT: l32r a9, .LCPI3_1
; CHECK-NEXT: memw
; CHECK-NEXT: s32i a8, a9, 0
; CHECK-NEXT: ret
entry:
%a = load volatile i32, ptr @x_i32, align 4
store volatile i32 %a, ptr @y_i32, align 4
ret void
}

define void @test_i64() {
; CHECK-LABEL: test_i64:
; CHECK: l32r a8, .LCPI4_0
; CHECK-NEXT: memw
; CHECK-NEXT: l32i a9, a8, 0
; CHECK-NEXT: memw
; CHECK-NEXT: l32i a8, a8, 4
; CHECK-NEXT: l32r a10, .LCPI4_1
; CHECK-NEXT: memw
; CHECK-NEXT: s32i a8, a10, 4
; CHECK-NEXT: memw
; CHECK-NEXT: s32i a9, a10, 0
; CHECK-NEXT: ret
entry:
%a = load volatile i64, ptr @x_i64, align 4
store volatile i64 %a, ptr @y_i64, align 4
ret void
}

define void @test_float() {
; CHECK-LABEL: test_float:
; CHECK: l32r a8, .LCPI5_0
; CHECK-NEXT: memw
; CHECK-NEXT: l32i a8, a8, 0
; CHECK-NEXT: l32r a9, .LCPI5_1
; CHECK-NEXT: memw
; CHECK-NEXT: s32i a8, a9, 0
; CHECK-NEXT: ret
entry:
%a = load volatile float, ptr @x_float, align 4
store volatile float %a, ptr @y_float, align 4
ret void
}

define void @test_double() {
; CHECK-LABEL: test_double:
; CHECK: l32r a8, .LCPI6_0
; CHECK-NEXT: memw
; CHECK-NEXT: l32i a9, a8, 0
; CHECK-NEXT: memw
; CHECK-NEXT: l32i a8, a8, 4
; CHECK-NEXT: l32r a10, .LCPI6_1
; CHECK-NEXT: memw
; CHECK-NEXT: s32i a8, a10, 4
; CHECK-NEXT: memw
; CHECK-NEXT: s32i a9, a10, 0
; CHECK-NEXT: ret
entry:
%a = load volatile double, ptr @x_double, align 4
store volatile double %a, ptr @y_double, align 4
ret void
}

define void @test_vec() {
; CHECK-LABEL: test_vec:
; CHECK: l32r a8, .LCPI7_0
; CHECK-NEXT: memw
; CHECK-NEXT: l32i a9, a8, 0
; CHECK-NEXT: memw
; CHECK-NEXT: l32i a10, a8, 4
; CHECK-NEXT: memw
; CHECK-NEXT: l32i a11, a8, 8
; CHECK-NEXT: memw
; CHECK-NEXT: l32i a8, a8, 12
; CHECK-NEXT: l32r a7, .LCPI7_1
; CHECK-NEXT: memw
; CHECK-NEXT: s32i a8, a7, 12
; CHECK-NEXT: memw
; CHECK-NEXT: s32i a11, a7, 8
; CHECK-NEXT: memw
; CHECK-NEXT: s32i a10, a7, 4
; CHECK-NEXT: memw
; CHECK-NEXT: s32i a9, a7, 0
; CHECK-NEXT: ret
entry:
%a = load volatile <4 x i32>, ptr @x_vec, align 4
store volatile <4 x i32> %a, ptr @y_vec, align 4
ret void
}

0 comments on commit 6bc11f4

Please sign in to comment.