Skip to content

Commit

Permalink
fix unittest
Browse files Browse the repository at this point in the history
  • Loading branch information
yeonbok committed Feb 13, 2025
1 parent 8a5d2fe commit 0855be7
Show file tree
Hide file tree
Showing 7 changed files with 18 additions and 56 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -7927,9 +7927,7 @@ TEST_P(convolution_grouped_gpu, base) {
auto outputs = network.execute();

auto out_mem = outputs.at("conv").get_memory();
auto out_lockable = engine.allocate_memory(out_mem->get_layout());
out_mem->copy_to(get_test_stream(), *out_lockable, true);
cldnn::mem_lock<float> out_ptr(out_lockable, get_test_stream());
cldnn::mem_lock<float, mem_lock_type::read> out_ptr(out_mem, get_test_stream());
auto out_lay = out_mem->get_layout();

ASSERT_EQ(out_mem->get_layout().format, input_data_format);
Expand Down Expand Up @@ -10632,14 +10630,9 @@ TEST_P(conv_dyn_test, convolution_gpu_bfyx_os_iyx_osv16_no_bias) {
config.set_property(ov::intel_gpu::force_implementations(ov::intel_gpu::ImplForcingMap{ { "conv", conv_impl_ref } }));

auto output_memory_ref = calculate_ref(input, weights, config);
auto output_ref_lockable = get_test_engine().allocate_memory(output_memory_ref->get_layout());
output_memory_ref->copy_to(get_test_stream(), *output_ref_lockable, true);
cldnn::mem_lock<float> output_ptr_ref(output_ref_lockable, get_test_stream());

auto output_lockable = get_test_engine().allocate_memory(output_memory->get_layout());
output_memory->copy_to(get_test_stream(), *output_lockable, true);
cldnn::mem_lock<float> output_ptr(output_lockable, get_test_stream());

cldnn::mem_lock<float, mem_lock_type::read> output_ptr_ref(output_memory_ref, get_test_stream());
cldnn::mem_lock<float, mem_lock_type::read> output_ptr(output_memory, get_test_stream());

ASSERT_EQ(outputs.at("conv").get_layout(), output_memory_ref->get_layout());
for (size_t i = 0; i < output_ptr.size(); i++) {
Expand All @@ -10665,13 +10658,8 @@ TEST_P(conv_dyn_test, convolution_gpu_bfyx_os_iyx_osv16_no_bias) {

auto output_memory = outputs.at("conv").get_memory();
auto output_memory_ref = calculate_ref(input, weights, config);
auto output_ref_lockable = get_test_engine().allocate_memory(output_memory_ref->get_layout());
output_memory_ref->copy_to(get_test_stream(), *output_ref_lockable, true);
cldnn::mem_lock<float> output_ptr_ref(output_ref_lockable, get_test_stream());

auto output_lockable = get_test_engine().allocate_memory(output_memory->get_layout());
output_memory->copy_to(get_test_stream(), *output_lockable, true);
cldnn::mem_lock<float> output_ptr(output_lockable, get_test_stream());
cldnn::mem_lock<float, mem_lock_type::read> output_ptr_ref(output_memory_ref, get_test_stream());
cldnn::mem_lock<float, mem_lock_type::read> output_ptr(output_memory, get_test_stream());

ASSERT_EQ(outputs.at("conv").get_layout(), output_memory_ref->get_layout());
for (size_t i = 0; i < output_ptr.size(); i++) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2701,12 +2701,10 @@ class deconvolution_random_test_base {
}

auto out_mem = result.at("deconv").get_memory();
auto output_lockable = get_test_engine().allocate_memory(out_mem->get_layout());
out_mem->copy_to(get_test_stream(), *output_lockable, true);

// Compare results
{
cldnn::mem_lock<OutputT> ptr(output_lockable, get_test_stream());
cldnn::mem_lock<OutputT, mem_lock_type::read> ptr(out_mem, get_test_stream());

auto b = static_cast<size_t>(out_mem->get_layout().batch());
auto of = static_cast<size_t>(out_mem->get_layout().feature());
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -155,12 +155,8 @@ class dynamic_quantization_gpu_tests: public ::testing::Test {
std::cout << "Outputs number: " << ref_output_buffers.size() << "\n";

for (size_t i = 0; i < ref_output_buffers.size(); i++) {
auto output_lockable = get_test_engine().allocate_memory(output_buffers[i]->get_layout());
output_buffers[i]->copy_to(get_test_stream(), *output_lockable, true);
cldnn::mem_lock<ov::float16> output_ptr(output_lockable, get_test_stream());
auto output_lockable_ref = get_test_engine().allocate_memory(ref_output_buffers[i]->get_layout());
ref_output_buffers[i]->copy_to(get_test_stream(), *output_lockable_ref, true);
cldnn::mem_lock<ov::float16> output_ptr_ref(output_lockable_ref, get_test_stream());
cldnn::mem_lock<ov::float16, mem_lock_type::read> output_ptr(output_buffers[i], get_test_stream());
cldnn::mem_lock<ov::float16, mem_lock_type::read> output_ptr_ref(ref_output_buffers[i], get_test_stream());

for (size_t i = 0; i < output_ptr_ref.size(); ++i) {
auto abs_diff = std::abs(output_ptr_ref[i] - output_ptr[i]);
Expand Down
12 changes: 3 additions & 9 deletions src/plugins/intel_gpu/tests/unit/test_cases/eltwise_gpu_test.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -3814,9 +3814,7 @@ TEST(eltwise_gpu_f16, bfyx_and_fs_b_yx_fsv32_output_padding) {

auto golden_outputs = golden_network.execute();
auto golden_output = golden_outputs.at("eltwise").get_memory();
auto output_lockable_golden = engine.allocate_memory(golden_output->get_layout());
golden_output->copy_to(get_test_stream(), *output_lockable_golden, true);
cldnn::mem_lock<ov::float16> golden_ptr(output_lockable_golden, get_test_stream());
cldnn::mem_lock<ov::float16, mem_lock_type::read> golden_ptr(golden_output, get_test_stream());
// GOLDEN BFYX ELTWISE - END
// MIXED INPUT, FS_B_YX_FSV32 OUTPUT
topology FS_B_YX_FSV32_OUTPUT_topology;
Expand All @@ -3836,9 +3834,7 @@ TEST(eltwise_gpu_f16, bfyx_and_fs_b_yx_fsv32_output_padding) {

auto FS_B_YX_FSV32_OUTPUT_outputs = FS_B_YX_FSV32_OUTPUT_network.execute();
auto FS_B_YX_FSV32_OUTPUT_output = FS_B_YX_FSV32_OUTPUT_outputs.at("reorderOutput").get_memory();
auto output_lockable_fsv32 = engine.allocate_memory(FS_B_YX_FSV32_OUTPUT_output->get_layout());
FS_B_YX_FSV32_OUTPUT_output->copy_to(get_test_stream(), *output_lockable_fsv32, true);
cldnn::mem_lock<ov::float16> FS_B_YX_FSV32_OUTPUT_ptr(output_lockable_fsv32, get_test_stream());
cldnn::mem_lock<ov::float16, mem_lock_type::read> FS_B_YX_FSV32_OUTPUT_ptr(FS_B_YX_FSV32_OUTPUT_output, get_test_stream());
// MIXED INPUT, FS_B_YX_FSV32 OUTPUT - END
// MIXED INPUT, BYXF OUTPUT
topology BYXF_OUTPUT_topology;
Expand All @@ -3858,9 +3854,7 @@ TEST(eltwise_gpu_f16, bfyx_and_fs_b_yx_fsv32_output_padding) {

auto BYXF_OUTPUT_outputs = BYXF_OUTPUT_network.execute();
auto BYXF_OUTPUT_output = BYXF_OUTPUT_outputs.at("reorderOutput").get_memory();
auto output_lockable_byxf = engine.allocate_memory(BYXF_OUTPUT_output->get_layout());
BYXF_OUTPUT_output->copy_to(get_test_stream(), *output_lockable_byxf, true);
cldnn::mem_lock<ov::float16> BYXF_OUTPUT_ptr(output_lockable_byxf, get_test_stream());
cldnn::mem_lock<ov::float16, mem_lock_type::read> BYXF_OUTPUT_ptr(BYXF_OUTPUT_output, get_test_stream());
// MIXED INPUT, BYXF OUTPUT - END

ASSERT_EQ(golden_ptr.size(), FS_B_YX_FSV32_OUTPUT_ptr.size());
Expand Down
12 changes: 3 additions & 9 deletions src/plugins/intel_gpu/tests/unit/test_cases/pooling_gpu_test.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -231,9 +231,7 @@ TEST(pooling_forward_gpu, basic_max_byxf_f32_wsiz3x3_wstr1x1_i1x3x3x8_nopad) {

auto output_prim = outputs.begin()->second.get_memory();

auto output_lockable = engine.allocate_memory(output_prim->get_layout());
output_prim->copy_to(get_test_stream(), *output_lockable, true);
cldnn::mem_lock<float> output_ptr (output_lockable, get_test_stream());
cldnn::mem_lock<float, mem_lock_type::read> output_ptr (output_prim, get_test_stream());
ASSERT_EQ(4.0f, output_ptr[3]);
}

Expand Down Expand Up @@ -1456,9 +1454,7 @@ TEST(pooling_forward_gpu, b_fs_yx_fsv4)
auto searchC = outputs.find("pool_GOLD");
ASSERT_FALSE(searchC == outputs.end());
auto output = outputs.begin()->second.get_memory();
auto output_lockable = engine.allocate_memory(output->get_layout());
output->copy_to(get_test_stream(), *output_lockable, true);
cldnn::mem_lock<char> output_ptr(output_lockable, get_test_stream());
cldnn::mem_lock<char, mem_lock_type::read> output_ptr(output, get_test_stream());
vGoldOutput.reserve(output_ptr.size());
for (size_t i = 0; i < output_ptr.size(); i++)
vGoldOutput.push_back(output_ptr[i]);
Expand Down Expand Up @@ -1507,9 +1503,7 @@ TEST(pooling_forward_gpu, b_fs_yx_fsv4)
auto searchC = outputs.find("reorder_UnSwizzelled");
ASSERT_FALSE(searchC == outputs.end());
auto output = outputs.begin()->second.get_memory();
auto output_lockable = engine.allocate_memory(output->get_layout());
output->copy_to(get_test_stream(), *output_lockable, true);
cldnn::mem_lock<char> output_ptr(output_lockable, get_test_stream());
cldnn::mem_lock<char, mem_lock_type::read> output_ptr(output, get_test_stream());
vTestOutput.reserve(output_ptr.size());
for (size_t i = 0; i < output_ptr.size(); i++)
vTestOutput.push_back(output_ptr[i]);
Expand Down
12 changes: 3 additions & 9 deletions src/plugins/intel_gpu/tests/unit/test_cases/reorder_gpu_test.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -2514,9 +2514,7 @@ TEST(reorder_gpu_f32, bfzyx_to_bfyx_padded) {
auto output1 = outputs.at("reshape1").get_memory();
auto output2 = outputs.at("reshape2").get_memory();

auto output0_lockable = engine.allocate_memory(output0->get_layout());
output0->copy_to(get_test_stream(), *output0_lockable, true);
cldnn::mem_lock<float> output_ptr0(output0_lockable, get_test_stream());
cldnn::mem_lock<float, mem_lock_type::read> output_ptr0(output0, get_test_stream());
for (int b = 0; b < b_crop; ++b) {
for (int f = 0; f < f_crop; ++f) {
for (int z = 0; z < z_crop; ++z) {
Expand All @@ -2531,9 +2529,7 @@ TEST(reorder_gpu_f32, bfzyx_to_bfyx_padded) {
}
}

auto output1_lockable = engine.allocate_memory(output1->get_layout());
output1->copy_to(get_test_stream(), *output1_lockable, true);
cldnn::mem_lock<float> output_ptr1(output1_lockable, get_test_stream());
cldnn::mem_lock<float, mem_lock_type::read> output_ptr1(output1, get_test_stream());
for (int b = 0; b < b_crop; ++b) {
for (int f = 0; f < f_crop; ++f) {
for (int z = 0; z < z_crop; ++z) {
Expand All @@ -2548,9 +2544,7 @@ TEST(reorder_gpu_f32, bfzyx_to_bfyx_padded) {
}
}

auto output2_lockable = engine.allocate_memory(output2->get_layout());
output2->copy_to(get_test_stream(), *output2_lockable, true);
cldnn::mem_lock<float> output_ptr2(output2_lockable, get_test_stream());
cldnn::mem_lock<float, mem_lock_type::read> output_ptr2(output2, get_test_stream());
for (int b = 0; b < b_crop; ++b) {
for (int f = 0; f < f_crop; ++f) {
for (int z = 0; z < z_crop; ++z) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -97,9 +97,7 @@ void generic_reshape_test(format fmt, tensor const& input_size, tensor const& re
reshape_size.add(tensor(default_fmt, lower_sizes, 0)).add(tensor(default_fmt, upper_sizes, 0)));

{
auto output_lockable = engine.allocate_memory(output->get_layout());
output->copy_to(get_test_stream(), *output_lockable, true);
cldnn::mem_lock<const ElemType> output_ptr(output_lockable, get_test_stream());
cldnn::mem_lock<const ElemType, mem_lock_type::read> output_ptr(output, get_test_stream());
auto output_itr = output_ptr.begin();

auto sizes = reshape_size.sizes(fmt);
Expand Down

0 comments on commit 0855be7

Please sign in to comment.