1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
|
#if GOOGLE_CUDA
#include "tensorflow/core/common_runtime/gpu/gpu_region_allocator.h"
#include <algorithm>
#include <vector>
#include "tensorflow/core/platform/port.h"
#include "tensorflow/core/lib/gtl/inlined_vector.h"
#include "tensorflow/core/platform/logging.h"
#include "tensorflow/core/common_runtime/gpu/gpu_init.h"
#include "tensorflow/stream_executor/stream_executor.h"
#include <gtest/gtest.h>
namespace gpu = ::perftools::gputools;
namespace tensorflow {
namespace {
TEST(GPURegionAllocatorTest, Simple) {
GPURegionAllocator a(0, 1 << 26);
std::vector<void*> ptrs;
for (int s = 1; s < 1024; s++) {
void* raw = a.AllocateRaw(1, s);
ptrs.push_back(raw);
}
std::sort(ptrs.begin(), ptrs.end());
for (int i = 0; i < ptrs.size(); i++) {
if (i > 0) {
CHECK_NE(ptrs[i], ptrs[i - 1]); // No dups
}
a.DeallocateRaw(ptrs[i]);
}
float* t1 = a.Allocate<float>(1024);
double* t2 = a.Allocate<double>(1048576);
a.Deallocate(t1);
a.Deallocate(t2);
}
TEST(GPURegionAllocatorTest, CheckMemLeak) {
EXPECT_DEATH(
{
GPURegionAllocator a(0, 1 << 26);
float* t1 = a.Allocate<float>(1024);
if (t1) {
LOG(INFO) << "Not deallocating";
}
},
"");
}
TEST(GPURegionAllocatorTest, TracksSizes) {
GPURegionAllocator a(0, 1 << 26);
EXPECT_EQ(true, a.TracksAllocationSizes());
}
TEST(GPURegionAllocatorTest, AllocatedVsRequested) {
GPURegionAllocator a(0, 1 << 26);
float* t1 = a.Allocate<float>(1);
EXPECT_EQ(sizeof(float), a.RequestedSize(t1));
// Minimum allocation size if 256
EXPECT_EQ(256, a.AllocatedSize(t1));
a.Deallocate(t1);
}
} // namespace
} // namespace tensorflow
#endif // GOOGLE_CUDA
|