aosp12/bionic/tests/leak_test.cpp

163 lines
4.1 KiB
C++

/*
* Copyright (C) 2017 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <err.h>
#include <inttypes.h>
#include <pthread.h>
#include <sched.h>
#include <stdio.h>
#include <string.h>
#include <sys/mman.h>
#include <sys/user.h>
#include <unistd.h>
#include <gtest/gtest.h>
#include <chrono>
#include <thread>
#include <vector>
#include <android-base/macros.h>
#include <android-base/threads.h>
#include "utils.h"
using namespace std::chrono_literals;
static void WaitUntilAllThreadsExited(pid_t* tids, size_t tid_count) {
// Wait until all children have exited.
bool alive = true;
while (alive) {
alive = false;
for (size_t i = 0; i < tid_count; ++i) {
if (tids[i] != 0) {
if (tgkill(getpid(), tids[i], 0) == 0) {
alive = true;
} else {
EXPECT_EQ(errno, ESRCH);
tids[i] = 0; // Skip in next loop.
}
}
}
sched_yield();
}
}
class LeakChecker {
public:
LeakChecker() {
// Avoid resizing and using memory later.
// 64Ki is the default limit on VMAs per process.
maps_.reserve(64*1024);
Reset();
}
~LeakChecker() {
Check();
}
void Reset() {
previous_size_ = GetMappingSize();
}
void DumpTo(std::ostream& os) const {
os << previous_size_;
}
private:
size_t previous_size_;
std::vector<map_record> maps_;
void Check() {
auto current_size = GetMappingSize();
if (current_size > previous_size_) {
FAIL() << "increase in process map size: " << previous_size_ << " -> " << current_size;
}
}
size_t GetMappingSize() {
if (!Maps::parse_maps(&maps_)) {
err(1, "failed to parse maps");
}
size_t result = 0;
for (const map_record& map : maps_) {
result += map.addr_end - map.addr_start;
}
return result;
}
};
std::ostream& operator<<(std::ostream& os, const LeakChecker& lc) {
lc.DumpTo(os);
return os;
}
// http://b/36045112
TEST(pthread_leak, join) {
SKIP_WITH_NATIVE_BRIDGE; // http://b/37920774
LeakChecker lc;
for (int i = 0; i < 100; ++i) {
pthread_t thread;
ASSERT_EQ(0, pthread_create(&thread, nullptr, [](void*) -> void* { return nullptr; }, nullptr));
ASSERT_EQ(0, pthread_join(thread, nullptr));
}
}
// http://b/36045112
TEST(pthread_leak, detach) {
SKIP_WITH_NATIVE_BRIDGE; // http://b/37920774
LeakChecker lc;
// Ancient devices with only 2 cores need a lower limit.
// http://b/129924384 and https://issuetracker.google.com/142210680.
const int thread_count = (sysconf(_SC_NPROCESSORS_CONF) > 2) ? 100 : 50;
for (size_t pass = 0; pass < 1; ++pass) {
struct thread_data { pthread_barrier_t* barrier; pid_t* tid; } threads[thread_count];
pthread_barrier_t barrier;
ASSERT_EQ(pthread_barrier_init(&barrier, nullptr, thread_count + 1), 0);
// Start child threads.
pid_t tids[thread_count];
for (int i = 0; i < thread_count; ++i) {
threads[i] = {&barrier, &tids[i]};
const auto thread_function = +[](void* ptr) -> void* {
thread_data* data = static_cast<thread_data*>(ptr);
*data->tid = gettid();
pthread_barrier_wait(data->barrier);
return nullptr;
};
pthread_t thread;
ASSERT_EQ(0, pthread_create(&thread, nullptr, thread_function, &threads[i]));
ASSERT_EQ(0, pthread_detach(thread));
}
pthread_barrier_wait(&barrier);
ASSERT_EQ(pthread_barrier_destroy(&barrier), 0);
WaitUntilAllThreadsExited(tids, thread_count);
// TODO(b/158573595): the test is flaky without the warmup pass.
if (pass == 0) lc.Reset();
}
}