Magisk/native/jni/magiskhide/proc_monitor.cpp

410 lines
9.8 KiB
C++
Raw Normal View History

2017-04-06 00:12:29 +02:00
#include <stdlib.h>
#include <stdio.h>
#include <string.h>
#include <unistd.h>
2018-07-13 16:14:32 +02:00
#include <fcntl.h>
2017-04-06 00:12:29 +02:00
#include <signal.h>
#include <pthread.h>
#include <sys/ptrace.h>
#include <sys/inotify.h>
2017-04-06 00:12:29 +02:00
#include <sys/types.h>
#include <sys/wait.h>
2017-07-10 17:39:33 +02:00
#include <sys/mount.h>
#include <vector>
2019-04-13 08:43:43 +02:00
#include <bitset>
2017-04-06 00:12:29 +02:00
2020-03-09 09:50:30 +01:00
#include <logging.hpp>
#include <utils.hpp>
2020-03-09 09:50:30 +01:00
#include "magiskhide.hpp"
2016-12-30 19:44:24 +01:00
2019-01-20 05:59:37 +01:00
using namespace std;
static int inotify_fd = -1;
static void term_thread(int sig = SIGTERMTHRD);
static void new_zygote(int pid);
/**********************
* All data structures
**********************/
set<pair<string, string>> hide_set; /* set of <pkg, process> pair */
static map<int, struct stat> zygote_map; /* zygote pid -> mnt ns */
static map<int, vector<string_view>> uid_proc_map; /* uid -> list of process */
pthread_mutex_t monitor_lock;
#define PID_MAX 32768
2019-04-13 08:43:43 +02:00
static bitset<PID_MAX> attaches; /* true if pid is monitored */
/********
* Utils
********/
2017-01-01 11:54:13 +01:00
static inline int read_ns(const int pid, struct stat *st) {
2017-07-02 19:02:11 +02:00
char path[32];
sprintf(path, "/proc/%d/ns/mnt", pid);
2018-07-11 17:41:38 +02:00
return stat(path, st);
2017-07-02 19:02:11 +02:00
}
2019-03-09 05:53:53 +01:00
static int parse_ppid(int pid) {
char path[32];
int ppid;
sprintf(path, "/proc/%d/stat", pid);
FILE *stat = fopen(path, "re");
if (stat == nullptr)
return -1;
/* PID COMM STATE PPID ..... */
fscanf(stat, "%*d %*s %*c %d", &ppid);
fclose(stat);
return ppid;
}
2019-05-26 03:17:25 +02:00
static inline long xptrace(int request, pid_t pid, void *addr, void *data) {
long ret = ptrace(request, pid, addr, data);
2019-05-26 03:17:25 +02:00
if (ret < 0)
PLOGE("ptrace %d", pid);
return ret;
}
2019-05-26 03:17:25 +02:00
static inline long xptrace(int request, pid_t pid, void *addr = nullptr, intptr_t data = 0) {
return xptrace(request, pid, addr, reinterpret_cast<void *>(data));
}
void update_uid_map() {
2019-09-26 05:55:39 +02:00
mutex_guard lock(monitor_lock);
uid_proc_map.clear();
2019-06-04 08:32:49 +02:00
string data_path(APP_DATA_DIR);
data_path += "/0/";
size_t len = data_path.length();
struct stat st;
for (auto &hide : hide_set) {
data_path.erase(data_path.begin() + len, data_path.end());
data_path += hide.first;
if (stat(data_path.data(), &st))
continue;
uid_proc_map[st.st_uid].emplace_back(hide.second);
}
}
static void check_zygote() {
2019-05-28 01:27:19 +02:00
crawl_procfs([](int pid) -> bool {
char buf[512];
snprintf(buf, sizeof(buf), "/proc/%d/cmdline", pid);
if (FILE *f = fopen(buf, "re")) {
2019-05-28 01:27:19 +02:00
fgets(buf, sizeof(buf), f);
if (strncmp(buf, "zygote", 6) == 0 && parse_ppid(pid) == 1)
new_zygote(pid);
fclose(f);
}
return true;
});
}
#define APP_PROC "/system/bin/app_process"
static void setup_inotify() {
inotify_fd = xinotify_init1(IN_CLOEXEC);
if (inotify_fd < 0)
term_thread();
// Setup inotify asynchronous I/O
fcntl(inotify_fd, F_SETFL, O_ASYNC);
struct f_owner_ex ex = {
2019-05-26 11:35:12 +02:00
.type = F_OWNER_TID,
.pid = gettid()
};
fcntl(inotify_fd, F_SETOWN_EX, &ex);
// Monitor packages.xml
inotify_add_watch(inotify_fd, "/data/system", IN_CLOSE_WRITE);
// Monitor app_process
if (access(APP_PROC "32", F_OK) == 0) {
inotify_add_watch(inotify_fd, APP_PROC "32", IN_ACCESS);
if (access(APP_PROC "64", F_OK) == 0)
inotify_add_watch(inotify_fd, APP_PROC "64", IN_ACCESS);
} else {
inotify_add_watch(inotify_fd, APP_PROC, IN_ACCESS);
}
}
/************************
* Async signal handlers
************************/
static void inotify_event(int) {
/* Make sure we can actually read stuffs
* or else the whole thread will be blocked.*/
struct pollfd pfd = {
.fd = inotify_fd,
.events = POLLIN,
.revents = 0
};
if (poll(&pfd, 1, 0) <= 0)
return; // Nothing to read
char buf[512];
auto event = reinterpret_cast<struct inotify_event *>(buf);
read(inotify_fd, buf, sizeof(buf));
2019-05-28 01:27:19 +02:00
if ((event->mask & IN_CLOSE_WRITE) && event->name == "packages.xml"sv)
update_uid_map();
check_zygote();
2019-02-14 06:52:59 +01:00
}
static void check_zygote(int) {
check_zygote();
}
// Workaround for the lack of pthread_cancel
static void term_thread(int) {
LOGD("proc_monitor: cleaning up\n");
uid_proc_map.clear();
zygote_map.clear();
hide_set.clear();
2019-04-13 08:43:43 +02:00
attaches.reset();
// Misc
set_hide_state(false);
pthread_mutex_destroy(&monitor_lock);
close(inotify_fd);
inotify_fd = -1;
LOGD("proc_monitor: terminate\n");
pthread_exit(nullptr);
}
/******************
* Ptrace Madness
******************/
/* Ptrace is super tricky, preserve all excessive logging in code
* but disable when actually building for usage (you won't want
* your logcat spammed with new thread events from all apps) */
2019-05-26 06:42:24 +02:00
//#define PTRACE_LOG(fmt, args...) LOGD("PID=[%d] " fmt, pid, ##args)
#define PTRACE_LOG(...)
static void detach_pid(int pid, int signal = 0) {
2019-05-26 11:35:12 +02:00
attaches[pid] = false;
2019-06-27 09:28:34 +02:00
ptrace(PTRACE_DETACH, pid, 0, signal);
PTRACE_LOG("detach\n");
}
static bool check_pid(int pid) {
char path[128];
char cmdline[1024];
2019-06-27 09:28:34 +02:00
struct stat st;
sprintf(path, "/proc/%d", pid);
if (stat(path, &st)) {
// Process killed unexpectedly, ignore
detach_pid(pid);
return true;
}
// UID hasn't changed
if (st.st_uid == 0)
return false;
sprintf(path, "/proc/%d/cmdline", pid);
2019-06-27 09:28:34 +02:00
if (FILE *f; (f = fopen(path, "re"))) {
fgets(cmdline, sizeof(cmdline), f);
fclose(f);
} else {
// Process killed unexpectedly, ignore
detach_pid(pid);
return true;
}
2019-09-17 07:50:38 +02:00
if (cmdline == "zygote"sv || cmdline == "zygote32"sv || cmdline == "zygote64"sv ||
cmdline == "usap32"sv || cmdline == "usap64"sv)
return false;
int uid = st.st_uid % 100000;
auto it = uid_proc_map.find(uid);
if (it != uid_proc_map.end()) {
for (auto &s : it->second) {
if (s == cmdline) {
// Double check whether ns is separated
read_ns(pid, &st);
bool mnt_ns = true;
for (auto &zit : zygote_map) {
if (zit.second.st_ino == st.st_ino &&
zit.second.st_dev == st.st_dev) {
mnt_ns = false;
break;
}
}
// For some reason ns is not separated, abort
if (!mnt_ns)
break;
/* Finally this is our target!
* Detach from ptrace but should still remain stopped.
* The hide daemon will resume the process. */
PTRACE_LOG("target found\n");
LOGI("proc_monitor: [%s] PID=[%d] UID=[%d]\n", cmdline, pid, uid);
detach_pid(pid, SIGSTOP);
2020-05-18 00:01:20 +02:00
hide_daemon(pid);
return true;
}
}
}
PTRACE_LOG("[%s] not our target\n", cmdline);
detach_pid(pid);
return true;
}
2019-06-27 09:28:34 +02:00
static bool is_process(int pid) {
char buf[128];
char key[32];
int tgid;
sprintf(buf, "/proc/%d/status", pid);
unique_ptr<FILE, decltype(&fclose)> fp(fopen(buf, "re"), &fclose);
// PID is dead
if (!fp)
return false;
while (fgets(buf, sizeof(buf), fp.get())) {
sscanf(buf, "%s", key);
if (key == "Tgid:"sv) {
sscanf(buf, "%*s %d", &tgid);
return tgid == pid;
}
}
return false;
}
static void new_zygote(int pid) {
struct stat st;
if (read_ns(pid, &st))
return;
2019-05-26 03:30:43 +02:00
auto it = zygote_map.find(pid);
if (it != zygote_map.end()) {
// Update namespace info
it->second = st;
return;
}
LOGD("proc_monitor: ptrace zygote PID=[%d]\n", pid);
zygote_map[pid] = st;
xptrace(PTRACE_ATTACH, pid);
waitpid(pid, nullptr, __WALL | __WNOTHREAD);
xptrace(PTRACE_SETOPTIONS, pid, nullptr,
PTRACE_O_TRACEFORK | PTRACE_O_TRACEVFORK | PTRACE_O_TRACEEXIT);
xptrace(PTRACE_CONT, pid);
}
#define WEVENT(s) (((s) & 0xffff0000) >> 16)
2019-03-10 07:14:41 +01:00
#define DETACH_AND_CONT { detach = true; continue; }
void proc_monitor() {
// Unblock some signals
sigset_t block_set;
sigemptyset(&block_set);
sigaddset(&block_set, SIGTERMTHRD);
sigaddset(&block_set, SIGIO);
2019-01-20 05:59:37 +01:00
pthread_sigmask(SIG_UNBLOCK, &block_set, nullptr);
2019-01-20 05:59:37 +01:00
struct sigaction act{};
act.sa_handler = term_thread;
sigaction(SIGTERMTHRD, &act, nullptr);
act.sa_handler = inotify_event;
sigaction(SIGIO, &act, nullptr);
act.sa_handler = check_zygote;
sigaction(SIGZYGOTE, &act, nullptr);
setup_inotify();
2017-05-07 21:11:14 +02:00
// First find existing zygotes
check_zygote();
int status;
for (;;) {
2019-03-10 07:14:41 +01:00
const int pid = waitpid(-1, &status, __WALL | __WNOTHREAD);
2019-05-26 03:17:25 +02:00
if (pid < 0) {
if (errno == ECHILD) {
/* This mean we have nothing to wait, sleep
* and wait till signal interruption */
LOGD("proc_monitor: nothing to monitor, wait for signal\n");
2019-05-26 11:35:12 +02:00
struct timespec ts = {
2019-05-26 03:17:25 +02:00
.tv_sec = INT_MAX,
.tv_nsec = 0
};
2019-05-26 11:35:12 +02:00
nanosleep(&ts, nullptr);
2019-05-26 03:17:25 +02:00
}
2019-02-16 08:49:36 +01:00
continue;
2019-05-26 03:17:25 +02:00
}
2019-03-10 07:14:41 +01:00
bool detach = false;
2019-09-26 05:55:39 +02:00
run_finally f([&] {
2019-06-27 09:28:34 +02:00
if (detach)
2019-03-10 07:14:41 +01:00
// Non of our business now
2019-06-27 09:28:34 +02:00
detach_pid(pid);
2019-03-10 07:14:41 +01:00
});
2019-05-26 11:35:12 +02:00
2019-06-27 09:28:34 +02:00
if (!WIFSTOPPED(status) /* Ignore if not ptrace-stop */)
2019-03-11 17:44:50 +01:00
DETACH_AND_CONT;
2019-05-26 11:35:12 +02:00
2019-03-11 17:44:50 +01:00
if (WSTOPSIG(status) == SIGTRAP && WEVENT(status)) {
unsigned long msg;
xptrace(PTRACE_GETEVENTMSG, pid, nullptr, &msg);
if (zygote_map.count(pid)) {
// Zygote event
switch (WEVENT(status)) {
case PTRACE_EVENT_FORK:
case PTRACE_EVENT_VFORK:
PTRACE_LOG("zygote forked: [%d]\n", msg);
attaches[msg] = true;
break;
case PTRACE_EVENT_EXIT:
PTRACE_LOG("zygote exited with status: [%d]\n", msg);
2019-05-26 11:35:12 +02:00
[[fallthrough]];
default:
2019-03-11 17:44:50 +01:00
zygote_map.erase(pid);
DETACH_AND_CONT;
}
2019-03-11 17:44:50 +01:00
} else {
switch (WEVENT(status)) {
case PTRACE_EVENT_CLONE:
PTRACE_LOG("create new threads: [%d]\n", msg);
if (attaches[pid] && check_pid(pid))
continue;
break;
case PTRACE_EVENT_EXEC:
case PTRACE_EVENT_EXIT:
2019-05-26 11:35:12 +02:00
PTRACE_LOG("exit or execve\n");
[[fallthrough]];
2019-03-11 17:44:50 +01:00
default:
2019-05-26 11:35:12 +02:00
DETACH_AND_CONT;
}
2019-03-11 17:44:50 +01:00
}
2019-05-26 11:35:12 +02:00
xptrace(PTRACE_CONT, pid);
2019-03-11 17:44:50 +01:00
} else if (WSTOPSIG(status) == SIGSTOP) {
2019-06-27 09:28:34 +02:00
if (!attaches[pid]) {
// Double check if this is actually a process
attaches[pid] = is_process(pid);
}
if (attaches[pid]) {
// This is a process, continue monitoring
PTRACE_LOG("SIGSTOP from child\n");
xptrace(PTRACE_SETOPTIONS, pid, nullptr,
PTRACE_O_TRACECLONE | PTRACE_O_TRACEEXEC | PTRACE_O_TRACEEXIT);
xptrace(PTRACE_CONT, pid);
} else {
// This is a thread, do NOT monitor
PTRACE_LOG("SIGSTOP from thread\n");
DETACH_AND_CONT;
}
} else {
2019-03-11 17:44:50 +01:00
// Not caused by us, resend signal
xptrace(PTRACE_CONT, pid, nullptr, WSTOPSIG(status));
PTRACE_LOG("signal [%d]\n", WSTOPSIG(status));
}
2017-06-02 22:31:01 +02:00
}
}