/*-*- Mode: C; c-basic-offset: 8; indent-tabs-mode: nil -*-*/
/***
This file is part of systemd.
Copyright 2010 Lennart Poettering
systemd is free software; you can redistribute it and/or modify it
under the terms of the GNU Lesser General Public License as published by
the Free Software Foundation; either version 2.1 of the License, or
(at your option) any later version.
systemd is distributed in the hope that it will be useful, but
WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public License
along with systemd; If not, see .
***/
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#ifdef HAVE_FANOTIFY_INIT
#include
#endif
#include
#include "missing.h"
#include "util.h"
#include "set.h"
#include "ioprio.h"
#include "readahead-common.h"
#include "virt.h"
/* fixme:
*
* - detect ssd on btrfs/lvm...
* - read ahead directories
* - gzip?
* - remount rw?
* - handle files where nothing is in mincore
* - does ioprio_set work with fadvise()?
*/
static ReadaheadShared *shared = NULL;
static usec_t starttime;
/* Avoid collisions with the NULL pointer */
#define SECTOR_TO_PTR(s) ULONG_TO_PTR((s)+1)
#define PTR_TO_SECTOR(p) (PTR_TO_ULONG(p)-1)
static int btrfs_defrag(int fd) {
struct btrfs_ioctl_vol_args data = { .fd = fd };
return ioctl(fd, BTRFS_IOC_DEFRAG, &data);
}
static int pack_file(FILE *pack, const char *fn, bool on_btrfs) {
struct stat st;
void *start = MAP_FAILED;
uint8_t *vec;
uint32_t b, c;
uint64_t inode;
size_t l, pages;
bool mapped;
int r = 0, fd = -1, k;
assert(pack);
assert(fn);
fd = open(fn, O_RDONLY|O_CLOEXEC|O_NOATIME|O_NOCTTY|O_NOFOLLOW);
if (fd < 0) {
if (errno == ENOENT)
return 0;
if (errno == EPERM || errno == EACCES)
return 0;
log_warning("open(%s) failed: %m", fn);
r = -errno;
goto finish;
}
k = file_verify(fd, fn, arg_file_size_max, &st);
if (k <= 0) {
r = k;
goto finish;
}
if (on_btrfs)
btrfs_defrag(fd);
l = PAGE_ALIGN(st.st_size);
start = mmap(NULL, l, PROT_READ, MAP_SHARED, fd, 0);
if (start == MAP_FAILED) {
log_warning("mmap(%s) failed: %m", fn);
r = -errno;
goto finish;
}
pages = l / page_size();
vec = alloca(pages);
memset(vec, 0, pages);
if (mincore(start, l, vec) < 0) {
log_warning("mincore(%s) failed: %m", fn);
r = -errno;
goto finish;
}
fputs(fn, pack);
fputc('\n', pack);
/* Store the inode, so that we notice when the file is deleted */
inode = (uint64_t) st.st_ino;
fwrite(&inode, sizeof(inode), 1, pack);
mapped = false;
for (c = 0; c < pages; c++) {
bool new_mapped = !!(vec[c] & 1);
if (!mapped && new_mapped)
b = c;
else if (mapped && !new_mapped) {
fwrite(&b, sizeof(b), 1, pack);
fwrite(&c, sizeof(c), 1, pack);
log_debug("%s: page %u to %u", fn, b, c);
}
mapped = new_mapped;
}
/* We don't write any range data if we should read the entire file */
if (mapped && b > 0) {
fwrite(&b, sizeof(b), 1, pack);
fwrite(&c, sizeof(c), 1, pack);
log_debug("%s: page %u to %u", fn, b, c);
}
/* End marker */
b = 0;
fwrite(&b, sizeof(b), 1, pack);
fwrite(&b, sizeof(b), 1, pack);
finish:
if (start != MAP_FAILED)
munmap(start, l);
if (fd >= 0)
close_nointr_nofail(fd);
return r;
}
static unsigned long fd_first_block(int fd) {
struct {
struct fiemap fiemap;
struct fiemap_extent extent;
} data = {
.fiemap.fm_length = ~0ULL,
.fiemap.fm_extent_count = 1,
};
if (ioctl(fd, FS_IOC_FIEMAP, &data) < 0)
return 0;
if (data.fiemap.fm_mapped_extents <= 0)
return 0;
if (data.fiemap.fm_extents[0].fe_flags & FIEMAP_EXTENT_UNKNOWN)
return 0;
return (unsigned long) data.fiemap.fm_extents[0].fe_physical;
}
struct item {
const char *path;
unsigned long block;
unsigned long bin;
};
static int qsort_compare(const void *a, const void *b) {
const struct item *i, *j;
i = a;
j = b;
/* sort by bin first */
if (i->bin < j->bin)
return -1;
if (i->bin > j->bin)
return 1;
/* then sort by sector */
if (i->block < j->block)
return -1;
if (i->block > j->block)
return 1;
return strcmp(i->path, j->path);
}
static int collect(const char *root) {
enum {
FD_FANOTIFY, /* Get the actual fs events */
FD_SIGNAL,
FD_INOTIFY, /* We get notifications to quit early via this fd */
_FD_MAX
};
struct pollfd pollfd[_FD_MAX] = {};
int fanotify_fd = -1, signal_fd = -1, inotify_fd = -1, r = 0;
pid_t my_pid;
Hashmap *files = NULL;
Iterator i;
char *p, *q;
sigset_t mask;
FILE *pack = NULL;
char *pack_fn_new = NULL, *pack_fn = NULL;
bool on_ssd, on_btrfs;
struct statfs sfs;
usec_t not_after;
uint64_t previous_block_readahead;
bool previous_block_readahead_set = false;
assert(root);
if (asprintf(&pack_fn, "%s/.readahead", root) < 0) {
r = log_oom();
goto finish;
}
starttime = now(CLOCK_MONOTONIC);
/* If there's no pack file yet we lower the kernel readahead
* so that mincore() is accurate. If there is a pack file
* already we assume it is accurate enough so that kernel
* readahead is never triggered. */
previous_block_readahead_set =
access(pack_fn, F_OK) < 0 &&
block_get_readahead(root, &previous_block_readahead) >= 0 &&
block_set_readahead(root, 8*1024) >= 0;
if (ioprio_set(IOPRIO_WHO_PROCESS, getpid(), IOPRIO_PRIO_VALUE(IOPRIO_CLASS_IDLE, 0)) < 0)
log_warning("Failed to set IDLE IO priority class: %m");
assert_se(sigemptyset(&mask) == 0);
sigset_add_many(&mask, SIGINT, SIGTERM, -1);
assert_se(sigprocmask(SIG_SETMASK, &mask, NULL) == 0);
if ((signal_fd = signalfd(-1, &mask, SFD_NONBLOCK|SFD_CLOEXEC)) < 0) {
log_error("signalfd(): %m");
r = -errno;
goto finish;
}
files = hashmap_new(string_hash_func, string_compare_func);
if (!files) {
log_error("Failed to allocate set.");
r = -ENOMEM;
goto finish;
}
fanotify_fd = fanotify_init(FAN_CLOEXEC|FAN_NONBLOCK, O_RDONLY|O_LARGEFILE|O_CLOEXEC|O_NOATIME);
if (fanotify_fd < 0) {
log_error("Failed to create fanotify object: %m");
r = -errno;
goto finish;
}
if (fanotify_mark(fanotify_fd, FAN_MARK_ADD|FAN_MARK_MOUNT, FAN_OPEN, AT_FDCWD, root) < 0) {
log_error("Failed to mark %s: %m", root);
r = -errno;
goto finish;
}
inotify_fd = open_inotify();
if (inotify_fd < 0) {
r = inotify_fd;
goto finish;
}
not_after = now(CLOCK_MONOTONIC) + arg_timeout;
my_pid = getpid();
pollfd[FD_FANOTIFY].fd = fanotify_fd;
pollfd[FD_FANOTIFY].events = POLLIN;
pollfd[FD_SIGNAL].fd = signal_fd;
pollfd[FD_SIGNAL].events = POLLIN;
pollfd[FD_INOTIFY].fd = inotify_fd;
pollfd[FD_INOTIFY].events = POLLIN;
sd_notify(0,
"READY=1\n"
"STATUS=Collecting readahead data");
log_debug("Collecting...");
if (access("/run/systemd/readahead/cancel", F_OK) >= 0) {
log_debug("Collection canceled");
r = -ECANCELED;
goto finish;
}
if (access("/run/systemd/readahead/done", F_OK) >= 0) {
log_debug("Got termination request");
goto done;
}
for (;;) {
union {
struct fanotify_event_metadata metadata;
char buffer[4096];
} data;
ssize_t n;
struct fanotify_event_metadata *m;
usec_t t;
int h;
if (hashmap_size(files) > arg_files_max) {
log_debug("Reached maximum number of read ahead files, ending collection.");
break;
}
t = now(CLOCK_MONOTONIC);
if (t >= not_after) {
log_debug("Reached maximum collection time, ending collection.");
break;
}
if ((h = poll(pollfd, _FD_MAX, (int) ((not_after - t) / USEC_PER_MSEC))) < 0) {
if (errno == EINTR)
continue;
log_error("poll(): %m");
r = -errno;
goto finish;
}
if (h == 0) {
log_debug("Reached maximum collection time, ending collection.");
break;
}
if (pollfd[FD_SIGNAL].revents) {
log_debug("Got signal.");
break;
}
if (pollfd[FD_INOTIFY].revents) {
uint8_t inotify_buffer[sizeof(struct inotify_event) + FILENAME_MAX];
struct inotify_event *e;
if ((n = read(inotify_fd, &inotify_buffer, sizeof(inotify_buffer))) < 0) {
if (errno == EINTR || errno == EAGAIN)
continue;
log_error("Failed to read inotify event: %m");
r = -errno;
goto finish;
}
e = (struct inotify_event*) inotify_buffer;
while (n > 0) {
size_t step;
if ((e->mask & IN_CREATE) && streq(e->name, "cancel")) {
log_debug("Collection canceled");
r = -ECANCELED;
goto finish;
}
if ((e->mask & IN_CREATE) && streq(e->name, "done")) {
log_debug("Got termination request");
goto done;
}
step = sizeof(struct inotify_event) + e->len;
assert(step <= (size_t) n);
e = (struct inotify_event*) ((uint8_t*) e + step);
n -= step;
}
}
if ((n = read(fanotify_fd, &data, sizeof(data))) < 0) {
if (errno == EINTR || errno == EAGAIN)
continue;
/* fanotify sometimes returns EACCES on read()
* where it shouldn't. For now let's just
* ignore it here (which is safe), but
* eventually this should be
* dropped when the kernel is fixed.
*
* https://bugzilla.redhat.com/show_bug.cgi?id=707577 */
if (errno == EACCES)
continue;
log_error("Failed to read event: %m");
r = -errno;
goto finish;
}
for (m = &data.metadata; FAN_EVENT_OK(m, n); m = FAN_EVENT_NEXT(m, n)) {
char fn[PATH_MAX];
int k;
if (m->fd < 0)
goto next_iteration;
if (m->pid == my_pid)
goto next_iteration;
__sync_synchronize();
if (m->pid == shared->replay)
goto next_iteration;
snprintf(fn, sizeof(fn), "/proc/self/fd/%i", m->fd);
char_array_0(fn);
if ((k = readlink_malloc(fn, &p)) >= 0) {
if (startswith(p, "/tmp") ||
endswith(p, " (deleted)") ||
hashmap_get(files, p))
/* Not interesting, or
* already read */
free(p);
else {
unsigned long ul;
usec_t entrytime;
struct item *entry;
entry = new0(struct item, 1);
if (!entry) {
r = log_oom();
goto finish;
}
ul = fd_first_block(m->fd);
entrytime = now(CLOCK_MONOTONIC);
entry->block = ul;
entry->path = strdup(p);
if (!entry->path) {
free(entry);
r = log_oom();
goto finish;
}
entry->bin = (entrytime - starttime) / 2000000;
if ((k = hashmap_put(files, p, entry)) < 0) {
log_warning("set_put() failed: %s", strerror(-k));
free(p);
}
}
} else
log_warning("readlink(%s) failed: %s", fn, strerror(-k));
next_iteration:
if (m->fd >= 0)
close_nointr_nofail(m->fd);
}
}
done:
if (fanotify_fd >= 0) {
close_nointr_nofail(fanotify_fd);
fanotify_fd = -1;
}
log_debug("Writing Pack File...");
on_ssd = fs_on_ssd(root) > 0;
log_debug("On SSD: %s", yes_no(on_ssd));
on_btrfs = statfs(root, &sfs) >= 0 && (unsigned) sfs.f_type == BTRFS_SUPER_MAGIC;
log_debug("On btrfs: %s", yes_no(on_btrfs));
if (asprintf(&pack_fn_new, "%s/.readahead.new", root) < 0) {
r = log_oom();
goto finish;
}
pack = fopen(pack_fn_new, "we");
if (!pack) {
log_error("Failed to open pack file: %m");
r = -errno;
goto finish;
}
fputs(CANONICAL_HOST READAHEAD_PACK_FILE_VERSION, pack);
putc(on_ssd ? 'S' : 'R', pack);
if (on_ssd || on_btrfs) {
/* On SSD or on btrfs, just write things out in the
* order the files were accessed. */
HASHMAP_FOREACH_KEY(q, p, files, i)
pack_file(pack, p, on_btrfs);
} else {
struct item *ordered, *j;
unsigned k, n;
/* On rotating media, order things by the block
* numbers */
log_debug("Ordering...");
n = hashmap_size(files);
if (!(ordered = new(struct item, n))) {
r = log_oom();
goto finish;
}
j = ordered;
HASHMAP_FOREACH_KEY(q, p, files, i) {
memcpy(j, q, sizeof(struct item));
j++;
}
assert(ordered + n == j);
qsort(ordered, n, sizeof(struct item), qsort_compare);
for (k = 0; k < n; k++)
pack_file(pack, ordered[k].path, on_btrfs);
free(ordered);
}
log_debug("Finalizing...");
fflush(pack);
if (ferror(pack)) {
log_error("Failed to write pack file.");
r = -EIO;
goto finish;
}
if (rename(pack_fn_new, pack_fn) < 0) {
log_error("Failed to rename readahead file: %m");
r = -errno;
goto finish;
}
fclose(pack);
pack = NULL;
log_debug("Done.");
finish:
if (fanotify_fd >= 0)
close_nointr_nofail(fanotify_fd);
if (signal_fd >= 0)
close_nointr_nofail(signal_fd);
if (inotify_fd >= 0)
close_nointr_nofail(inotify_fd);
if (pack) {
fclose(pack);
unlink(pack_fn_new);
}
free(pack_fn_new);
free(pack_fn);
while ((p = hashmap_steal_first_key(files)))
free(p);
hashmap_free(files);
if (previous_block_readahead_set) {
uint64_t bytes;
/* Restore the original kernel readahead setting if we
* changed it, and nobody has overwritten it since
* yet. */
if (block_get_readahead(root, &bytes) >= 0 && bytes == 8*1024)
block_set_readahead(root, previous_block_readahead);
}
return r;
}
int main_collect(const char *root) {
if (!root)
root = "/";
/* Skip this step on read-only media. Note that we check the
* underlying block device here, not he read-only flag of the
* file system on top, since that one is most likely mounted
* read-only anyway at boot, even if the underlying block
* device is theoretically writable. */
if (fs_on_read_only(root) > 0) {
log_info("Disabling readahead collector due to read-only media.");
return EXIT_SUCCESS;
}
if (!enough_ram()) {
log_info("Disabling readahead collector due to low memory.");
return EXIT_SUCCESS;
}
shared = shared_get();
if (!shared)
return EXIT_FAILURE;
shared->collect = getpid();
__sync_synchronize();
if (collect(root) < 0)
return EXIT_FAILURE;
return EXIT_SUCCESS;
}