#include "replace-var.h"
#include "fileio.h"
-#define JOURNAL_FILES_MAX 1024
+#define JOURNAL_FILES_MAX 7168
#define JOURNAL_FILES_RECHECK_USEC (2 * USEC_PER_SEC)
j->current_file = NULL;
j->current_field = 0;
- HASHMAP_FOREACH(f, j->files, i)
- f->current_offset = 0;
+ ORDERED_HASHMAP_FOREACH(f, j->files, i)
+ journal_file_reset_location(f);
}
static void reset_location(sd_journal *j) {
l->seqnum_set = l->realtime_set = l->monotonic_set = l->xor_hash_set = true;
}
-static void set_location(sd_journal *j, LocationType type, JournalFile *f, Object *o,
- direction_t direction, uint64_t offset) {
+static void set_location(sd_journal *j, JournalFile *f, Object *o) {
assert(j);
- assert(type == LOCATION_DISCRETE || type == LOCATION_SEEK);
assert(f);
assert(o);
- init_location(&j->current_location, type, f, o);
+ init_location(&j->current_location, LOCATION_DISCRETE, f, o);
j->current_file = f;
j->current_field = 0;
- f->last_direction = direction;
- f->current_offset = offset;
+ /* Let f know its candidate entry was picked. */
+ assert(f->location_type == LOCATION_SEEK);
+ f->location_type = LOCATION_DISCRETE;
}
static int match_is_valid(const void *data, size_t size) {
detach_location(j);
}
-static int compare_entry_order(JournalFile *af, Object *_ao,
- JournalFile *bf, uint64_t bp) {
-
- uint64_t a, b;
- Object *ao, *bo;
- int r;
-
- assert(af);
- assert(bf);
- assert(_ao);
-
- /* The mmap cache might invalidate the object from the first
- * file if we look at the one from the second file. Hence
- * temporarily copy the header of the first one, and look at
- * that only. */
- ao = alloca(offsetof(EntryObject, items));
- memcpy(ao, _ao, offsetof(EntryObject, items));
-
- r = journal_file_move_to_object(bf, OBJECT_ENTRY, bp, &bo);
- if (r < 0)
- return strcmp(af->path, bf->path);
-
- /* We operate on two different files here, hence we can access
- * two objects at the same time, which we normally can't.
- *
- * If contents and timestamps match, these entries are
- * identical, even if the seqnum does not match */
-
- if (sd_id128_equal(ao->entry.boot_id, bo->entry.boot_id) &&
- ao->entry.monotonic == bo->entry.monotonic &&
- ao->entry.realtime == bo->entry.realtime &&
- ao->entry.xor_hash == bo->entry.xor_hash)
- return 0;
-
- if (sd_id128_equal(af->header->seqnum_id, bf->header->seqnum_id)) {
-
- /* If this is from the same seqnum source, compare
- * seqnums */
- a = le64toh(ao->entry.seqnum);
- b = le64toh(bo->entry.seqnum);
-
- if (a < b)
- return -1;
- if (a > b)
- return 1;
-
- /* Wow! This is weird, different data but the same
- * seqnums? Something is borked, but let's make the
- * best of it and compare by time. */
- }
-
- if (sd_id128_equal(ao->entry.boot_id, bo->entry.boot_id)) {
-
- /* If the boot id matches, compare monotonic time */
- a = le64toh(ao->entry.monotonic);
- b = le64toh(bo->entry.monotonic);
-
- if (a < b)
- return -1;
- if (a > b)
- return 1;
- }
-
- /* Otherwise, compare UTC time */
- a = le64toh(ao->entry.realtime);
- b = le64toh(bo->entry.realtime);
-
- if (a < b)
- return -1;
- if (a > b)
- return 1;
-
- /* Finally, compare by contents */
- a = le64toh(ao->entry.xor_hash);
- b = le64toh(bo->entry.xor_hash);
-
- if (a < b)
- return -1;
- if (a > b)
- return 1;
-
- return 0;
-}
-
-_pure_ static int compare_with_location(JournalFile *af, Object *ao, Location *l) {
- uint64_t a;
-
- assert(af);
- assert(ao);
+_pure_ static int compare_with_location(JournalFile *f, Location *l) {
+ assert(f);
assert(l);
+ assert(f->location_type == LOCATION_SEEK);
assert(l->type == LOCATION_DISCRETE || l->type == LOCATION_SEEK);
if (l->monotonic_set &&
- sd_id128_equal(ao->entry.boot_id, l->boot_id) &&
+ sd_id128_equal(f->current_boot_id, l->boot_id) &&
l->realtime_set &&
- le64toh(ao->entry.realtime) == l->realtime &&
+ f->current_realtime == l->realtime &&
l->xor_hash_set &&
- le64toh(ao->entry.xor_hash) == l->xor_hash)
+ f->current_xor_hash == l->xor_hash)
return 0;
if (l->seqnum_set &&
- sd_id128_equal(af->header->seqnum_id, l->seqnum_id)) {
-
- a = le64toh(ao->entry.seqnum);
+ sd_id128_equal(f->header->seqnum_id, l->seqnum_id)) {
- if (a < l->seqnum)
+ if (f->current_seqnum < l->seqnum)
return -1;
- if (a > l->seqnum)
+ if (f->current_seqnum > l->seqnum)
return 1;
}
if (l->monotonic_set &&
- sd_id128_equal(ao->entry.boot_id, l->boot_id)) {
+ sd_id128_equal(f->current_boot_id, l->boot_id)) {
- a = le64toh(ao->entry.monotonic);
-
- if (a < l->monotonic)
+ if (f->current_monotonic < l->monotonic)
return -1;
- if (a > l->monotonic)
+ if (f->current_monotonic > l->monotonic)
return 1;
}
if (l->realtime_set) {
- a = le64toh(ao->entry.realtime);
-
- if (a < l->realtime)
+ if (f->current_realtime < l->realtime)
return -1;
- if (a > l->realtime)
+ if (f->current_realtime > l->realtime)
return 1;
}
if (l->xor_hash_set) {
- a = le64toh(ao->entry.xor_hash);
- if (a < l->xor_hash)
+ if (f->current_xor_hash < l->xor_hash)
return -1;
- if (a > l->xor_hash)
+ if (f->current_xor_hash > l->xor_hash)
return 1;
}
/* No matches is simple */
if (j->current_location.type == LOCATION_HEAD)
- return journal_file_next_entry(f, NULL, 0, DIRECTION_DOWN, ret, offset);
+ return journal_file_next_entry(f, 0, DIRECTION_DOWN, ret, offset);
if (j->current_location.type == LOCATION_TAIL)
- return journal_file_next_entry(f, NULL, 0, DIRECTION_UP, ret, offset);
+ return journal_file_next_entry(f, 0, DIRECTION_UP, ret, offset);
if (j->current_location.seqnum_set && sd_id128_equal(j->current_location.seqnum_id, f->header->seqnum_id))
return journal_file_move_to_entry_by_seqnum(f, j->current_location.seqnum, direction, ret, offset);
if (j->current_location.monotonic_set) {
if (j->current_location.realtime_set)
return journal_file_move_to_entry_by_realtime(f, j->current_location.realtime, direction, ret, offset);
- return journal_file_next_entry(f, NULL, 0, direction, ret, offset);
+ return journal_file_next_entry(f, 0, direction, ret, offset);
} else
return find_location_for_match(j, j->level0, f, direction, ret, offset);
}
Object **ret,
uint64_t *offset) {
- Object *c;
- uint64_t cp;
-
assert(j);
assert(f);
assert(ret);
assert(offset);
- c = *ret;
- cp = *offset;
-
/* No matches is easy. We simple advance the file
* pointer by one. */
if (!j->level0)
- return journal_file_next_entry(f, c, cp, direction, ret, offset);
+ return journal_file_next_entry(f, f->current_offset, direction, ret, offset);
/* If we have a match then we look for the next matching entry
* with an offset at least one step larger */
- return next_for_match(j, j->level0, f, direction == DIRECTION_DOWN ? cp+1 : cp-1, direction, ret, offset);
+ return next_for_match(j, j->level0, f,
+ direction == DIRECTION_DOWN ? f->current_offset + 1
+ : f->current_offset - 1,
+ direction, ret, offset);
}
-static int next_beyond_location(sd_journal *j, JournalFile *f, direction_t direction, Object **ret, uint64_t *offset) {
+static int next_beyond_location(sd_journal *j, JournalFile *f, direction_t direction) {
Object *c;
- uint64_t cp;
+ uint64_t cp, n_entries;
int r;
assert(j);
assert(f);
if (f->last_direction == direction && f->current_offset > 0) {
- cp = f->current_offset;
+ /* If we hit EOF before, recheck if any new entries arrived. */
+ n_entries = le64toh(f->header->n_entries);
+ if (f->location_type == LOCATION_TAIL && n_entries == f->last_n_entries)
+ return 0;
+ f->last_n_entries = n_entries;
- r = journal_file_move_to_object(f, OBJECT_ENTRY, cp, &c);
- if (r < 0)
- return r;
+ /* LOCATION_SEEK here means we did the work in a previous
+ * iteration and the current location already points to a
+ * candidate entry. */
+ if (f->location_type != LOCATION_SEEK) {
+ r = next_with_matches(j, f, direction, &c, &cp);
+ if (r <= 0)
+ return r;
- r = next_with_matches(j, f, direction, &c, &cp);
- if (r <= 0)
- return r;
+ journal_file_save_location(f, direction, c, cp);
+ }
} else {
r = find_location_with_matches(j, f, direction, &c, &cp);
if (r <= 0)
return r;
+
+ journal_file_save_location(f, direction, c, cp);
}
/* OK, we found the spot, now let's advance until an entry
if (j->current_location.type == LOCATION_DISCRETE) {
int k;
- k = compare_with_location(f, c, &j->current_location);
- if (direction == DIRECTION_DOWN)
- found = k > 0;
- else
- found = k < 0;
+ k = compare_with_location(f, &j->current_location);
+
+ found = direction == DIRECTION_DOWN ? k > 0 : k < 0;
} else
found = true;
- if (found) {
- if (ret)
- *ret = c;
- if (offset)
- *offset = cp;
+ if (found)
return 1;
- }
r = next_with_matches(j, f, direction, &c, &cp);
if (r <= 0)
return r;
+
+ journal_file_save_location(f, direction, c, cp);
}
}
static int real_journal_next(sd_journal *j, direction_t direction) {
JournalFile *f, *new_file = NULL;
- uint64_t new_offset = 0;
- uint64_t p = 0;
Iterator i;
Object *o;
int r;
assert_return(j, -EINVAL);
assert_return(!journal_pid_changed(j), -ECHILD);
- HASHMAP_FOREACH(f, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(f, j->files, i) {
bool found;
- r = next_beyond_location(j, f, direction, &o, &p);
+ r = next_beyond_location(j, f, direction);
if (r < 0) {
- log_debug("Can't iterate through %s, ignoring: %s", f->path, strerror(-r));
+ log_debug_errno(r, "Can't iterate through %s, ignoring: %m", f->path);
remove_file_real(j, f);
continue;
- } else if (r == 0)
+ } else if (r == 0) {
+ f->location_type = LOCATION_TAIL;
continue;
+ }
if (!new_file)
found = true;
else {
int k;
- k = compare_entry_order(f, o, new_file, new_offset);
+ k = journal_file_compare_locations(f, new_file);
found = direction == DIRECTION_DOWN ? k < 0 : k > 0;
}
- if (found) {
+ if (found)
new_file = f;
- new_offset = p;
- }
}
if (!new_file)
return 0;
- r = journal_file_move_to_object(new_file, OBJECT_ENTRY, new_offset, &o);
+ r = journal_file_move_to_object(new_file, OBJECT_ENTRY, new_file->current_offset, &o);
if (r < 0)
return r;
- set_location(j, LOCATION_DISCRETE, new_file, o, direction, new_offset);
+ set_location(j, new_file, o);
return 1;
}
assert(j);
assert(path);
- if (hashmap_get(j->files, path))
+ if (ordered_hashmap_get(j->files, path))
return 0;
- if (hashmap_size(j->files) >= JOURNAL_FILES_MAX) {
+ if (ordered_hashmap_size(j->files) >= JOURNAL_FILES_MAX) {
log_warning("Too many open journal files, not adding %s.", path);
return set_put_error(j, -ETOOMANYREFS);
}
/* journal_file_dump(f); */
- r = hashmap_put(j->files, f->path, f);
+ r = ordered_hashmap_put(j->files, f->path, f);
if (r < 0) {
journal_file_close(f);
return r;
if (!path)
return -ENOMEM;
- f = hashmap_get(j->files, path);
+ f = ordered_hashmap_get(j->files, path);
if (!f)
return 0;
assert(j);
assert(f);
- hashmap_remove(j->files, f->path);
+ ordered_hashmap_remove(j->files, f->path);
log_debug("File %s removed.", f->path);
if (j->unique_file == f) {
/* Jump to the next unique_file or NULL if that one was last */
- j->unique_file = hashmap_next(j->files, j->unique_file->path);
+ j->unique_file = ordered_hashmap_next(j->files, j->unique_file->path);
j->unique_offset = 0;
if (!j->unique_file)
j->unique_file_lost = true;
d = opendir(path);
if (!d) {
- log_debug("Failed to open %s: %m", path);
+ log_debug_errno(errno, "Failed to open %s: %m", path);
if (errno == ENOENT)
return 0;
return -errno;
de = readdir(d);
if (!de && errno != 0) {
r = -errno;
- log_debug("Failed to read directory %s: %m", m->path);
+ log_debug_errno(errno, "Failed to read directory %s: %m", m->path);
return r;
}
if (!de)
dirent_is_file_with_suffix(de, ".journal~")) {
r = add_file(j, m->path, de->d_name);
if (r < 0) {
- log_debug("Failed to add file %s/%s: %s",
- m->path, de->d_name, strerror(-r));
+ log_debug_errno(r, "Failed to add file %s/%s: %m",
+ m->path, de->d_name);
r = set_put_error(j, r);
if (r < 0)
return r;
de = readdir(d);
if (!de && errno != 0) {
r = -errno;
- log_debug("Failed to read directory %s: %m", m->path);
+ log_debug_errno(errno, "Failed to read directory %s: %m", m->path);
return r;
}
if (!de)
dirent_is_file_with_suffix(de, ".journal~")) {
r = add_file(j, m->path, de->d_name);
if (r < 0) {
- log_debug("Failed to add file %s/%s: %s",
- m->path, de->d_name, strerror(-r));
+ log_debug_errno(r, "Failed to add file %s/%s: %m",
+ m->path, de->d_name);
r = set_put_error(j, r);
if (r < 0)
return r;
r = add_directory(j, m->path, de->d_name);
if (r < 0)
- log_debug("Failed to add directory %s/%s: %s", m->path, de->d_name, strerror(-r));
+ log_debug_errno(r, "Failed to add directory %s/%s: %m", m->path, de->d_name);
}
}
* "root" directories. We don't expect errors here, so we
* treat them as fatal. */
- HASHMAP_FOREACH(f, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(f, j->files, i) {
_cleanup_free_ char *dir;
int r;
goto fail;
}
- j->files = hashmap_new(&string_hash_ops);
+ j->files = ordered_hashmap_new(&string_hash_ops);
j->directories_by_path = hashmap_new(&string_hash_ops);
j->mmap = mmap_cache_new();
if (!j->files || !j->directories_by_path || !j->mmap)
assert_return(machine, -EINVAL);
assert_return(ret, -EINVAL);
assert_return((flags & ~(SD_JOURNAL_LOCAL_ONLY|SD_JOURNAL_SYSTEM)) == 0, -EINVAL);
- assert_return(filename_is_safe(machine), -EINVAL);
+ assert_return(machine_name_is_valid(machine), -EINVAL);
p = strappenda("/run/systemd/machines/", machine);
r = parse_env_file(p, NEWLINE, "ROOT", &root, "CLASS", &class, NULL);
STRV_FOREACH(path, paths) {
r = add_any_file(j, *path);
if (r < 0) {
- log_error("Failed to open %s: %s", *path, strerror(-r));
+ log_error_errno(r, "Failed to open %s: %m", *path);
goto fail;
}
}
sd_journal_flush_matches(j);
- while ((f = hashmap_steal_first(j->files)))
+ while ((f = ordered_hashmap_steal_first(j->files)))
journal_file_close(f);
- hashmap_free(j->files);
+ ordered_hashmap_free(j->files);
while ((d = hashmap_first(j->directories_by_path)))
remove_directory(j, d);
if (e->mask & (IN_CREATE|IN_MOVED_TO|IN_MODIFY|IN_ATTRIB)) {
r = add_file(j, d->path, e->name);
if (r < 0) {
- log_debug("Failed to add file %s/%s: %s",
- d->path, e->name, strerror(-r));
+ log_debug_errno(r, "Failed to add file %s/%s: %m",
+ d->path, e->name);
set_put_error(j, r);
}
r = remove_file(j, d->path, e->name);
if (r < 0)
- log_debug("Failed to remove file %s/%s: %s", d->path, e->name, strerror(-r));
+ log_debug_errno(r, "Failed to remove file %s/%s: %m", d->path, e->name);
}
} else if (!d->is_root && e->len == 0) {
if (e->mask & (IN_DELETE_SELF|IN_MOVE_SELF|IN_UNMOUNT)) {
r = remove_directory(j, d);
if (r < 0)
- log_debug("Failed to remove directory %s: %s", d->path, strerror(-r));
+ log_debug_errno(r, "Failed to remove directory %s: %m", d->path);
}
if (e->mask & (IN_CREATE|IN_MOVED_TO|IN_MODIFY|IN_ATTRIB)) {
r = add_directory(j, d->path, e->name);
if (r < 0)
- log_debug("Failed to add directory %s/%s: %s", d->path, e->name, strerror(-r));
+ log_debug_errno(r, "Failed to add directory %s/%s: %m", d->path, e->name);
}
}
}
_public_ int sd_journal_process(sd_journal *j) {
- uint8_t buffer[sizeof(struct inotify_event) + FILENAME_MAX] _alignas_(struct inotify_event);
bool got_something = false;
assert_return(j, -EINVAL);
j->last_process_usec = now(CLOCK_MONOTONIC);
for (;;) {
+ union inotify_event_buffer buffer;
struct inotify_event *e;
ssize_t l;
- l = read(j->inotify_fd, buffer, sizeof(buffer));
+ l = read(j->inotify_fd, &buffer, sizeof(buffer));
if (l < 0) {
if (errno == EAGAIN || errno == EINTR)
return got_something ? determine_change(j) : SD_JOURNAL_NOP;
got_something = true;
- e = (struct inotify_event*) buffer;
- while (l > 0) {
- size_t step;
-
+ FOREACH_INOTIFY_EVENT(e, buffer, l)
process_inotify_event(j, e);
-
- step = sizeof(struct inotify_event) + e->len;
- assert(step <= (size_t) l);
-
- e = (struct inotify_event*) ((uint8_t*) e + step);
- l -= step;
- }
}
}
assert_return(from || to, -EINVAL);
assert_return(from != to, -EINVAL);
- HASHMAP_FOREACH(f, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(f, j->files, i) {
usec_t fr, t;
r = journal_file_get_cutoff_realtime_usec(f, &fr, &t);
assert_return(from || to, -EINVAL);
assert_return(from != to, -EINVAL);
- HASHMAP_FOREACH(f, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(f, j->files, i) {
usec_t fr, t;
r = journal_file_get_cutoff_monotonic_usec(f, boot_id, &fr, &t);
assert(j);
- HASHMAP_FOREACH(f, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(f, j->files, i) {
if (newline)
putchar('\n');
else
assert_return(!journal_pid_changed(j), -ECHILD);
assert_return(bytes, -EINVAL);
- HASHMAP_FOREACH(f, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(f, j->files, i) {
struct stat st;
if (fstat(f->fd, &st) < 0)
if (j->unique_file_lost)
return 0;
- j->unique_file = hashmap_first(j->files);
+ j->unique_file = ordered_hashmap_first(j->files);
if (!j->unique_file)
return 0;
size_t ol;
bool found;
int r;
- void *release_cookie;
/* Proceed to next data object in the field's linked list */
if (j->unique_offset == 0) {
/* We reached the end of the list? Then start again, with the next file */
if (j->unique_offset == 0) {
- j->unique_file = hashmap_next(j->files, j->unique_file->path);
+ j->unique_file = ordered_hashmap_next(j->files, j->unique_file->path);
if (!j->unique_file)
return 0;
continue;
}
- /* We do not use the type context here, but 0 instead,
- * so that we can look at this data object at the same
+ /* We do not use OBJECT_DATA context here, but OBJECT_UNUSED
+ * instead, so that we can look at this data object at the same
* time as one on another file */
- r = journal_file_move_to_object(j->unique_file, 0, j->unique_offset, &o);
+ r = journal_file_move_to_object(j->unique_file, OBJECT_UNUSED, j->unique_offset, &o);
if (r < 0)
return r;
return -EBADMSG;
}
- r = journal_file_object_keep(j->unique_file, o, j->unique_offset, &release_cookie);
- if (r < 0)
- return r;
-
r = return_data(j, j->unique_file, o, &odata, &ol);
if (r < 0)
return r;
* object by checking if it exists in the earlier
* traversed files. */
found = false;
- HASHMAP_FOREACH(of, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(of, j->files, i) {
Object *oo;
uint64_t op;
found = true;
}
- r = journal_file_object_release(j->unique_file, release_cookie);
- if (r < 0)
- return r;
-
if (found)
continue;