j->current_file = NULL;
j->current_field = 0;
- HASHMAP_FOREACH(f, j->files, i)
+ ORDERED_HASHMAP_FOREACH(f, j->files, i)
f->current_offset = 0;
}
assert_return(j, -EINVAL);
assert_return(!journal_pid_changed(j), -ECHILD);
- HASHMAP_FOREACH(f, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(f, j->files, i) {
bool found;
r = next_beyond_location(j, f, direction, &o, &p);
if (r < 0) {
- log_debug("Can't iterate through %s, ignoring: %s", f->path, strerror(-r));
+ log_debug_errno(r, "Can't iterate through %s, ignoring: %m", f->path);
remove_file_real(j, f);
continue;
} else if (r == 0)
assert(j);
assert(path);
- if (hashmap_get(j->files, path))
+ if (ordered_hashmap_get(j->files, path))
return 0;
- if (hashmap_size(j->files) >= JOURNAL_FILES_MAX) {
+ if (ordered_hashmap_size(j->files) >= JOURNAL_FILES_MAX) {
log_warning("Too many open journal files, not adding %s.", path);
return set_put_error(j, -ETOOMANYREFS);
}
/* journal_file_dump(f); */
- r = hashmap_put(j->files, f->path, f);
+ r = ordered_hashmap_put(j->files, f->path, f);
if (r < 0) {
journal_file_close(f);
return r;
if (!path)
return -ENOMEM;
- f = hashmap_get(j->files, path);
+ f = ordered_hashmap_get(j->files, path);
if (!f)
return 0;
assert(j);
assert(f);
- hashmap_remove(j->files, f->path);
+ ordered_hashmap_remove(j->files, f->path);
log_debug("File %s removed.", f->path);
if (j->unique_file == f) {
/* Jump to the next unique_file or NULL if that one was last */
- j->unique_file = hashmap_next(j->files, j->unique_file->path);
+ j->unique_file = ordered_hashmap_next(j->files, j->unique_file->path);
j->unique_offset = 0;
if (!j->unique_file)
j->unique_file_lost = true;
d = opendir(path);
if (!d) {
- log_debug("Failed to open %s: %m", path);
+ log_debug_errno(errno, "Failed to open %s: %m", path);
if (errno == ENOENT)
return 0;
return -errno;
de = readdir(d);
if (!de && errno != 0) {
r = -errno;
- log_debug("Failed to read directory %s: %m", m->path);
+ log_debug_errno(errno, "Failed to read directory %s: %m", m->path);
return r;
}
if (!de)
dirent_is_file_with_suffix(de, ".journal~")) {
r = add_file(j, m->path, de->d_name);
if (r < 0) {
- log_debug("Failed to add file %s/%s: %s",
- m->path, de->d_name, strerror(-r));
+ log_debug_errno(r, "Failed to add file %s/%s: %m",
+ m->path, de->d_name);
r = set_put_error(j, r);
if (r < 0)
return r;
de = readdir(d);
if (!de && errno != 0) {
r = -errno;
- log_debug("Failed to read directory %s: %m", m->path);
+ log_debug_errno(errno, "Failed to read directory %s: %m", m->path);
return r;
}
if (!de)
dirent_is_file_with_suffix(de, ".journal~")) {
r = add_file(j, m->path, de->d_name);
if (r < 0) {
- log_debug("Failed to add file %s/%s: %s",
- m->path, de->d_name, strerror(-r));
+ log_debug_errno(r, "Failed to add file %s/%s: %m",
+ m->path, de->d_name);
r = set_put_error(j, r);
if (r < 0)
return r;
r = add_directory(j, m->path, de->d_name);
if (r < 0)
- log_debug("Failed to add directory %s/%s: %s", m->path, de->d_name, strerror(-r));
+ log_debug_errno(r, "Failed to add directory %s/%s: %m", m->path, de->d_name);
}
}
* "root" directories. We don't expect errors here, so we
* treat them as fatal. */
- HASHMAP_FOREACH(f, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(f, j->files, i) {
_cleanup_free_ char *dir;
int r;
goto fail;
}
- j->files = hashmap_new(&string_hash_ops);
+ j->files = ordered_hashmap_new(&string_hash_ops);
j->directories_by_path = hashmap_new(&string_hash_ops);
j->mmap = mmap_cache_new();
if (!j->files || !j->directories_by_path || !j->mmap)
STRV_FOREACH(path, paths) {
r = add_any_file(j, *path);
if (r < 0) {
- log_error("Failed to open %s: %s", *path, strerror(-r));
+ log_error_errno(r, "Failed to open %s: %m", *path);
goto fail;
}
}
sd_journal_flush_matches(j);
- while ((f = hashmap_steal_first(j->files)))
+ while ((f = ordered_hashmap_steal_first(j->files)))
journal_file_close(f);
- hashmap_free(j->files);
+ ordered_hashmap_free(j->files);
while ((d = hashmap_first(j->directories_by_path)))
remove_directory(j, d);
if (e->mask & (IN_CREATE|IN_MOVED_TO|IN_MODIFY|IN_ATTRIB)) {
r = add_file(j, d->path, e->name);
if (r < 0) {
- log_debug("Failed to add file %s/%s: %s",
- d->path, e->name, strerror(-r));
+ log_debug_errno(r, "Failed to add file %s/%s: %m",
+ d->path, e->name);
set_put_error(j, r);
}
r = remove_file(j, d->path, e->name);
if (r < 0)
- log_debug("Failed to remove file %s/%s: %s", d->path, e->name, strerror(-r));
+ log_debug_errno(r, "Failed to remove file %s/%s: %m", d->path, e->name);
}
} else if (!d->is_root && e->len == 0) {
if (e->mask & (IN_DELETE_SELF|IN_MOVE_SELF|IN_UNMOUNT)) {
r = remove_directory(j, d);
if (r < 0)
- log_debug("Failed to remove directory %s: %s", d->path, strerror(-r));
+ log_debug_errno(r, "Failed to remove directory %s: %m", d->path);
}
if (e->mask & (IN_CREATE|IN_MOVED_TO|IN_MODIFY|IN_ATTRIB)) {
r = add_directory(j, d->path, e->name);
if (r < 0)
- log_debug("Failed to add directory %s/%s: %s", d->path, e->name, strerror(-r));
+ log_debug_errno(r, "Failed to add directory %s/%s: %m", d->path, e->name);
}
}
assert_return(from || to, -EINVAL);
assert_return(from != to, -EINVAL);
- HASHMAP_FOREACH(f, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(f, j->files, i) {
usec_t fr, t;
r = journal_file_get_cutoff_realtime_usec(f, &fr, &t);
assert_return(from || to, -EINVAL);
assert_return(from != to, -EINVAL);
- HASHMAP_FOREACH(f, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(f, j->files, i) {
usec_t fr, t;
r = journal_file_get_cutoff_monotonic_usec(f, boot_id, &fr, &t);
assert(j);
- HASHMAP_FOREACH(f, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(f, j->files, i) {
if (newline)
putchar('\n');
else
assert_return(!journal_pid_changed(j), -ECHILD);
assert_return(bytes, -EINVAL);
- HASHMAP_FOREACH(f, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(f, j->files, i) {
struct stat st;
if (fstat(f->fd, &st) < 0)
if (j->unique_file_lost)
return 0;
- j->unique_file = hashmap_first(j->files);
+ j->unique_file = ordered_hashmap_first(j->files);
if (!j->unique_file)
return 0;
/* We reached the end of the list? Then start again, with the next file */
if (j->unique_offset == 0) {
- j->unique_file = hashmap_next(j->files, j->unique_file->path);
+ j->unique_file = ordered_hashmap_next(j->files, j->unique_file->path);
if (!j->unique_file)
return 0;
* object by checking if it exists in the earlier
* traversed files. */
found = false;
- HASHMAP_FOREACH(of, j->files, i) {
+ ORDERED_HASHMAP_FOREACH(of, j->files, i) {
Object *oo;
uint64_t op;