summaryrefslogtreecommitdiffstats
path: root/debuginfod
diff options
context:
space:
mode:
authorMark Wielaard <[email protected]>2020-06-11 00:06:30 +0200
committerMark Wielaard <[email protected]>2020-06-11 03:58:27 +0200
commit49f13584d60322578c19b6118393ab04236ca7bf (patch)
tree88c79fbf73c54734c51b6b4a9a2a154e229f30c6 /debuginfod
parentc0d643e7d91fc002c9fecd83277c62a0e56ef76f (diff)
parent2c7c40373b68968cce20a60a28234e2a2cbc55cb (diff)
Merge tag 'elfutils-0.178' into mjw/RH-DTSdts-0.178
elfutils 0.178 release Adopt ebl backends loading from trunk.
Diffstat (limited to 'debuginfod')
-rw-r--r--debuginfod/ChangeLog70
-rw-r--r--debuginfod/Makefile.am119
-rw-r--r--debuginfod/debuginfod-client.c754
-rw-r--r--debuginfod/debuginfod-find.c145
-rw-r--r--debuginfod/debuginfod.cxx2688
-rw-r--r--debuginfod/debuginfod.h85
-rw-r--r--debuginfod/libdebuginfod.map10
7 files changed, 3871 insertions, 0 deletions
diff --git a/debuginfod/ChangeLog b/debuginfod/ChangeLog
new file mode 100644
index 00000000..8aa29443
--- /dev/null
+++ b/debuginfod/ChangeLog
@@ -0,0 +1,70 @@
+2019-11-26 Mark Wielaard <[email protected]>
+
+ * Makefile.am (BUILD_STATIC): Add needed libraries for libdw and
+ libdebuginfod.
+
+2019-11-25 Frank Ch. Eigler <[email protected]>
+
+ * debuginfod.cxx (groom): Add a sqlite3_db_release_memory()
+ at the end of periodic grooming to try to shrink the process.
+
+2019-11-24 Mark Wielaard <[email protected]>
+
+ * debuginfod.cxx (test_webapi_sleep): Removed.
+ (handler_cb): Don't check test_webapi_sleep and sleep.
+ (main): Don't set test_webapi_sleep.
+
+2019-11-24 Mark Wielaard <[email protected]>
+
+ * debuginfod.cxx (add_metric): New function.
+ (scan_source_file_path): Record metrics for
+ found_executable_total, found_debuginfo_total and
+ found_sourcerefs_total.
+ (scan_source_rpm_path): Likewise.
+
+2019-11-07 Frank Ch. Eigler <[email protected]>
+
+ * debuginfod.cxx: Add /metrics endpoint. Add numerous
+ calls to new functions inc_metric/set_metric to populate
+ threadsafe map containing stats. Add http content-type
+ response headers throughout.
+ (thread_main_*): Simplify counter/timer flow.
+ (main): Reorder web service shutdown to leave http running
+ as long as possible.
+ * debuginfod.8: Document it, add security caution.
+
+2019-11-06 Frank Ch. Eigler <[email protected]>
+
+ * debuginfod.cxx: Add new -L (symlink-following) mode.
+ * debuginfod.8: Document it.
+
+2019-11-04 Frank Ch. Eigler <[email protected]>
+
+ * debuginfo-client.c (debuginfod_set_progressfn): New function
+ for progress/interrupt callback.
+ (debuginfod_clean_cache, debuginfod_query_server): Call it.
+ * debuginfo.h: Declare it.
+ * debuginfod_set_progressfn.3, *_find_debuginfo.3: Document it.
+ * Makefile.am: Install it.
+ * libdebuginfod.map: Export it all under ELFUTILS_0.178 symversion.
+
+ * debuginfod-find.c: Add -v option to activate progress cb.
+ * debuginfod-find.1: Document it.
+ * debuginfod.cxx: Add $DEBUGINFOD_TEST_WEBAPI_SLEEP env var
+ to insert sleep in webapi callbacks, to help manual testing.
+
+2019-10-28 Frank Ch. Eigler <[email protected]>
+
+ * debuginfod.cxx: New file: debuginfod server.
+ * debuginfod.8: New file: man page.
+ * Makefile.am: Build it.
+
+2019-10-28 Aaron Merey <[email protected]>
+
+ * debuginfod-client.c: New file: debuginfod client library.
+ * debuginfod.h: New file: header for same.
+ * libdebuginfod.map: New file: govern its solib exports.
+ * debuginfod-find.c: New file: command line frontend.
+ * debuginfod-find.1, debuginfod_find_source.3,
+ debuginfod_find_executable.3, debuginfod_find_debuginfo.3:
+ New man pages.
diff --git a/debuginfod/Makefile.am b/debuginfod/Makefile.am
new file mode 100644
index 00000000..7ae74e06
--- /dev/null
+++ b/debuginfod/Makefile.am
@@ -0,0 +1,119 @@
+## Makefile.am for libdebuginfod library subdirectory in elfutils.
+##
+## Process this file with automake to create Makefile.in
+##
+## Copyright (C) 2019 Red Hat, Inc.
+## This file is part of elfutils.
+##
+## This file is free software; you can redistribute it and/or modify
+## it under the terms of either
+##
+## * the GNU Lesser General Public License as published by the Free
+## Software Foundation; either version 3 of the License, or (at
+## your option) any later version
+##
+## or
+##
+## * the GNU General Public License as published by the Free
+## Software Foundation; either version 2 of the License, or (at
+## your option) any later version
+##
+## or both in parallel, as here.
+##
+## elfutils is distributed in the hope that it will be useful, but
+## WITHOUT ANY WARRANTY; without even the implied warranty of
+## MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+## General Public License for more details.
+##
+## You should have received copies of the GNU General Public License and
+## the GNU Lesser General Public License along with this program. If
+## not, see <http://www.gnu.org/licenses/>.
+##
+include $(top_srcdir)/config/eu.am
+AM_CPPFLAGS += -I$(srcdir) -I$(srcdir)/../libelf -I$(srcdir)/../libebl \
+ -I$(srcdir)/../libdw -I$(srcdir)/../libdwelf \
+ $(libmicrohttpd_CFLAGS) $(libcurl_CFLAGS) $(sqlite3_CFLAGS) \
+ $(libarchive_CFLAGS)
+VERSION = 1
+
+# Disable eu- prefixing for artifacts (binaries & man pages) in this
+# directory, since they do not conflict with binutils tools.
+program_prefix=
+program_transform_name = s,x,x,
+
+if BUILD_STATIC
+libasm = ../libasm/libasm.a
+libdw = ../libdw/libdw.a -lz $(zip_LIBS) $(libelf) $(libebl) -ldl -lpthread
+libelf = ../libelf/libelf.a -lz
+libdebuginfod = ./libdebuginfod.a $(libcurl_LIBS)
+else
+libasm = ../libasm/libasm.so
+libdw = ../libdw/libdw.so
+libelf = ../libelf/libelf.so
+libdebuginfod = ./libdebuginfod.so
+endif
+libebl = ../libebl/libebl.a
+libeu = ../lib/libeu.a
+
+AM_LDFLAGS = -Wl,-rpath-link,../libelf:../libdw:.
+
+bin_PROGRAMS = debuginfod debuginfod-find
+debuginfod_SOURCES = debuginfod.cxx
+debuginfod_LDADD = $(libdw) $(libelf) $(libeu) $(libdebuginfod) $(libmicrohttpd_LIBS) $(libcurl_LIBS) $(sqlite3_LIBS) $(libarchive_LIBS) -lpthread -ldl
+
+debuginfod_find_SOURCES = debuginfod-find.c
+debuginfod_find_LDADD = $(libeu) $(libdebuginfod)
+
+lib_LIBRARIES = libdebuginfod.a
+noinst_LIBRARIES = libdebuginfod_pic.a
+
+libdebuginfod_a_CFLAGS = -fPIC $(AM_CFLAGS)
+libdebuginfod_a_SOURCES = debuginfod-client.c
+libdebuginfod_pic_a_SOURCES = debuginfod-client.c
+am_libdebuginfod_pic_a_OBJECTS = $(libdebuginfod_a_SOURCES:.c=.os)
+
+pkginclude_HEADERS = debuginfod.h
+
+libdebuginfod_so_LIBS = libdebuginfod_pic.a
+libdebuginfod_so_LDLIBS = $(libcurl_LIBS)
+libdebuginfod.so$(EXEEXT): $(srcdir)/libdebuginfod.map $(libdebuginfod_so_LIBS)
+ $(AM_V_CCLD)$(LINK) $(dso_LDFLAGS) -o $@ \
+ -Wl,--soname,$@.$(VERSION) \
+ -Wl,--version-script,$<,--no-undefined \
+ -Wl,--whole-archive $(libdebuginfod_so_LIBS) -Wl,--no-whole-archive \
+ $(libdebuginfod_so_LDLIBS)
+ @$(textrel_check)
+ $(AM_V_at)ln -fs $@ $@.$(VERSION)
+
+install: install-am libdebuginfod.so
+ $(mkinstalldirs) $(DESTDIR)$(libdir)
+ $(INSTALL_PROGRAM) libdebuginfod.so $(DESTDIR)$(libdir)/libdebuginfod-$(PACKAGE_VERSION).so
+ ln -fs libdebuginfod-$(PACKAGE_VERSION).so $(DESTDIR)$(libdir)/libdebuginfod.so.$(VERSION)
+ ln -fs libdebuginfod.so.$(VERSION) $(DESTDIR)$(libdir)/libdebuginfod.so
+
+uninstall: uninstall-am
+ rm -f $(DESTDIR)$(libdir)/libdebuginfod-$(PACKAGE_VERSION).so
+ rm -f $(DESTDIR)$(libdir)/libdebuginfod.so.$(VERSION)
+ rm -f $(DESTDIR)$(libdir)/libdebuginfod.so
+ rmdir --ignore-fail-on-non-empty $(DESTDIR)$(includedir)/elfutils
+
+EXTRA_DIST = libdebuginfod.map
+MOSTLYCLEANFILES = $(am_libdebuginfod_pic_a_OBJECTS) libdebuginfod.so.$(VERSION)
+CLEANFILES += $(am_libdebuginfod_pic_a_OBJECTS) libdebuginfod.so
+
+# automake std-options override: arrange to pass LD_LIBRARY_PATH
+installcheck-binPROGRAMS: $(bin_PROGRAMS)
+ bad=0; pid=$$$$; list="$(bin_PROGRAMS)"; for p in $$list; do \
+ case ' $(AM_INSTALLCHECK_STD_OPTIONS_EXEMPT) ' in \
+ *" $$p "* | *" $(srcdir)/$$p "*) continue;; \
+ esac; \
+ f=`echo "$$p" | \
+ sed 's,^.*/,,;s/$(EXEEXT)$$//;$(transform);s/$$/$(EXEEXT)/'`; \
+ for opt in --help --version; do \
+ if LD_LIBRARY_PATH=$(DESTDIR)$(libdir) \
+ $(DESTDIR)$(bindir)/$$f $$opt > c$${pid}_.out 2> c$${pid}_.err \
+ && test -n "`cat c$${pid}_.out`" \
+ && test -z "`cat c$${pid}_.err`"; then :; \
+ else echo "$$f does not support $$opt" 1>&2; bad=1; fi; \
+ done; \
+ done; rm -f c$${pid}_.???; exit $$bad
diff --git a/debuginfod/debuginfod-client.c b/debuginfod/debuginfod-client.c
new file mode 100644
index 00000000..6e62b86c
--- /dev/null
+++ b/debuginfod/debuginfod-client.c
@@ -0,0 +1,754 @@
+/* Retrieve ELF / DWARF / source files from the debuginfod.
+ Copyright (C) 2019 Red Hat, Inc.
+ This file is part of elfutils.
+
+ This file is free software; you can redistribute it and/or modify
+ it under the terms of either
+
+ * the GNU Lesser General Public License as published by the Free
+ Software Foundation; either version 3 of the License, or (at
+ your option) any later version
+
+ or
+
+ * the GNU General Public License as published by the Free
+ Software Foundation; either version 2 of the License, or (at
+ your option) any later version
+
+ or both in parallel, as here.
+
+ elfutils is distributed in the hope that it will be useful, but
+ WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ General Public License for more details.
+
+ You should have received copies of the GNU General Public License and
+ the GNU Lesser General Public License along with this program. If
+ not, see <http://www.gnu.org/licenses/>. */
+
+
+/* cargo-cult from libdwfl linux-kernel-modules.c */
+/* In case we have a bad fts we include this before config.h because it
+ can't handle _FILE_OFFSET_BITS.
+ Everything we need here is fine if its declarations just come first.
+ Also, include sys/types.h before fts. On some systems fts.h is not self
+ contained. */
+#ifdef BAD_FTS
+ #include <sys/types.h>
+ #include <fts.h>
+#endif
+
+#include "config.h"
+#include "debuginfod.h"
+#include <assert.h>
+#include <dirent.h>
+#include <stdio.h>
+#include <stdlib.h>
+#include <errno.h>
+#include <unistd.h>
+#include <errno.h>
+#include <fcntl.h>
+#include <fts.h>
+#include <string.h>
+#include <stdbool.h>
+#include <linux/limits.h>
+#include <time.h>
+#include <utime.h>
+#include <sys/syscall.h>
+#include <sys/types.h>
+#include <sys/stat.h>
+#include <curl/curl.h>
+
+/* If fts.h is included before config.h, its indirect inclusions may not
+ give us the right LFS aliases of these functions, so map them manually. */
+#ifdef BAD_FTS
+ #ifdef _FILE_OFFSET_BITS
+ #define open open64
+ #define fopen fopen64
+ #endif
+#else
+ #include <sys/types.h>
+ #include <fts.h>
+#endif
+
+struct debuginfod_client
+{
+ /* Progress/interrupt callback function. */
+ debuginfod_progressfn_t progressfn;
+
+ /* Can contain all other context, like cache_path, server_urls,
+ timeout or other info gotten from environment variables, the
+ handle data, etc. So those don't have to be reparsed and
+ recreated on each request. */
+};
+
+/* The cache_clean_interval_s file within the debuginfod cache specifies
+ how frequently the cache should be cleaned. The file's st_mtime represents
+ the time of last cleaning. */
+static const char *cache_clean_interval_filename = "cache_clean_interval_s";
+static const time_t cache_clean_default_interval_s = 86400; /* 1 day */
+
+/* The cache_max_unused_age_s file within the debuginfod cache specifies the
+ the maximum time since last access that a file will remain in the cache. */
+static const char *cache_max_unused_age_filename = "max_unused_age_s";
+static const time_t cache_default_max_unused_age_s = 604800; /* 1 week */
+
+/* Location of the cache of files downloaded from debuginfods.
+ The default parent directory is $HOME, or '/' if $HOME doesn't exist. */
+static const char *cache_default_name = ".debuginfod_client_cache";
+static const char *cache_path_envvar = DEBUGINFOD_CACHE_PATH_ENV_VAR;
+
+/* URLs of debuginfods, separated by url_delim.
+ This env var must be set for debuginfod-client to run. */
+static const char *server_urls_envvar = DEBUGINFOD_URLS_ENV_VAR;
+static const char *url_delim = " ";
+static const char url_delim_char = ' ';
+
+/* Timeout for debuginfods, in seconds.
+ This env var must be set for debuginfod-client to run. */
+static const char *server_timeout_envvar = DEBUGINFOD_TIMEOUT_ENV_VAR;
+static int server_timeout = 5;
+
+/* Data associated with a particular CURL easy handle. Passed to
+ the write callback. */
+struct handle_data
+{
+ /* Cache file to be written to in case query is successful. */
+ int fd;
+
+ /* URL queried by this handle. */
+ char url[PATH_MAX];
+
+ /* This handle. */
+ CURL *handle;
+
+ /* Pointer to handle that should write to fd. Initially points to NULL,
+ then points to the first handle that begins writing the target file
+ to the cache. Used to ensure that a file is not downloaded from
+ multiple servers unnecessarily. */
+ CURL **target_handle;
+};
+
+static size_t
+debuginfod_write_callback (char *ptr, size_t size, size_t nmemb, void *data)
+{
+ ssize_t count = size * nmemb;
+
+ struct handle_data *d = (struct handle_data*)data;
+
+ /* Indicate to other handles that they can abort their transfer. */
+ if (*d->target_handle == NULL)
+ *d->target_handle = d->handle;
+
+ /* If this handle isn't the target handle, abort transfer. */
+ if (*d->target_handle != d->handle)
+ return -1;
+
+ return (size_t) write(d->fd, (void*)ptr, count);
+}
+
+/* Create the cache and interval file if they do not already exist.
+ Return 0 if cache and config file are initialized, otherwise return
+ the appropriate error code. */
+static int
+debuginfod_init_cache (char *cache_path, char *interval_path, char *maxage_path)
+{
+ struct stat st;
+
+ /* If the cache and config file already exist then we are done. */
+ if (stat(cache_path, &st) == 0 && stat(interval_path, &st) == 0)
+ return 0;
+
+ /* Create the cache and config files as necessary. */
+ if (stat(cache_path, &st) != 0 && mkdir(cache_path, 0777) < 0)
+ return -errno;
+
+ int fd = -1;
+
+ /* init cleaning interval config file. */
+ fd = open(interval_path, O_CREAT | O_RDWR, 0666);
+ if (fd < 0)
+ return -errno;
+
+ if (dprintf(fd, "%ld", cache_clean_default_interval_s) < 0)
+ return -errno;
+
+ /* init max age config file. */
+ if (stat(maxage_path, &st) != 0
+ && (fd = open(maxage_path, O_CREAT | O_RDWR, 0666)) < 0)
+ return -errno;
+
+ if (dprintf(fd, "%ld", cache_default_max_unused_age_s) < 0)
+ return -errno;
+
+ return 0;
+}
+
+
+/* Delete any files that have been unmodied for a period
+ longer than $DEBUGINFOD_CACHE_CLEAN_INTERVAL_S. */
+static int
+debuginfod_clean_cache(debuginfod_client *c,
+ char *cache_path, char *interval_path,
+ char *max_unused_path)
+{
+ struct stat st;
+ FILE *interval_file;
+ FILE *max_unused_file;
+
+ if (stat(interval_path, &st) == -1)
+ {
+ /* Create new interval file. */
+ interval_file = fopen(interval_path, "w");
+
+ if (interval_file == NULL)
+ return -errno;
+
+ int rc = fprintf(interval_file, "%ld", cache_clean_default_interval_s);
+ fclose(interval_file);
+
+ if (rc < 0)
+ return -errno;
+ }
+
+ /* Check timestamp of interval file to see whether cleaning is necessary. */
+ time_t clean_interval;
+ interval_file = fopen(interval_path, "r");
+ if (fscanf(interval_file, "%ld", &clean_interval) != 1)
+ clean_interval = cache_clean_default_interval_s;
+ fclose(interval_file);
+
+ if (time(NULL) - st.st_mtime < clean_interval)
+ /* Interval has not passed, skip cleaning. */
+ return 0;
+
+ /* Read max unused age value from config file. */
+ time_t max_unused_age;
+ max_unused_file = fopen(max_unused_path, "r");
+ if (max_unused_file)
+ {
+ if (fscanf(max_unused_file, "%ld", &max_unused_age) != 1)
+ max_unused_age = cache_default_max_unused_age_s;
+ fclose(max_unused_file);
+ }
+ else
+ max_unused_age = cache_default_max_unused_age_s;
+
+ char * const dirs[] = { cache_path, NULL, };
+
+ FTS *fts = fts_open(dirs, 0, NULL);
+ if (fts == NULL)
+ return -errno;
+
+ FTSENT *f;
+ long files = 0;
+ while ((f = fts_read(fts)) != NULL)
+ {
+ files++;
+ if (c->progressfn) /* inform/check progress callback */
+ if ((c->progressfn) (c, files, 0))
+ break;
+
+ switch (f->fts_info)
+ {
+ case FTS_F:
+ /* delete file if max_unused_age has been met or exceeded. */
+ /* XXX consider extra effort to clean up old tmp files */
+ if (time(NULL) - f->fts_statp->st_atime >= max_unused_age)
+ unlink (f->fts_path);
+ break;
+
+ case FTS_DP:
+ /* Remove if empty. */
+ (void) rmdir (f->fts_path);
+ break;
+
+ default:
+ ;
+ }
+ }
+ fts_close(fts);
+
+ /* Update timestamp representing when the cache was last cleaned. */
+ utime (interval_path, NULL);
+ return 0;
+}
+
+
+#define MAX_BUILD_ID_BYTES 64
+
+
+/* Query each of the server URLs found in $DEBUGINFOD_URLS for the file
+ with the specified build-id, type (debuginfo, executable or source)
+ and filename. filename may be NULL. If found, return a file
+ descriptor for the target, otherwise return an error code.
+*/
+static int
+debuginfod_query_server (debuginfod_client *c,
+ const unsigned char *build_id,
+ int build_id_len,
+ const char *type,
+ const char *filename,
+ char **path)
+{
+ char *urls_envvar;
+ char *server_urls;
+ char cache_path[PATH_MAX];
+ char maxage_path[PATH_MAX*3]; /* These *3 multipliers are to shut up gcc -Wformat-truncation */
+ char interval_path[PATH_MAX*4];
+ char target_cache_dir[PATH_MAX*2];
+ char target_cache_path[PATH_MAX*4];
+ char target_cache_tmppath[PATH_MAX*5];
+ char suffix[PATH_MAX*2];
+ char build_id_bytes[MAX_BUILD_ID_BYTES * 2 + 1];
+
+ /* Copy lowercase hex representation of build_id into buf. */
+ if ((build_id_len >= MAX_BUILD_ID_BYTES) ||
+ (build_id_len == 0 &&
+ sizeof(build_id_bytes) > MAX_BUILD_ID_BYTES*2 + 1))
+ return -EINVAL;
+ if (build_id_len == 0) /* expect clean hexadecimal */
+ strcpy (build_id_bytes, (const char *) build_id);
+ else
+ for (int i = 0; i < build_id_len; i++)
+ sprintf(build_id_bytes + (i * 2), "%02x", build_id[i]);
+
+ if (filename != NULL)
+ {
+ if (filename[0] != '/') // must start with /
+ return -EINVAL;
+
+ /* copy the filename to suffix, s,/,#,g */
+ unsigned q = 0;
+ for (unsigned fi=0; q < PATH_MAX-1; fi++)
+ switch (filename[fi])
+ {
+ case '\0':
+ suffix[q] = '\0';
+ q = PATH_MAX-1; /* escape for loop too */
+ break;
+ case '/': /* escape / to prevent dir escape */
+ suffix[q++]='#';
+ suffix[q++]='#';
+ break;
+ case '#': /* escape # to prevent /# vs #/ collisions */
+ suffix[q++]='#';
+ suffix[q++]='_';
+ break;
+ default:
+ suffix[q++]=filename[fi];
+ }
+ suffix[q] = '\0';
+ /* If the DWARF filenames are super long, this could exceed
+ PATH_MAX and truncate/collide. Oh well, that'll teach
+ them! */
+ }
+ else
+ suffix[0] = '\0';
+
+ /* set paths needed to perform the query
+
+ example format
+ cache_path: $HOME/.debuginfod_cache
+ target_cache_dir: $HOME/.debuginfod_cache/0123abcd
+ target_cache_path: $HOME/.debuginfod_cache/0123abcd/debuginfo
+ target_cache_path: $HOME/.debuginfod_cache/0123abcd/source#PATH#TO#SOURCE ?
+ */
+
+ if (getenv(cache_path_envvar))
+ strcpy(cache_path, getenv(cache_path_envvar));
+ else
+ {
+ if (getenv("HOME"))
+ sprintf(cache_path, "%s/%s", getenv("HOME"), cache_default_name);
+ else
+ sprintf(cache_path, "/%s", cache_default_name);
+ }
+
+ /* avoid using snprintf here due to compiler warning. */
+ snprintf(target_cache_dir, sizeof(target_cache_dir), "%s/%s", cache_path, build_id_bytes);
+ snprintf(target_cache_path, sizeof(target_cache_path), "%s/%s%s", target_cache_dir, type, suffix);
+ snprintf(target_cache_tmppath, sizeof(target_cache_tmppath), "%s.XXXXXX", target_cache_path);
+
+ /* XXX combine these */
+ snprintf(interval_path, sizeof(interval_path), "%s/%s", cache_path, cache_clean_interval_filename);
+ snprintf(maxage_path, sizeof(maxage_path), "%s/%s", cache_path, cache_max_unused_age_filename);
+ int rc = debuginfod_init_cache(cache_path, interval_path, maxage_path);
+ if (rc != 0)
+ goto out;
+ rc = debuginfod_clean_cache(c, cache_path, interval_path, maxage_path);
+ if (rc != 0)
+ goto out;
+
+ /* If the target is already in the cache then we are done. */
+ int fd = open (target_cache_path, O_RDONLY);
+ if (fd >= 0)
+ {
+ /* Success!!!! */
+ if (path != NULL)
+ *path = strdup(target_cache_path);
+ return fd;
+ }
+
+
+ urls_envvar = getenv(server_urls_envvar);
+ if (urls_envvar == NULL || urls_envvar[0] == '\0')
+ {
+ rc = -ENOSYS;
+ goto out;
+ }
+
+ if (getenv(server_timeout_envvar))
+ server_timeout = atoi (getenv(server_timeout_envvar));
+
+ /* make a copy of the envvar so it can be safely modified. */
+ server_urls = strdup(urls_envvar);
+ if (server_urls == NULL)
+ {
+ rc = -ENOMEM;
+ goto out;
+ }
+ /* thereafter, goto out0 on error*/
+
+ /* create target directory in cache if not found. */
+ struct stat st;
+ if (stat(target_cache_dir, &st) == -1 && mkdir(target_cache_dir, 0700) < 0)
+ {
+ rc = -errno;
+ goto out0;
+ }
+
+ /* NB: write to a temporary file first, to avoid race condition of
+ multiple clients checking the cache, while a partially-written or empty
+ file is in there, being written from libcurl. */
+ fd = mkstemp (target_cache_tmppath);
+ if (fd < 0)
+ {
+ rc = -errno;
+ goto out0;
+ }
+
+ /* Count number of URLs. */
+ int num_urls = 0;
+ for (int i = 0; server_urls[i] != '\0'; i++)
+ if (server_urls[i] != url_delim_char
+ && (i == 0 || server_urls[i - 1] == url_delim_char))
+ num_urls++;
+
+ /* Tracks which handle should write to fd. Set to the first
+ handle that is ready to write the target file to the cache. */
+ CURL *target_handle = NULL;
+ struct handle_data *data = malloc(sizeof(struct handle_data) * num_urls);
+
+ /* Initalize handle_data with default values. */
+ for (int i = 0; i < num_urls; i++)
+ {
+ data[i].handle = NULL;
+ data[i].fd = -1;
+ }
+
+ CURLM *curlm = curl_multi_init();
+ if (curlm == NULL)
+ {
+ rc = -ENETUNREACH;
+ goto out0;
+ }
+ /* thereafter, goto out1 on error. */
+
+ char *strtok_saveptr;
+ char *server_url = strtok_r(server_urls, url_delim, &strtok_saveptr);
+
+ /* Initialize each handle. */
+ for (int i = 0; i < num_urls && server_url != NULL; i++)
+ {
+ data[i].fd = fd;
+ data[i].target_handle = &target_handle;
+ data[i].handle = curl_easy_init();
+
+ if (data[i].handle == NULL)
+ {
+ rc = -ENETUNREACH;
+ goto out1;
+ }
+
+ /* Build handle url. Tolerate both http://foo:999 and
+ http://foo:999/ forms */
+ char *slashbuildid;
+ if (strlen(server_url) > 1 && server_url[strlen(server_url)-1] == '/')
+ slashbuildid = "buildid";
+ else
+ slashbuildid = "/buildid";
+
+ if (filename) /* must start with / */
+ snprintf(data[i].url, PATH_MAX, "%s%s/%s/%s%s", server_url,
+ slashbuildid, build_id_bytes, type, filename);
+ else
+ snprintf(data[i].url, PATH_MAX, "%s%s/%s/%s", server_url,
+ slashbuildid, build_id_bytes, type);
+
+ curl_easy_setopt(data[i].handle, CURLOPT_URL, data[i].url);
+ curl_easy_setopt(data[i].handle,
+ CURLOPT_WRITEFUNCTION,
+ debuginfod_write_callback);
+ curl_easy_setopt(data[i].handle, CURLOPT_WRITEDATA, (void*)&data[i]);
+ curl_easy_setopt(data[i].handle, CURLOPT_TIMEOUT, (long) server_timeout);
+ curl_easy_setopt(data[i].handle, CURLOPT_FILETIME, (long) 1);
+ curl_easy_setopt(data[i].handle, CURLOPT_FOLLOWLOCATION, (long) 1);
+ curl_easy_setopt(data[i].handle, CURLOPT_FAILONERROR, (long) 1);
+ curl_easy_setopt(data[i].handle, CURLOPT_NOSIGNAL, (long) 1);
+ curl_easy_setopt(data[i].handle, CURLOPT_AUTOREFERER, (long) 1);
+ curl_easy_setopt(data[i].handle, CURLOPT_ACCEPT_ENCODING, "");
+ curl_easy_setopt(data[i].handle, CURLOPT_USERAGENT, (void*) PACKAGE_STRING);
+
+ curl_multi_add_handle(curlm, data[i].handle);
+ server_url = strtok_r(NULL, url_delim, &strtok_saveptr);
+ }
+
+ /* Query servers in parallel. */
+ int still_running;
+ long loops = 0;
+ do
+ {
+ CURLMcode curl_res;
+
+ if (c->progressfn) /* inform/check progress callback */
+ {
+ loops ++;
+ long pa = loops; /* default params for progress callback */
+ long pb = 0;
+ if (target_handle) /* we've committed to a server; report its download progress */
+ {
+#ifdef CURLINFO_SIZE_DOWNLOAD_T
+ curl_off_t dl;
+ curl_res = curl_easy_getinfo(target_handle,
+ CURLINFO_SIZE_DOWNLOAD_T,
+ &dl);
+ if (curl_res == 0 && dl >= 0)
+ pa = (dl > LONG_MAX ? LONG_MAX : (long)dl);
+#else
+ double dl;
+ curl_res = curl_easy_getinfo(target_handle,
+ CURLINFO_SIZE_DOWNLOAD,
+ &dl);
+ if (curl_res == 0)
+ pa = (dl > LONG_MAX ? LONG_MAX : (long)dl);
+#endif
+
+#ifdef CURLINFO_CURLINFO_CONTENT_LENGTH_DOWNLOAD_T
+ curl_off_t cl;
+ curl_res = curl_easy_getinfo(target_handle,
+ CURLINFO_CONTENT_LENGTH_DOWNLOAD_T,
+ &cl);
+ if (curl_res == 0 && cl >= 0)
+ pb = (cl > LONG_MAX ? LONG_MAX : (long)cl);
+#else
+ double cl;
+ curl_res = curl_easy_getinfo(target_handle,
+ CURLINFO_CONTENT_LENGTH_DOWNLOAD,
+ &cl);
+ if (curl_res == 0)
+ pb = (cl > LONG_MAX ? LONG_MAX : (long)cl);
+#endif
+ }
+
+ if ((*c->progressfn) (c, pa, pb))
+ break;
+ }
+
+ /* Wait 1 second, the minimum DEBUGINFOD_TIMEOUT. */
+ curl_multi_wait(curlm, NULL, 0, 1000, NULL);
+
+ /* If the target file has been found, abort the other queries. */
+ if (target_handle != NULL)
+ for (int i = 0; i < num_urls; i++)
+ if (data[i].handle != target_handle)
+ curl_multi_remove_handle(curlm, data[i].handle);
+
+ curl_res = curl_multi_perform(curlm, &still_running);
+ if (curl_res != CURLM_OK)
+ {
+ switch (curl_res)
+ {
+ case CURLM_CALL_MULTI_PERFORM: continue;
+ case CURLM_OUT_OF_MEMORY: rc = -ENOMEM; break;
+ default: rc = -ENETUNREACH; break;
+ }
+ goto out1;
+ }
+ } while (still_running);
+
+ /* Check whether a query was successful. If so, assign its handle
+ to verified_handle. */
+ int num_msg;
+ rc = -ENOENT;
+ CURL *verified_handle = NULL;
+ do
+ {
+ CURLMsg *msg;
+
+ msg = curl_multi_info_read(curlm, &num_msg);
+ if (msg != NULL && msg->msg == CURLMSG_DONE)
+ {
+ if (msg->data.result != CURLE_OK)
+ {
+ /* Unsucessful query, determine error code. */
+ switch (msg->data.result)
+ {
+ case CURLE_COULDNT_RESOLVE_HOST: rc = -EHOSTUNREACH; break; // no NXDOMAIN
+ case CURLE_URL_MALFORMAT: rc = -EINVAL; break;
+ case CURLE_COULDNT_CONNECT: rc = -ECONNREFUSED; break;
+ case CURLE_REMOTE_ACCESS_DENIED: rc = -EACCES; break;
+ case CURLE_WRITE_ERROR: rc = -EIO; break;
+ case CURLE_OUT_OF_MEMORY: rc = -ENOMEM; break;
+ case CURLE_TOO_MANY_REDIRECTS: rc = -EMLINK; break;
+ case CURLE_SEND_ERROR: rc = -ECONNRESET; break;
+ case CURLE_RECV_ERROR: rc = -ECONNRESET; break;
+ case CURLE_OPERATION_TIMEDOUT: rc = -ETIME; break;
+ default: rc = -ENOENT; break;
+ }
+ }
+ else
+ {
+ /* Query completed without an error. Confirm that the
+ response code is 200 and set verified_handle. */
+ long resp_code = 500;
+ CURLcode curl_res;
+
+ curl_res = curl_easy_getinfo(target_handle,
+ CURLINFO_RESPONSE_CODE,
+ &resp_code);
+
+ if (curl_res == CURLE_OK
+ && resp_code == 200
+ && msg->easy_handle != NULL)
+ {
+ verified_handle = msg->easy_handle;
+ break;
+ }
+ }
+ }
+ } while (num_msg > 0);
+
+ if (verified_handle == NULL)
+ goto out1;
+
+ /* we've got one!!!! */
+ time_t mtime;
+ CURLcode curl_res = curl_easy_getinfo(verified_handle, CURLINFO_FILETIME, (void*) &mtime);
+ if (curl_res != CURLE_OK)
+ mtime = time(NULL); /* fall back to current time */
+
+ struct timeval tvs[2];
+ tvs[0].tv_sec = tvs[1].tv_sec = mtime;
+ tvs[0].tv_usec = tvs[1].tv_usec = 0;
+ (void) futimes (fd, tvs); /* best effort */
+
+ /* rename tmp->real */
+ rc = rename (target_cache_tmppath, target_cache_path);
+ if (rc < 0)
+ {
+ rc = -errno;
+ goto out1;
+ /* Perhaps we need not give up right away; could retry or something ... */
+ }
+
+ /* Success!!!! */
+ for (int i = 0; i < num_urls; i++)
+ curl_easy_cleanup(data[i].handle);
+
+ curl_multi_cleanup (curlm);
+ free (data);
+ free (server_urls);
+ /* don't close fd - we're returning it */
+ /* don't unlink the tmppath; it's already been renamed. */
+ if (path != NULL)
+ *path = strdup(target_cache_path);
+
+ return fd;
+
+/* error exits */
+ out1:
+ for (int i = 0; i < num_urls; i++)
+ curl_easy_cleanup(data[i].handle);
+
+ curl_multi_cleanup(curlm);
+ unlink (target_cache_tmppath);
+ (void) rmdir (target_cache_dir); /* nop if not empty */
+ free(data);
+ close (fd);
+
+ out0:
+ free (server_urls);
+
+ out:
+ return rc;
+}
+
+/* See debuginfod.h */
+debuginfod_client *
+debuginfod_begin (void)
+{
+ debuginfod_client *client;
+ size_t size = sizeof (struct debuginfod_client);
+ client = (debuginfod_client *) malloc (size);
+ if (client != NULL)
+ client->progressfn = NULL;
+ return client;
+}
+
+void
+debuginfod_end (debuginfod_client *client)
+{
+ free (client);
+}
+
+int
+debuginfod_find_debuginfo (debuginfod_client *client,
+ const unsigned char *build_id, int build_id_len,
+ char **path)
+{
+ return debuginfod_query_server(client, build_id, build_id_len,
+ "debuginfo", NULL, path);
+}
+
+
+/* See debuginfod.h */
+int
+debuginfod_find_executable(debuginfod_client *client,
+ const unsigned char *build_id, int build_id_len,
+ char **path)
+{
+ return debuginfod_query_server(client, build_id, build_id_len,
+ "executable", NULL, path);
+}
+
+/* See debuginfod.h */
+int debuginfod_find_source(debuginfod_client *client,
+ const unsigned char *build_id, int build_id_len,
+ const char *filename, char **path)
+{
+ return debuginfod_query_server(client, build_id, build_id_len,
+ "source", filename, path);
+}
+
+
+void
+debuginfod_set_progressfn(debuginfod_client *client,
+ debuginfod_progressfn_t fn)
+{
+ client->progressfn = fn;
+}
+
+
+/* NB: these are thread-unsafe. */
+__attribute__((constructor)) attribute_hidden void libdebuginfod_ctor(void)
+{
+ curl_global_init(CURL_GLOBAL_DEFAULT);
+}
+
+/* NB: this is very thread-unsafe: it breaks other threads that are still in libcurl */
+__attribute__((destructor)) attribute_hidden void libdebuginfod_dtor(void)
+{
+ /* ... so don't do this: */
+ /* curl_global_cleanup(); */
+}
diff --git a/debuginfod/debuginfod-find.c b/debuginfod/debuginfod-find.c
new file mode 100644
index 00000000..8bd3a3db
--- /dev/null
+++ b/debuginfod/debuginfod-find.c
@@ -0,0 +1,145 @@
+/* Command-line frontend for retrieving ELF / DWARF / source files
+ from the debuginfod.
+ Copyright (C) 2019 Red Hat, Inc.
+ This file is part of elfutils.
+
+ This file is free software; you can redistribute it and/or modify
+ it under the terms of the GNU General Public License as published by
+ the Free Software Foundation; either version 3 of the License, or
+ (at your option) any later version.
+
+ elfutils is distributed in the hope that it will be useful, but
+ WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ General Public License for more details.
+
+ You should have received copies of the GNU General Public License and
+ the GNU Lesser General Public License along with this program. If
+ not, see <http://www.gnu.org/licenses/>. */
+
+#include "config.h"
+#include "printversion.h"
+#include "debuginfod.h"
+#include <errno.h>
+#include <stdio.h>
+#include <stdlib.h>
+#include <string.h>
+#include <argp.h>
+
+
+/* Name and version of program. */
+ARGP_PROGRAM_VERSION_HOOK_DEF = print_version;
+
+/* Bug report address. */
+ARGP_PROGRAM_BUG_ADDRESS_DEF = PACKAGE_BUGREPORT;
+
+/* Short description of program. */
+static const char doc[] = N_("Request debuginfo-related content "
+ "from debuginfods listed in $" DEBUGINFOD_URLS_ENV_VAR ".");
+
+/* Strings for arguments in help texts. */
+static const char args_doc[] = N_("debuginfo BUILDID\n"
+ "executable BUILDID\n"
+ "source BUILDID /FILENAME");
+
+/* Definitions of arguments for argp functions. */
+static const struct argp_option options[] =
+ {
+ { "verbose", 'v', NULL, 0, "Increase verbosity.", 0 },
+ { NULL, 0, NULL, 0, NULL, 0 }
+ };
+
+/* debuginfod connection handle. */
+static debuginfod_client *client;
+
+int progressfn(debuginfod_client *c __attribute__((__unused__)),
+ long a, long b)
+{
+ fprintf (stderr, "Progress %ld / %ld\n", a, b);
+ return 0;
+}
+
+
+static error_t parse_opt (int key, char *arg, struct argp_state *state)
+{
+ (void) arg;
+ (void) state;
+ switch (key)
+ {
+ case 'v': debuginfod_set_progressfn (client, & progressfn); break;
+ default: return ARGP_ERR_UNKNOWN;
+ }
+ return 0;
+}
+
+
+/* Data structure to communicate with argp functions. */
+static struct argp argp =
+ {
+ options, parse_opt, args_doc, doc, NULL, NULL, NULL
+ };
+
+
+
+int
+main(int argc, char** argv)
+{
+ client = debuginfod_begin ();
+ if (client == NULL)
+ {
+ fprintf(stderr, "Couldn't create debuginfod client context\n");
+ return 1;
+ }
+
+ int remaining;
+ (void) argp_parse (&argp, argc, argv, ARGP_IN_ORDER|ARGP_NO_ARGS, &remaining, NULL);
+
+ if (argc < 2 || remaining+1 == argc) /* no arguments or at least two non-option words */
+ {
+ argp_help (&argp, stderr, ARGP_HELP_USAGE, argv[0]);
+ return 1;
+ }
+
+ int rc;
+ char *cache_name;
+
+ /* Check whether FILETYPE is valid and call the appropriate
+ debuginfod_find_* function. If FILETYPE is "source"
+ then ensure a FILENAME was also supplied as an argument. */
+ if (strcmp(argv[remaining], "debuginfo") == 0)
+ rc = debuginfod_find_debuginfo(client,
+ (unsigned char *)argv[remaining+1], 0,
+ &cache_name);
+ else if (strcmp(argv[remaining], "executable") == 0)
+ rc = debuginfod_find_executable(client,
+ (unsigned char *)argv[remaining+1], 0,
+ &cache_name);
+ else if (strcmp(argv[remaining], "source") == 0)
+ {
+ if (remaining+2 == argc || argv[3][0] != '/')
+ {
+ fprintf(stderr, "If FILETYPE is \"source\" then absolute /FILENAME must be given\n");
+ return 1;
+ }
+ rc = debuginfod_find_source(client, (unsigned char *)argv[remaining+1],
+ 0, argv[remaining+2], &cache_name);
+ }
+ else
+ {
+ argp_help (&argp, stderr, ARGP_HELP_USAGE, argv[0]);
+ return 1;
+ }
+
+ if (rc < 0)
+ {
+ fprintf(stderr, "Server query failed: %s\n", strerror(-rc));
+ return 1;
+ }
+
+ printf("%s\n", cache_name);
+
+ free (cache_name);
+ debuginfod_end (client);
+
+ return 0;
+}
diff --git a/debuginfod/debuginfod.cxx b/debuginfod/debuginfod.cxx
new file mode 100644
index 00000000..aa7ffcf6
--- /dev/null
+++ b/debuginfod/debuginfod.cxx
@@ -0,0 +1,2688 @@
+/* Debuginfo-over-http server.
+ Copyright (C) 2019 Red Hat, Inc.
+ This file is part of elfutils.
+
+ This file is free software; you can redistribute it and/or modify
+ it under the terms of the GNU General Public License as published by
+ the Free Software Foundation; either version 3 of the License, or
+ (at your option) any later version.
+
+ elfutils is distributed in the hope that it will be useful, but
+ WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ GNU General Public License for more details.
+
+ You should have received a copy of the GNU General Public License
+ along with this program. If not, see <http://www.gnu.org/licenses/>. */
+
+
+/* cargo-cult from libdwfl linux-kernel-modules.c */
+/* In case we have a bad fts we include this before config.h because it
+ can't handle _FILE_OFFSET_BITS.
+ Everything we need here is fine if its declarations just come first.
+ Also, include sys/types.h before fts. On some systems fts.h is not self
+ contained. */
+#ifdef BAD_FTS
+ #include <sys/types.h>
+ #include <fts.h>
+#endif
+
+#ifdef HAVE_CONFIG_H
+ #include "config.h"
+#endif
+
+extern "C" {
+#include "printversion.h"
+}
+
+#include "debuginfod.h"
+#include <dwarf.h>
+
+#include <argp.h>
+#ifdef __GNUC__
+#undef __attribute__ /* glibc bug - rhbz 1763325 */
+#endif
+
+#include <unistd.h>
+#include <stdlib.h>
+#include <error.h>
+// #include <libintl.h> // not until it supports C++ << better
+#include <locale.h>
+#include <pthread.h>
+#include <signal.h>
+#include <sys/stat.h>
+#include <sys/time.h>
+#include <unistd.h>
+#include <fcntl.h>
+#include <netdb.h>
+
+
+/* If fts.h is included before config.h, its indirect inclusions may not
+ give us the right LFS aliases of these functions, so map them manually. */
+#ifdef BAD_FTS
+ #ifdef _FILE_OFFSET_BITS
+ #define open open64
+ #define fopen fopen64
+ #endif
+#else
+ #include <sys/types.h>
+ #include <fts.h>
+#endif
+
+#include <cstring>
+#include <vector>
+#include <set>
+#include <map>
+#include <string>
+#include <iostream>
+#include <iomanip>
+#include <ostream>
+#include <sstream>
+#include <mutex>
+#include <condition_variable>
+#include <thread>
+// #include <regex> // on rhel7 gcc 4.8, not competent
+#include <regex.h>
+// #include <algorithm>
+using namespace std;
+
+#include <gelf.h>
+#include <libdwelf.h>
+
+#include <microhttpd.h>
+#include <curl/curl.h>
+#include <archive.h>
+#include <archive_entry.h>
+#include <sqlite3.h>
+
+#ifdef __linux__
+#include <sys/syscall.h>
+#endif
+
+#ifdef __linux__
+#define tid() syscall(SYS_gettid)
+#else
+#define tid() pthread_self()
+#endif
+
+
+// Roll this identifier for every sqlite schema incompatiblity.
+#define BUILDIDS "buildids9"
+
+#if SQLITE_VERSION_NUMBER >= 3008000
+#define WITHOUT_ROWID "without rowid"
+#else
+#define WITHOUT_ROWID ""
+#endif
+
+static const char DEBUGINFOD_SQLITE_DDL[] =
+ "pragma foreign_keys = on;\n"
+ "pragma synchronous = 0;\n" // disable fsync()s - this cache is disposable across a machine crash
+ "pragma journal_mode = wal;\n" // https://sqlite.org/wal.html
+ "pragma wal_checkpoint = truncate;\n" // clean out any preexisting wal file
+ "pragma journal_size_limit = 0;\n" // limit steady state file (between grooming, which also =truncate's)
+ "pragma auto_vacuum = incremental;\n" // https://sqlite.org/pragma.html
+ "pragma busy_timeout = 1000;\n" // https://sqlite.org/pragma.html
+ // NB: all these are overridable with -D option
+
+ // Normalization table for interning file names
+ "create table if not exists " BUILDIDS "_files (\n"
+ " id integer primary key not null,\n"
+ " name text unique not null\n"
+ " );\n"
+ // Normalization table for interning buildids
+ "create table if not exists " BUILDIDS "_buildids (\n"
+ " id integer primary key not null,\n"
+ " hex text unique not null);\n"
+ // Track the completion of scanning of a given file & sourcetype at given time
+ "create table if not exists " BUILDIDS "_file_mtime_scanned (\n"
+ " mtime integer not null,\n"
+ " file integer not null,\n"
+ " size integer not null,\n" // in bytes
+ " sourcetype text(1) not null\n"
+ " check (sourcetype IN ('F', 'R')),\n"
+ " foreign key (file) references " BUILDIDS "_files(id) on update cascade on delete cascade,\n"
+ " primary key (file, mtime, sourcetype)\n"
+ " ) " WITHOUT_ROWID ";\n"
+ "create table if not exists " BUILDIDS "_f_de (\n"
+ " buildid integer not null,\n"
+ " debuginfo_p integer not null,\n"
+ " executable_p integer not null,\n"
+ " file integer not null,\n"
+ " mtime integer not null,\n"
+ " foreign key (file) references " BUILDIDS "_files(id) on update cascade on delete cascade,\n"
+ " foreign key (buildid) references " BUILDIDS "_buildids(id) on update cascade on delete cascade,\n"
+ " primary key (buildid, file, mtime)\n"
+ " ) " WITHOUT_ROWID ";\n"
+ "create table if not exists " BUILDIDS "_f_s (\n"
+ " buildid integer not null,\n"
+ " artifactsrc integer not null,\n"
+ " file integer not null,\n" // NB: not necessarily entered into _mtime_scanned
+ " mtime integer not null,\n"
+ " foreign key (file) references " BUILDIDS "_files(id) on update cascade on delete cascade,\n"
+ " foreign key (artifactsrc) references " BUILDIDS "_files(id) on update cascade on delete cascade,\n"
+ " foreign key (buildid) references " BUILDIDS "_buildids(id) on update cascade on delete cascade,\n"
+ " primary key (buildid, artifactsrc, file, mtime)\n"
+ " ) " WITHOUT_ROWID ";\n"
+ "create table if not exists " BUILDIDS "_r_de (\n"
+ " buildid integer not null,\n"
+ " debuginfo_p integer not null,\n"
+ " executable_p integer not null,\n"
+ " file integer not null,\n"
+ " mtime integer not null,\n"
+ " content integer not null,\n"
+ " foreign key (file) references " BUILDIDS "_files(id) on update cascade on delete cascade,\n"
+ " foreign key (content) references " BUILDIDS "_files(id) on update cascade on delete cascade,\n"
+ " foreign key (buildid) references " BUILDIDS "_buildids(id) on update cascade on delete cascade,\n"
+ " primary key (buildid, debuginfo_p, executable_p, file, content, mtime)\n"
+ " ) " WITHOUT_ROWID ";\n"
+ "create table if not exists " BUILDIDS "_r_sref (\n" // outgoing dwarf sourcefile references from rpm
+ " buildid integer not null,\n"
+ " artifactsrc integer not null,\n"
+ " foreign key (artifactsrc) references " BUILDIDS "_files(id) on update cascade on delete cascade,\n"
+ " foreign key (buildid) references " BUILDIDS "_buildids(id) on update cascade on delete cascade,\n"
+ " primary key (buildid, artifactsrc)\n"
+ " ) " WITHOUT_ROWID ";\n"
+ "create table if not exists " BUILDIDS "_r_sdef (\n" // rpm contents that may satisfy sref
+ " file integer not null,\n"
+ " mtime integer not null,\n"
+ " content integer not null,\n"
+ " foreign key (file) references " BUILDIDS "_files(id) on update cascade on delete cascade,\n"
+ " foreign key (content) references " BUILDIDS "_files(id) on update cascade on delete cascade,\n"
+ " primary key (content, file, mtime)\n"
+ " ) " WITHOUT_ROWID ";\n"
+ // create views to glue together some of the above tables, for webapi D queries
+ "create view if not exists " BUILDIDS "_query_d as \n"
+ "select\n"
+ " b.hex as buildid, n.mtime, 'F' as sourcetype, f0.name as source0, n.mtime as mtime, null as source1\n"
+ " from " BUILDIDS "_buildids b, " BUILDIDS "_files f0, " BUILDIDS "_f_de n\n"
+ " where b.id = n.buildid and f0.id = n.file and n.debuginfo_p = 1\n"
+ "union all select\n"
+ " b.hex as buildid, n.mtime, 'R' as sourcetype, f0.name as source0, n.mtime as mtime, f1.name as source1\n"
+ " from " BUILDIDS "_buildids b, " BUILDIDS "_files f0, " BUILDIDS "_files f1, " BUILDIDS "_r_de n\n"
+ " where b.id = n.buildid and f0.id = n.file and f1.id = n.content and n.debuginfo_p = 1\n"
+ ";"
+ // ... and for E queries
+ "create view if not exists " BUILDIDS "_query_e as \n"
+ "select\n"
+ " b.hex as buildid, n.mtime, 'F' as sourcetype, f0.name as source0, n.mtime as mtime, null as source1\n"
+ " from " BUILDIDS "_buildids b, " BUILDIDS "_files f0, " BUILDIDS "_f_de n\n"
+ " where b.id = n.buildid and f0.id = n.file and n.executable_p = 1\n"
+ "union all select\n"
+ " b.hex as buildid, n.mtime, 'R' as sourcetype, f0.name as source0, n.mtime as mtime, f1.name as source1\n"
+ " from " BUILDIDS "_buildids b, " BUILDIDS "_files f0, " BUILDIDS "_files f1, " BUILDIDS "_r_de n\n"
+ " where b.id = n.buildid and f0.id = n.file and f1.id = n.content and n.executable_p = 1\n"
+ ";"
+ // ... and for S queries
+ "create view if not exists " BUILDIDS "_query_s as \n"
+ "select\n"
+ " b.hex as buildid, fs.name as artifactsrc, 'F' as sourcetype, f0.name as source0, n.mtime as mtime, null as source1, null as source0ref\n"
+ " from " BUILDIDS "_buildids b, " BUILDIDS "_files f0, " BUILDIDS "_files fs, " BUILDIDS "_f_s n\n"
+ " where b.id = n.buildid and f0.id = n.file and fs.id = n.artifactsrc\n"
+ "union all select\n"
+ " b.hex as buildid, f1.name as artifactsrc, 'R' as sourcetype, f0.name as source0, sd.mtime as mtime, f1.name as source1, fsref.name as source0ref\n"
+ " from " BUILDIDS "_buildids b, " BUILDIDS "_files f0, " BUILDIDS "_files f1, " BUILDIDS "_files fsref, "
+ " " BUILDIDS "_r_sdef sd, " BUILDIDS "_r_sref sr, " BUILDIDS "_r_de sde\n"
+ " where b.id = sr.buildid and f0.id = sd.file and fsref.id = sde.file and f1.id = sd.content\n"
+ " and sr.artifactsrc = sd.content and sde.buildid = sr.buildid\n"
+ ";"
+ // and for startup overview counts
+ "drop view if exists " BUILDIDS "_stats;\n"
+ "create view if not exists " BUILDIDS "_stats as\n"
+ " select 'file d/e' as label,count(*) as quantity from " BUILDIDS "_f_de\n"
+ "union all select 'file s',count(*) from " BUILDIDS "_f_s\n"
+ "union all select 'rpm d/e',count(*) from " BUILDIDS "_r_de\n"
+ "union all select 'rpm sref',count(*) from " BUILDIDS "_r_sref\n"
+ "union all select 'rpm sdef',count(*) from " BUILDIDS "_r_sdef\n"
+ "union all select 'buildids',count(*) from " BUILDIDS "_buildids\n"
+ "union all select 'filenames',count(*) from " BUILDIDS "_files\n"
+ "union all select 'files scanned (#)',count(*) from " BUILDIDS "_file_mtime_scanned\n"
+ "union all select 'files scanned (mb)',coalesce(sum(size)/1024/1024,0) from " BUILDIDS "_file_mtime_scanned\n"
+#if SQLITE_VERSION_NUMBER >= 3016000
+ "union all select 'index db size (mb)',page_count*page_size/1024/1024 as size FROM pragma_page_count(), pragma_page_size()\n"
+#endif
+ ";\n"
+
+// schema change history & garbage collection
+//
+// XXX: we could have migration queries here to bring prior-schema
+// data over instead of just dropping it.
+//
+// buildids9: widen the mtime_scanned table
+ "" // <<< we are here
+// buildids8: slim the sref table
+ "drop table if exists buildids8_f_de;\n"
+ "drop table if exists buildids8_f_s;\n"
+ "drop table if exists buildids8_r_de;\n"
+ "drop table if exists buildids8_r_sref;\n"
+ "drop table if exists buildids8_r_sdef;\n"
+ "drop table if exists buildids8_file_mtime_scanned;\n"
+ "drop table if exists buildids8_files;\n"
+ "drop table if exists buildids8_buildids;\n"
+// buildids7: separate _norm table into dense subtype tables
+ "drop table if exists buildids7_f_de;\n"
+ "drop table if exists buildids7_f_s;\n"
+ "drop table if exists buildids7_r_de;\n"
+ "drop table if exists buildids7_r_sref;\n"
+ "drop table if exists buildids7_r_sdef;\n"
+ "drop table if exists buildids7_file_mtime_scanned;\n"
+ "drop table if exists buildids7_files;\n"
+ "drop table if exists buildids7_buildids;\n"
+// buildids6: drop bolo/rfolo again, represent sources / rpmcontents in main table
+ "drop table if exists buildids6_norm;\n"
+ "drop table if exists buildids6_files;\n"
+ "drop table if exists buildids6_buildids;\n"
+ "drop view if exists buildids6;\n"
+// buildids5: redefine srcfile1 column to be '.'-less (for rpms)
+ "drop table if exists buildids5_norm;\n"
+ "drop table if exists buildids5_files;\n"
+ "drop table if exists buildids5_buildids;\n"
+ "drop table if exists buildids5_bolo;\n"
+ "drop table if exists buildids5_rfolo;\n"
+ "drop view if exists buildids5;\n"
+// buildids4: introduce rpmfile RFOLO
+ "drop table if exists buildids4_norm;\n"
+ "drop table if exists buildids4_files;\n"
+ "drop table if exists buildids4_buildids;\n"
+ "drop table if exists buildids4_bolo;\n"
+ "drop table if exists buildids4_rfolo;\n"
+ "drop view if exists buildids4;\n"
+// buildids3*: split out srcfile BOLO
+ "drop table if exists buildids3_norm;\n"
+ "drop table if exists buildids3_files;\n"
+ "drop table if exists buildids3_buildids;\n"
+ "drop table if exists buildids3_bolo;\n"
+ "drop view if exists buildids3;\n"
+// buildids2: normalized buildid and filenames into interning tables;
+ "drop table if exists buildids2_norm;\n"
+ "drop table if exists buildids2_files;\n"
+ "drop table if exists buildids2_buildids;\n"
+ "drop view if exists buildids2;\n"
+ // buildids1: made buildid and artifacttype NULLable, to represent cached-negative
+// lookups from sources, e.g. files or rpms that contain no buildid-indexable content
+ "drop table if exists buildids1;\n"
+// buildids: original
+ "drop table if exists buildids;\n"
+ ;
+
+static const char DEBUGINFOD_SQLITE_CLEANUP_DDL[] =
+ "pragma wal_checkpoint = truncate;\n" // clean out any preexisting wal file
+ ;
+
+
+
+
+/* Name and version of program. */
+/* ARGP_PROGRAM_VERSION_HOOK_DEF = print_version; */ // not this simple for C++
+
+/* Bug report address. */
+ARGP_PROGRAM_BUG_ADDRESS_DEF = PACKAGE_BUGREPORT;
+
+/* Definitions of arguments for argp functions. */
+static const struct argp_option options[] =
+ {
+ { NULL, 0, NULL, 0, "Scanners:", 1 },
+ { "scan-file-dir", 'F', NULL, 0, "Enable ELF/DWARF file scanning threads.", 0 },
+ { "scan-rpm-dir", 'R', NULL, 0, "Enable RPM scanning threads.", 0 },
+ // "source-oci-imageregistry" ...
+
+ { NULL, 0, NULL, 0, "Options:", 2 },
+ { "logical", 'L', NULL, 0, "Follow symlinks, default=ignore.", 0 },
+ { "rescan-time", 't', "SECONDS", 0, "Number of seconds to wait between rescans, 0=disable.", 0 },
+ { "groom-time", 'g', "SECONDS", 0, "Number of seconds to wait between database grooming, 0=disable.", 0 },
+ { "maxigroom", 'G', NULL, 0, "Run a complete database groom/shrink pass at startup.", 0 },
+ { "concurrency", 'c', "NUM", 0, "Limit scanning thread concurrency to NUM.", 0 },
+ { "include", 'I', "REGEX", 0, "Include files matching REGEX, default=all.", 0 },
+ { "exclude", 'X', "REGEX", 0, "Exclude files matching REGEX, default=none.", 0 },
+ { "port", 'p', "NUM", 0, "HTTP port to listen on, default 8002.", 0 },
+ { "database", 'd', "FILE", 0, "Path to sqlite database.", 0 },
+ { "ddl", 'D', "SQL", 0, "Apply extra sqlite ddl/pragma to connection.", 0 },
+ { "verbose", 'v', NULL, 0, "Increase verbosity.", 0 },
+
+ { NULL, 0, NULL, 0, NULL, 0 }
+ };
+
+/* Short description of program. */
+static const char doc[] = "Serve debuginfo-related content across HTTP from files under PATHs.";
+
+/* Strings for arguments in help texts. */
+static const char args_doc[] = "[PATH ...]";
+
+/* Prototype for option handler. */
+static error_t parse_opt (int key, char *arg, struct argp_state *state);
+
+/* Data structure to communicate with argp functions. */
+static struct argp argp =
+ {
+ options, parse_opt, args_doc, doc, NULL, NULL, NULL
+ };
+
+
+static string db_path;
+static sqlite3 *db;
+static unsigned verbose;
+static volatile sig_atomic_t interrupted = 0;
+static volatile sig_atomic_t sigusr1 = 0;
+static volatile sig_atomic_t sigusr2 = 0;
+static unsigned http_port = 8002;
+static unsigned rescan_s = 300;
+static unsigned groom_s = 86400;
+static unsigned maxigroom = false;
+static unsigned concurrency = std::thread::hardware_concurrency() ?: 1;
+static set<string> source_paths;
+static bool scan_files = false;
+static bool scan_rpms = false;
+static vector<string> extra_ddl;
+static regex_t file_include_regex;
+static regex_t file_exclude_regex;
+static bool traverse_logical;
+
+static void set_metric(const string& key, int64_t value);
+// static void inc_metric(const string& key);
+static void set_metric(const string& metric,
+ const string& lname, const string& lvalue,
+ int64_t value);
+static void inc_metric(const string& metric,
+ const string& lname, const string& lvalue);
+static void add_metric(const string& metric,
+ const string& lname, const string& lvalue,
+ int64_t value);
+
+/* Handle program arguments. */
+static error_t
+parse_opt (int key, char *arg,
+ struct argp_state *state __attribute__ ((unused)))
+{
+ int rc;
+ switch (key)
+ {
+ case 'v': verbose ++; break;
+ case 'd': db_path = string(arg); break;
+ case 'p': http_port = (unsigned) atoi(arg);
+ if (http_port > 65535) argp_failure(state, 1, EINVAL, "port number");
+ break;
+ case 'F': scan_files = true; break;
+ case 'R': scan_rpms = true; break;
+ case 'L':
+ traverse_logical = true;
+ break;
+ case 'D': extra_ddl.push_back(string(arg)); break;
+ case 't':
+ rescan_s = (unsigned) atoi(arg);
+ break;
+ case 'g':
+ groom_s = (unsigned) atoi(arg);
+ break;
+ case 'G':
+ maxigroom = true;
+ break;
+ case 'c':
+ concurrency = (unsigned) atoi(arg);
+ if (concurrency < 1) concurrency = 1;
+ break;
+ case 'I':
+ // NB: no problem with unconditional free here - an earlier failed regcomp would exit program
+ regfree (&file_include_regex);
+ rc = regcomp (&file_include_regex, arg, REG_EXTENDED|REG_NOSUB);
+ if (rc != 0)
+ argp_failure(state, 1, EINVAL, "regular expession");
+ break;
+ case 'X':
+ regfree (&file_exclude_regex);
+ rc = regcomp (&file_exclude_regex, arg, REG_EXTENDED|REG_NOSUB);
+ if (rc != 0)
+ argp_failure(state, 1, EINVAL, "regular expession");
+ break;
+ case ARGP_KEY_ARG:
+ source_paths.insert(string(arg));
+ break;
+ // case 'h': argp_state_help (state, stderr, ARGP_HELP_LONG|ARGP_HELP_EXIT_OK);
+ default: return ARGP_ERR_UNKNOWN;
+ }
+
+ return 0;
+}
+
+
+////////////////////////////////////////////////////////////////////////
+
+
+// represent errors that may get reported to an ostream and/or a libmicrohttpd connection
+
+struct reportable_exception
+{
+ int code;
+ string message;
+
+ reportable_exception(int c, const string& m): code(c), message(m) {}
+ reportable_exception(const string& m): code(503), message(m) {}
+ reportable_exception(): code(503), message() {}
+
+ void report(ostream& o) const; // defined under obatched() class below
+
+ int mhd_send_response(MHD_Connection* c) const {
+ MHD_Response* r = MHD_create_response_from_buffer (message.size(),
+ (void*) message.c_str(),
+ MHD_RESPMEM_MUST_COPY);
+ MHD_add_response_header (r, "Content-Type", "text/plain");
+ int rc = MHD_queue_response (c, code, r);
+ MHD_destroy_response (r);
+ return rc;
+ }
+};
+
+
+struct sqlite_exception: public reportable_exception
+{
+ sqlite_exception(int rc, const string& msg):
+ reportable_exception(string("sqlite3 error: ") + msg + ": " + string(sqlite3_errstr(rc) ?: "?")) {}
+};
+
+struct libc_exception: public reportable_exception
+{
+ libc_exception(int rc, const string& msg):
+ reportable_exception(string("libc error: ") + msg + ": " + string(strerror(rc) ?: "?")) {}
+};
+
+
+struct archive_exception: public reportable_exception
+{
+ archive_exception(const string& msg):
+ reportable_exception(string("libarchive error: ") + msg) {}
+ archive_exception(struct archive* a, const string& msg):
+ reportable_exception(string("libarchive error: ") + msg + ": " + string(archive_error_string(a) ?: "?")) {}
+};
+
+
+struct elfutils_exception: public reportable_exception
+{
+ elfutils_exception(int rc, const string& msg):
+ reportable_exception(string("elfutils error: ") + msg + ": " + string(elf_errmsg(rc) ?: "?")) {}
+};
+
+
+////////////////////////////////////////////////////////////////////////
+
+// a c++ counting-semaphore class ... since we're c++11 not c++20
+
+class semaphore
+{
+public:
+ semaphore (unsigned c=1): count(c) {}
+ inline void notify () {
+ unique_lock<mutex> lock(mtx);
+ count++;
+ cv.notify_one();
+ }
+ inline void wait() {
+ unique_lock<mutex> lock(mtx);
+ while (count == 0)
+ cv.wait(lock);
+ count--;
+ }
+private:
+ mutex mtx;
+ condition_variable cv;
+ unsigned count;
+};
+
+
+class semaphore_borrower
+{
+public:
+ semaphore_borrower(semaphore* s): sem(s) { sem->wait(); }
+ ~semaphore_borrower() { sem->notify(); }
+private:
+ semaphore* sem;
+};
+
+
+////////////////////////////////////////////////////////////////////////
+
+
+// Print a standard timestamp.
+static ostream&
+timestamp (ostream &o)
+{
+ char datebuf[80];
+ char *now2 = NULL;
+ time_t now_t = time(NULL);
+ struct tm *now = gmtime (&now_t);
+ if (now)
+ {
+ (void) strftime (datebuf, sizeof (datebuf), "%c", now);
+ now2 = datebuf;
+ }
+
+ return o << "[" << (now2 ? now2 : "") << "] "
+ << "(" << getpid () << "/" << tid() << "): ";
+}
+
+
+// A little class that impersonates an ostream to the extent that it can
+// take << streaming operations. It batches up the bits into an internal
+// stringstream until it is destroyed; then flushes to the original ostream.
+// It adds a timestamp
+class obatched
+{
+private:
+ ostream& o;
+ stringstream stro;
+ static mutex lock;
+public:
+ obatched(ostream& oo, bool timestamp_p = true): o(oo)
+ {
+ if (timestamp_p)
+ timestamp(stro);
+ }
+ ~obatched()
+ {
+ unique_lock<mutex> do_not_cross_the_streams(obatched::lock);
+ o << stro.str();
+ o.flush();
+ }
+ operator ostream& () { return stro; }
+ template <typename T> ostream& operator << (const T& t) { stro << t; return stro; }
+};
+mutex obatched::lock; // just the one, since cout/cerr iostreams are not thread-safe
+
+
+void reportable_exception::report(ostream& o) const {
+ obatched(o) << message << endl;
+}
+
+
+////////////////////////////////////////////////////////////////////////
+
+
+// RAII style sqlite prepared-statement holder that matches { } block lifetime
+
+struct sqlite_ps
+{
+private:
+ sqlite3* db;
+ const string nickname;
+ const string sql;
+ sqlite3_stmt *pp;
+
+ sqlite_ps(const sqlite_ps&); // make uncopyable
+ sqlite_ps& operator=(const sqlite_ps &); // make unassignable
+
+public:
+ sqlite_ps (sqlite3* d, const string& n, const string& s): db(d), nickname(n), sql(s) {
+ if (verbose > 4)
+ obatched(clog) << nickname << " prep " << sql << endl;
+ int rc = sqlite3_prepare_v2 (db, sql.c_str(), -1 /* to \0 */, & this->pp, NULL);
+ if (rc != SQLITE_OK)
+ throw sqlite_exception(rc, "prepare " + sql);
+ }
+
+ sqlite_ps& reset()
+ {
+ sqlite3_reset(this->pp);
+ return *this;
+ }
+
+ sqlite_ps& bind(int parameter, const string& str)
+ {
+ if (verbose > 4)
+ obatched(clog) << nickname << " bind " << parameter << "=" << str << endl;
+ int rc = sqlite3_bind_text (this->pp, parameter, str.c_str(), -1, SQLITE_TRANSIENT);
+ if (rc != SQLITE_OK)
+ throw sqlite_exception(rc, "sqlite3 bind");
+ return *this;
+ }
+
+ sqlite_ps& bind(int parameter, int64_t value)
+ {
+ if (verbose > 4)
+ obatched(clog) << nickname << " bind " << parameter << "=" << value << endl;
+ int rc = sqlite3_bind_int64 (this->pp, parameter, value);
+ if (rc != SQLITE_OK)
+ throw sqlite_exception(rc, "sqlite3 bind");
+ return *this;
+ }
+
+ sqlite_ps& bind(int parameter)
+ {
+ if (verbose > 4)
+ obatched(clog) << nickname << " bind " << parameter << "=" << "NULL" << endl;
+ int rc = sqlite3_bind_null (this->pp, parameter);
+ if (rc != SQLITE_OK)
+ throw sqlite_exception(rc, "sqlite3 bind");
+ return *this;
+ }
+
+
+ void step_ok_done() {
+ int rc = sqlite3_step (this->pp);
+ if (verbose > 4)
+ obatched(clog) << nickname << " step-ok-done(" << sqlite3_errstr(rc) << ") " << sql << endl;
+ if (rc != SQLITE_OK && rc != SQLITE_DONE && rc != SQLITE_ROW)
+ throw sqlite_exception(rc, "sqlite3 step");
+ (void) sqlite3_reset (this->pp);
+ }
+
+
+ int step() {
+ int rc = sqlite3_step (this->pp);
+ if (verbose > 4)
+ obatched(clog) << nickname << " step(" << sqlite3_errstr(rc) << ") " << sql << endl;
+ return rc;
+ }
+
+
+
+ ~sqlite_ps () { sqlite3_finalize (this->pp); }
+ operator sqlite3_stmt* () { return this->pp; }
+};
+
+
+////////////////////////////////////////////////////////////////////////
+
+// RAII style templated autocloser
+
+template <class Payload, class Ignore>
+struct defer_dtor
+{
+public:
+ typedef Ignore (*dtor_fn) (Payload);
+
+private:
+ Payload p;
+ dtor_fn fn;
+
+public:
+ defer_dtor(Payload _p, dtor_fn _fn): p(_p), fn(_fn) {}
+ ~defer_dtor() { (void) (*fn)(p); }
+
+private:
+ defer_dtor(const defer_dtor<Payload,Ignore>&); // make uncopyable
+ defer_dtor& operator=(const defer_dtor<Payload,Ignore> &); // make unassignable
+};
+
+
+
+////////////////////////////////////////////////////////////////////////
+
+
+
+
+
+static string
+conninfo (struct MHD_Connection * conn)
+{
+ char hostname[256]; // RFC1035
+ char servname[256];
+ int sts = -1;
+
+ if (conn == 0)
+ return "internal";
+
+ /* Look up client address data. */
+ const union MHD_ConnectionInfo *u = MHD_get_connection_info (conn,
+ MHD_CONNECTION_INFO_CLIENT_ADDRESS);
+ struct sockaddr *so = u ? u->client_addr : 0;
+
+ if (so && so->sa_family == AF_INET) {
+ sts = getnameinfo (so, sizeof (struct sockaddr_in), hostname, sizeof (hostname), servname,
+ sizeof (servname), NI_NUMERICHOST | NI_NUMERICSERV);
+ } else if (so && so->sa_family == AF_INET6) {
+ sts = getnameinfo (so, sizeof (struct sockaddr_in6), hostname, sizeof (hostname),
+ servname, sizeof (servname), NI_NUMERICHOST | NI_NUMERICSERV);
+ }
+ if (sts != 0) {
+ hostname[0] = servname[0] = '\0';
+ }
+
+ return string(hostname) + string(":") + string(servname);
+}
+
+
+
+////////////////////////////////////////////////////////////////////////
+
+static void
+add_mhd_last_modified (struct MHD_Response *resp, time_t mtime)
+{
+ struct tm *now = gmtime (&mtime);
+ if (now != NULL)
+ {
+ char datebuf[80];
+ size_t rc = strftime (datebuf, sizeof (datebuf), "%a, %d %b %Y %T GMT", now);
+ if (rc > 0 && rc < sizeof (datebuf))
+ (void) MHD_add_response_header (resp, "Last-Modified", datebuf);
+ }
+
+ (void) MHD_add_response_header (resp, "Cache-Control", "public");
+}
+
+
+
+static struct MHD_Response*
+handle_buildid_f_match (int64_t b_mtime,
+ const string& b_source0,
+ int *result_fd)
+{
+ int fd = open(b_source0.c_str(), O_RDONLY);
+ if (fd < 0)
+ {
+ if (verbose)
+ obatched(clog) << "cannot open " << b_source0 << endl;
+ // if still missing, a periodic groom pass will delete this buildid record
+ return 0;
+ }
+
+ // NB: use manual close(2) in error case instead of defer_dtor, because
+ // in the normal case, we want to hand the fd over to libmicrohttpd for
+ // file transfer.
+
+ struct stat s;
+ int rc = fstat(fd, &s);
+ if (rc < 0)
+ {
+ if (verbose)
+ clog << "cannot fstat " << b_source0 << endl;
+ close(fd);
+ return 0;
+ }
+
+ if ((int64_t) s.st_mtime != b_mtime)
+ {
+ if (verbose)
+ obatched(clog) << "mtime mismatch for " << b_source0 << endl;
+ close(fd);
+ return 0;
+ }
+
+ inc_metric ("http_responses_total","result","file");
+ struct MHD_Response* r = MHD_create_response_from_fd ((uint64_t) s.st_size, fd);
+ if (r == 0)
+ {
+ if (verbose)
+ obatched(clog) << "cannot create fd-response for " << b_source0 << endl;
+ close(fd);
+ }
+ else
+ {
+ MHD_add_response_header (r, "Content-Type", "application/octet-stream");
+ add_mhd_last_modified (r, s.st_mtime);
+ if (verbose > 1)
+ obatched(clog) << "serving file " << b_source0 << endl;
+ /* libmicrohttpd will close it. */
+ if (result_fd)
+ *result_fd = fd;
+ }
+
+ return r;
+}
+
+
+// quote all questionable characters of str for safe passage through a sh -c expansion.
+static string
+shell_escape(const string& str)
+{
+ string y;
+ for (auto&& x : str)
+ {
+ if (! isalnum(x) && x != '/')
+ y += "\\";
+ y += x;
+ }
+ return y;
+}
+
+
+static struct MHD_Response*
+handle_buildid_r_match (int64_t b_mtime,
+ const string& b_source0,
+ const string& b_source1,
+ int *result_fd)
+{
+ struct stat fs;
+ int rc = stat (b_source0.c_str(), &fs);
+ if (rc != 0)
+ throw libc_exception (errno, string("stat ") + b_source0);
+
+ if ((int64_t) fs.st_mtime != b_mtime)
+ {
+ if (verbose)
+ obatched(clog) << "mtime mismatch for " << b_source0 << endl;
+ return 0;
+ }
+
+ string popen_cmd = string("rpm2cpio " + shell_escape(b_source0));
+ FILE* fp = popen (popen_cmd.c_str(), "r"); // "e" O_CLOEXEC?
+ if (fp == NULL)
+ throw libc_exception (errno, string("popen ") + popen_cmd);
+ defer_dtor<FILE*,int> fp_closer (fp, pclose);
+
+ struct archive *a;
+ a = archive_read_new();
+ if (a == NULL)
+ throw archive_exception("cannot create archive reader");
+ defer_dtor<struct archive*,int> archive_closer (a, archive_read_free);
+
+ rc = archive_read_support_format_cpio(a);
+ if (rc != ARCHIVE_OK)
+ throw archive_exception(a, "cannot select cpio format");
+ rc = archive_read_support_filter_all(a);
+ if (rc != ARCHIVE_OK)
+ throw archive_exception(a, "cannot select all filters");
+
+ rc = archive_read_open_FILE (a, fp);
+ if (rc != ARCHIVE_OK)
+ throw archive_exception(a, "cannot open archive from rpm2cpio pipe");
+
+ while(1) // parse cpio archive entries
+ {
+ struct archive_entry *e;
+ rc = archive_read_next_header (a, &e);
+ if (rc != ARCHIVE_OK)
+ break;
+
+ if (! S_ISREG(archive_entry_mode (e))) // skip non-files completely
+ continue;
+
+ string fn = archive_entry_pathname (e);
+ if (fn != string(".")+b_source1)
+ continue;
+
+ // extract this file to a temporary file
+ char tmppath[PATH_MAX] = "/tmp/debuginfod.XXXXXX"; // XXX: $TMP_DIR etc.
+ int fd = mkstemp (tmppath);
+ if (fd < 0)
+ throw libc_exception (errno, "cannot create temporary file");
+ unlink (tmppath); // unlink now so OS will release the file as soon as we close the fd
+
+ rc = archive_read_data_into_fd (a, fd);
+ if (rc != ARCHIVE_OK)
+ {
+ close (fd);
+ throw archive_exception(a, "cannot extract file");
+ }
+
+ inc_metric ("http_responses_total","result","rpm");
+ struct MHD_Response* r = MHD_create_response_from_fd (archive_entry_size(e), fd);
+ if (r == 0)
+ {
+ if (verbose)
+ obatched(clog) << "cannot create fd-response for " << b_source0 << endl;
+ close(fd);
+ break; // assume no chance of better luck around another iteration
+ }
+ else
+ {
+ MHD_add_response_header (r, "Content-Type", "application/octet-stream");
+ add_mhd_last_modified (r, archive_entry_mtime(e));
+ if (verbose > 1)
+ obatched(clog) << "serving rpm " << b_source0 << " file " << b_source1 << endl;
+ /* libmicrohttpd will close it. */
+ if (result_fd)
+ *result_fd = fd;
+ return r;
+ }
+ }
+
+ // XXX: rpm/file not found: delete this R entry?
+ return 0;
+}
+
+
+static struct MHD_Response*
+handle_buildid_match (int64_t b_mtime,
+ const string& b_stype,
+ const string& b_source0,
+ const string& b_source1,
+ int *result_fd)
+{
+ if (b_stype == "F")
+ return handle_buildid_f_match(b_mtime, b_source0, result_fd);
+ else if (b_stype == "R")
+ return handle_buildid_r_match(b_mtime, b_source0, b_source1, result_fd);
+ else
+ return 0;
+}
+
+
+static int
+debuginfod_find_progress (debuginfod_client *, long a, long b)
+{
+ if (verbose > 4)
+ obatched(clog) << "federated debuginfod progress=" << a << "/" << b << endl;
+
+ return interrupted;
+}
+
+
+static struct MHD_Response* handle_buildid (const string& buildid /* unsafe */,
+ const string& artifacttype /* unsafe */,
+ const string& suffix /* unsafe */,
+ int *result_fd
+ )
+{
+ // validate artifacttype
+ string atype_code;
+ if (artifacttype == "debuginfo") atype_code = "D";
+ else if (artifacttype == "executable") atype_code = "E";
+ else if (artifacttype == "source") atype_code = "S";
+ else throw reportable_exception("invalid artifacttype");
+
+ if (atype_code == "S" && suffix == "")
+ throw reportable_exception("invalid source suffix");
+
+ // validate buildid
+ if ((buildid.size() < 2) || // not empty
+ (buildid.size() % 2) || // even number
+ (buildid.find_first_not_of("0123456789abcdef") != string::npos)) // pure tasty lowercase hex
+ throw reportable_exception("invalid buildid");
+
+ if (verbose > 1)
+ obatched(clog) << "searching for buildid=" << buildid << " artifacttype=" << artifacttype
+ << " suffix=" << suffix << endl;
+
+ sqlite_ps *pp = 0;
+
+ if (atype_code == "D")
+ {
+ pp = new sqlite_ps (db, "mhd-query-d",
+ "select mtime, sourcetype, source0, source1 from " BUILDIDS "_query_d where buildid = ? "
+ "order by mtime desc");
+ pp->reset();
+ pp->bind(1, buildid);
+ }
+ else if (atype_code == "E")
+ {
+ pp = new sqlite_ps (db, "mhd-query-e",
+ "select mtime, sourcetype, source0, source1 from " BUILDIDS "_query_e where buildid = ? "
+ "order by mtime desc");
+ pp->reset();
+ pp->bind(1, buildid);
+ }
+ else if (atype_code == "S")
+ {
+ pp = new sqlite_ps (db, "mhd-query-s",
+ "select mtime, sourcetype, source0, source1 from " BUILDIDS "_query_s where buildid = ? and artifactsrc = ? "
+ "order by sharedprefix(source0,source0ref) desc, mtime desc");
+ pp->reset();
+ pp->bind(1, buildid);
+ pp->bind(2, suffix);
+ }
+ unique_ptr<sqlite_ps> ps_closer(pp); // release pp if exception or return
+
+ // consume all the rows
+ while (1)
+ {
+ int rc = pp->step();
+ if (rc == SQLITE_DONE) break;
+ if (rc != SQLITE_ROW)
+ throw sqlite_exception(rc, "step");
+
+ int64_t b_mtime = sqlite3_column_int64 (*pp, 0);
+ string b_stype = string((const char*) sqlite3_column_text (*pp, 1) ?: ""); /* by DDL may not be NULL */
+ string b_source0 = string((const char*) sqlite3_column_text (*pp, 2) ?: ""); /* may be NULL */
+ string b_source1 = string((const char*) sqlite3_column_text (*pp, 3) ?: ""); /* may be NULL */
+
+ if (verbose > 1)
+ obatched(clog) << "found mtime=" << b_mtime << " stype=" << b_stype
+ << " source0=" << b_source0 << " source1=" << b_source1 << endl;
+
+ // Try accessing the located match.
+ // XXX: in case of multiple matches, attempt them in parallel?
+ auto r = handle_buildid_match (b_mtime, b_stype, b_source0, b_source1, result_fd);
+ if (r)
+ return r;
+ }
+
+ // We couldn't find it in the database. Last ditch effort
+ // is to defer to other debuginfo servers.
+
+ int fd = -1;
+ debuginfod_client *client = debuginfod_begin ();
+ if (client != NULL)
+ {
+ debuginfod_set_progressfn (client, & debuginfod_find_progress);
+
+ if (artifacttype == "debuginfo")
+ fd = debuginfod_find_debuginfo (client,
+ (const unsigned char*) buildid.c_str(),
+ 0, NULL);
+ else if (artifacttype == "executable")
+ fd = debuginfod_find_executable (client,
+ (const unsigned char*) buildid.c_str(),
+ 0, NULL);
+ else if (artifacttype == "source")
+ fd = debuginfod_find_source (client,
+ (const unsigned char*) buildid.c_str(),
+ 0, suffix.c_str(), NULL);
+ }
+ else
+ fd = -errno; /* Set by debuginfod_begin. */
+ debuginfod_end (client);
+
+ if (fd >= 0)
+ {
+ inc_metric ("http_responses_total","result","upstream");
+ struct stat s;
+ int rc = fstat (fd, &s);
+ if (rc == 0)
+ {
+ auto r = MHD_create_response_from_fd ((uint64_t) s.st_size, fd);
+ if (r)
+ {
+ MHD_add_response_header (r, "Content-Type", "application/octet-stream");
+ add_mhd_last_modified (r, s.st_mtime);
+ if (verbose > 1)
+ obatched(clog) << "serving file from upstream debuginfod/cache" << endl;
+ if (result_fd)
+ *result_fd = fd;
+ return r; // NB: don't close fd; libmicrohttpd will
+ }
+ }
+ close (fd);
+ }
+ else if (fd != -ENOSYS) // no DEBUGINFOD_URLS configured
+ throw libc_exception(-fd, "upstream debuginfod query failed");
+
+ throw reportable_exception(MHD_HTTP_NOT_FOUND, "not found");
+}
+
+
+////////////////////////////////////////////////////////////////////////
+
+static map<string,int64_t> metrics; // arbitrary data for /metrics query
+// NB: store int64_t since all our metrics are integers; prometheus accepts double
+static mutex metrics_lock;
+
+// utility function for assembling prometheus-compatible
+// name="escaped-value" strings
+// https://prometheus.io/docs/instrumenting/exposition_formats/
+static string
+metric_label(const string& name, const string& value)
+{
+ string x = name + "=\"";
+ for (auto&& c : value)
+ switch(c)
+ {
+ case '\\': x += "\\\\"; break;
+ case '\"': x += "\\\""; break;
+ case '\n': x += "\\n"; break;
+ default: x += c; break;
+ }
+ x += "\"";
+ return x;
+}
+
+
+// add prometheus-format metric name + label tuple (if any) + value
+
+static void
+set_metric(const string& metric, int64_t value)
+{
+ unique_lock<mutex> lock(metrics_lock);
+ metrics[metric] = value;
+}
+#if 0 /* unused */
+static void
+inc_metric(const string& metric)
+{
+ unique_lock<mutex> lock(metrics_lock);
+ metrics[metric] ++;
+}
+#endif
+static void
+set_metric(const string& metric,
+ const string& lname, const string& lvalue,
+ int64_t value)
+{
+ string key = (metric + "{" + metric_label(lname, lvalue) + "}");
+ unique_lock<mutex> lock(metrics_lock);
+ metrics[key] = value;
+}
+
+static void
+inc_metric(const string& metric,
+ const string& lname, const string& lvalue)
+{
+ string key = (metric + "{" + metric_label(lname, lvalue) + "}");
+ unique_lock<mutex> lock(metrics_lock);
+ metrics[key] ++;
+}
+static void
+add_metric(const string& metric,
+ const string& lname, const string& lvalue,
+ int64_t value)
+{
+ string key = (metric + "{" + metric_label(lname, lvalue) + "}");
+ unique_lock<mutex> lock(metrics_lock);
+ metrics[key] += value;
+}
+
+
+// and more for higher arity labels if needed
+
+
+static struct MHD_Response*
+handle_metrics ()
+{
+ stringstream o;
+ {
+ unique_lock<mutex> lock(metrics_lock);
+ for (auto&& i : metrics)
+ o << i.first << " " << i.second << endl;
+ }
+ const string& os = o.str();
+ MHD_Response* r = MHD_create_response_from_buffer (os.size(),
+ (void*) os.c_str(),
+ MHD_RESPMEM_MUST_COPY);
+ MHD_add_response_header (r, "Content-Type", "text/plain");
+ return r;
+}
+
+
+////////////////////////////////////////////////////////////////////////
+
+
+/* libmicrohttpd callback */
+static int
+handler_cb (void * /*cls*/,
+ struct MHD_Connection *connection,
+ const char *url,
+ const char *method,
+ const char * /*version*/,
+ const char * /*upload_data*/,
+ size_t * /*upload_data_size*/,
+ void ** /*con_cls*/)
+{
+ struct MHD_Response *r = NULL;
+ string url_copy = url;
+
+ if (verbose)
+ obatched(clog) << conninfo(connection) << " " << method << " " << url << endl;
+
+ try
+ {
+ if (string(method) != "GET")
+ throw reportable_exception(400, "we support GET only");
+
+ /* Start decoding the URL. */
+ size_t slash1 = url_copy.find('/', 1);
+ string url1 = url_copy.substr(0, slash1); // ok even if slash1 not found
+
+ if (slash1 != string::npos && url1 == "/buildid")
+ {
+ size_t slash2 = url_copy.find('/', slash1+1);
+ if (slash2 == string::npos)
+ throw reportable_exception("/buildid/ webapi error, need buildid");
+
+ string buildid = url_copy.substr(slash1+1, slash2-slash1-1);
+
+ size_t slash3 = url_copy.find('/', slash2+1);
+ string artifacttype, suffix;
+ if (slash3 == string::npos)
+ {
+ artifacttype = url_copy.substr(slash2+1);
+ suffix = "";
+ }
+ else
+ {
+ artifacttype = url_copy.substr(slash2+1, slash3-slash2-1);
+ suffix = url_copy.substr(slash3); // include the slash in the suffix
+ }
+
+ inc_metric("http_requests_total", "type", artifacttype);
+ r = handle_buildid(buildid, artifacttype, suffix, 0); // NB: don't care about result-fd
+ }
+ else if (url1 == "/metrics")
+ {
+ inc_metric("http_requests_total", "type", "metrics");
+ r = handle_metrics();
+ }
+ else
+ throw reportable_exception("webapi error, unrecognized /operation");
+
+ if (r == 0)
+ throw reportable_exception("internal error, missing response");
+
+ int rc = MHD_queue_response (connection, MHD_HTTP_OK, r);
+ MHD_destroy_response (r);
+ return rc;
+ }
+ catch (const reportable_exception& e)
+ {
+ inc_metric("http_responses_total","result","error");
+ e.report(clog);
+ return e.mhd_send_response (connection);
+ }
+}
+
+
+////////////////////////////////////////////////////////////////////////
+// borrowed originally from src/nm.c get_local_names()
+
+static void
+dwarf_extract_source_paths (Elf *elf, set<string>& debug_sourcefiles)
+ noexcept // no exceptions - so we can simplify the altdbg resource release at end
+{
+ Dwarf* dbg = dwarf_begin_elf (elf, DWARF_C_READ, NULL);
+ if (dbg == NULL)
+ return;
+
+ Dwarf* altdbg = NULL;
+ int altdbg_fd = -1;
+
+ // DWZ handling: if we have an unsatisfied debug-alt-link, add an
+ // empty string into the outgoing sourcefiles set, so the caller
+ // should know that our data is incomplete.
+ const char *alt_name_p;
+ const void *alt_build_id; // elfutils-owned memory
+ ssize_t sz = dwelf_dwarf_gnu_debugaltlink (dbg, &alt_name_p, &alt_build_id);
+ if (sz > 0) // got one!
+ {
+ string buildid;
+ unsigned char* build_id_bytes = (unsigned char*) alt_build_id;
+ for (ssize_t idx=0; idx<sz; idx++)
+ {
+ buildid += "0123456789abcdef"[build_id_bytes[idx] >> 4];
+ buildid += "0123456789abcdef"[build_id_bytes[idx] & 0xf];
+ }
+
+ if (verbose > 3)
+ obatched(clog) << "Need altdebug buildid=" << buildid << endl;
+
+ // but is it unsatisfied the normal elfutils ways?
+ Dwarf* alt = dwarf_getalt (dbg);
+ if (alt == NULL)
+ {
+ // Yup, unsatisfied the normal way. Maybe we can satisfy it
+ // from our own debuginfod database.
+ int alt_fd;
+ struct MHD_Response *r = 0;
+ try
+ {
+ r = handle_buildid (buildid, "debuginfo", "", &alt_fd);
+ }
+ catch (const reportable_exception& e)
+ {
+ // swallow exceptions
+ }
+
+ // NB: this is not actually recursive! This invokes the web-query
+ // path, which cannot get back into the scan code paths.
+ if (r)
+ {
+ // Found it!
+ altdbg_fd = dup(alt_fd); // ok if this fails, downstream failures ok
+ alt = altdbg = dwarf_begin (altdbg_fd, DWARF_C_READ);
+ // NB: must close this dwarf and this fd at the bottom of the function!
+ MHD_destroy_response (r); // will close alt_fd
+ if (alt)
+ dwarf_setalt (dbg, alt);
+ }
+ }
+ else
+ {
+ // NB: dwarf_setalt(alt) inappropriate - already done!
+ // NB: altdbg will stay 0 so nothing tries to redundantly dealloc.
+ }
+
+ if (alt)
+ {
+ if (verbose > 3)
+ obatched(clog) << "Resolved altdebug buildid=" << buildid << endl;
+ }
+ else // (alt == NULL) - signal possible presence of poor debuginfo
+ {
+ debug_sourcefiles.insert("");
+ if (verbose > 3)
+ obatched(clog) << "Unresolved altdebug buildid=" << buildid << endl;
+ }
+ }
+
+ Dwarf_Off offset = 0;
+ Dwarf_Off old_offset;
+ size_t hsize;
+
+ while (dwarf_nextcu (dbg, old_offset = offset, &offset, &hsize, NULL, NULL, NULL) == 0)
+ {
+ Dwarf_Die cudie_mem;
+ Dwarf_Die *cudie = dwarf_offdie (dbg, old_offset + hsize, &cudie_mem);
+
+ if (cudie == NULL)
+ continue;
+ if (dwarf_tag (cudie) != DW_TAG_compile_unit)
+ continue;
+
+ const char *cuname = dwarf_diename(cudie) ?: "unknown";
+
+ Dwarf_Files *files;
+ size_t nfiles;
+ if (dwarf_getsrcfiles (cudie, &files, &nfiles) != 0)
+ continue;
+
+ // extract DW_AT_comp_dir to resolve relative file names
+ const char *comp_dir = "";
+ const char *const *dirs;
+ size_t ndirs;
+ if (dwarf_getsrcdirs (files, &dirs, &ndirs) == 0 &&
+ dirs[0] != NULL)
+ comp_dir = dirs[0];
+ if (comp_dir == NULL)
+ comp_dir = "";
+
+ if (verbose > 3)
+ obatched(clog) << "searching for sources for cu=" << cuname << " comp_dir=" << comp_dir
+ << " #files=" << nfiles << " #dirs=" << ndirs << endl;
+
+ if (comp_dir[0] == '\0' && cuname[0] != '/')
+ {
+ // This is a common symptom for dwz-compressed debug files,
+ // where the altdebug file cannot be resolved.
+ if (verbose > 3)
+ obatched(clog) << "skipping cu=" << cuname << " due to empty comp_dir" << endl;
+ continue;
+ }
+
+ for (size_t f = 1; f < nfiles; f++)
+ {
+ const char *hat = dwarf_filesrc (files, f, NULL, NULL);
+ if (hat == NULL)
+ continue;
+
+ if (string(hat) == "<built-in>") // gcc intrinsics, don't bother record
+ continue;
+
+ string waldo;
+ if (hat[0] == '/') // absolute
+ waldo = (string (hat));
+ else if (comp_dir[0] != '\0') // comp_dir relative
+ waldo = (string (comp_dir) + string("/") + string (hat));
+ else
+ {
+ obatched(clog) << "skipping hat=" << hat << " due to empty comp_dir" << endl;
+ continue;
+ }
+
+ // NB: this is the 'waldo' that a dbginfo client will have
+ // to supply for us to give them the file The comp_dir
+ // prefixing is a definite complication. Otherwise we'd
+ // have to return a setof comp_dirs (one per CU!) with
+ // corresponding filesrc[] names, instead of one absolute
+ // resoved set. Maybe we'll have to do that anyway. XXX
+
+ if (verbose > 4)
+ obatched(clog) << waldo
+ << (debug_sourcefiles.find(waldo)==debug_sourcefiles.end() ? " new" : " dup") << endl;
+
+ debug_sourcefiles.insert (waldo);
+ }
+ }
+
+ dwarf_end(dbg);
+ if (altdbg)
+ dwarf_end(altdbg);
+ if (altdbg_fd >= 0)
+ close(altdbg_fd);
+}
+
+
+
+static void
+elf_classify (int fd, bool &executable_p, bool &debuginfo_p, string &buildid, set<string>& debug_sourcefiles)
+{
+ Elf *elf = elf_begin (fd, ELF_C_READ_MMAP_PRIVATE, NULL);
+ if (elf == NULL)
+ return;
+
+ try // catch our types of errors and clean up the Elf* object
+ {
+ if (elf_kind (elf) != ELF_K_ELF)
+ {
+ elf_end (elf);
+ return;
+ }
+
+ GElf_Ehdr ehdr_storage;
+ GElf_Ehdr *ehdr = gelf_getehdr (elf, &ehdr_storage);
+ if (ehdr == NULL)
+ {
+ elf_end (elf);
+ return;
+ }
+ auto elf_type = ehdr->e_type;
+
+ const void *build_id; // elfutils-owned memory
+ ssize_t sz = dwelf_elf_gnu_build_id (elf, & build_id);
+ if (sz <= 0)
+ {
+ // It's not a diagnostic-worthy error for an elf file to lack build-id.
+ // It might just be very old.
+ elf_end (elf);
+ return;
+ }
+
+ // build_id is a raw byte array; convert to hexadecimal *lowercase*
+ unsigned char* build_id_bytes = (unsigned char*) build_id;
+ for (ssize_t idx=0; idx<sz; idx++)
+ {
+ buildid += "0123456789abcdef"[build_id_bytes[idx] >> 4];
+ buildid += "0123456789abcdef"[build_id_bytes[idx] & 0xf];
+ }
+
+ // now decide whether it's an executable - namely, any allocatable section has
+ // PROGBITS;
+ if (elf_type == ET_EXEC || elf_type == ET_DYN)
+ {
+ size_t shnum;
+ int rc = elf_getshdrnum (elf, &shnum);
+ if (rc < 0)
+ throw elfutils_exception(rc, "getshdrnum");
+
+ executable_p = false;
+ for (size_t sc = 0; sc < shnum; sc++)
+ {
+ Elf_Scn *scn = elf_getscn (elf, sc);
+ if (scn == NULL)
+ continue;
+
+ GElf_Shdr shdr_mem;
+ GElf_Shdr *shdr = gelf_getshdr (scn, &shdr_mem);
+ if (shdr == NULL)
+ continue;
+
+ // allocated (loadable / vm-addr-assigned) section with available content?
+ if ((shdr->sh_type == SHT_PROGBITS) && (shdr->sh_flags & SHF_ALLOC))
+ {
+ if (verbose > 4)
+ obatched(clog) << "executable due to SHF_ALLOC SHT_PROGBITS sc=" << sc << endl;
+ executable_p = true;
+ break; // no need to keep looking for others
+ }
+ } // iterate over sections
+ } // executable_p classification
+
+ // now decide whether it's a debuginfo - namely, if it has any .debug* or .zdebug* sections
+ // logic mostly stolen from [email protected]'s elfclassify drafts
+ size_t shstrndx;
+ int rc = elf_getshdrstrndx (elf, &shstrndx);
+ if (rc < 0)
+ throw elfutils_exception(rc, "getshdrstrndx");
+
+ Elf_Scn *scn = NULL;
+ while (true)
+ {
+ scn = elf_nextscn (elf, scn);
+ if (scn == NULL)
+ break;
+ GElf_Shdr shdr_storage;
+ GElf_Shdr *shdr = gelf_getshdr (scn, &shdr_storage);
+ if (shdr == NULL)
+ break;
+ const char *section_name = elf_strptr (elf, shstrndx, shdr->sh_name);
+ if (section_name == NULL)
+ break;
+ if (strncmp(section_name, ".debug_line", 11) == 0 ||
+ strncmp(section_name, ".zdebug_line", 12) == 0)
+ {
+ debuginfo_p = true;
+ dwarf_extract_source_paths (elf, debug_sourcefiles);
+ break; // expecting only one .*debug_line, so no need to look for others
+ }
+ else if (strncmp(section_name, ".debug_", 7) == 0 ||
+ strncmp(section_name, ".zdebug_", 8) == 0)
+ {
+ debuginfo_p = true;
+ // NB: don't break; need to parse .debug_line for sources
+ }
+ }
+ }
+ catch (const reportable_exception& e)
+ {
+ e.report(clog);
+ }
+ elf_end (elf);
+}
+
+
+static semaphore* scan_concurrency_sem = 0; // used to implement -c load limiting
+
+
+static void
+scan_source_file_path (const string& dir)
+{
+ obatched(clog) << "fts/file traversing " << dir << endl;
+
+ struct timeval tv_start, tv_end;
+ gettimeofday (&tv_start, NULL);
+
+ sqlite_ps ps_upsert_buildids (db, "file-buildids-intern", "insert or ignore into " BUILDIDS "_buildids VALUES (NULL, ?);");
+ sqlite_ps ps_upsert_files (db, "file-files-intern", "insert or ignore into " BUILDIDS "_files VALUES (NULL, ?);");
+ sqlite_ps ps_upsert_de (db, "file-de-upsert",
+ "insert or ignore into " BUILDIDS "_f_de "
+ "(buildid, debuginfo_p, executable_p, file, mtime) "
+ "values ((select id from " BUILDIDS "_buildids where hex = ?),"
+ " ?,?,"
+ " (select id from " BUILDIDS "_files where name = ?), ?);");
+ sqlite_ps ps_upsert_s (db, "file-s-upsert",
+ "insert or ignore into " BUILDIDS "_f_s "
+ "(buildid, artifactsrc, file, mtime) "
+ "values ((select id from " BUILDIDS "_buildids where hex = ?),"
+ " (select id from " BUILDIDS "_files where name = ?),"
+ " (select id from " BUILDIDS "_files where name = ?),"
+ " ?);");
+ sqlite_ps ps_query (db, "file-negativehit-find",
+ "select 1 from " BUILDIDS "_file_mtime_scanned where sourcetype = 'F' and file = (select id from " BUILDIDS "_files where name = ?) and mtime = ?;");
+ sqlite_ps ps_scan_done (db, "file-scanned",
+ "insert or ignore into " BUILDIDS "_file_mtime_scanned (sourcetype, file, mtime, size)"
+ "values ('F', (select id from " BUILDIDS "_files where name = ?), ?, ?);");
+
+
+ char * const dirs[] = { (char*) dir.c_str(), NULL };
+
+ unsigned fts_scanned=0, fts_regex=0, fts_cached=0, fts_debuginfo=0, fts_executable=0, fts_sourcefiles=0;
+
+ FTS *fts = fts_open (dirs,
+ (traverse_logical ? FTS_LOGICAL : FTS_PHYSICAL|FTS_XDEV)
+ | FTS_NOCHDIR /* multithreaded */,
+ NULL);
+ if (fts == NULL)
+ {
+ obatched(cerr) << "cannot fts_open " << dir << endl;
+ return;
+ }
+
+ FTSENT *f;
+ while ((f = fts_read (fts)) != NULL)
+ {
+ semaphore_borrower handle_one_file (scan_concurrency_sem);
+
+ fts_scanned ++;
+ if (interrupted)
+ break;
+
+ if (verbose > 2)
+ obatched(clog) << "fts/file traversing " << f->fts_path << endl;
+
+ try
+ {
+ /* Found a file. Convert it to an absolute path, so
+ the buildid database does not have relative path
+ names that are unresolvable from a subsequent run
+ in a different cwd. */
+ char *rp = realpath(f->fts_path, NULL);
+ if (rp == NULL)
+ continue; // ignore dangling symlink or such
+ string rps = string(rp);
+ free (rp);
+
+ bool ri = !regexec (&file_include_regex, rps.c_str(), 0, 0, 0);
+ bool rx = !regexec (&file_exclude_regex, rps.c_str(), 0, 0, 0);
+ if (!ri || rx)
+ {
+ if (verbose > 3)
+ obatched(clog) << "fts/file skipped by regex " << (!ri ? "I" : "") << (rx ? "X" : "") << endl;
+ fts_regex ++;
+ continue;
+ }
+
+ switch (f->fts_info)
+ {
+ case FTS_D:
+ break;
+
+ case FTS_DP:
+ break;
+
+ case FTS_F:
+ {
+ /* See if we know of it already. */
+ int rc = ps_query
+ .reset()
+ .bind(1, rps)
+ .bind(2, f->fts_statp->st_mtime)
+ .step();
+ ps_query.reset();
+ if (rc == SQLITE_ROW) // i.e., a result, as opposed to DONE (no results)
+ // no need to recheck a file/version we already know
+ // specifically, no need to elf-begin a file we already determined is non-elf
+ // (so is stored with buildid=NULL)
+ {
+ fts_cached ++;
+ continue;
+ }
+
+ bool executable_p = false, debuginfo_p = false; // E and/or D
+ string buildid;
+ set<string> sourcefiles;
+
+ int fd = open (rps.c_str(), O_RDONLY);
+ try
+ {
+ if (fd >= 0)
+ elf_classify (fd, executable_p, debuginfo_p, buildid, sourcefiles);
+ else
+ throw libc_exception(errno, string("open ") + rps);
+ inc_metric ("scanned_total","source","file");
+ }
+
+ // NB: we catch exceptions here too, so that we can
+ // cache the corrupt-elf case (!executable_p &&
+ // !debuginfo_p) just below, just as if we had an
+ // EPERM error from open(2).
+
+ catch (const reportable_exception& e)
+ {
+ e.report(clog);
+ }
+
+ if (fd >= 0)
+ close (fd);
+
+ // register this file name in the interning table
+ ps_upsert_files
+ .reset()
+ .bind(1, rps)
+ .step_ok_done();
+
+ if (buildid == "")
+ {
+ // no point storing an elf file without buildid
+ executable_p = false;
+ debuginfo_p = false;
+ }
+ else
+ {
+ // register this build-id in the interning table
+ ps_upsert_buildids
+ .reset()
+ .bind(1, buildid)
+ .step_ok_done();
+ }
+
+ if (executable_p)
+ fts_executable ++;
+ if (debuginfo_p)
+ fts_debuginfo ++;
+ if (executable_p || debuginfo_p)
+ {
+ ps_upsert_de
+ .reset()
+ .bind(1, buildid)
+ .bind(2, debuginfo_p ? 1 : 0)
+ .bind(3, executable_p ? 1 : 0)
+ .bind(4, rps)
+ .bind(5, f->fts_statp->st_mtime)
+ .step_ok_done();
+ }
+ if (executable_p)
+ inc_metric("found_executable_total","source","files");
+ if (debuginfo_p)
+ inc_metric("found_debuginfo_total","source","files");
+
+ if (sourcefiles.size() && buildid != "")
+ {
+ fts_sourcefiles += sourcefiles.size();
+
+ for (auto&& dwarfsrc : sourcefiles)
+ {
+ char *srp = realpath(dwarfsrc.c_str(), NULL);
+ if (srp == NULL) // also if DWZ unresolved dwarfsrc=""
+ continue; // unresolvable files are not a serious problem
+ // throw libc_exception(errno, "fts/file realpath " + srcpath);
+ string srps = string(srp);
+ free (srp);
+
+ struct stat sfs;
+ rc = stat(srps.c_str(), &sfs);
+ if (rc != 0)
+ continue;
+
+ if (verbose > 2)
+ obatched(clog) << "recorded buildid=" << buildid << " file=" << srps
+ << " mtime=" << sfs.st_mtime
+ << " as source " << dwarfsrc << endl;
+
+ ps_upsert_files
+ .reset()
+ .bind(1, srps)
+ .step_ok_done();
+
+ // register the dwarfsrc name in the interning table too
+ ps_upsert_files
+ .reset()
+ .bind(1, dwarfsrc)
+ .step_ok_done();
+
+ ps_upsert_s
+ .reset()
+ .bind(1, buildid)
+ .bind(2, dwarfsrc)
+ .bind(3, srps)
+ .bind(4, sfs.st_mtime)
+ .step_ok_done();
+
+ inc_metric("found_sourcerefs_total","source","files");
+ }
+ }
+
+ ps_scan_done
+ .reset()
+ .bind(1, rps)
+ .bind(2, f->fts_statp->st_mtime)
+ .bind(3, f->fts_statp->st_size)
+ .step_ok_done();
+
+ if (verbose > 2)
+ obatched(clog) << "recorded buildid=" << buildid << " file=" << rps
+ << " mtime=" << f->fts_statp->st_mtime << " atype="
+ << (executable_p ? "E" : "")
+ << (debuginfo_p ? "D" : "") << endl;
+ }
+ break;
+
+ case FTS_ERR:
+ case FTS_NS:
+ throw libc_exception(f->fts_errno, string("fts/file traversal ") + string(f->fts_path));
+
+ default:
+ case FTS_SL: /* ignore symlinks; seen in non-L mode only */
+ break;
+ }
+
+ if ((verbose && f->fts_info == FTS_DP) ||
+ (verbose > 1 && f->fts_info == FTS_F))
+ obatched(clog) << "fts/file traversing " << rps << ", scanned=" << fts_scanned
+ << ", regex-skipped=" << fts_regex
+ << ", cached=" << fts_cached << ", debuginfo=" << fts_debuginfo
+ << ", executable=" << fts_executable << ", source=" << fts_sourcefiles << endl;
+ }
+ catch (const reportable_exception& e)
+ {
+ e.report(clog);
+ }
+ }
+ fts_close (fts);
+
+ gettimeofday (&tv_end, NULL);
+ double deltas = (tv_end.tv_sec - tv_start.tv_sec) + (tv_end.tv_usec - tv_start.tv_usec)*0.000001;
+
+ obatched(clog) << "fts/file traversed " << dir << " in " << deltas << "s, scanned=" << fts_scanned
+ << ", regex-skipped=" << fts_regex
+ << ", cached=" << fts_cached << ", debuginfo=" << fts_debuginfo
+ << ", executable=" << fts_executable << ", source=" << fts_sourcefiles << endl;
+}
+
+
+static void*
+thread_main_scan_source_file_path (void* arg)
+{
+ string dir = string((const char*) arg);
+
+ unsigned rescan_timer = 0;
+ sig_atomic_t forced_rescan_count = 0;
+ set_metric("thread_timer_max", "file", dir, rescan_s);
+ set_metric("thread_tid", "file", dir, tid());
+ while (! interrupted)
+ {
+ set_metric("thread_timer", "file", dir, rescan_timer);
+ set_metric("thread_forced_total", "file", dir, forced_rescan_count);
+ if (rescan_s && rescan_timer > rescan_s)
+ rescan_timer = 0;
+ if (sigusr1 != forced_rescan_count)
+ {
+ forced_rescan_count = sigusr1;
+ rescan_timer = 0;
+ }
+ if (rescan_timer == 0)
+ try
+ {
+ set_metric("thread_working", "file", dir, time(NULL));
+ inc_metric("thread_work_total", "file", dir);
+ scan_source_file_path (dir);
+ set_metric("thread_working", "file", dir, 0);
+ }
+ catch (const sqlite_exception& e)
+ {
+ obatched(cerr) << e.message << endl;
+ }
+ sleep (1);
+ rescan_timer ++;
+ }
+
+ return 0;
+}
+
+
+////////////////////////////////////////////////////////////////////////
+
+
+
+
+// Analyze given *.rpm file of given age; record buildids / exec/debuginfo-ness of its
+// constituent files with given upsert statements.
+static void
+rpm_classify (const string& rps, sqlite_ps& ps_upsert_buildids, sqlite_ps& ps_upsert_files,
+ sqlite_ps& ps_upsert_de, sqlite_ps& ps_upsert_sref, sqlite_ps& ps_upsert_sdef,
+ time_t mtime,
+ unsigned& fts_executable, unsigned& fts_debuginfo, unsigned& fts_sref, unsigned& fts_sdef,
+ bool& fts_sref_complete_p)
+{
+ string popen_cmd = string("rpm2cpio " + shell_escape(rps));
+ FILE* fp = popen (popen_cmd.c_str(), "r"); // "e" O_CLOEXEC?
+ if (fp == NULL)
+ throw libc_exception (errno, string("popen ") + popen_cmd);
+ defer_dtor<FILE*,int> fp_closer (fp, pclose);
+
+ struct archive *a;
+ a = archive_read_new();
+ if (a == NULL)
+ throw archive_exception("cannot create archive reader");
+ defer_dtor<struct archive*,int> archive_closer (a, archive_read_free);
+
+ int rc = archive_read_support_format_cpio(a);
+ if (rc != ARCHIVE_OK)
+ throw archive_exception(a, "cannot select cpio format");
+ rc = archive_read_support_filter_all(a);
+ if (rc != ARCHIVE_OK)
+ throw archive_exception(a, "cannot select all filters");
+
+ rc = archive_read_open_FILE (a, fp);
+ if (rc != ARCHIVE_OK)
+ throw archive_exception(a, "cannot open archive from rpm2cpio pipe");
+
+ if (verbose > 3)
+ obatched(clog) << "rpm2cpio|libarchive scanning " << rps << endl;
+
+ while(1) // parse cpio archive entries
+ {
+ try
+ {
+ struct archive_entry *e;
+ rc = archive_read_next_header (a, &e);
+ if (rc != ARCHIVE_OK)
+ break;
+
+ if (! S_ISREG(archive_entry_mode (e))) // skip non-files completely
+ continue;
+
+ string fn = archive_entry_pathname (e);
+ if (fn.size() > 1 && fn[0] == '.')
+ fn = fn.substr(1); // trim off the leading '.'
+
+ if (verbose > 3)
+ obatched(clog) << "rpm2cpio|libarchive checking " << fn << endl;
+
+ // extract this file to a temporary file
+ const char *tmpdir_env = getenv ("TMPDIR") ?: "/tmp";
+ char* tmppath = NULL;
+ rc = asprintf (&tmppath, "%s/debuginfod.XXXXXX", tmpdir_env);
+ if (rc < 0)
+ throw libc_exception (ENOMEM, "cannot allocate tmppath");
+ defer_dtor<void*,void> tmmpath_freer (tmppath, free);
+ int fd = mkstemp (tmppath);
+ if (fd < 0)
+ throw libc_exception (errno, "cannot create temporary file");
+ unlink (tmppath); // unlink now so OS will release the file as soon as we close the fd
+ defer_dtor<int,int> minifd_closer (fd, close);
+
+ rc = archive_read_data_into_fd (a, fd);
+ if (rc != ARCHIVE_OK)
+ throw archive_exception(a, "cannot extract file");
+
+ // finally ... time to run elf_classify on this bad boy and update the database
+ bool executable_p = false, debuginfo_p = false;
+ string buildid;
+ set<string> sourcefiles;
+ elf_classify (fd, executable_p, debuginfo_p, buildid, sourcefiles);
+ // NB: might throw
+
+ if (buildid != "") // intern buildid
+ {
+ ps_upsert_buildids
+ .reset()
+ .bind(1, buildid)
+ .step_ok_done();
+ }
+
+ ps_upsert_files // register this rpm constituent file name in interning table
+ .reset()
+ .bind(1, fn)
+ .step_ok_done();
+
+ if (sourcefiles.size() > 0) // sref records needed
+ {
+ // NB: we intern each source file once. Once raw, as it
+ // appears in the DWARF file list coming back from
+ // elf_classify() - because it'll end up in the
+ // _norm.artifactsrc column. We don't also put another
+ // version with a '.' at the front, even though that's
+ // how rpm/cpio packs names, because we hide that from
+ // the database for storage efficiency.
+
+ for (auto&& s : sourcefiles)
+ {
+ if (s == "")
+ {
+ fts_sref_complete_p = false;
+ continue;
+ }
+
+ ps_upsert_files
+ .reset()
+ .bind(1, s)
+ .step_ok_done();
+
+ ps_upsert_sref
+ .reset()
+ .bind(1, buildid)
+ .bind(2, s)
+ .step_ok_done();
+
+ fts_sref ++;
+ }
+ }
+
+ if (executable_p)
+ fts_executable ++;
+ if (debuginfo_p)
+ fts_debuginfo ++;
+
+ if (executable_p || debuginfo_p)
+ {
+ ps_upsert_de
+ .reset()
+ .bind(1, buildid)
+ .bind(2, debuginfo_p ? 1 : 0)
+ .bind(3, executable_p ? 1 : 0)
+ .bind(4, rps)
+ .bind(5, mtime)
+ .bind(6, fn)
+ .step_ok_done();
+ }
+ else // potential source - sdef record
+ {
+ fts_sdef ++;
+ ps_upsert_sdef
+ .reset()
+ .bind(1, rps)
+ .bind(2, mtime)
+ .bind(3, fn)
+ .step_ok_done();
+ }
+
+ if ((verbose > 2) && (executable_p || debuginfo_p))
+ obatched(clog) << "recorded buildid=" << buildid << " rpm=" << rps << " file=" << fn
+ << " mtime=" << mtime << " atype="
+ << (executable_p ? "E" : "")
+ << (debuginfo_p ? "D" : "")
+ << " sourcefiles=" << sourcefiles.size() << endl;
+
+ }
+ catch (const reportable_exception& e)
+ {
+ e.report(clog);
+ }
+ }
+}
+
+
+
+// scan for *.rpm files
+static void
+scan_source_rpm_path (const string& dir)
+{
+ obatched(clog) << "fts/rpm traversing " << dir << endl;
+
+ sqlite_ps ps_upsert_buildids (db, "rpm-buildid-intern", "insert or ignore into " BUILDIDS "_buildids VALUES (NULL, ?);");
+ sqlite_ps ps_upsert_files (db, "rpm-file-intern", "insert or ignore into " BUILDIDS "_files VALUES (NULL, ?);");
+ sqlite_ps ps_upsert_de (db, "rpm-de-insert",
+ "insert or ignore into " BUILDIDS "_r_de (buildid, debuginfo_p, executable_p, file, mtime, content) values ("
+ "(select id from " BUILDIDS "_buildids where hex = ?), ?, ?, "
+ "(select id from " BUILDIDS "_files where name = ?), ?, "
+ "(select id from " BUILDIDS "_files where name = ?));");
+ sqlite_ps ps_upsert_sref (db, "rpm-sref-insert",
+ "insert or ignore into " BUILDIDS "_r_sref (buildid, artifactsrc) values ("
+ "(select id from " BUILDIDS "_buildids where hex = ?), "
+ "(select id from " BUILDIDS "_files where name = ?));");
+ sqlite_ps ps_upsert_sdef (db, "rpm-sdef-insert",
+ "insert or ignore into " BUILDIDS "_r_sdef (file, mtime, content) values ("
+ "(select id from " BUILDIDS "_files where name = ?), ?,"
+ "(select id from " BUILDIDS "_files where name = ?));");
+ sqlite_ps ps_query (db, "rpm-negativehit-query",
+ "select 1 from " BUILDIDS "_file_mtime_scanned where "
+ "sourcetype = 'R' and file = (select id from " BUILDIDS "_files where name = ?) and mtime = ?;");
+ sqlite_ps ps_scan_done (db, "rpm-scanned",
+ "insert or ignore into " BUILDIDS "_file_mtime_scanned (sourcetype, file, mtime, size)"
+ "values ('R', (select id from " BUILDIDS "_files where name = ?), ?, ?);");
+
+ char * const dirs[] = { (char*) dir.c_str(), NULL };
+
+ struct timeval tv_start, tv_end;
+ gettimeofday (&tv_start, NULL);
+ unsigned fts_scanned=0, fts_regex=0, fts_cached=0, fts_debuginfo=0;
+ unsigned fts_executable=0, fts_rpm = 0, fts_sref=0, fts_sdef=0;
+
+ FTS *fts = fts_open (dirs,
+ (traverse_logical ? FTS_LOGICAL : FTS_PHYSICAL|FTS_XDEV)
+ | FTS_NOCHDIR /* multithreaded */,
+ NULL);
+ if (fts == NULL)
+ {
+ obatched(cerr) << "cannot fts_open " << dir << endl;
+ return;
+ }
+
+ FTSENT *f;
+ while ((f = fts_read (fts)) != NULL)
+ {
+ semaphore_borrower handle_one_file (scan_concurrency_sem);
+
+ fts_scanned ++;
+ if (interrupted)
+ break;
+
+ if (verbose > 2)
+ obatched(clog) << "fts/rpm traversing " << f->fts_path << endl;
+
+ try
+ {
+ /* Found a file. Convert it to an absolute path, so
+ the buildid database does not have relative path
+ names that are unresolvable from a subsequent run
+ in a different cwd. */
+ char *rp = realpath(f->fts_path, NULL);
+ if (rp == NULL)
+ continue; // ignore dangling symlink or such
+ string rps = string(rp);
+ free (rp);
+
+ bool ri = !regexec (&file_include_regex, rps.c_str(), 0, 0, 0);
+ bool rx = !regexec (&file_exclude_regex, rps.c_str(), 0, 0, 0);
+ if (!ri || rx)
+ {
+ if (verbose > 3)
+ obatched(clog) << "fts/rpm skipped by regex " << (!ri ? "I" : "") << (rx ? "X" : "") << endl;
+ fts_regex ++;
+ continue;
+ }
+
+ switch (f->fts_info)
+ {
+ case FTS_D:
+ break;
+
+ case FTS_DP:
+ break;
+
+ case FTS_F:
+ {
+ // heuristic: reject if file name does not end with ".rpm"
+ // (alternative: try opening with librpm etc., caching)
+ string suffix = ".rpm";
+ if (rps.size() < suffix.size() ||
+ rps.substr(rps.size()-suffix.size()) != suffix)
+ continue;
+ fts_rpm ++;
+
+ /* See if we know of it already. */
+ int rc = ps_query
+ .reset()
+ .bind(1, rps)
+ .bind(2, f->fts_statp->st_mtime)
+ .step();
+ ps_query.reset();
+ if (rc == SQLITE_ROW) // i.e., a result, as opposed to DONE (no results)
+ // no need to recheck a file/version we already know
+ // specifically, no need to parse this rpm again, since we already have
+ // it as a D or E or S record,
+ // (so is stored with buildid=NULL)
+ {
+ fts_cached ++;
+ continue;
+ }
+
+ // intern the rpm file name
+ ps_upsert_files
+ .reset()
+ .bind(1, rps)
+ .step_ok_done();
+
+ // extract the rpm contents via popen("rpm2cpio") | libarchive | loop-of-elf_classify()
+ unsigned my_fts_executable = 0, my_fts_debuginfo = 0, my_fts_sref = 0, my_fts_sdef = 0;
+ bool my_fts_sref_complete_p = true;
+ try
+ {
+ rpm_classify (rps,
+ ps_upsert_buildids, ps_upsert_files,
+ ps_upsert_de, ps_upsert_sref, ps_upsert_sdef, // dalt
+ f->fts_statp->st_mtime,
+ my_fts_executable, my_fts_debuginfo, my_fts_sref, my_fts_sdef,
+ my_fts_sref_complete_p);
+ inc_metric ("scanned_total","source","rpm");
+ add_metric("found_debuginfo_total","source","rpm",
+ my_fts_debuginfo);
+ add_metric("found_executable_total","source","rpm",
+ my_fts_executable);
+ add_metric("found_sourcerefs_total","source","rpm",
+ my_fts_sref);
+ }
+ catch (const reportable_exception& e)
+ {
+ e.report(clog);
+ }
+
+ if (verbose > 2)
+ obatched(clog) << "scanned rpm=" << rps
+ << " mtime=" << f->fts_statp->st_mtime
+ << " executables=" << my_fts_executable
+ << " debuginfos=" << my_fts_debuginfo
+ << " srefs=" << my_fts_sref
+ << " sdefs=" << my_fts_sdef
+ << endl;
+
+ fts_executable += my_fts_executable;
+ fts_debuginfo += my_fts_debuginfo;
+ fts_sref += my_fts_sref;
+ fts_sdef += my_fts_sdef;
+
+ if (my_fts_sref_complete_p) // leave incomplete?
+ ps_scan_done
+ .reset()
+ .bind(1, rps)
+ .bind(2, f->fts_statp->st_mtime)
+ .bind(3, f->fts_statp->st_size)
+ .step_ok_done();
+ }
+ break;
+
+ case FTS_ERR:
+ case FTS_NS:
+ throw libc_exception(f->fts_errno, string("fts/rpm traversal ") + string(f->fts_path));
+
+ default:
+ case FTS_SL: /* ignore symlinks; seen in non-L mode only */
+ break;
+ }
+
+ if ((verbose && f->fts_info == FTS_DP) ||
+ (verbose > 1 && f->fts_info == FTS_F))
+ obatched(clog) << "fts/rpm traversing " << rps << ", scanned=" << fts_scanned
+ << ", regex-skipped=" << fts_regex
+ << ", rpm=" << fts_rpm << ", cached=" << fts_cached << ", debuginfo=" << fts_debuginfo
+ << ", executable=" << fts_executable
+ << ", sourcerefs=" << fts_sref << ", sourcedefs=" << fts_sdef << endl;
+ }
+ catch (const reportable_exception& e)
+ {
+ e.report(clog);
+ }
+ }
+ fts_close (fts);
+
+ gettimeofday (&tv_end, NULL);
+ double deltas = (tv_end.tv_sec - tv_start.tv_sec) + (tv_end.tv_usec - tv_start.tv_usec)*0.000001;
+
+ obatched(clog) << "fts/rpm traversed " << dir << " in " << deltas << "s, scanned=" << fts_scanned
+ << ", regex-skipped=" << fts_regex
+ << ", rpm=" << fts_rpm << ", cached=" << fts_cached << ", debuginfo=" << fts_debuginfo
+ << ", executable=" << fts_executable
+ << ", sourcerefs=" << fts_sref << ", sourcedefs=" << fts_sdef << endl;
+}
+
+
+
+static void*
+thread_main_scan_source_rpm_path (void* arg)
+{
+ string dir = string((const char*) arg);
+
+ unsigned rescan_timer = 0;
+ sig_atomic_t forced_rescan_count = 0;
+ set_metric("thread_timer_max", "rpm", dir, rescan_s);
+ set_metric("thread_tid", "rpm", dir, tid());
+ while (! interrupted)
+ {
+ set_metric("thread_timer", "rpm", dir, rescan_timer);
+ set_metric("thread_forced_total", "rpm", dir, forced_rescan_count);
+ if (rescan_s && rescan_timer > rescan_s)
+ rescan_timer = 0;
+ if (sigusr1 != forced_rescan_count)
+ {
+ forced_rescan_count = sigusr1;
+ rescan_timer = 0;
+ }
+ if (rescan_timer == 0)
+ try
+ {
+ set_metric("thread_working", "rpm", dir, time(NULL));
+ inc_metric("thread_work_total", "rpm", dir);
+ scan_source_rpm_path (dir);
+ set_metric("thread_working", "rpm", dir, 0);
+ }
+ catch (const sqlite_exception& e)
+ {
+ obatched(cerr) << e.message << endl;
+ }
+ sleep (1);
+ rescan_timer ++;
+ }
+
+ return 0;
+}
+
+
+////////////////////////////////////////////////////////////////////////
+
+static void
+database_stats_report()
+{
+ sqlite_ps ps_query (db, "database-overview",
+ "select label,quantity from " BUILDIDS "_stats");
+
+ obatched(clog) << "database record counts:" << endl;
+ while (1)
+ {
+ int rc = sqlite3_step (ps_query);
+ if (rc == SQLITE_DONE) break;
+ if (rc != SQLITE_ROW)
+ throw sqlite_exception(rc, "step");
+
+ obatched(clog)
+ << right << setw(20) << ((const char*) sqlite3_column_text(ps_query, 0) ?: (const char*) "NULL")
+ << " "
+ << (sqlite3_column_text(ps_query, 1) ?: (const unsigned char*) "NULL")
+ << endl;
+
+ set_metric("groom", "statistic",
+ ((const char*) sqlite3_column_text(ps_query, 0) ?: (const char*) "NULL"),
+ (sqlite3_column_double(ps_query, 1)));
+ }
+}
+
+
+// Do a round of database grooming that might take many minutes to run.
+void groom()
+{
+ obatched(clog) << "grooming database" << endl;
+
+ struct timeval tv_start, tv_end;
+ gettimeofday (&tv_start, NULL);
+
+ // scan for files that have disappeared
+ sqlite_ps files (db, "check old files", "select s.mtime, s.file, f.name from "
+ BUILDIDS "_file_mtime_scanned s, " BUILDIDS "_files f "
+ "where f.id = s.file");
+ sqlite_ps files_del_f_de (db, "nuke f_de", "delete from " BUILDIDS "_f_de where file = ? and mtime = ?");
+ sqlite_ps files_del_r_de (db, "nuke r_de", "delete from " BUILDIDS "_r_de where file = ? and mtime = ?");
+ sqlite_ps files_del_scan (db, "nuke f_m_s", "delete from " BUILDIDS "_file_mtime_scanned "
+ "where file = ? and mtime = ?");
+ files.reset();
+ while(1)
+ {
+ int rc = files.step();
+ if (rc != SQLITE_ROW)
+ break;
+
+ int64_t mtime = sqlite3_column_int64 (files, 0);
+ int64_t fileid = sqlite3_column_int64 (files, 1);
+ const char* filename = ((const char*) sqlite3_column_text (files, 2) ?: "");
+ struct stat s;
+ rc = stat(filename, &s);
+ if (rc < 0 || (mtime != (int64_t) s.st_mtime))
+ {
+ if (verbose > 2)
+ obatched(clog) << "groom: forgetting file=" << filename << " mtime=" << mtime << endl;
+ files_del_f_de.reset().bind(1,fileid).bind(2,mtime).step_ok_done();
+ files_del_r_de.reset().bind(1,fileid).bind(2,mtime).step_ok_done();
+ files_del_scan.reset().bind(1,fileid).bind(2,mtime).step_ok_done();
+ }
+ }
+ files.reset();
+
+ // delete buildids with no references in _r_de or _f_de tables;
+ // cascades to _r_sref & _f_s records
+ sqlite_ps buildids_del (db, "nuke orphan buildids",
+ "delete from " BUILDIDS "_buildids "
+ "where not exists (select 1 from " BUILDIDS "_f_de d where " BUILDIDS "_buildids.id = d.buildid) "
+ "and not exists (select 1 from " BUILDIDS "_r_de d where " BUILDIDS "_buildids.id = d.buildid)");
+ buildids_del.reset().step_ok_done();
+
+ // NB: "vacuum" is too heavy for even daily runs: it rewrites the entire db, so is done as maxigroom -G
+ sqlite_ps g1 (db, "incremental vacuum", "pragma incremental_vacuum");
+ g1.reset().step_ok_done();
+ sqlite_ps g2 (db, "optimize", "pragma optimize");
+ g2.reset().step_ok_done();
+ sqlite_ps g3 (db, "wal checkpoint", "pragma wal_checkpoint=truncate");
+ g3.reset().step_ok_done();
+
+ database_stats_report();
+
+ sqlite3_db_release_memory(db); // shrink the process if possible
+
+ gettimeofday (&tv_end, NULL);
+ double deltas = (tv_end.tv_sec - tv_start.tv_sec) + (tv_end.tv_usec - tv_start.tv_usec)*0.000001;
+
+ obatched(clog) << "groomed database in " << deltas << "s" << endl;
+}
+
+
+static void*
+thread_main_groom (void* /*arg*/)
+{
+ unsigned groom_timer = 0;
+ sig_atomic_t forced_groom_count = 0;
+ set_metric("thread_timer_max", "role", "groom", groom_s);
+ set_metric("thread_tid", "role", "groom", tid());
+ while (! interrupted)
+ {
+ set_metric("thread_timer", "role", "groom", groom_timer);
+ set_metric("thread_forced_total", "role", "groom", forced_groom_count);
+ if (groom_s && groom_timer > groom_s)
+ groom_timer = 0;
+ if (sigusr2 != forced_groom_count)
+ {
+ forced_groom_count = sigusr2;
+ groom_timer = 0;
+ }
+ if (groom_timer == 0)
+ try
+ {
+ set_metric("thread_working", "role", "groom", time(NULL));
+ inc_metric("thread_work_total", "role", "groom");
+ groom ();
+ set_metric("thread_working", "role", "groom", 0);
+ }
+ catch (const sqlite_exception& e)
+ {
+ obatched(cerr) << e.message << endl;
+ }
+ sleep (1);
+ groom_timer ++;
+ }
+
+ return 0;
+}
+
+
+////////////////////////////////////////////////////////////////////////
+
+
+static void
+signal_handler (int /* sig */)
+{
+ interrupted ++;
+
+ if (db)
+ sqlite3_interrupt (db);
+
+ // NB: don't do anything else in here
+}
+
+static void
+sigusr1_handler (int /* sig */)
+{
+ sigusr1 ++;
+ // NB: don't do anything else in here
+}
+
+static void
+sigusr2_handler (int /* sig */)
+{
+ sigusr2 ++;
+ // NB: don't do anything else in here
+}
+
+
+
+
+
+// A user-defined sqlite function, to score the sharedness of the
+// prefix of two strings. This is used to compare candidate debuginfo
+// / source-rpm names, so that the closest match
+// (directory-topology-wise closest) is found. This is important in
+// case the same sref (source file name) is in many -debuginfo or
+// -debugsource RPMs, such as when multiple versions/releases of the
+// same package are in the database.
+
+static void sqlite3_sharedprefix_fn (sqlite3_context* c, int argc, sqlite3_value** argv)
+{
+ if (argc != 2)
+ sqlite3_result_error(c, "expect 2 string arguments", -1);
+ else if ((sqlite3_value_type(argv[0]) != SQLITE_TEXT) ||
+ (sqlite3_value_type(argv[1]) != SQLITE_TEXT))
+ sqlite3_result_null(c);
+ else
+ {
+ const unsigned char* a = sqlite3_value_text (argv[0]);
+ const unsigned char* b = sqlite3_value_text (argv[1]);
+ int i = 0;
+ while (*a++ == *b++)
+ i++;
+ sqlite3_result_int (c, i);
+ }
+}
+
+
+int
+main (int argc, char *argv[])
+{
+ (void) setlocale (LC_ALL, "");
+ (void) bindtextdomain (PACKAGE_TARNAME, LOCALEDIR);
+ (void) textdomain (PACKAGE_TARNAME);
+
+ /* Tell the library which version we are expecting. */
+ elf_version (EV_CURRENT);
+
+ /* Set computed default values. */
+ db_path = string(getenv("HOME") ?: "/") + string("/.debuginfod.sqlite"); /* XDG? */
+ int rc = regcomp (& file_include_regex, ".*", REG_EXTENDED|REG_NOSUB); // match everything
+ if (rc != 0)
+ error (EXIT_FAILURE, 0, "regcomp failure: %d", rc);
+ rc = regcomp (& file_exclude_regex, "^$", REG_EXTENDED|REG_NOSUB); // match nothing
+ if (rc != 0)
+ error (EXIT_FAILURE, 0, "regcomp failure: %d", rc);
+
+ /* Parse and process arguments. */
+ int remaining;
+ argp_program_version_hook = print_version; // this works
+ (void) argp_parse (&argp, argc, argv, ARGP_IN_ORDER, &remaining, NULL);
+ if (remaining != argc)
+ error (EXIT_FAILURE, 0,
+ "unexpected argument: %s", argv[remaining]);
+
+ if (!scan_rpms && !scan_files && source_paths.size()>0)
+ obatched(clog) << "warning: without -F and/or -R, ignoring PATHs" << endl;
+
+ (void) signal (SIGPIPE, SIG_IGN); // microhttpd can generate it incidentally, ignore
+ (void) signal (SIGINT, signal_handler); // ^C
+ (void) signal (SIGHUP, signal_handler); // EOF
+ (void) signal (SIGTERM, signal_handler); // systemd
+ (void) signal (SIGUSR1, sigusr1_handler); // end-user
+ (void) signal (SIGUSR2, sigusr2_handler); // end-user
+
+ // do this before any threads start
+ scan_concurrency_sem = new semaphore(concurrency);
+
+ /* Get database ready. */
+ rc = sqlite3_open_v2 (db_path.c_str(), &db, (SQLITE_OPEN_READWRITE
+ |SQLITE_OPEN_CREATE
+ |SQLITE_OPEN_FULLMUTEX), /* thread-safe */
+ NULL);
+ if (rc == SQLITE_CORRUPT)
+ {
+ (void) unlink (db_path.c_str());
+ error (EXIT_FAILURE, 0,
+ "cannot open %s, deleted database: %s", db_path.c_str(), sqlite3_errmsg(db));
+ }
+ else if (rc)
+ {
+ error (EXIT_FAILURE, 0,
+ "cannot open %s, consider deleting database: %s", db_path.c_str(), sqlite3_errmsg(db));
+ }
+
+ obatched(clog) << "opened database " << db_path << endl;
+ obatched(clog) << "sqlite version " << sqlite3_version << endl;
+
+ // add special string-prefix-similarity function used in rpm sref/sdef resolution
+ rc = sqlite3_create_function(db, "sharedprefix", 2, SQLITE_UTF8, NULL,
+ & sqlite3_sharedprefix_fn, NULL, NULL);
+ if (rc != SQLITE_OK)
+ error (EXIT_FAILURE, 0,
+ "cannot create sharedprefix( function: %s", sqlite3_errmsg(db));
+
+ if (verbose > 3)
+ obatched(clog) << "ddl: " << DEBUGINFOD_SQLITE_DDL << endl;
+ rc = sqlite3_exec (db, DEBUGINFOD_SQLITE_DDL, NULL, NULL, NULL);
+ if (rc != SQLITE_OK)
+ {
+ error (EXIT_FAILURE, 0,
+ "cannot run database schema ddl: %s", sqlite3_errmsg(db));
+ }
+
+ // Start httpd server threads. Separate pool for IPv4 and IPv6, in
+ // case the host only has one protocol stack.
+ MHD_Daemon *d4 = MHD_start_daemon (MHD_USE_THREAD_PER_CONNECTION
+#if MHD_VERSION >= 0x00095300
+ | MHD_USE_INTERNAL_POLLING_THREAD
+#else
+ | MHD_USE_SELECT_INTERNALLY
+#endif
+ | MHD_USE_DEBUG, /* report errors to stderr */
+ http_port,
+ NULL, NULL, /* default accept policy */
+ handler_cb, NULL, /* handler callback */
+ MHD_OPTION_END);
+ MHD_Daemon *d6 = MHD_start_daemon (MHD_USE_THREAD_PER_CONNECTION
+#if MHD_VERSION >= 0x00095300
+ | MHD_USE_INTERNAL_POLLING_THREAD
+#else
+ | MHD_USE_SELECT_INTERNALLY
+#endif
+ | MHD_USE_IPv6
+ | MHD_USE_DEBUG, /* report errors to stderr */
+ http_port,
+ NULL, NULL, /* default accept policy */
+ handler_cb, NULL, /* handler callback */
+ MHD_OPTION_END);
+
+ if (d4 == NULL && d6 == NULL) // neither ipv4 nor ipv6? boo
+ {
+ sqlite3 *database = db;
+ db = 0; // for signal_handler not to freak
+ sqlite3_close (database);
+ error (EXIT_FAILURE, 0, "cannot start http server at port %d", http_port);
+ }
+
+ obatched(clog) << "started http server on "
+ << (d4 != NULL ? "IPv4 " : "")
+ << (d6 != NULL ? "IPv6 " : "")
+ << "port=" << http_port << endl;
+
+ // add maxigroom sql if -G given
+ if (maxigroom)
+ {
+ obatched(clog) << "maxigrooming database, please wait." << endl;
+ extra_ddl.push_back("create index if not exists " BUILDIDS "_r_sref_arc on " BUILDIDS "_r_sref(artifactsrc);");
+ extra_ddl.push_back("delete from " BUILDIDS "_r_sdef where not exists (select 1 from " BUILDIDS "_r_sref b where " BUILDIDS "_r_sdef.content = b.artifactsrc);");
+ extra_ddl.push_back("drop index if exists " BUILDIDS "_r_sref_arc;");
+
+ // NB: we don't maxigroom the _files interning table. It'd require a temp index on all the
+ // tables that have file foreign-keys, which is a lot.
+
+ // NB: with =delete, may take up 3x disk space total during vacuum process
+ // vs. =off (only 2x but may corrupt database if program dies mid-vacuum)
+ // vs. =wal (>3x observed, but safe)
+ extra_ddl.push_back("pragma journal_mode=delete;");
+ extra_ddl.push_back("vacuum;");
+ extra_ddl.push_back("pragma journal_mode=wal;");
+ }
+
+ // run extra -D sql if given
+ for (auto&& i: extra_ddl)
+ {
+ if (verbose > 1)
+ obatched(clog) << "extra ddl:\n" << i << endl;
+ rc = sqlite3_exec (db, i.c_str(), NULL, NULL, NULL);
+ if (rc != SQLITE_OK && rc != SQLITE_DONE && rc != SQLITE_ROW)
+ error (0, 0,
+ "warning: cannot run database extra ddl %s: %s", i.c_str(), sqlite3_errmsg(db));
+ }
+
+ if (maxigroom)
+ obatched(clog) << "maxigroomed database" << endl;
+
+
+ obatched(clog) << "search concurrency " << concurrency << endl;
+ obatched(clog) << "rescan time " << rescan_s << endl;
+ obatched(clog) << "groom time " << groom_s << endl;
+ const char* du = getenv(DEBUGINFOD_URLS_ENV_VAR);
+ if (du && du[0] != '\0') // set to non-empty string?
+ obatched(clog) << "upstream debuginfod servers: " << du << endl;
+
+ vector<pthread_t> source_file_scanner_threads;
+ vector<pthread_t> source_rpm_scanner_threads;
+ pthread_t groom_thread;
+
+ rc = pthread_create (& groom_thread, NULL, thread_main_groom, NULL);
+ if (rc < 0)
+ error (0, 0, "warning: cannot spawn thread (%d) to groom database\n", rc);
+
+ if (scan_files) for (auto&& it : source_paths)
+ {
+ pthread_t pt;
+ rc = pthread_create (& pt, NULL, thread_main_scan_source_file_path, (void*) it.c_str());
+ if (rc < 0)
+ error (0, 0, "warning: cannot spawn thread (%d) to scan source files %s\n", rc, it.c_str());
+ else
+ source_file_scanner_threads.push_back(pt);
+ }
+
+ if (scan_rpms) for (auto&& it : source_paths)
+ {
+ pthread_t pt;
+ rc = pthread_create (& pt, NULL, thread_main_scan_source_rpm_path, (void*) it.c_str());
+ if (rc < 0)
+ error (0, 0, "warning: cannot spawn thread (%d) to scan source rpms %s\n", rc, it.c_str());
+ else
+ source_rpm_scanner_threads.push_back(pt);
+ }
+
+ /* Trivial main loop! */
+ set_metric("ready", 1);
+ while (! interrupted)
+ pause ();
+ set_metric("ready", 0);
+
+ if (verbose)
+ obatched(clog) << "stopping" << endl;
+
+ /* Join any source scanning threads. */
+ for (auto&& it : source_file_scanner_threads)
+ pthread_join (it, NULL);
+ for (auto&& it : source_rpm_scanner_threads)
+ pthread_join (it, NULL);
+ pthread_join (groom_thread, NULL);
+
+ /* Stop all the web service threads. */
+ if (d4) MHD_stop_daemon (d4);
+ if (d6) MHD_stop_daemon (d6);
+
+ /* With all threads known dead, we can clean up the global resources. */
+ delete scan_concurrency_sem;
+ rc = sqlite3_exec (db, DEBUGINFOD_SQLITE_CLEANUP_DDL, NULL, NULL, NULL);
+ if (rc != SQLITE_OK)
+ {
+ error (0, 0,
+ "warning: cannot run database cleanup ddl: %s", sqlite3_errmsg(db));
+ }
+
+ // NB: no problem with unconditional free here - an earlier failed regcomp would exit program
+ (void) regfree (& file_include_regex);
+ (void) regfree (& file_exclude_regex);
+
+ sqlite3 *database = db;
+ db = 0; // for signal_handler not to freak
+ (void) sqlite3_close (database);
+
+ return 0;
+}
diff --git a/debuginfod/debuginfod.h b/debuginfod/debuginfod.h
new file mode 100644
index 00000000..6b1b1cc3
--- /dev/null
+++ b/debuginfod/debuginfod.h
@@ -0,0 +1,85 @@
+/* External declarations for the libdebuginfod client library.
+ Copyright (C) 2019 Red Hat, Inc.
+ This file is part of elfutils.
+
+ This file is free software; you can redistribute it and/or modify
+ it under the terms of either
+
+ * the GNU Lesser General Public License as published by the Free
+ Software Foundation; either version 3 of the License, or (at
+ your option) any later version
+
+ or
+
+ * the GNU General Public License as published by the Free
+ Software Foundation; either version 2 of the License, or (at
+ your option) any later version
+
+ or both in parallel, as here.
+
+ elfutils is distributed in the hope that it will be useful, but
+ WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ General Public License for more details.
+
+ You should have received copies of the GNU General Public License and
+ the GNU Lesser General Public License along with this program. If
+ not, see <http://www.gnu.org/licenses/>. */
+
+#ifndef _DEBUGINFOD_CLIENT_H
+#define _DEBUGINFOD_CLIENT_H 1
+
+/* Names of environment variables that control the client logic. */
+#define DEBUGINFOD_URLS_ENV_VAR "DEBUGINFOD_URLS"
+#define DEBUGINFOD_CACHE_PATH_ENV_VAR "DEBUGINFOD_CACHE_PATH"
+#define DEBUGINFOD_TIMEOUT_ENV_VAR "DEBUGINFOD_TIMEOUT"
+
+/* Handle for debuginfod-client connection. */
+typedef struct debuginfod_client debuginfod_client;
+
+#ifdef __cplusplus
+extern "C" {
+#endif
+
+/* Create a handle for a new debuginfod-client session. */
+debuginfod_client *debuginfod_begin (void);
+
+/* Query the urls contained in $DEBUGINFOD_URLS for a file with
+ the specified type and build id. If build_id_len == 0, the
+ build_id is supplied as a lowercase hexadecimal string; otherwise
+ it is a binary blob of given legnth.
+
+ If successful, return a file descriptor to the target, otherwise
+ return a posix error code. If successful, set *path to a
+ strdup'd copy of the name of the same file in the cache.
+ Caller must free() it later. */
+
+int debuginfod_find_debuginfo (debuginfod_client *client,
+ const unsigned char *build_id,
+ int build_id_len,
+ char **path);
+
+int debuginfod_find_executable (debuginfod_client *client,
+ const unsigned char *build_id,
+ int build_id_len,
+ char **path);
+
+int debuginfod_find_source (debuginfod_client *client,
+ const unsigned char *build_id,
+ int build_id_len,
+ const char *filename,
+ char **path);
+
+typedef int (*debuginfod_progressfn_t)(debuginfod_client *c, long a, long b);
+void debuginfod_set_progressfn(debuginfod_client *c,
+ debuginfod_progressfn_t fn);
+
+/* Release debuginfod client connection context handle. */
+void debuginfod_end (debuginfod_client *client);
+
+#ifdef __cplusplus
+}
+#endif
+
+
+#endif /* _DEBUGINFOD_CLIENT_H */
diff --git a/debuginfod/libdebuginfod.map b/debuginfod/libdebuginfod.map
new file mode 100644
index 00000000..0d26f93e
--- /dev/null
+++ b/debuginfod/libdebuginfod.map
@@ -0,0 +1,10 @@
+ELFUTILS_0 { };
+ELFUTILS_0.178 {
+ global:
+ debuginfod_begin;
+ debuginfod_end;
+ debuginfod_find_debuginfo;
+ debuginfod_find_executable;
+ debuginfod_find_source;
+ debuginfod_set_progressfn;
+} ELFUTILS_0;