Tim Serong [Tue, 5 Jul 2016 12:01:25 +0000 (22:01 +1000)]
cmake: Really add FCGI_INCLUDE_DIR to include_directories for rgw
Commit 3cf6c53 was incorrect. FCGI_INCLUDE_DIR doesn't need to be
set in src/rgw/CMakeLists.txt, but it does need to be set for the
rgw_a target in src/CMakeLists.txt, as well as for the
ceph_test_librgw_file and ceph_test_librgw_file_nfsns targets in
src/test/CMakeLists.txt. I can only assume that I must not have
done a completely clean rebuild at some point when testing a
reworked version of the earlier commit :-/
This is only a problem for distros that keep the FCGI headers in
/usr/include/fastcgi/ (e.g.: SUSE).
This commit also removes a redundant include of <fcgiapp.h>
test: cephtool/test.sh: fix expect_false with pipe
this fixes failures like,
/home/jenkins-build/build/workspace/ceph-pull-requests/qa/workunits/cephtool/test.sh:
line 32: ceph osd blacklist ls | grep 192.168.0.1: command not found
where the failure is not the "failure" we are expecting.
in our tests, following command
expect_false "ceph osd blacklist ls | grep 192.168.0.1"
is designed to to verify that "ceph osd blacklist ls | grep 192.168.0.1"
fails with non-zero return code. but expect_false() evaluates the command
line using plain "$@", which will send the arguments direct to the shell,
and $0 is "ceph auth get client.xx | grep caps | grep mon", which does
not exist and is not built-in command. so we need to check the grep
command instead.
for multiple piped command line, use
expect_false sh <<< "echo foo | grep bar | grep baz"
Loic Dachary [Thu, 26 May 2016 07:38:47 +0000 (09:38 +0200)]
ceph-disk: partprobe should block udev induced BLKRRPART
Wrap partprobe with flock to stop udev from issuing BLKRRPART because
this is racy and frequently fails with a message like:
Error: Error informing the kernel about modifications to partition
/dev/vdc1 -- Device or resource busy. This means Linux won't know about
any changes you made to /dev/vdc1 until you reboot -- so you shouldn't
mount it or use it in any way before rebooting.
Opening a device (/dev/vdc for instance) in write mode indirectly
triggers a BLKRRPART ioctl from udev (starting version 214 and up)
when the device is closed (see below for the udev release note).
However, if udev fails to acquire an exclusive lock (with
flock(fd, LOCK_EX|LOCK_NB); ) the BLKRRPART ioctl is not issued.
Acquiring an exclusive lock before running the process that opens the
device in write mode is therefore an effective way to control this
behavior.
git clone git://anonscm.debian.org/pkg-systemd/systemd.git
systemd/NEWS:
CHANGES WITH 214:
* As an experimental feature, udev now tries to lock the
disk device node (flock(LOCK_SH|LOCK_NB)) while it
executes events for the disk or any of its partitions.
Applications like partitioning programs can lock the
disk device node (flock(LOCK_EX)) and claim temporary
device ownership that way; udev will entirely skip all event
handling for this disk and its partitions. If the disk
was opened for writing, the close will trigger a partition
table rescan in udev's "watch" facility, and if needed
synthesize "change" events for the disk and all its partitions.
This is now unconditionally enabled, and if it turns out to
cause major problems, we might turn it on only for specific
devices, or might need to disable it entirely. Device Mapper
devices are excluded from this logic.
Kefu Chai [Fri, 17 Jun 2016 05:58:55 +0000 (13:58 +0800)]
msg/simple: set close on exec on server sockets
mds execv() when handling the "respawn" command, to avoid fd leakage,
and enormous CLOSE_WAIT connections after respawning, we need to set
FD_CLOEXEC flag for the socket fds.
Kefu Chai [Thu, 16 Jun 2016 17:17:05 +0000 (01:17 +0800)]
msg/async: set close on exec on server sockets
mds execv() when handling the "respawn" command, to avoid fd leakage,
and enormous CLOSE_WAIT connections after respawning, we need to set
FD_CLOEXEC flag for the socket fds.
os/bluestore: add compression required ratio to enable/disable compression
Require the net gain of compression at least to be at a specified ratio,
otherwise we don't compress.
Ask for compressing at least 12.5% off, by default.
This is for the sake of performance because if the compression turns out
to be meaningless(saving little space), we can simply shut it down, as we
know the compression/decompression can be rather CPU-consuming.
crush: reset bucket->h.items[i] when removing tree item
* crush: so we don't see the reference after the removing, this keeps
check_item_loc() happy, and move_bucket() use check_item_loc() to see if
the removed bucket disappears after the removal.
* test: also add unittest_crush_wrapper::CrushWrapper.insert_item
cmake: remove unnecessary linked libs from libcephfs
* some of the libs shares the same .cc which has static C++ variables. if
we link against the different libs sharing the same static C++
variables, and the dtor of the C++ variables has side-effects, among
other things, deallocates a memory chunk. then, we are in the trouble of
double free. so "osd" lib is removed.
* some of the libs are referenced by the linked lib, so no need to link
against them again. for example, BLKID_LIBRARIES are linked by
libcommon, so we can remove it from the linked libs list.
* and lib "os" and "cls_references_objs" are not used by libcephfs at all,
so remove them.
rgw: forward input data when forwarding set_bucket_version to master
Fixes: http://tracker.ceph.com/issues/16494
Needed to keep input data around to be forwarded correctly. Also, master
does not send any data back, so don't try to parse anything.
Sage Weil [Fri, 24 Jun 2016 13:43:52 +0000 (09:43 -0400)]
os/bluestore/BlueFS: make _sync_and_flush_log smarter
If we know what event we need to wait for, only wait long enough for it
to flush. This helps the situation where another thread flushed what we
needed, and more dirty stuff was added to log_t, but we don't need to
wait for that too for our caller to be happy.
Sage Weil [Fri, 24 Jun 2016 13:23:21 +0000 (09:23 -0400)]
os/bluestore: drop lock while we flush the log
Handle cases where we have multiple racing threads trying to flush the
log by only allowing one concurrent log flush to be in progress at a time,
and behave if, after flushing, there are no more dirty records to flush.
Sage Weil [Thu, 23 Jun 2016 13:39:31 +0000 (09:39 -0400)]
os/bluestore/BlueFS: drop lock while waiting for user io to complete
_flush_wait is safe to call without a lock, as long as our reference is
stable. Rename it wait_for_aio() to be more clear about what it does and
the fact that it doesn't require a lock.
Sage Weil [Thu, 23 Jun 2016 13:33:09 +0000 (09:33 -0400)]
os/bluestore/BlueFS: track dirty by log_seq, log_seq_stable
Note when we dirty a file, and clean it only if that seq has been
committed. Currently this is always the case because we don't drop the
lock, but that will change shortly.