From: John Wilkins Date: Sat, 17 Aug 2013 00:31:43 +0000 (-0700) Subject: doc: Removed old mkcephfs references. X-Git-Tag: v0.68~57 X-Git-Url: http://git-server-git.apps.pok.os.sepia.ceph.com/?a=commitdiff_plain;h=50cc2efccf11ce04bd9cd93e636074c32c427393;p=ceph.git doc: Removed old mkcephfs references. Signed-off-by: John Wilkins --- diff --git a/doc/rados/deployment/ceph-deploy-transition.rst b/doc/rados/deployment/ceph-deploy-transition.rst deleted file mode 100644 index 74e28c4fdc03..000000000000 --- a/doc/rados/deployment/ceph-deploy-transition.rst +++ /dev/null @@ -1,35 +0,0 @@ -============================== - Transitioning to ceph-deploy -============================== - -If you have an existing cluster that you deployed with ``mkcephfs``, -you will need to make a few changes to your configuration to -ensure that your cluster will work with ``ceph-deploy``. - -Monitor Keyring -=============== - -You will need to add ``caps mon = "allow *"`` to your monitor keyring if it is -not already in the keyring. By default, the monitor keyring is located under -``/var/lib/ceph/mon/ceph-$id/keyring``. When you have added the ``caps`` -setting, your monitor keyring should look something like this:: - - [mon.] - key = AQBJIHhRuHCwDRAAZjBTSJcIBIoGpdOR9ToiyQ== - caps mon = "allow *" - -Adding ``caps mon = "allow *"`` will ease the transition from ``mkcephfs`` to -``ceph-deploy`` by allowing ``ceph-create-keys`` to use the ``mon.`` keyring -file in ``$mon_data`` and get the caps it needs. - - -Use Default Paths -================= - -Under the ``/var/lib/ceph`` directory, the ``mon`` and ``osd`` directories need -to use the default paths. - -- **OSDs**: The path should be ``/var/lib/ceph/osd/ceph-$id`` -- **MON**: The path should be ``/var/lib/ceph/mon/ceph-$id`` - -Under those directories, the keyring should be in a file named ``keyring``. \ No newline at end of file diff --git a/doc/rados/deployment/mkcephfs.rst b/doc/rados/deployment/mkcephfs.rst deleted file mode 100644 index 168922822902..000000000000 --- a/doc/rados/deployment/mkcephfs.rst +++ /dev/null @@ -1,155 +0,0 @@ -============================= - Deploying with ``mkcephfs`` -============================= - -The ``mkcephfs`` tool is the old method of deploying new Ceph -clusters. It is now deprecated in favor of ``ceph-deploy``, which has -better support for modifying an existing cluster. - - -Enable Login to Cluster Hosts as ``root`` -========================================= - -To deploy with ``mkcephfs``, you will need to be able to login as ``root`` -on each host without a password. For each host, perform the following:: - - sudo passwd root - -Enter a password for the root user. - -On the admin host, generate an ``ssh`` key without specifying a passphrase -and use the default locations. :: - - sudo -i - ssh-keygen - Generating public/private key pair. - Enter file in which to save the key (/root/.ssh/id_rsa): - Enter passphrase (empty for no passphrase): - Enter same passphrase again: - Your identification has been saved in /root/.ssh/id_rsa. - Your public key has been saved in /root/.ssh/id_rsa.pub. - -Modify your ``/root/.ssh/config`` file to login as ``root``, as follows:: - - Host myserver01 - Hostname myserver01.fully-qualified-domain.com - User root - Host myserver02 - Hostname myserver02.fully-qualified-domain.com - User root - -You may use RSA or DSA keys. Once you generate your keys, copy them to each -OSD host. For example:: - - ssh-copy-id root@myserver01 - ssh-copy-id root@myserver02 - -Copy Configuration File to All Hosts -==================================== - -Ceph's ``mkcephfs`` deployment script does not copy the configuration file you -created from the Administration host to the OSD Cluster hosts. Copy the -configuration file you created (*i.e.,* ``mycluster.conf`` in the example below) -from the Administration host to ``etc/ceph/ceph.conf`` on each OSD Cluster host -if you are using ``mkcephfs`` to deploy Ceph. - -:: - - sudo ssh myserver01 tee /etc/ceph/ceph.conf < /etc/ceph/ceph.conf - sudo ssh myserver02 tee /etc/ceph/ceph.conf < /etc/ceph/ceph.conf - sudo ssh myserver03 tee /etc/ceph/ceph.conf < /etc/ceph/ceph.conf - - -Create the Default Directories -============================== - -The ``mkcephfs`` deployment script does not create the default server -directories. Create server directories for each instance of a Ceph daemon (if -you haven't done so already). The ``host`` variables in the ``ceph.conf`` file -determine which host runs each instance of a Ceph daemon. Using the exemplary -``ceph.conf`` file, you would perform the following: - -On ``myserver01``:: - - sudo mkdir /var/lib/ceph/osd/ceph-0 - sudo mkdir /var/lib/ceph/mon/ceph-a - -On ``myserver02``:: - - sudo mkdir /var/lib/ceph/osd/ceph-1 - sudo mkdir /var/lib/ceph/mon/ceph-b - -On ``myserver03``:: - - sudo mkdir /var/lib/ceph/osd/ceph-2 - sudo mkdir /var/lib/ceph/mon/ceph-c - sudo mkdir /var/lib/ceph/mds/ceph-a - - -Mount Disks to the Data Directories -=================================== - -If you are running multiple OSDs per host and one hard disk per OSD, you should -mount the disk under the OSD data directory (if you haven't done so already). -When mounting disks in this manner, there is no need for an entry in -``/etc/fstab``. - -.. versionadded:: 0.56 - -For Bobtail (v 0.56) and beyond, you may specify the file system type, filesystem -options, and mount options. Add the following to the ``[global]`` section of your -Ceph configuration file, and replace the values in braces with appropriate values:: - - osd mkfs type = {fs-type} - osd mkfs options {fs-type} = {mkfs options} # default for xfs is "-f" - osd mount options {fs-type} = {mount options} # default mount option is "rw, noatime" - -For example:: - - osd mkfs type = btrfs - osd mkfs options btrfs = -m raid0 - osd mount options btrfs = rw, noatime - -For each ``[osd.n]`` section of your configuration file, specify the storage device. -For example:: - - [osd.1] - devs = /dev/sda - [osd.2] - devs = /dev/sdb - - -Run ``mkcephfs`` -================ - -Once you have copied your Ceph Configuration to the OSD Cluster hosts -and created the default directories, you may deploy Ceph with the -``mkcephfs`` script. - -.. note:: ``mkcephfs`` is a quick bootstrapping tool. It does not handle more - complex operations, such as upgrades. - -To run ``mkcephfs``, execute the following:: - - cd /etc/ceph - sudo mkcephfs -a -c /etc/ceph/ceph.conf -k ceph.keyring - -.. note:: For ``mkcephfs`` to use the ``mkfs`` configuration options, you MUST - specify a ``devs`` entry for each OSD. - -The script adds an admin key to the ``ceph.keyring``, which is analogous to a -root password. See `Authentication`_ when running with ``cephx`` enabled. To -start the cluster, execute the following:: - - sudo service ceph -a start - -See `Operating a Cluster`_ for details. Also see `man mkcephfs`_. - -.. _Authentication: ../authentication -.. _Operating a Cluster: ../../operations/ -.. _man mkcephfs: ../../../man/8/mkcephfs - -.. toctree:: - :hidden: - - ../../../man/8/mkcephfs diff --git a/doc/start/quick-start.rst b/doc/start/quick-start.rst deleted file mode 100644 index a36921ceffef..000000000000 --- a/doc/start/quick-start.rst +++ /dev/null @@ -1,179 +0,0 @@ -====================== - 5-minute Quick Start -====================== - -Thank you for trying Ceph! Petabyte-scale data clusters are quite an -undertaking. Before delving deeper into Ceph, we recommend setting up a two-node -demo cluster to explore some of the functionality. The Ceph **5-Minute Quick -Start** deploys a Ceph object store cluster on one server machine and a Ceph -client on a separate machine, each with a recent Debian/Ubuntu operating system. -The intent of this **Quick Start** is to help you exercise Ceph object store -functionality without the configuration and deployment overhead associated with -a production-ready object store cluster. Once you complete this quick start, you -may exercise Ceph commands on the command line. You may also proceed to the -quick start guides for block devices, CephFS filesystems, and the RESTful -gateway. - -.. ditaa:: - /----------------\ /----------------\ - | Client Machine |<------->| Server Machine | - | cCCC | | cCCC | - +----------------+ +----------------+ - | Ceph Commands | | ceph - mon | - +----------------+ +----------------+ - | Block Device | | ceph - osd | - +----------------+ +----------------+ - | Ceph FS | | ceph - mds | - \----------------/ \----------------/ - - -Install Debian/Ubuntu -===================== - -Install a recent release of Debian or Ubuntu (e.g., 12.04 precise) on your -Ceph server machine and your client machine. - - -Add Ceph Packages -================= - -To get the latest Ceph packages, add a release key to :abbr:`APT (Advanced -Package Tool)`, add a source location to the ``/etc/apt/sources.list`` on your -Ceph server and client machines, update your systems and install Ceph. :: - - wget -q -O- 'https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc' | sudo apt-key add - - echo deb http://ceph.com/debian/ $(lsb_release -sc) main | sudo tee /etc/apt/sources.list.d/ceph.list - sudo apt-get update && sudo apt-get install ceph - -Check the Ceph version you are using and make a note of it so that you have the -correct settings in your configuration file:: - - ceph -v - -If ``ceph -v`` reflects an earlier version from what you installed, your -``ceph-common`` library may be using the version distributed with the kernel. -Once you've installed Ceph, you may also update and upgrade your packages to -ensure you have the latest ``ceph-common`` library installed. :: - - sudo apt-get update && sudo apt-get upgrade - -If you want to use a version other than the current release, see `Installing -Debian/Ubuntu Packages`_ for further details. - -.. _Installing Debian/Ubuntu Packages: ../../install/debian - - -Add a Configuration File -======================== - -The example configuration file will configure Ceph to operate a monitor, two OSD -daemons and one metadata server on your Ceph server machine. To add a -configuration file to Ceph, we suggest copying the contents of the example file -below to an editor. Then, follow the steps below to modify it. - -.. literalinclude:: ceph.conf - :language: ini - -#. Open a command line on your Ceph server machine and execute ``hostname -s`` - to retrieve the name of your Ceph server machine. - -#. Replace ``{hostname}`` in the sample configuration file with your host name. - -#. Execute ``ifconfig`` on the command line of your Ceph server machine to - retrieve the IP address of your Ceph server machine. - -#. Replace ``{ip-address}`` in the sample configuration file with the IP - address of your Ceph server host. - -#. Save the contents to ``/etc/ceph/ceph.conf`` on Ceph server host. - -#. Copy the configuration file to ``/etc/ceph/ceph.conf`` on your client host. :: - - sudo scp {user}@{server-machine}:/etc/ceph/ceph.conf /etc/ceph/ceph.conf - -.. tip:: Ensure the ``ceph.conf`` file has appropriate permissions set - (e.g. ``chmod 644``) on your client machine. - -.. versionadded:: 0.55 - -Ceph v0.55 and above have authentication enabled by default. You should -explicitly enable or disable authentication with version 0.55 and above. -The example configuration provides ``auth`` entries for authentication. -For details on Ceph authentication see `Cephx Configuration Reference`_ -and `Cephx Guide`_. - -.. _Cephx Guide: ../../rados/operations/authentication -.. _Cephx Configuration Reference: ../../rados/configuration/auth-config-ref - - -Deploy the Configuration -======================== - -You must perform the following steps to deploy the configuration. - -#. On your Ceph server host, create a directory for each daemon. For the - example configuration, execute the following:: - - sudo mkdir -p /var/lib/ceph/osd/ceph-0 - sudo mkdir -p /var/lib/ceph/osd/ceph-1 - sudo mkdir -p /var/lib/ceph/mon/ceph-a - sudo mkdir -p /var/lib/ceph/mds/ceph-a - - -#. Execute the following on the Ceph server host:: - - cd /etc/ceph - sudo mkcephfs -a -c /etc/ceph/ceph.conf -k ceph.keyring - - -Among other things, ``mkcephfs`` will deploy Ceph and generate a -``client.admin`` user and key. For Bobtail and subsequent versions (v 0.56 and -after), the ``mkcephfs`` script will create and mount the filesystem for you -provided you specify ``osd mkfs`` ``osd mount`` and ``devs`` settings in your -Ceph configuration file. - - -Start Ceph -========== - -Once you have deployed the configuration, start Ceph from the command line of -your server machine. :: - - sudo service ceph -a start - -Check the health of your Ceph cluster to ensure it is ready. :: - - sudo ceph health - -When your cluster echoes back ``HEALTH_OK``, you may begin using Ceph. - - -Copy The Keyring to The Client -============================== - -The next step you must perform is to copy ``/etc/ceph/ceph.keyring``, which -contains the ``client.admin`` key, from the server machine to the client -machine. If you don't perform this step, you will not be able to use the Ceph -command line, as the example Ceph configuration requires authentication. :: - - sudo scp {user}@{server-machine}:/etc/ceph/ceph.keyring /etc/ceph/ceph.keyring - -.. tip:: Ensure the ``ceph.keyring`` file has appropriate permissions set - (e.g., ``chmod 644``) on your client machine. - - -Proceed to Other Quick Starts -============================= - -Once you have Ceph running with both a client and a server, you -may proceed to the other Quick Start guides. - -#. For Ceph block devices, proceed to `Block Device Quick Start`_. - -#. For the CephFS filesystem, proceed to `CephFS Quick Start`_. - -#. For the RESTful Gateway, proceed to `Gateway Quick Start`_. - -.. _Block Device Quick Start: ../quick-rbd -.. _CephFS Quick Start: ../quick-cephfs -.. _Gateway Quick Start: ../quick-rgw \ No newline at end of file