An object store is a collection of resources and services that work together to serve HTTP requests to PUT and GET objects. Rook will automate the configuration of the Ceph resources and services that are necessary to start and maintain a highly available, durable, and performant object store.
The Ceph object store supports S3 and Swift APIs and a multitude of features such as replication of object stores between different zones. The Rook object store is designed to support all of these features, though will take some time to implement them. We welcome contributions! In the meantime, features that are not yet implemented can be configured by using the Rook toolbox to run the radosgw-admin
and other tools for advanced configuration.
A Rook storage cluster must be configured and running in Kubernetes. In this example, it is assumed the cluster is in the rook
namespace.
When the storage admin is ready to create an object storage, the admin will specify his desired configuration settings in a yaml file such as the following object-store.yaml
. This example is a simple object store with metadata that is replicated across different hosts, and the data is erasure coded across multiple devices in the cluster.
apiVersion: ceph.rook.io/v1alpha1
kind: CephObjectStore
metadata:
name: my-store
namespace: rook-ceph
spec:
metadataPool:
failureDomain: host
replicated:
size: 3
dataPool:
failureDomain: device
erasureCoded:
dataChunks: 6
codingChunks: 2
gateway:
port: 80
securePort: 443
instances: 3
Now create the object store.
kubectl create -f object-store.yaml
At this point the Rook operator recognizes that a new object store resource needs to be configured. The operator will create all of the resources to start the object store.
.rgw.root
, my-store.rgw.control
, my-store.rgw.meta
, my-store.rgw.log
, my-store.rgw.buckets.index
)my-store.rgw.buckets.data
)When the RGW pods start, the object store is ready to receive the http or https requests as configured.
The object store settings are exposed to Rook as a Custom Resource Definition (CRD). The CRD is the Kubernetes-native means by which the Rook operator can watch for new resources. The operator stays in a control loop to watch for a new object store, changes to an existing object store, or requests to delete an object store.
The pools are the backing data store for the object store and are created with specific names to be private to an object store. Pools can be configured with all of the settings that can be specified in the Pool CRD. The underlying schema for pools defined by a pool CRD is the same as the schema under the metadataPool
and dataPool
elements of the object store CRD. All metadata pools are created with the same settings, while the data pool can be created with independent settings. The metadata pools must use replication, while the data pool can use replication or erasure coding.
If preservePoolsOnDelete
is set to 'true' the pools used to support the object store will remain when the object store will be deleted. This is a security measure to avoid accidental loss of data. It is set to 'false' by default. If not specified is also deemed as 'false'.
metadataPool:
failureDomain: host
replicated:
size: 3
dataPool:
failureDomain: device
erasureCoded:
dataChunks: 6
codingChunks: 2
preservePoolsOnDelete: true
If there is a zone
section in object-store configuration, then the pool section in the ceph-object-zone resource will be used to define the pools.
The gateway settings correspond to the RGW service.
type
: Can be s3
. In the future support for swift
can be added.sslCertificateRef
: If specified, this is the name of the Kubernetes secret that contains the SSL
certificate to be used for secure connections to the object store. The secret must be in the same
namespace as the Rook cluster. If it is an opaque Kubernetes Secret, Rook will look in the secret
provided at the cert
key name. The value of the cert
key must be in the format expected by the
RGW
service:
"The server key, server certificate, and any other CA or intermediate certificates be supplied in
one file. Each of these items must be in pem form." If the certificate is not specified, SSL will
not be configured.caBundleRef
: If specified, this is the name of the Kubernetes secret (type opaque
) that contains ca-bundle to use. The secret must be in the same namespace as the Rook cluster. Rook will look in the secret provided at the cabundle
key name.port
: The service port where the RGW service will be listening (http)securePort
: The service port where the RGW service will be listening (https)instances
: The number of RGW pods that will be started for this object storeplacement
: The rgw pods can be given standard Kubernetes placement restrictions with nodeAffinity
, tolerations
, podAffinity
, podAntiAffinity
, and topologySpreadConstraints
similar to placement defined for daemons configured by the cluster CRD.The RGW service can be configured to listen on both http and https by specifying both port
and securePort
.
gateway:
sslCertificateRef: my-ssl-cert-secret
securePort: 443
instances: 1
By default, the object store will be created independently from any other object stores and replication to another object store will not be configured. This done by creating a new Ceph realm, zone group, and zone all with the name of the new object store.
If desired to configure the object store to replicate and sync data amongst object-store or Ceph clusters, the zone
section would be required.
This section enables the object store to be part of a specified ceph-object-zone.
Specifying this section also ensures that the pool section in the ceph-object-zone is used for the object-store. If pools are specified for the object-store they are neither created nor deleted.
name
: name of the ceph-object-zone the object store is in. This name must be of a ceph-object-zone resource not just of a zone that has been already created.
zone:
name: "name"
The Ceph Object Gateway supports integrating with LDAP for authenticating and creating users, please refer here. This means that the rgw backend user
is also required to be part of groups in the LDAP server otherwise, authentication will fail. The rgw backend user
can be generated from CephObjectStoreUser
or ObjectBucketClaim
CRDs. For the both resources credentials are saved in Kubernetes Secrets which may not be valid with LDAP Server
, user need to follow the steps mentioned here.The following settings need to be configured in the RGW server:
rgw ldap binddn =
rgw ldap secret = /etc/ceph/ldap/bindpass.secret
rgw ldap uri =
rgw ldap searchdn =
rgw ldap dnattr =
rgw ldap searchfilter =
rgw s3 auth use ldap = true
So the CRD for the Ceph Object Store will be modified to include the above changes:
spec:
security
ldap:
config:
uri: ldaps://ldap-server:636
binddn: "uid=ceph,cn=users,cn=accounts,dc=example,dc=com"
searchdn: "cn=users,cn=accounts,dc=example,dc=com"
dnattr: "uid"
searchfilter: "memberof=cn=s3,cn=groups,cn=accounts,dc=example,dc=com"
credential:
volumeSource:
secret:
secretName: object-my-store-ldap-creds
defaultMode: 0600 #required
The config
section includes options used for RGW wrt LDAP server. These options are strongly typed rather than string map approach since very less chance to modify in future.
uri
: It specifies the address of LDAP server to use.* binddn
: The bind domain for the service account used by RGW server.searchdn
: The search domain where can it look for the user details.dnattr
: The attribute being used in the constructed search filter to match a username, this can either be uid
or cn
.searchfilter
: A generic search filter. If dnattr
is set, this filter is &()
'd together with the automatically constructed filter.The credential
defines where the password for accessing ldap server should be sourced from
volumeSource
: this is a standard Kubernetes VolumeSource for the Kerberos keytab file like
what is normally used to configure Volumes for a Pod. For example, a Secret or HostPath.
There are two requirements for the source's content:
subPath: password
. For example, in a Secret, the
data item must be named password
, or items
must be defined to select the key and
give it path password
. A HostPath directory must have the password
file.The CA bundle for ldap can be added to the caBundleRef
option in Gateway
settings:
spec:
gateway:
caBundleRef: #ldaps-cabundle