# Copyright (c) 2012 Red Hat, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or # implied. # See the License for the specific language governing permissions and # limitations under the License. import logging import os, fcntl, time from ConfigParser import ConfigParser from swift.common.utils import TRUE_VALUES from gluster.swift.common.fs_utils import mkdirs # # Read the fs.conf file once at startup (module load) # _fs_conf = ConfigParser() MOUNT_IP = 'localhost' REMOTE_CLUSTER = False OBJECT_ONLY = False if _fs_conf.read(os.path.join('/etc/swift', 'fs.conf')): try: MOUNT_IP = _fs_conf.get('DEFAULT', 'mount_ip', 'localhost') except (NoSectionError, NoOptionError): pass try: REMOTE_CLUSTER = _fs_conf.get('DEFAULT', 'remote_cluster', False) in TRUE_VALUES except (NoSectionError, NoOptionError): pass try: OBJECT_ONLY = _fs_conf.get('DEFAULT', 'object_only', "no") in TRUE_VALUES except (NoSectionError, NoOptionError): pass NAME = 'glusterfs' def _busy_wait(full_mount_path): # Iterate for definite number of time over a given # interval for successful mount for i in range(0, 5): if os.path.ismount(os.path.join(full_mount_path)): return True time.sleep(2) logging.error('Busy wait for mount timed out for mount %s', full_mount_path) return False def mount(root, drive): # FIXME: Possible thundering herd problem here el = _get_export_list() for export in el: if drive == export: break else: logging.error('No export found in %r matching drive %s', el, drive) return False # NOTE: root is typically the default value of /mnt/gluster-object full_mount_path = os.path.join(root, drive) if not os.path.isdir(full_mount_path): mkdirs(full_mount_path) pid_dir = "/var/lib/glusterd/vols/%s/run/" % drive pid_file = os.path.join(pid_dir, 'swift.pid'); if not os.path.exists(pid_dir): mkdirs(pid_dir) fd = os.open(pid_file, os.O_CREAT|os.O_RDWR) with os.fdopen(fd, 'r+b') as f: try: fcntl.lockf(f, fcntl.LOCK_EX|fcntl.LOCK_NB) except: ex = sys.exc_info()[1] if isinstance(ex, IOError) and ex.errno in (EACCES, EAGAIN): # This means that some other process is mounting the # filesystem, so wait for the mount process to complete return _busy_wait(full_mount_path) mnt_cmd = 'mount -t glusterfs %s:%s %s' % (MOUNT_IP, export, \ full_mount_path) if os.system(mnt_cmd) or not _busy_wait(full_mount_path): logging.error('Mount failed %s: %s', NAME, mnt_cmd) return False return True def unmount(full_mount_path): # FIXME: Possible thundering herd problem here umnt_cmd = 'umount %s 2>> /dev/null' % full_mount_path if os.system(umnt_cmd): logging.error('Unable to unmount %s %s' % (full_mount_path, NAME)) def _get_export_list(): if REMOTE_CLUSTER: cmnd = 'gluster --remote-host=%s volume info' % MOUNT_IP else: cmnd = 'gluster volume info' export_list = [] if os.system(cmnd + ' >> /dev/null'): if REMOTE_CLUSTER: logging.error('Getting volume info failed for %s, make sure '\ 'gluster --remote-host=%s works', NAME, MOUNT_IP) else: logging.error('Getting volume info failed for %s', NAME) else: fp = os.popen(cmnd) while True: item = fp.readline() if not item: break item = item.strip('\n').strip(' ') if item.lower().startswith('volume name:'): export_list.append(item.split(':')[1].strip(' ')) return export_list