frostfs-s3-gw/docs/shared-backend/DESIGN.md

5.7 KiB

Introduction Slack

This feature allows MinIO to serve a shared NAS drive across multiple MinIO instances. There are no special configuration changes required to enable this feature. Access to files stored on NAS volume are locked and synchronized by default.

Motivation

Since MinIO instances serve the purpose of a single tenant there is an increasing requirement where users want to run multiple MinIO instances on a same backend which is managed by an existing NAS (NFS, GlusterFS, Other distributed filesystems) rather than a local disk. This feature is implemented also with minimal disruption in mind for the user and overall UI.

Restrictions

  • A PutObject() is blocked and waits if another GetObject() is in progress.
  • A CompleteMultipartUpload() is blocked and waits if another PutObject() or GetObject() is in progress.
  • Cannot run FS mode as a remote disk RPC.

How To Run?

Running MinIO instances on shared backend is no different than running on a stand-alone disk. There are no special configuration changes required to enable this feature. Access to files stored on NAS volume are locked and synchronized by default. Following examples will clarify this further for each operating system of your choice:

Ubuntu 16.04 LTS

Example 1: Start MinIO instance on a shared backend mounted and available at /path/to/nfs-volume.

On linux server1

minio gateway nas /path/to/nfs-volume

On linux server2

minio gateway nas /path/to/nfs-volume

Windows 2012 Server

Example 1: Start MinIO instance on a shared backend mounted and available at \\remote-server\cifs.

On windows server1

minio.exe gateway nas \\remote-server\cifs\data

On windows server2

minio.exe gateway nas \\remote-server\cifs\data

Alternatively if \\remote-server\cifs is mounted as D:\ drive.

On windows server1

minio.exe gateway nas D:\data

On windows server2

minio.exe gateway nas D:\data

Architecture

POSIX/Win32 Locks

Lock process

With in the same MinIO instance locking is handled by existing in-memory namespace locks (sync.RWMutex et. al). To synchronize locks between many MinIO instances we leverage POSIX fcntl() locks on Unixes and on Windows LockFileEx() Win32 API. Requesting write lock block if there are any read locks held by neighboring MinIO instance on the same path. So does the read lock if there are any active write locks in-progress.

Unlock process

Unlocking happens on filesystems locks by just closing the file descriptor (fd) which was initially requested for lock operation. Closing the fd tells the kernel to relinquish all the locks held on the path by the current process. This gets trickier when there are many readers on the same path by the same process, it would mean that closing an fd relinquishes locks for all concurrent readers as well. To properly manage this situation a simple fd reference count is implemented, the same fd is shared between many readers. When readers start closing on the fd we start reducing the reference count, once reference count has reached zero we can be sure that there are no more readers active. So we proceed and close the underlying file descriptor which would relinquish the read lock held on the path.

This doesn't apply for the writes because there is always one writer and many readers for any unique object.

Handling Concurrency.

An example here shows how the contention is handled with GetObject().

GetObject() holds a read lock on fs.json.

	fsMetaPath := pathJoin(fs.fsPath, minioMetaBucket, bucketMetaPrefix, bucket, object, fsMetaJSONFile)
	rlk, err := fs.rwPool.Open(fsMetaPath)
	if err != nil {
		return toObjectErr(err, bucket, object)
	}
	defer rlk.Close()

... you can perform other operations here ...

	_, err = io.CopyBuffer(writer, reader, buf)

... after successful copy operation unlocks the read lock ...

A concurrent PutObject is requested on the same object, PutObject() attempts a write lock on fs.json.

	fsMetaPath := pathJoin(fs.fsPath, minioMetaBucket, bucketMetaPrefix, bucket, object, fsMetaJSONFile)
	wlk, err := fs.rwPool.Create(fsMetaPath)
	if err != nil {
		return ObjectInfo{}, toObjectErr(err, bucket, object)
	}
	// This close will allow for locks to be synchronized on `fs.json`.
	defer wlk.Close()

Now from the above snippet the following code one can notice that until the GetObject() returns writing to the client. Following portion of the code will block.

	wlk, err := fs.rwPool.Create(fsMetaPath)

This restriction is needed so that corrupted data is not returned to the client in between I/O. The logic works vice-versa as well an on-going PutObject(), GetObject() would wait for the PutObject() to complete.

Caveats (concurrency)

Consider for example 3 servers sharing the same backend

On minio1

  • DeleteObject(object1) --> lock acquired on fs.json while object1 is being deleted.

On minio2

  • PutObject(object1) --> lock waiting until DeleteObject finishes.

On minio3

  • PutObject(object1) --> (concurrent request during PutObject minio2 checking if fs.json exists)

Once lock is acquired the minio2 validates if the file really exists to avoid obtaining lock on an fd which is already deleted. But this situation calls for a race with a third server which is also attempting to write the same file before the minio2 can validate if the file exists. It might be potentially possible fs.json is created so the lock acquired by minio2 might be invalid and can lead to a potential inconsistency.

This is a known problem and cannot be solved by POSIX fcntl locks. These are considered to be the limits of shared filesystem.