forked from luck/tmp_suning_uos_patched
f7ab093f74
OrangeFS (formerly PVFS) is an lgpl licensed userspace networked parallel file system. OrangeFS can be accessed through included system utilities, user integration libraries, MPI-IO and can be used by the Hadoop ecosystem as an alternative to the HDFS filesystem. OrangeFS is used widely for parallel science, data analytics and engineering applications. While applications often don't require Orangefs to be mounted into the VFS, users do like to be able to access their files in the normal way. The Orangefs kernel client allows Orangefs filesystems to be mounted as a VFS. The kernel client communicates with a userspace daemon which in turn communicates with the Orangefs server daemons that implement the filesystem. The server daemons (there's almost always more than one) need not be running on the same host as the kernel client. Orangefs filesystems can also be mounted with FUSE, and we ship code and instructions to facilitate that, but most of our users report preferring to use our kernel module instead. Further, as an example of a problem we can't solve with fuse, we have in the works a not-yet-ready-for-prime-time version of a file_operations lock function that accounts for the server daemons being distributed across more than one running kernel. Many people and organizations, including Clemson University, Argonne National Laboratories and Acxiom Corporation have helped to create what has become Orangefs over more than twenty years. Some of the more recent contributors to the kernel client include: Mike Marshall Christoph Hellwig Randy Martin Becky Ligon Walt Ligon Michael Moore Rob Ross Phil Carnes Signed-off-by: Mike Marshall <hubcap@omnibond.com>
77 lines
2.1 KiB
C
77 lines
2.1 KiB
C
/*
|
|
* (C) 2001 Clemson University and The University of Chicago
|
|
*
|
|
* See COPYING in top-level directory.
|
|
*/
|
|
|
|
#ifndef __PVFS2_BUFMAP_H
|
|
#define __PVFS2_BUFMAP_H
|
|
|
|
/* used to describe mapped buffers */
|
|
struct pvfs_bufmap_desc {
|
|
void *uaddr; /* user space address pointer */
|
|
struct page **page_array; /* array of mapped pages */
|
|
int array_count; /* size of above arrays */
|
|
struct list_head list_link;
|
|
};
|
|
|
|
struct pvfs2_bufmap;
|
|
|
|
struct pvfs2_bufmap *pvfs2_bufmap_ref(void);
|
|
void pvfs2_bufmap_unref(struct pvfs2_bufmap *bufmap);
|
|
|
|
/*
|
|
* pvfs_bufmap_size_query is now an inline function because buffer
|
|
* sizes are not hardcoded
|
|
*/
|
|
int pvfs_bufmap_size_query(void);
|
|
|
|
int pvfs_bufmap_shift_query(void);
|
|
|
|
int pvfs_bufmap_initialize(struct PVFS_dev_map_desc *user_desc);
|
|
|
|
int get_bufmap_init(void);
|
|
|
|
void pvfs_bufmap_finalize(void);
|
|
|
|
int pvfs_bufmap_get(struct pvfs2_bufmap **mapp, int *buffer_index);
|
|
|
|
void pvfs_bufmap_put(struct pvfs2_bufmap *bufmap, int buffer_index);
|
|
|
|
int readdir_index_get(struct pvfs2_bufmap **mapp, int *buffer_index);
|
|
|
|
void readdir_index_put(struct pvfs2_bufmap *bufmap, int buffer_index);
|
|
|
|
int pvfs_bufmap_copy_iovec_from_user(struct pvfs2_bufmap *bufmap,
|
|
int buffer_index,
|
|
const struct iovec *iov,
|
|
unsigned long nr_segs,
|
|
size_t size);
|
|
|
|
int pvfs_bufmap_copy_iovec_from_kernel(struct pvfs2_bufmap *bufmap,
|
|
int buffer_index,
|
|
const struct iovec *iov,
|
|
unsigned long nr_segs,
|
|
size_t size);
|
|
|
|
int pvfs_bufmap_copy_to_user_iovec(struct pvfs2_bufmap *bufmap,
|
|
int buffer_index,
|
|
const struct iovec *iov,
|
|
unsigned long nr_segs,
|
|
size_t size);
|
|
|
|
int pvfs_bufmap_copy_to_kernel_iovec(struct pvfs2_bufmap *bufmap,
|
|
int buffer_index,
|
|
const struct iovec *iov,
|
|
unsigned long nr_segs,
|
|
size_t size);
|
|
|
|
size_t pvfs_bufmap_copy_to_user_task_iovec(struct task_struct *tsk,
|
|
struct iovec *iovec,
|
|
unsigned long nr_segs,
|
|
struct pvfs2_bufmap *bufmap,
|
|
int buffer_index,
|
|
size_t bytes_to_be_copied);
|
|
|
|
#endif /* __PVFS2_BUFMAP_H */
|