123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276 |
- package filesys
- import (
- "context"
- "os"
- "sort"
- "time"
- "github.com/chrislusf/seaweedfs/weed/filer2"
- "github.com/chrislusf/seaweedfs/weed/glog"
- "github.com/chrislusf/seaweedfs/weed/pb/filer_pb"
- "github.com/seaweedfs/fuse"
- "github.com/seaweedfs/fuse/fs"
- )
- const blockSize = 512
- var _ = fs.Node(&File{})
- var _ = fs.NodeOpener(&File{})
- var _ = fs.NodeFsyncer(&File{})
- var _ = fs.NodeSetattrer(&File{})
- var _ = fs.NodeGetxattrer(&File{})
- var _ = fs.NodeSetxattrer(&File{})
- var _ = fs.NodeRemovexattrer(&File{})
- var _ = fs.NodeListxattrer(&File{})
- var _ = fs.NodeForgetter(&File{})
- type File struct {
- Name string
- dir *Dir
- wfs *WFS
- entry *filer_pb.Entry
- entryViewCache []filer2.VisibleInterval
- isOpen int
- }
- func (file *File) fullpath() filer2.FullPath {
- return filer2.NewFullPath(file.dir.Path, file.Name)
- }
- func (file *File) Attr(ctx context.Context, attr *fuse.Attr) error {
- glog.V(4).Infof("file Attr %s, open:%v, existing attr: %+v", file.fullpath(), file.isOpen, attr)
- if file.isOpen <= 0 {
- if err := file.maybeLoadEntry(ctx); err != nil {
- return err
- }
- }
- attr.Inode = file.fullpath().AsInode()
- attr.Valid = time.Second
- attr.Mode = os.FileMode(file.entry.Attributes.FileMode)
- attr.Size = filer2.TotalSize(file.entry.Chunks)
- if file.isOpen > 0 {
- attr.Size = file.entry.Attributes.FileSize
- glog.V(4).Infof("file Attr %s, open:%v, size: %d", file.fullpath(), file.isOpen, attr.Size)
- }
- attr.Crtime = time.Unix(file.entry.Attributes.Crtime, 0)
- attr.Mtime = time.Unix(file.entry.Attributes.Mtime, 0)
- attr.Gid = file.entry.Attributes.Gid
- attr.Uid = file.entry.Attributes.Uid
- attr.Blocks = attr.Size/blockSize + 1
- attr.BlockSize = uint32(file.wfs.option.ChunkSizeLimit)
- return nil
- }
- func (file *File) Getxattr(ctx context.Context, req *fuse.GetxattrRequest, resp *fuse.GetxattrResponse) error {
- // glog.V(4).Infof("file Getxattr %s", file.fullpath())
- if err := file.maybeLoadEntry(ctx); err != nil {
- return err
- }
- return getxattr(file.entry, req, resp)
- }
- func (file *File) Open(ctx context.Context, req *fuse.OpenRequest, resp *fuse.OpenResponse) (fs.Handle, error) {
- glog.V(4).Infof("file %v open %+v", file.fullpath(), req)
- file.isOpen++
- handle := file.wfs.AcquireHandle(file, req.Uid, req.Gid)
- resp.Handle = fuse.HandleID(handle.handle)
- glog.V(3).Infof("%v file open handle id = %d", file.fullpath(), handle.handle)
- return handle, nil
- }
- func (file *File) Setattr(ctx context.Context, req *fuse.SetattrRequest, resp *fuse.SetattrResponse) error {
- glog.V(3).Infof("%v file setattr %+v, old:%+v", file.fullpath(), req, file.entry.Attributes)
- if err := file.maybeLoadEntry(ctx); err != nil {
- return err
- }
- if req.Valid.Size() {
- glog.V(3).Infof("%v file setattr set size=%v", file.fullpath(), req.Size)
- if req.Size < filer2.TotalSize(file.entry.Chunks) {
- // fmt.Printf("truncate %v \n", fullPath)
- var chunks []*filer_pb.FileChunk
- for _, chunk := range file.entry.Chunks {
- int64Size := int64(chunk.Size)
- if chunk.Offset+int64Size > int64(req.Size) {
- int64Size = int64(req.Size) - chunk.Offset
- }
- if int64Size > 0 {
- chunks = append(chunks, chunk)
- }
- }
- file.entry.Chunks = chunks
- file.entryViewCache = nil
- }
- file.entry.Attributes.FileSize = req.Size
- }
- if req.Valid.Mode() {
- file.entry.Attributes.FileMode = uint32(req.Mode)
- }
- if req.Valid.Uid() {
- file.entry.Attributes.Uid = req.Uid
- }
- if req.Valid.Gid() {
- file.entry.Attributes.Gid = req.Gid
- }
- if req.Valid.Crtime() {
- file.entry.Attributes.Crtime = req.Crtime.Unix()
- }
- if req.Valid.Mtime() {
- file.entry.Attributes.Mtime = req.Mtime.Unix()
- }
- if file.isOpen > 0 {
- return nil
- }
- file.wfs.cacheDelete(file.fullpath())
- return file.saveEntry()
- }
- func (file *File) Setxattr(ctx context.Context, req *fuse.SetxattrRequest) error {
- glog.V(4).Infof("file Setxattr %s: %s", file.fullpath(), req.Name)
- if err := file.maybeLoadEntry(ctx); err != nil {
- return err
- }
- if err := setxattr(file.entry, req); err != nil {
- return err
- }
- file.wfs.cacheDelete(file.fullpath())
- return file.saveEntry()
- }
- func (file *File) Removexattr(ctx context.Context, req *fuse.RemovexattrRequest) error {
- glog.V(4).Infof("file Removexattr %s: %s", file.fullpath(), req.Name)
- if err := file.maybeLoadEntry(ctx); err != nil {
- return err
- }
- if err := removexattr(file.entry, req); err != nil {
- return err
- }
- file.wfs.cacheDelete(file.fullpath())
- return file.saveEntry()
- }
- func (file *File) Listxattr(ctx context.Context, req *fuse.ListxattrRequest, resp *fuse.ListxattrResponse) error {
- glog.V(4).Infof("file Listxattr %s", file.fullpath())
- if err := file.maybeLoadEntry(ctx); err != nil {
- return err
- }
- if err := listxattr(file.entry, req, resp); err != nil {
- return err
- }
- return nil
- }
- func (file *File) Fsync(ctx context.Context, req *fuse.FsyncRequest) error {
- // fsync works at OS level
- // write the file chunks to the filerGrpcAddress
- glog.V(3).Infof("%s/%s fsync file %+v", file.dir.Path, file.Name, req)
- return nil
- }
- func (file *File) Forget() {
- glog.V(3).Infof("Forget file %s/%s", file.dir.Path, file.Name)
- file.wfs.forgetNode(filer2.NewFullPath(file.dir.Path, file.Name))
- }
- func (file *File) maybeLoadEntry(ctx context.Context) error {
- if file.entry == nil || file.isOpen <= 0 {
- entry, err := file.wfs.maybeLoadEntry(file.dir.Path, file.Name)
- if err != nil {
- return err
- }
- if entry != nil {
- file.setEntry(entry)
- }
- }
- return nil
- }
- func (file *File) addChunks(chunks []*filer_pb.FileChunk) {
- sort.Slice(chunks, func(i, j int) bool {
- return chunks[i].Mtime < chunks[j].Mtime
- })
- var newVisibles []filer2.VisibleInterval
- for _, chunk := range chunks {
- newVisibles = filer2.MergeIntoVisibles(file.entryViewCache, newVisibles, chunk)
- t := file.entryViewCache[:0]
- file.entryViewCache = newVisibles
- newVisibles = t
- }
- glog.V(3).Infof("%s existing %d chunks adds %d more", file.fullpath(), len(file.entry.Chunks), len(chunks))
- file.entry.Chunks = append(file.entry.Chunks, chunks...)
- }
- func (file *File) setEntry(entry *filer_pb.Entry) {
- file.entry = entry
- file.entryViewCache = filer2.NonOverlappingVisibleIntervals(file.entry.Chunks)
- }
- func (file *File) saveEntry() error {
- return file.wfs.WithFilerClient(func(client filer_pb.SeaweedFilerClient) error {
- request := &filer_pb.UpdateEntryRequest{
- Directory: file.dir.Path,
- Entry: file.entry,
- }
- glog.V(1).Infof("save file entry: %v", request)
- _, err := client.UpdateEntry(context.Background(), request)
- if err != nil {
- glog.V(0).Infof("UpdateEntry file %s/%s: %v", file.dir.Path, file.Name, err)
- return fuse.EIO
- }
- return nil
- })
- }
|