package shell import ( "context" "flag" "fmt" "io" "path/filepath" "strings" "time" "github.com/seaweedfs/seaweedfs/weed/filer" "github.com/seaweedfs/seaweedfs/weed/pb/filer_pb" "github.com/seaweedfs/seaweedfs/weed/util" ) func init() { Commands = append(Commands, &commandRemoteUncache{}) } type commandRemoteUncache struct { } func (c *commandRemoteUncache) Name() string { return "remote.uncache" } func (c *commandRemoteUncache) Help() string { return `keep the metadata but remote cache the file content for mounted directories or files This is designed to run regularly. So you can add it to some cronjob. If a file is not synchronized with the remote copy, the file will be skipped to avoid loss of data. remote.uncache -dir=/xxx remote.uncache -dir=/xxx/some/sub/dir remote.uncache -dir=/xxx/some/sub/dir -include=*.pdf remote.uncache -dir=/xxx/some/sub/dir -exclude=*.txt remote.uncache -minSize=1024000 # uncache files larger than 100K remote.uncache -minAge=3600 # uncache files older than 1 hour ` } func (c *commandRemoteUncache) Do(args []string, commandEnv *CommandEnv, writer io.Writer) (err error) { remoteUncacheCommand := flag.NewFlagSet(c.Name(), flag.ContinueOnError) dir := remoteUncacheCommand.String("dir", "", "a directory in filer") fileFiler := newFileFilter(remoteUncacheCommand) if err = remoteUncacheCommand.Parse(args); err != nil { return nil } mappings, listErr := filer.ReadMountMappings(commandEnv.option.GrpcDialOption, commandEnv.option.FilerAddress) if listErr != nil { return listErr } if *dir != "" { var localMountedDir string for k := range mappings.Mappings { if strings.HasPrefix(*dir, k) { localMountedDir = k } } if localMountedDir == "" { jsonPrintln(writer, mappings) fmt.Fprintf(writer, "%s is not mounted\n", *dir) return nil } // pull content from remote if err = c.uncacheContentData(commandEnv, writer, util.FullPath(*dir), fileFiler); err != nil { return fmt.Errorf("uncache content data: %v", err) } return nil } for key, _ := range mappings.Mappings { if err := c.uncacheContentData(commandEnv, writer, util.FullPath(key), fileFiler); err != nil { return err } } return nil } func (c *commandRemoteUncache) uncacheContentData(commandEnv *CommandEnv, writer io.Writer, dirToCache util.FullPath, fileFilter *FileFilter) error { return recursivelyTraverseDirectory(commandEnv, dirToCache, func(dir util.FullPath, entry *filer_pb.Entry) bool { if !mayHaveCachedToLocal(entry) { return true // true means recursive traversal should continue } if !fileFilter.matches(entry) { return true } if entry.RemoteEntry.LastLocalSyncTsNs/1e9 < entry.Attributes.Mtime { return true // should not uncache an entry that is not synchronized with remote } entry.RemoteEntry.LastLocalSyncTsNs = 0 entry.Chunks = nil fmt.Fprintf(writer, "Uncache %+v ... ", dir.Child(entry.Name)) err := commandEnv.WithFilerClient(false, func(client filer_pb.SeaweedFilerClient) error { _, updateErr := client.UpdateEntry(context.Background(), &filer_pb.UpdateEntryRequest{ Directory: string(dir), Entry: entry, }) return updateErr }) if err != nil { fmt.Fprintf(writer, "uncache %+v: %v\n", dir.Child(entry.Name), err) return false } fmt.Fprintf(writer, "Done\n") return true }) } type FileFilter struct { include *string exclude *string minSize *int64 maxSize *int64 minAge *int64 maxAge *int64 } func newFileFilter(remoteMountCommand *flag.FlagSet) (ff *FileFilter) { ff = &FileFilter{} ff.include = remoteMountCommand.String("include", "", "pattens of file names, e.g., *.pdf, *.html, ab?d.txt") ff.exclude = remoteMountCommand.String("exclude", "", "pattens of file names, e.g., *.pdf, *.html, ab?d.txt") ff.minSize = remoteMountCommand.Int64("minSize", -1, "minimum file size in bytes") ff.maxSize = remoteMountCommand.Int64("maxSize", -1, "maximum file size in bytes") ff.minAge = remoteMountCommand.Int64("minAge", -1, "minimum file age in seconds") ff.maxAge = remoteMountCommand.Int64("maxAge", -1, "maximum file age in seconds") return } func (ff *FileFilter) matches(entry *filer_pb.Entry) bool { if *ff.include != "" { if ok, _ := filepath.Match(*ff.include, entry.Name); !ok { return false } } if *ff.exclude != "" { if ok, _ := filepath.Match(*ff.exclude, entry.Name); ok { return false } } if *ff.minSize != -1 { if int64(entry.Attributes.FileSize) < *ff.minSize { return false } } if *ff.maxSize != -1 { if int64(entry.Attributes.FileSize) > *ff.maxSize { return false } } if *ff.minAge != -1 { if entry.Attributes.Crtime + *ff.minAge > time.Now().Unix() { return false } } if *ff.maxAge != -1 { if entry.Attributes.Crtime + *ff.maxAge < time.Now().Unix() { return false } } return true }