123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182 |
- package shell
- import (
- "context"
- "flag"
- "fmt"
- "io"
- "path/filepath"
- "strings"
- "time"
- "github.com/seaweedfs/seaweedfs/weed/filer"
- "github.com/seaweedfs/seaweedfs/weed/pb/filer_pb"
- "github.com/seaweedfs/seaweedfs/weed/util"
- )
- func init() {
- Commands = append(Commands, &commandRemoteUncache{})
- }
- type commandRemoteUncache struct {
- }
- func (c *commandRemoteUncache) Name() string {
- return "remote.uncache"
- }
- func (c *commandRemoteUncache) Help() string {
- return `keep the metadata but remote cache the file content for mounted directories or files
- This is designed to run regularly. So you can add it to some cronjob.
- If a file is not synchronized with the remote copy, the file will be skipped to avoid loss of data.
- remote.uncache -dir=/xxx
- remote.uncache -dir=/xxx/some/sub/dir
- remote.uncache -dir=/xxx/some/sub/dir -include=*.pdf
- remote.uncache -dir=/xxx/some/sub/dir -exclude=*.txt
- remote.uncache -minSize=1024000 # uncache files larger than 100K
- remote.uncache -minAge=3600 # uncache files older than 1 hour
- `
- }
- func (c *commandRemoteUncache) HasTag(CommandTag) bool {
- return false
- }
- func (c *commandRemoteUncache) Do(args []string, commandEnv *CommandEnv, writer io.Writer) (err error) {
- remoteUncacheCommand := flag.NewFlagSet(c.Name(), flag.ContinueOnError)
- dir := remoteUncacheCommand.String("dir", "", "a directory in filer")
- fileFiler := newFileFilter(remoteUncacheCommand)
- if err = remoteUncacheCommand.Parse(args); err != nil {
- return nil
- }
- mappings, listErr := filer.ReadMountMappings(commandEnv.option.GrpcDialOption, commandEnv.option.FilerAddress)
- if listErr != nil {
- return listErr
- }
- if *dir != "" {
- var localMountedDir string
- for k := range mappings.Mappings {
- if strings.HasPrefix(*dir, k) {
- localMountedDir = k
- }
- }
- if localMountedDir == "" {
- jsonPrintln(writer, mappings)
- fmt.Fprintf(writer, "%s is not mounted\n", *dir)
- return nil
- }
- // pull content from remote
- if err = c.uncacheContentData(commandEnv, writer, util.FullPath(*dir), fileFiler); err != nil {
- return fmt.Errorf("uncache content data: %v", err)
- }
- return nil
- }
- for key, _ := range mappings.Mappings {
- if err := c.uncacheContentData(commandEnv, writer, util.FullPath(key), fileFiler); err != nil {
- return err
- }
- }
- return nil
- }
- func (c *commandRemoteUncache) uncacheContentData(commandEnv *CommandEnv, writer io.Writer, dirToCache util.FullPath, fileFilter *FileFilter) error {
- return recursivelyTraverseDirectory(commandEnv, dirToCache, func(dir util.FullPath, entry *filer_pb.Entry) bool {
- if !mayHaveCachedToLocal(entry) {
- return true // true means recursive traversal should continue
- }
- if !fileFilter.matches(entry) {
- return true
- }
- if entry.RemoteEntry.LastLocalSyncTsNs/1e9 < entry.Attributes.Mtime {
- return true // should not uncache an entry that is not synchronized with remote
- }
- entry.RemoteEntry.LastLocalSyncTsNs = 0
- entry.Chunks = nil
- fmt.Fprintf(writer, "Uncache %+v ... ", dir.Child(entry.Name))
- err := commandEnv.WithFilerClient(false, func(client filer_pb.SeaweedFilerClient) error {
- _, updateErr := client.UpdateEntry(context.Background(), &filer_pb.UpdateEntryRequest{
- Directory: string(dir),
- Entry: entry,
- })
- return updateErr
- })
- if err != nil {
- fmt.Fprintf(writer, "uncache %+v: %v\n", dir.Child(entry.Name), err)
- return false
- }
- fmt.Fprintf(writer, "Done\n")
- return true
- })
- }
- type FileFilter struct {
- include *string
- exclude *string
- minSize *int64
- maxSize *int64
- minAge *int64
- maxAge *int64
- }
- func newFileFilter(remoteMountCommand *flag.FlagSet) (ff *FileFilter) {
- ff = &FileFilter{}
- ff.include = remoteMountCommand.String("include", "", "pattens of file names, e.g., *.pdf, *.html, ab?d.txt")
- ff.exclude = remoteMountCommand.String("exclude", "", "pattens of file names, e.g., *.pdf, *.html, ab?d.txt")
- ff.minSize = remoteMountCommand.Int64("minSize", -1, "minimum file size in bytes")
- ff.maxSize = remoteMountCommand.Int64("maxSize", -1, "maximum file size in bytes")
- ff.minAge = remoteMountCommand.Int64("minAge", -1, "minimum file age in seconds")
- ff.maxAge = remoteMountCommand.Int64("maxAge", -1, "maximum file age in seconds")
- return
- }
- func (ff *FileFilter) matches(entry *filer_pb.Entry) bool {
- if *ff.include != "" {
- if ok, _ := filepath.Match(*ff.include, entry.Name); !ok {
- return false
- }
- }
- if *ff.exclude != "" {
- if ok, _ := filepath.Match(*ff.exclude, entry.Name); ok {
- return false
- }
- }
- if *ff.minSize != -1 {
- if int64(entry.Attributes.FileSize) < *ff.minSize {
- return false
- }
- }
- if *ff.maxSize != -1 {
- if int64(entry.Attributes.FileSize) > *ff.maxSize {
- return false
- }
- }
- if *ff.minAge != -1 {
- if entry.Attributes.Crtime+*ff.minAge > time.Now().Unix() {
- return false
- }
- }
- if *ff.maxAge != -1 {
- if entry.Attributes.Crtime+*ff.maxAge < time.Now().Unix() {
- return false
- }
- }
- return true
- }
|