123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108 |
- package command
- import (
- "encoding/json"
- "fmt"
- "os"
- "path/filepath"
- "github.com/chrislusf/seaweedfs/weed/operation"
- "github.com/chrislusf/seaweedfs/weed/security"
- "github.com/chrislusf/seaweedfs/weed/util"
- )
- var (
- upload UploadOptions
- )
- type UploadOptions struct {
- master *string
- dir *string
- include *string
- replication *string
- collection *string
- dataCenter *string
- ttl *string
- maxMB *int
- }
- func init() {
- cmdUpload.Run = runUpload // break init cycle
- cmdUpload.IsDebug = cmdUpload.Flag.Bool("debug", false, "verbose debug information")
- upload.master = cmdUpload.Flag.String("master", "localhost:9333", "SeaweedFS master location")
- upload.dir = cmdUpload.Flag.String("dir", "", "Upload the whole folder recursively if specified.")
- upload.include = cmdUpload.Flag.String("include", "", "pattens of files to upload, e.g., *.pdf, *.html, ab?d.txt, works together with -dir")
- upload.replication = cmdUpload.Flag.String("replication", "", "replication type")
- upload.collection = cmdUpload.Flag.String("collection", "", "optional collection name")
- upload.dataCenter = cmdUpload.Flag.String("dataCenter", "", "optional data center name")
- upload.ttl = cmdUpload.Flag.String("ttl", "", "time to live, e.g.: 1m, 1h, 1d, 1M, 1y")
- upload.maxMB = cmdUpload.Flag.Int("maxMB", 32, "split files larger than the limit")
- }
- var cmdUpload = &Command{
- UsageLine: "upload -master=localhost:9333 file1 [file2 file3]\n weed upload -master=localhost:9333 -dir=one_directory -include=*.pdf",
- Short: "upload one or a list of files",
- Long: `upload one or a list of files, or batch upload one whole folder recursively.
- If uploading a list of files:
- It uses consecutive file keys for the list of files.
- e.g. If the file1 uses key k, file2 can be read via k_1
- If uploading a whole folder recursively:
- All files under the folder and subfolders will be uploaded, each with its own file key.
- Optional parameter "-include" allows you to specify the file name patterns.
- If "maxMB" is set to a positive number, files larger than it would be split into chunks and uploaded separately.
- The list of file ids of those chunks would be stored in an additional chunk, and this additional chunk's file id would be returned.
- `,
- }
- func runUpload(cmd *Command, args []string) bool {
- util.LoadConfiguration("security", false)
- grpcDialOption := security.LoadClientTLS(util.GetViper(), "grpc.client")
- if len(args) == 0 {
- if *upload.dir == "" {
- return false
- }
- filepath.Walk(*upload.dir, func(path string, info os.FileInfo, err error) error {
- if err == nil {
- if !info.IsDir() {
- if *upload.include != "" {
- if ok, _ := filepath.Match(*upload.include, filepath.Base(path)); !ok {
- return nil
- }
- }
- parts, e := operation.NewFileParts([]string{path})
- if e != nil {
- return e
- }
- results, e := operation.SubmitFiles(*upload.master, grpcDialOption, parts,
- *upload.replication, *upload.collection, *upload.dataCenter,
- *upload.ttl, *upload.maxMB)
- bytes, _ := json.Marshal(results)
- fmt.Println(string(bytes))
- if e != nil {
- return e
- }
- }
- } else {
- fmt.Println(err)
- }
- return err
- })
- } else {
- parts, e := operation.NewFileParts(args)
- if e != nil {
- fmt.Println(e.Error())
- }
- results, _ := operation.SubmitFiles(*upload.master, grpcDialOption, parts,
- *upload.replication, *upload.collection, *upload.dataCenter,
- *upload.ttl, *upload.maxMB)
- bytes, _ := json.Marshal(results)
- fmt.Println(string(bytes))
- }
- return true
- }
|