123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408 |
- package weed_server
- import (
- "context"
- "crypto/md5"
- "fmt"
- "hash"
- "io"
- "io/ioutil"
- "net/http"
- "os"
- "path"
- "strconv"
- "strings"
- "time"
- "github.com/chrislusf/seaweedfs/weed/filer"
- "github.com/chrislusf/seaweedfs/weed/glog"
- "github.com/chrislusf/seaweedfs/weed/operation"
- "github.com/chrislusf/seaweedfs/weed/pb/filer_pb"
- xhttp "github.com/chrislusf/seaweedfs/weed/s3api/http"
- "github.com/chrislusf/seaweedfs/weed/security"
- "github.com/chrislusf/seaweedfs/weed/stats"
- "github.com/chrislusf/seaweedfs/weed/storage/needle"
- "github.com/chrislusf/seaweedfs/weed/util"
- )
- func (fs *FilerServer) autoChunk(ctx context.Context, w http.ResponseWriter, r *http.Request, so *operation.StorageOption) {
- // autoChunking can be set at the command-line level or as a query param. Query param overrides command-line
- query := r.URL.Query()
- parsedMaxMB, _ := strconv.ParseInt(query.Get("maxMB"), 10, 32)
- maxMB := int32(parsedMaxMB)
- if maxMB <= 0 && fs.option.MaxMB > 0 {
- maxMB = int32(fs.option.MaxMB)
- }
- chunkSize := 1024 * 1024 * maxMB
- stats.FilerRequestCounter.WithLabelValues("chunk").Inc()
- start := time.Now()
- defer func() {
- stats.FilerRequestHistogram.WithLabelValues("chunk").Observe(time.Since(start).Seconds())
- }()
- var reply *FilerPostResult
- var err error
- var md5bytes []byte
- if r.Method == "POST" {
- if r.Header.Get("Content-Type") == "" && strings.HasSuffix(r.URL.Path, "/") {
- reply, err = fs.mkdir(ctx, w, r)
- } else {
- reply, md5bytes, err = fs.doPostAutoChunk(ctx, w, r, chunkSize, so)
- }
- } else {
- reply, md5bytes, err = fs.doPutAutoChunk(ctx, w, r, chunkSize, so)
- }
- if err != nil {
- if strings.HasPrefix(err.Error(), "read input:") {
- writeJsonError(w, r, 499, err)
- } else {
- writeJsonError(w, r, http.StatusInternalServerError, err)
- }
- } else if reply != nil {
- if len(md5bytes) > 0 {
- w.Header().Set("Content-MD5", util.Base64Encode(md5bytes))
- }
- writeJsonQuiet(w, r, http.StatusCreated, reply)
- }
- }
- func (fs *FilerServer) doPostAutoChunk(ctx context.Context, w http.ResponseWriter, r *http.Request, chunkSize int32, so *operation.StorageOption) (filerResult *FilerPostResult, md5bytes []byte, replyerr error) {
- multipartReader, multipartReaderErr := r.MultipartReader()
- if multipartReaderErr != nil {
- return nil, nil, multipartReaderErr
- }
- part1, part1Err := multipartReader.NextPart()
- if part1Err != nil {
- return nil, nil, part1Err
- }
- fileName := part1.FileName()
- if fileName != "" {
- fileName = path.Base(fileName)
- }
- contentType := part1.Header.Get("Content-Type")
- if contentType == "application/octet-stream" {
- contentType = ""
- }
- fileChunks, md5Hash, chunkOffset, err, smallContent := fs.uploadReaderToChunks(w, r, part1, chunkSize, fileName, contentType, so)
- if err != nil {
- return nil, nil, err
- }
- md5bytes = md5Hash.Sum(nil)
- filerResult, replyerr = fs.saveMetaData(ctx, r, fileName, contentType, so, md5bytes, fileChunks, chunkOffset, smallContent)
- return
- }
- func (fs *FilerServer) doPutAutoChunk(ctx context.Context, w http.ResponseWriter, r *http.Request, chunkSize int32, so *operation.StorageOption) (filerResult *FilerPostResult, md5bytes []byte, replyerr error) {
- fileName := path.Base(r.URL.Path)
- contentType := r.Header.Get("Content-Type")
- if contentType == "application/octet-stream" {
- contentType = ""
- }
- fileChunks, md5Hash, chunkOffset, err, smallContent := fs.uploadReaderToChunks(w, r, r.Body, chunkSize, fileName, contentType, so)
- if err != nil {
- return nil, nil, err
- }
- md5bytes = md5Hash.Sum(nil)
- filerResult, replyerr = fs.saveMetaData(ctx, r, fileName, contentType, so, md5bytes, fileChunks, chunkOffset, smallContent)
- return
- }
- func isAppend(r *http.Request) bool {
- return r.URL.Query().Get("op") == "append"
- }
- func (fs *FilerServer) saveMetaData(ctx context.Context, r *http.Request, fileName string, contentType string, so *operation.StorageOption, md5bytes []byte, fileChunks []*filer_pb.FileChunk, chunkOffset int64, content []byte) (filerResult *FilerPostResult, replyerr error) {
- // detect file mode
- modeStr := r.URL.Query().Get("mode")
- if modeStr == "" {
- modeStr = "0660"
- }
- mode, err := strconv.ParseUint(modeStr, 8, 32)
- if err != nil {
- glog.Errorf("Invalid mode format: %s, use 0660 by default", modeStr)
- mode = 0660
- }
- // fix the path
- path := r.URL.Path
- if strings.HasSuffix(path, "/") {
- if fileName != "" {
- path += fileName
- }
- }
- var entry *filer.Entry
- var mergedChunks []*filer_pb.FileChunk
- // when it is an append
- if isAppend(r) {
- existingEntry, findErr := fs.filer.FindEntry(ctx, util.FullPath(path))
- if findErr != nil && findErr != filer_pb.ErrNotFound {
- glog.V(0).Infof("failing to find %s: %v", path, findErr)
- }
- entry = existingEntry
- }
- if entry != nil {
- entry.Mtime = time.Now()
- entry.Md5 = nil
- // adjust chunk offsets
- for _, chunk := range fileChunks {
- chunk.Offset += int64(entry.FileSize)
- }
- mergedChunks = append(entry.Chunks, fileChunks...)
- entry.FileSize += uint64(chunkOffset)
- // TODO
- if len(entry.Content) > 0 {
- replyerr = fmt.Errorf("append to small file is not supported yet")
- return
- }
- } else {
- glog.V(4).Infoln("saving", path)
- mergedChunks = fileChunks
- entry = &filer.Entry{
- FullPath: util.FullPath(path),
- Attr: filer.Attr{
- Mtime: time.Now(),
- Crtime: time.Now(),
- Mode: os.FileMode(mode),
- Uid: OS_UID,
- Gid: OS_GID,
- Replication: so.Replication,
- Collection: so.Collection,
- TtlSec: so.TtlSeconds,
- DiskType: so.DiskType,
- Mime: contentType,
- Md5: md5bytes,
- FileSize: uint64(chunkOffset),
- },
- Content: content,
- }
- }
- // maybe compact entry chunks
- mergedChunks, replyerr = filer.MaybeManifestize(fs.saveAsChunk(so), mergedChunks)
- if replyerr != nil {
- glog.V(0).Infof("manifestize %s: %v", r.RequestURI, replyerr)
- return
- }
- entry.Chunks = mergedChunks
- filerResult = &FilerPostResult{
- Name: fileName,
- Size: int64(entry.FileSize),
- }
- if entry.Extended == nil {
- entry.Extended = make(map[string][]byte)
- }
- fs.saveAmzMetaData(r, entry)
- for k, v := range r.Header {
- if len(v) > 0 && strings.HasPrefix(k, needle.PairNamePrefix) {
- entry.Extended[k] = []byte(v[0])
- }
- }
- if dbErr := fs.filer.CreateEntry(ctx, entry, false, false, nil); dbErr != nil {
- fs.filer.DeleteChunks(fileChunks)
- replyerr = dbErr
- filerResult.Error = dbErr.Error()
- glog.V(0).Infof("failing to write %s to filer server : %v", path, dbErr)
- }
- return filerResult, replyerr
- }
- func (fs *FilerServer) uploadReaderToChunks(w http.ResponseWriter, r *http.Request, reader io.Reader, chunkSize int32, fileName, contentType string, so *operation.StorageOption) ([]*filer_pb.FileChunk, hash.Hash, int64, error, []byte) {
- var fileChunks []*filer_pb.FileChunk
- md5Hash := md5.New()
- var partReader = ioutil.NopCloser(io.TeeReader(reader, md5Hash))
- chunkOffset := int64(0)
- var smallContent []byte
- for {
- limitedReader := io.LimitReader(partReader, int64(chunkSize))
- data, err := ioutil.ReadAll(limitedReader)
- if err != nil {
- return nil, nil, 0, err, nil
- }
- if chunkOffset == 0 && !isAppend(r) {
- if len(data) < fs.option.SaveToFilerLimit || strings.HasPrefix(r.URL.Path, filer.DirectoryEtcRoot) && len(data) < 4*1024 {
- smallContent = data
- chunkOffset += int64(len(data))
- break
- }
- }
- dataReader := util.NewBytesReader(data)
- // retry to assign a different file id
- var fileId, urlLocation string
- var auth security.EncodedJwt
- var assignErr, uploadErr error
- var uploadResult *operation.UploadResult
- for i := 0; i < 3; i++ {
- // assign one file id for one chunk
- fileId, urlLocation, auth, assignErr = fs.assignNewFileInfo(so)
- if assignErr != nil {
- return nil, nil, 0, assignErr, nil
- }
- // upload the chunk to the volume server
- uploadResult, uploadErr, _ = fs.doUpload(urlLocation, w, r, dataReader, fileName, contentType, nil, auth)
- if uploadErr != nil {
- time.Sleep(251 * time.Millisecond)
- continue
- }
- break
- }
- if uploadErr != nil {
- return nil, nil, 0, uploadErr, nil
- }
- // if last chunk exhausted the reader exactly at the border
- if uploadResult.Size == 0 {
- break
- }
- // Save to chunk manifest structure
- fileChunks = append(fileChunks, uploadResult.ToPbFileChunk(fileId, chunkOffset))
- glog.V(4).Infof("uploaded %s chunk %d to %s [%d,%d)", fileName, len(fileChunks), fileId, chunkOffset, chunkOffset+int64(uploadResult.Size))
- // reset variables for the next chunk
- chunkOffset = chunkOffset + int64(uploadResult.Size)
- // if last chunk was not at full chunk size, but already exhausted the reader
- if int64(uploadResult.Size) < int64(chunkSize) {
- break
- }
- }
- return fileChunks, md5Hash, chunkOffset, nil, smallContent
- }
- func (fs *FilerServer) doUpload(urlLocation string, w http.ResponseWriter, r *http.Request, limitedReader io.Reader, fileName string, contentType string, pairMap map[string]string, auth security.EncodedJwt) (*operation.UploadResult, error, []byte) {
- stats.FilerRequestCounter.WithLabelValues("chunkUpload").Inc()
- start := time.Now()
- defer func() {
- stats.FilerRequestHistogram.WithLabelValues("chunkUpload").Observe(time.Since(start).Seconds())
- }()
- uploadResult, err, data := operation.Upload(urlLocation, fileName, fs.option.Cipher, limitedReader, false, contentType, pairMap, auth)
- if uploadResult != nil && uploadResult.RetryCount > 0 {
- stats.FilerRequestCounter.WithLabelValues("chunkUploadRetry").Add(float64(uploadResult.RetryCount))
- }
- return uploadResult, err, data
- }
- func (fs *FilerServer) saveAsChunk(so *operation.StorageOption) filer.SaveDataAsChunkFunctionType {
- return func(reader io.Reader, name string, offset int64) (*filer_pb.FileChunk, string, string, error) {
- // assign one file id for one chunk
- fileId, urlLocation, auth, assignErr := fs.assignNewFileInfo(so)
- if assignErr != nil {
- return nil, "", "", assignErr
- }
- // upload the chunk to the volume server
- uploadResult, uploadErr, _ := operation.Upload(urlLocation, name, fs.option.Cipher, reader, false, "", nil, auth)
- if uploadErr != nil {
- return nil, "", "", uploadErr
- }
- return uploadResult.ToPbFileChunk(fileId, offset), so.Collection, so.Replication, nil
- }
- }
- func (fs *FilerServer) mkdir(ctx context.Context, w http.ResponseWriter, r *http.Request) (filerResult *FilerPostResult, replyerr error) {
- // detect file mode
- modeStr := r.URL.Query().Get("mode")
- if modeStr == "" {
- modeStr = "0660"
- }
- mode, err := strconv.ParseUint(modeStr, 8, 32)
- if err != nil {
- glog.Errorf("Invalid mode format: %s, use 0660 by default", modeStr)
- mode = 0660
- }
- // fix the path
- path := r.URL.Path
- if strings.HasSuffix(path, "/") {
- path = path[:len(path)-1]
- }
- existingEntry, err := fs.filer.FindEntry(ctx, util.FullPath(path))
- if err == nil && existingEntry != nil {
- replyerr = fmt.Errorf("dir %s already exists", path)
- return
- }
- glog.V(4).Infoln("mkdir", path)
- entry := &filer.Entry{
- FullPath: util.FullPath(path),
- Attr: filer.Attr{
- Mtime: time.Now(),
- Crtime: time.Now(),
- Mode: os.FileMode(mode) | os.ModeDir,
- Uid: OS_UID,
- Gid: OS_GID,
- },
- }
- filerResult = &FilerPostResult{
- Name: util.FullPath(path).Name(),
- }
- if dbErr := fs.filer.CreateEntry(ctx, entry, false, false, nil); dbErr != nil {
- replyerr = dbErr
- filerResult.Error = dbErr.Error()
- glog.V(0).Infof("failing to create dir %s on filer server : %v", path, dbErr)
- }
- return filerResult, replyerr
- }
- func (fs *FilerServer) saveAmzMetaData(r *http.Request, entry *filer.Entry) {
- if sc := r.Header.Get(xhttp.AmzStorageClass); sc != "" {
- entry.Extended[xhttp.AmzStorageClass] = []byte(sc)
- }
- if tags := r.Header.Get(xhttp.AmzObjectTagging); tags != "" {
- for _, v := range strings.Split(tags, "&") {
- tag := strings.Split(v, "=")
- if len(tag) == 2 {
- entry.Extended[xhttp.AmzObjectTagging+"-"+tag[0]] = []byte(tag[1])
- }
- }
- }
- for header, values := range r.Header {
- if strings.HasPrefix(header, xhttp.AmzUserMetaPrefix) {
- for _, value := range values {
- entry.Extended[header] = []byte(value)
- }
- }
- }
- }
|