mirror of
https://github.com/seaweedfs/seaweedfs.git
synced 2025-04-05 20:52:50 +08:00
[filer] UploadRequestToChunks: terminate the upload of remaining chunks as early as possible might be more reasonable when one chunk upload goroutine encounters a failure. (#6610)
This commit is contained in:
parent
271505717e
commit
14cb8a24c6
@ -64,6 +64,16 @@ func (fs *FilerServer) uploadReaderToChunks(reader io.Reader, startOffset int64,
|
|||||||
// need to throttle used byte buffer
|
// need to throttle used byte buffer
|
||||||
bytesBufferLimitChan <- struct{}{}
|
bytesBufferLimitChan <- struct{}{}
|
||||||
|
|
||||||
|
// As long as there is an error in the upload of one chunk, it can be terminated early
|
||||||
|
// uploadErr may be modified in other go routines, lock is needed to avoid race condition
|
||||||
|
uploadErrLock.Lock()
|
||||||
|
if uploadErr != nil {
|
||||||
|
<-bytesBufferLimitChan
|
||||||
|
uploadErrLock.Unlock()
|
||||||
|
break
|
||||||
|
}
|
||||||
|
uploadErrLock.Unlock()
|
||||||
|
|
||||||
bytesBuffer := bufPool.Get().(*bytes.Buffer)
|
bytesBuffer := bufPool.Get().(*bytes.Buffer)
|
||||||
|
|
||||||
limitedReader := io.LimitReader(partReader, int64(chunkSize))
|
limitedReader := io.LimitReader(partReader, int64(chunkSize))
|
||||||
|
Loading…
Reference in New Issue
Block a user