Cleanup s3 backup uploads, add request timeouts
This commit is contained in:
parent
1239b1c0ca
commit
5475cb02c1
|
@ -7,10 +7,10 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
type BackupRemoteUploadResponse struct {
|
type BackupRemoteUploadResponse struct {
|
||||||
CompleteMultipartUpload string
|
CompleteMultipartUpload string `json:"complete_multipart_upload"`
|
||||||
AbortMultipartUpload string
|
AbortMultipartUpload string `json:"abort_multipart_upload"`
|
||||||
Parts []string
|
Parts []string `json:"parts"`
|
||||||
PartSize int64
|
PartSize int64 `json:"part_size"`
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *Request) GetBackupRemoteUploadURLs(backup string, size int64) (*BackupRemoteUploadResponse, error) {
|
func (r *Request) GetBackupRemoteUploadURLs(backup string, size int64) (*BackupRemoteUploadResponse, error) {
|
||||||
|
|
|
@ -11,6 +11,7 @@ import (
|
||||||
"net/http"
|
"net/http"
|
||||||
"os"
|
"os"
|
||||||
"strconv"
|
"strconv"
|
||||||
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
type S3Backup struct {
|
type S3Backup struct {
|
||||||
|
@ -87,85 +88,99 @@ func (s *S3Backup) generateRemoteRequest(rc io.ReadCloser) error {
|
||||||
r.Header.Add("Content-Length", strconv.Itoa(int(size)))
|
r.Header.Add("Content-Length", strconv.Itoa(int(size)))
|
||||||
r.Header.Add("Content-Type", "application/x-gzip")
|
r.Header.Add("Content-Type", "application/x-gzip")
|
||||||
|
|
||||||
|
// Limit the reader to the size of the part.
|
||||||
r.Body = Reader{io.LimitReader(rc, size)}
|
r.Body = Reader{io.LimitReader(rc, size)}
|
||||||
|
|
||||||
|
// This http request can block forever due to it not having a timeout,
|
||||||
|
// but we are uploading up to 5GB of data, so there is not really
|
||||||
|
// a good way to handle a timeout on this.
|
||||||
res, err := http.DefaultClient.Do(r)
|
res, err := http.DefaultClient.Do(r)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", err
|
return "", err
|
||||||
}
|
}
|
||||||
|
|
||||||
defer res.Body.Close()
|
defer res.Body.Close()
|
||||||
|
|
||||||
|
// Handle non-200 status codes.
|
||||||
if res.StatusCode != http.StatusOK {
|
if res.StatusCode != http.StatusOK {
|
||||||
return "", fmt.Errorf("failed to put S3 object part, %d:%s", res.StatusCode, res.Status)
|
return "", fmt.Errorf("failed to put S3 object part, %d:%s", res.StatusCode, res.Status)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Get the ETag from the uploaded part, this should be sent with the CompleteMultipartUpload request.
|
||||||
return res.Header.Get("ETag"), nil
|
return res.Header.Get("ETag"), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Keep track of errors from individual part uploads.
|
// Start assembling the body that will be sent as apart of the CompleteMultipartUpload request.
|
||||||
hasError := true
|
var completeUploadBody bytes.Buffer
|
||||||
defer func() {
|
completeUploadBody.WriteString("<CompleteMultipartUpload>\n")
|
||||||
if !hasError {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
r, err := http.NewRequest(http.MethodPost, urls.AbortMultipartUpload, nil)
|
|
||||||
if err != nil {
|
|
||||||
log.WithError(err).Warn("failed to create http request (AbortMultipartUpload)")
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
res, err := http.DefaultClient.Do(r)
|
|
||||||
if err != nil {
|
|
||||||
log.WithError(err).Warn("failed to make http request (AbortMultipartUpload)")
|
|
||||||
return
|
|
||||||
}
|
|
||||||
defer res.Body.Close()
|
|
||||||
|
|
||||||
if res.StatusCode != http.StatusOK {
|
|
||||||
log.Warnf("failed to abort S3 multipart upload, %d:%s", res.StatusCode, res.Status)
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
var completeBody bytes.Buffer
|
|
||||||
completeBody.WriteString("<CompleteMultipartUpload>\n")
|
|
||||||
|
|
||||||
partCount := len(urls.Parts)
|
partCount := len(urls.Parts)
|
||||||
for i, part := range urls.Parts {
|
for i, part := range urls.Parts {
|
||||||
var s int64
|
// Get the size for the current part.
|
||||||
|
var partSize int64
|
||||||
if i+1 < partCount {
|
if i+1 < partCount {
|
||||||
s = urls.PartSize
|
partSize = urls.PartSize
|
||||||
} else {
|
} else {
|
||||||
s = size - (int64(i) * urls.PartSize)
|
// This is the remaining size for the last part,
|
||||||
|
// there is not a minimum size limit for the last part.
|
||||||
|
partSize = size - (int64(i) * urls.PartSize)
|
||||||
}
|
}
|
||||||
|
|
||||||
etag, err := handlePart(part, s)
|
// Attempt to upload the part.
|
||||||
|
etag, err := handlePart(part, partSize)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
log.WithError(err).Warn("failed to upload part")
|
||||||
|
|
||||||
|
// Send an AbortMultipartUpload request.
|
||||||
|
if err := s.finishUpload(urls.AbortMultipartUpload, nil); err != nil {
|
||||||
|
log.WithError(err).Warn("failed to abort multipart backup upload")
|
||||||
|
}
|
||||||
|
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
completeBody.WriteString("\t<Part>\n")
|
// Add the part to the CompleteMultipartUpload body.
|
||||||
completeBody.WriteString("\t\t<ETag>\"" + etag + "\"</ETag>\n")
|
completeUploadBody.WriteString("\t<Part>\n")
|
||||||
completeBody.WriteString("\t\t<PartNumber>" + strconv.Itoa(i+1) + "</PartNumber>\n")
|
completeUploadBody.WriteString("\t\t<ETag>\"" + etag + "\"</ETag>\n")
|
||||||
completeBody.WriteString("\t</Part>\n")
|
completeUploadBody.WriteString("\t\t<PartNumber>" + strconv.Itoa(i+1) + "</PartNumber>\n")
|
||||||
|
completeUploadBody.WriteString("\t</Part>\n")
|
||||||
}
|
}
|
||||||
hasError = false
|
completeUploadBody.WriteString("</CompleteMultipartUpload>")
|
||||||
|
|
||||||
completeBody.WriteString("</CompleteMultipartUpload>")
|
// Send a CompleteMultipartUpload request.
|
||||||
|
if err := s.finishUpload(urls.CompleteMultipartUpload, &completeUploadBody); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
r, err := http.NewRequest(http.MethodPost, urls.CompleteMultipartUpload, &completeBody)
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// finishUpload sends a requests to the specified url to either complete or abort the upload.
|
||||||
|
func (s *S3Backup) finishUpload(url string, body io.Reader) error {
|
||||||
|
r, err := http.NewRequest(http.MethodPost, url, body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
res, err := http.DefaultClient.Do(r)
|
// Create a new http client with a 10 second timeout.
|
||||||
|
c := &http.Client{
|
||||||
|
Timeout: 10 * time.Second,
|
||||||
|
}
|
||||||
|
|
||||||
|
res, err := c.Do(r)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
defer res.Body.Close()
|
defer res.Body.Close()
|
||||||
|
|
||||||
|
// Handle non-200 status codes.
|
||||||
if res.StatusCode != http.StatusOK {
|
if res.StatusCode != http.StatusOK {
|
||||||
|
// If no body was sent, we were aborting the upload.
|
||||||
|
if body == nil {
|
||||||
|
return fmt.Errorf("failed to abort S3 multipart upload, %d:%s", res.StatusCode, res.Status)
|
||||||
|
}
|
||||||
|
|
||||||
|
// If a body was sent we were completing the upload.
|
||||||
|
// TODO: Attempt to send abort request?
|
||||||
return fmt.Errorf("failed to complete S3 multipart upload, %d:%s", res.StatusCode, res.Status)
|
return fmt.Errorf("failed to complete S3 multipart upload, %d:%s", res.StatusCode, res.Status)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue
Block a user