git-lfs/tq/basic_upload.go
brian m. carlson 00623425a2
tq: make Manifest an interface
Right now, any time we instantiate a Manifest object, we create an API
client, and when we create the API client, if we're using SSH, we try to
make a connection to the server.  However, we often instantiate a
Manifest object when performing various functionality such as smudging
data, which means that when a user creates an archive locally, they can
be prompted for an SSH password, which is undesirable.

Let's take a first step to fixing this by making Manifest an interface.
Right now, it has one concrete version, a concreteManifest, which can be
used to access the internals, and we provide methods to upgrade it from
the interface to the concrete type and determine whether it's upgraded
or not.  We attempt to upgrade it any time we need to access its
internals.  In the future, we'll also offer a lazyManifest, which is
lazy and will only instantiate the concreteManifest inside when we
attempt to upgrade it to the latter.  But for now, only implement the
concreteManifest to make it clearer what's changing.

Similarly, we make our TransferQueue upgradable so that we don't
upgrade its Manifest right away.

In both cases, we'll want to use the lazyManifest to delay the
instantiation of the API client (and hence the starting of the SSH
connection) in a future commit.
2023-03-23 16:55:57 +00:00

244 lines
6.5 KiB
Go

package tq
import (
"io"
"io/ioutil"
"net/http"
"os"
"path/filepath"
"strconv"
"strings"
"github.com/git-lfs/git-lfs/v3/config"
"github.com/git-lfs/git-lfs/v3/errors"
"github.com/git-lfs/git-lfs/v3/lfsapi"
"github.com/git-lfs/git-lfs/v3/tools"
"github.com/git-lfs/git-lfs/v3/tr"
)
const (
BasicAdapterName = "basic"
defaultContentType = "application/octet-stream"
)
// Adapter for basic uploads (non resumable)
type basicUploadAdapter struct {
*adapterBase
}
func (a *basicUploadAdapter) tempDir() string {
// Dedicated to this adapter rather than shared with basic download.
d := filepath.Join(os.TempDir(), "git-lfs-basic-temp")
if err := tools.MkdirAll(d, a.fs); err != nil {
return os.TempDir()
}
return d
}
func (a *basicUploadAdapter) WorkerStarting(workerNum int) (interface{}, error) {
return nil, nil
}
func (a *basicUploadAdapter) WorkerEnding(workerNum int, ctx interface{}) {
}
func (a *basicUploadAdapter) DoTransfer(ctx interface{}, t *Transfer, cb ProgressCallback, authOkFunc func()) error {
rel, err := t.Rel("upload")
if err != nil {
return err
}
if rel == nil {
return errors.Errorf(tr.Tr.Get("No upload action for object: %s", t.Oid))
}
req, err := a.newHTTPRequest("PUT", rel)
if err != nil {
return err
}
if req.Header.Get("Transfer-Encoding") == "chunked" {
req.TransferEncoding = []string{"chunked"}
} else {
req.Header.Set("Content-Length", strconv.FormatInt(t.Size, 10))
}
req.ContentLength = t.Size
f, err := os.OpenFile(t.Path, os.O_RDONLY, 0644)
if err != nil {
return errors.Wrap(err, tr.Tr.Get("basic upload"))
}
defer f.Close()
if err := a.setContentTypeFor(req, f); err != nil {
return err
}
// Ensure progress callbacks made while uploading
// Wrap callback to give name context
ccb := func(totalSize int64, readSoFar int64, readSinceLast int) error {
if cb != nil {
return cb(t.Name, totalSize, readSoFar, readSinceLast)
}
return nil
}
cbr := tools.NewFileBodyWithCallback(f, t.Size, ccb)
var reader lfsapi.ReadSeekCloser = cbr
// Signal auth was ok on first read; this frees up other workers to start
if authOkFunc != nil {
reader = newStartCallbackReader(reader, func() error {
authOkFunc()
return nil
})
}
req.Body = reader
req = a.apiClient.LogRequest(req, "lfs.data.upload")
res, err := a.makeRequest(t, req)
if err != nil {
if errors.IsUnprocessableEntityError(err) {
// If we got an HTTP 422, we do _not_ want to retry the
// request later below, because it is likely that the
// implementing server does not support non-standard
// Content-Type headers.
//
// Instead, return immediately and wait for the
// *tq.TransferQueue to report an error message.
return err
}
// We're about to return a retriable error, meaning that this
// transfer will either be retried, or it will fail.
//
// Either way, let's decrement the number of bytes that we've
// read _so far_, so that the next iteration doesn't re-transfer
// those bytes, according to the progress meter.
if perr := cbr.ResetProgress(); perr != nil {
err = errors.Wrap(err, perr.Error())
}
if res == nil {
// We encountered a network or similar error which caused us
// to not receive a response at all.
return errors.NewRetriableError(err)
}
if res.StatusCode == 429 {
retLaterErr := errors.NewRetriableLaterError(err, res.Header["Retry-After"][0])
if retLaterErr != nil {
return retLaterErr
}
}
return errors.NewRetriableError(err)
}
// A status code of 403 likely means that an authentication token for the
// upload has expired. This can be safely retried.
if res.StatusCode == 403 {
err = errors.New(tr.Tr.Get("Received status %d", res.StatusCode))
return errors.NewRetriableError(err)
}
if res.StatusCode > 299 {
return errors.Wrapf(nil, tr.Tr.Get("Invalid status for %s %s: %d",
req.Method,
strings.SplitN(req.URL.String(), "?", 2)[0],
res.StatusCode,
))
}
io.Copy(ioutil.Discard, res.Body)
res.Body.Close()
return verifyUpload(a.apiClient, a.remote, t)
}
func (a *adapterBase) setContentTypeFor(req *http.Request, r io.ReadSeeker) error {
uc := config.NewURLConfig(a.apiClient.GitEnv())
disabled := !uc.Bool("lfs", req.URL.String(), "contenttype", true)
if len(req.Header.Get("Content-Type")) != 0 {
return nil
}
var contentType string
if !disabled {
buffer := make([]byte, 512)
n, err := r.Read(buffer)
if err != nil && err != io.EOF {
return errors.Wrap(err, tr.Tr.Get("content type detection error"))
}
contentType = http.DetectContentType(buffer[:n])
if _, err := r.Seek(0, io.SeekStart); err != nil {
return errors.Wrap(err, tr.Tr.Get("content type rewind failure"))
}
}
if contentType == "" {
contentType = defaultContentType
}
req.Header.Set("Content-Type", contentType)
return nil
}
// startCallbackReader is a reader wrapper which calls a function as soon as the
// first Read() call is made. This callback is only made once
type startCallbackReader struct {
cb func() error
cbDone bool
lfsapi.ReadSeekCloser
}
func (s *startCallbackReader) Read(p []byte) (n int, err error) {
if !s.cbDone && s.cb != nil {
if err := s.cb(); err != nil {
return 0, err
}
s.cbDone = true
}
return s.ReadSeekCloser.Read(p)
}
func newStartCallbackReader(r lfsapi.ReadSeekCloser, cb func() error) *startCallbackReader {
return &startCallbackReader{
ReadSeekCloser: r,
cb: cb,
}
}
func configureBasicUploadAdapter(m *concreteManifest) {
m.RegisterNewAdapterFunc(BasicAdapterName, Upload, func(name string, dir Direction) Adapter {
switch dir {
case Upload:
bu := &basicUploadAdapter{newAdapterBase(m.fs, name, dir, nil)}
// self implements impl
bu.transferImpl = bu
return bu
case Download:
panic(tr.Tr.Get("Should never ask this function to download"))
}
return nil
})
}
func (a *basicUploadAdapter) makeRequest(t *Transfer, req *http.Request) (*http.Response, error) {
res, err := a.doHTTP(t, req)
if errors.IsAuthError(err) && len(req.Header.Get("Authorization")) == 0 {
// Construct a new body with just the raw file and no callbacks. Since
// all progress tracking happens when the net.http code copies our
// request body into a new request, we can safely make this request
// outside of the flow of the transfer adapter, and if it fails, the
// transfer progress will be rewound at the top level
f, _ := os.OpenFile(t.Path, os.O_RDONLY, 0644)
defer f.Close()
req.Body = tools.NewBodyWithCallback(f, t.Size, nil)
return a.makeRequest(t, req)
}
return res, err
}