2020-04-05 08:20:50 +02:00
|
|
|
// Copyright 2020 The Gitea Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a MIT-style
|
|
|
|
// license that can be found in the LICENSE file.
|
|
|
|
|
2016-12-26 02:16:37 +01:00
|
|
|
package lfs
|
|
|
|
|
|
|
|
import (
|
|
|
|
"crypto/sha256"
|
|
|
|
"encoding/hex"
|
|
|
|
"errors"
|
2020-10-31 21:51:48 +01:00
|
|
|
"fmt"
|
2021-03-06 21:21:56 +01:00
|
|
|
"hash"
|
2016-12-26 02:16:37 +01:00
|
|
|
"io"
|
|
|
|
"os"
|
2017-11-08 14:04:19 +01:00
|
|
|
|
|
|
|
"code.gitea.io/gitea/models"
|
2020-03-09 20:56:18 +01:00
|
|
|
"code.gitea.io/gitea/modules/log"
|
2020-09-08 17:45:10 +02:00
|
|
|
"code.gitea.io/gitea/modules/storage"
|
2016-12-26 02:16:37 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
var (
|
|
|
|
errHashMismatch = errors.New("Content hash does not match OID")
|
|
|
|
errSizeMismatch = errors.New("Content size does not match")
|
|
|
|
)
|
|
|
|
|
2020-10-31 21:51:48 +01:00
|
|
|
// ErrRangeNotSatisfiable represents an error which request range is not satisfiable.
|
|
|
|
type ErrRangeNotSatisfiable struct {
|
|
|
|
FromByte int64
|
|
|
|
}
|
|
|
|
|
|
|
|
func (err ErrRangeNotSatisfiable) Error() string {
|
|
|
|
return fmt.Sprintf("Requested range %d is not satisfiable", err.FromByte)
|
|
|
|
}
|
|
|
|
|
|
|
|
// IsErrRangeNotSatisfiable returns true if the error is an ErrRangeNotSatisfiable
|
|
|
|
func IsErrRangeNotSatisfiable(err error) bool {
|
|
|
|
_, ok := err.(ErrRangeNotSatisfiable)
|
|
|
|
return ok
|
|
|
|
}
|
|
|
|
|
2016-12-26 02:16:37 +01:00
|
|
|
// ContentStore provides a simple file system based storage.
|
|
|
|
type ContentStore struct {
|
2020-09-08 17:45:10 +02:00
|
|
|
storage.ObjectStorage
|
2016-12-26 02:16:37 +01:00
|
|
|
}
|
|
|
|
|
2017-03-15 01:52:01 +01:00
|
|
|
// Get takes a Meta object and retrieves the content from the store, returning
|
2016-12-26 02:16:37 +01:00
|
|
|
// it as an io.Reader. If fromByte > 0, the reader starts from that byte
|
|
|
|
func (s *ContentStore) Get(meta *models.LFSMetaObject, fromByte int64) (io.ReadCloser, error) {
|
2020-09-08 17:45:10 +02:00
|
|
|
f, err := s.Open(meta.RelativePath())
|
2016-12-26 02:16:37 +01:00
|
|
|
if err != nil {
|
2020-09-08 17:45:10 +02:00
|
|
|
log.Error("Whilst trying to read LFS OID[%s]: Unable to open Error: %v", meta.Oid, err)
|
2016-12-26 02:16:37 +01:00
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
if fromByte > 0 {
|
2020-10-31 21:51:48 +01:00
|
|
|
if fromByte >= meta.Size {
|
2021-04-03 14:44:12 +02:00
|
|
|
err = f.Close()
|
|
|
|
if err != nil {
|
|
|
|
log.Error("Whilst trying to read LFS OID[%s]: Unable to close Error: %v", meta.Oid, err)
|
|
|
|
}
|
2020-10-31 21:51:48 +01:00
|
|
|
return nil, ErrRangeNotSatisfiable{
|
|
|
|
FromByte: fromByte,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
_, err = f.Seek(fromByte, io.SeekStart)
|
2020-03-09 20:56:18 +01:00
|
|
|
if err != nil {
|
|
|
|
log.Error("Whilst trying to read LFS OID[%s]: Unable to seek to %d Error: %v", meta.Oid, fromByte, err)
|
2021-04-03 14:44:12 +02:00
|
|
|
errClose := f.Close()
|
|
|
|
if errClose != nil {
|
|
|
|
log.Error("Whilst trying to read LFS OID[%s]: Unable to close Error: %v", meta.Oid, errClose)
|
|
|
|
}
|
2020-03-09 20:56:18 +01:00
|
|
|
}
|
2016-12-26 02:16:37 +01:00
|
|
|
}
|
|
|
|
return f, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Put takes a Meta object and an io.Reader and writes the content to the store.
|
|
|
|
func (s *ContentStore) Put(meta *models.LFSMetaObject, r io.Reader) error {
|
2020-09-08 17:45:10 +02:00
|
|
|
p := meta.RelativePath()
|
2021-03-06 21:21:56 +01:00
|
|
|
|
|
|
|
// Wrap the provided reader with an inline hashing and size checker
|
|
|
|
wrappedRd := newHashingReader(meta.Size, meta.Oid, r)
|
|
|
|
|
|
|
|
// now pass the wrapped reader to Save - if there is a size mismatch or hash mismatch then
|
|
|
|
// the errors returned by the newHashingReader should percolate up to here
|
|
|
|
written, err := s.Save(p, wrappedRd)
|
2016-12-26 02:16:37 +01:00
|
|
|
if err != nil {
|
2020-09-08 17:45:10 +02:00
|
|
|
log.Error("Whilst putting LFS OID[%s]: Failed to copy to tmpPath: %s Error: %v", meta.Oid, p, err)
|
2016-12-26 02:16:37 +01:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2021-03-06 21:21:56 +01:00
|
|
|
// This shouldn't happen but it is sensible to test
|
2016-12-26 02:16:37 +01:00
|
|
|
if written != meta.Size {
|
2020-09-08 17:45:10 +02:00
|
|
|
if err := s.Delete(p); err != nil {
|
|
|
|
log.Error("Cleaning the LFS OID[%s] failed: %v", meta.Oid, err)
|
|
|
|
}
|
2016-12-26 02:16:37 +01:00
|
|
|
return errSizeMismatch
|
|
|
|
}
|
|
|
|
|
2020-03-09 20:56:18 +01:00
|
|
|
return nil
|
2016-12-26 02:16:37 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// Exists returns true if the object exists in the content store.
|
2020-09-08 17:45:10 +02:00
|
|
|
func (s *ContentStore) Exists(meta *models.LFSMetaObject) (bool, error) {
|
|
|
|
_, err := s.ObjectStorage.Stat(meta.RelativePath())
|
|
|
|
if err != nil {
|
|
|
|
if os.IsNotExist(err) {
|
|
|
|
return false, nil
|
|
|
|
}
|
|
|
|
return false, err
|
2016-12-26 02:16:37 +01:00
|
|
|
}
|
2020-09-08 17:45:10 +02:00
|
|
|
return true, nil
|
2016-12-26 02:16:37 +01:00
|
|
|
}
|
|
|
|
|
2017-11-08 14:04:19 +01:00
|
|
|
// Verify returns true if the object exists in the content store and size is correct.
|
|
|
|
func (s *ContentStore) Verify(meta *models.LFSMetaObject) (bool, error) {
|
2020-09-08 17:45:10 +02:00
|
|
|
p := meta.RelativePath()
|
|
|
|
fi, err := s.ObjectStorage.Stat(p)
|
|
|
|
if os.IsNotExist(err) || (err == nil && fi.Size() != meta.Size) {
|
2017-11-08 14:04:19 +01:00
|
|
|
return false, nil
|
|
|
|
} else if err != nil {
|
2020-09-08 17:45:10 +02:00
|
|
|
log.Error("Unable stat file: %s for LFS OID[%s] Error: %v", p, meta.Oid, err)
|
2017-11-08 14:04:19 +01:00
|
|
|
return false, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return true, nil
|
|
|
|
}
|
2021-03-06 21:21:56 +01:00
|
|
|
|
|
|
|
type hashingReader struct {
|
|
|
|
internal io.Reader
|
|
|
|
currentSize int64
|
|
|
|
expectedSize int64
|
|
|
|
hash hash.Hash
|
|
|
|
expectedHash string
|
|
|
|
}
|
|
|
|
|
|
|
|
func (r *hashingReader) Read(b []byte) (int, error) {
|
|
|
|
n, err := r.internal.Read(b)
|
|
|
|
|
|
|
|
if n > 0 {
|
|
|
|
r.currentSize += int64(n)
|
|
|
|
wn, werr := r.hash.Write(b[:n])
|
|
|
|
if wn != n || werr != nil {
|
|
|
|
return n, werr
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if err != nil && err == io.EOF {
|
|
|
|
if r.currentSize != r.expectedSize {
|
|
|
|
return n, errSizeMismatch
|
|
|
|
}
|
|
|
|
|
|
|
|
shaStr := hex.EncodeToString(r.hash.Sum(nil))
|
|
|
|
if shaStr != r.expectedHash {
|
|
|
|
return n, errHashMismatch
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return n, err
|
|
|
|
}
|
|
|
|
|
|
|
|
func newHashingReader(expectedSize int64, expectedHash string, reader io.Reader) *hashingReader {
|
|
|
|
return &hashingReader{
|
|
|
|
internal: reader,
|
|
|
|
expectedSize: expectedSize,
|
|
|
|
expectedHash: expectedHash,
|
|
|
|
hash: sha256.New(),
|
|
|
|
}
|
|
|
|
}
|