mirror of
https://codeberg.org/forgejo/forgejo
synced 2024-12-05 02:54:46 +01:00
b73d1ac1eb
Backport #23768 (no source code conflict, only some unrelated docs/test-ini conflicts) Some storages like: * https://developers.cloudflare.com/r2/api/s3/api/ * https://www.backblaze.com/b2/docs/s3_compatible_api.html They do not support "x-amz-checksum-algorithm" header But minio recently uses that header with CRC32C by default. So we have to tell minio to use legacy MD5 checksum.
163 lines
4.6 KiB
Go
163 lines
4.6 KiB
Go
// Copyright 2020 The Gitea Authors. All rights reserved.
|
|
// SPDX-License-Identifier: MIT
|
|
|
|
package lfs
|
|
|
|
import (
|
|
"crypto/sha256"
|
|
"encoding/hex"
|
|
"errors"
|
|
"hash"
|
|
"io"
|
|
"os"
|
|
|
|
"code.gitea.io/gitea/modules/log"
|
|
"code.gitea.io/gitea/modules/storage"
|
|
)
|
|
|
|
var (
|
|
// ErrHashMismatch occurs if the content has does not match OID
|
|
ErrHashMismatch = errors.New("Content hash does not match OID")
|
|
// ErrSizeMismatch occurs if the content size does not match
|
|
ErrSizeMismatch = errors.New("Content size does not match")
|
|
)
|
|
|
|
// ContentStore provides a simple file system based storage.
|
|
type ContentStore struct {
|
|
storage.ObjectStorage
|
|
}
|
|
|
|
// NewContentStore creates the default ContentStore
|
|
func NewContentStore() *ContentStore {
|
|
contentStore := &ContentStore{ObjectStorage: storage.LFS}
|
|
return contentStore
|
|
}
|
|
|
|
// Get takes a Meta object and retrieves the content from the store, returning
|
|
// it as an io.ReadSeekCloser.
|
|
func (s *ContentStore) Get(pointer Pointer) (storage.Object, error) {
|
|
f, err := s.Open(pointer.RelativePath())
|
|
if err != nil {
|
|
log.Error("Whilst trying to read LFS OID[%s]: Unable to open Error: %v", pointer.Oid, err)
|
|
return nil, err
|
|
}
|
|
return f, err
|
|
}
|
|
|
|
// Put takes a Meta object and an io.Reader and writes the content to the store.
|
|
func (s *ContentStore) Put(pointer Pointer, r io.Reader) error {
|
|
p := pointer.RelativePath()
|
|
|
|
// Wrap the provided reader with an inline hashing and size checker
|
|
wrappedRd := newHashingReader(pointer.Size, pointer.Oid, r)
|
|
|
|
// now pass the wrapped reader to Save - if there is a size mismatch or hash mismatch then
|
|
// the errors returned by the newHashingReader should percolate up to here
|
|
written, err := s.Save(p, wrappedRd, pointer.Size)
|
|
if err != nil {
|
|
log.Error("Whilst putting LFS OID[%s]: Failed to copy to tmpPath: %s Error: %v", pointer.Oid, p, err)
|
|
return err
|
|
}
|
|
|
|
// check again whether there is any error during the Save operation
|
|
// because some errors might be ignored by the Reader's caller
|
|
if wrappedRd.lastError != nil && !errors.Is(wrappedRd.lastError, io.EOF) {
|
|
err = wrappedRd.lastError
|
|
} else if written != pointer.Size {
|
|
err = ErrSizeMismatch
|
|
}
|
|
|
|
// if the upload failed, try to delete the file
|
|
if err != nil {
|
|
if errDel := s.Delete(p); errDel != nil {
|
|
log.Error("Cleaning the LFS OID[%s] failed: %v", pointer.Oid, errDel)
|
|
}
|
|
}
|
|
|
|
return err
|
|
}
|
|
|
|
// Exists returns true if the object exists in the content store.
|
|
func (s *ContentStore) Exists(pointer Pointer) (bool, error) {
|
|
_, err := s.ObjectStorage.Stat(pointer.RelativePath())
|
|
if err != nil {
|
|
if os.IsNotExist(err) {
|
|
return false, nil
|
|
}
|
|
return false, err
|
|
}
|
|
return true, nil
|
|
}
|
|
|
|
// Verify returns true if the object exists in the content store and size is correct.
|
|
func (s *ContentStore) Verify(pointer Pointer) (bool, error) {
|
|
p := pointer.RelativePath()
|
|
fi, err := s.ObjectStorage.Stat(p)
|
|
if os.IsNotExist(err) || (err == nil && fi.Size() != pointer.Size) {
|
|
return false, nil
|
|
} else if err != nil {
|
|
log.Error("Unable stat file: %s for LFS OID[%s] Error: %v", p, pointer.Oid, err)
|
|
return false, err
|
|
}
|
|
|
|
return true, nil
|
|
}
|
|
|
|
// ReadMetaObject will read a git_model.LFSMetaObject and return a reader
|
|
func ReadMetaObject(pointer Pointer) (io.ReadCloser, error) {
|
|
contentStore := NewContentStore()
|
|
return contentStore.Get(pointer)
|
|
}
|
|
|
|
type hashingReader struct {
|
|
internal io.Reader
|
|
currentSize int64
|
|
expectedSize int64
|
|
hash hash.Hash
|
|
expectedHash string
|
|
lastError error
|
|
}
|
|
|
|
// recordError records the last error during the Save operation
|
|
// Some callers of the Reader doesn't respect the returned "err"
|
|
// For example, MinIO's Put will ignore errors if the written size could equal to expected size
|
|
// So we must remember the error by ourselves,
|
|
// and later check again whether ErrSizeMismatch or ErrHashMismatch occurs during the Save operation
|
|
func (r *hashingReader) recordError(err error) error {
|
|
r.lastError = err
|
|
return err
|
|
}
|
|
|
|
func (r *hashingReader) Read(b []byte) (int, error) {
|
|
n, err := r.internal.Read(b)
|
|
|
|
if n > 0 {
|
|
r.currentSize += int64(n)
|
|
wn, werr := r.hash.Write(b[:n])
|
|
if wn != n || werr != nil {
|
|
return n, r.recordError(werr)
|
|
}
|
|
}
|
|
|
|
if errors.Is(err, io.EOF) || r.currentSize >= r.expectedSize {
|
|
if r.currentSize != r.expectedSize {
|
|
return n, r.recordError(ErrSizeMismatch)
|
|
}
|
|
|
|
shaStr := hex.EncodeToString(r.hash.Sum(nil))
|
|
if shaStr != r.expectedHash {
|
|
return n, r.recordError(ErrHashMismatch)
|
|
}
|
|
}
|
|
|
|
return n, r.recordError(err)
|
|
}
|
|
|
|
func newHashingReader(expectedSize int64, expectedHash string, reader io.Reader) *hashingReader {
|
|
return &hashingReader{
|
|
internal: reader,
|
|
expectedSize: expectedSize,
|
|
expectedHash: expectedHash,
|
|
hash: sha256.New(),
|
|
}
|
|
}
|