mirror of
https://codeberg.org/forgejo/forgejo.git
synced 2024-12-20 23:49:23 +00:00
5727056ea1
A feedback from discord: https://discord.com/channels/322538954119184384/561007778139734027/1090185427115319386 Some storages like: * https://developers.cloudflare.com/r2/api/s3/api/ * https://www.backblaze.com/b2/docs/s3_compatible_api.html They do not support "x-amz-checksum-algorithm" header But minio recently uses that header with CRC32C by default. So we have to tell minio to use legacy MD5 checksum. I guess this needs to be backported because IIRC we 1.19 and 1.20 are using similar minio package. The minio package code for SendContentMD5 looks like this: <details> <img width="755" alt="image" src="https://user-images.githubusercontent.com/2114189/228186768-4f2f6f67-62b9-4aee-9251-5af714ad9674.png"> </details>
165 lines
4.6 KiB
Go
165 lines
4.6 KiB
Go
// Copyright 2020 The Gitea Authors. All rights reserved.
|
|
// SPDX-License-Identifier: MIT
|
|
|
|
package lfs
|
|
|
|
import (
|
|
"encoding/hex"
|
|
"errors"
|
|
"hash"
|
|
"io"
|
|
"os"
|
|
|
|
"code.gitea.io/gitea/modules/log"
|
|
"code.gitea.io/gitea/modules/storage"
|
|
|
|
"github.com/minio/sha256-simd"
|
|
)
|
|
|
|
var (
|
|
// ErrHashMismatch occurs if the content has does not match OID
|
|
ErrHashMismatch = errors.New("Content hash does not match OID")
|
|
// ErrSizeMismatch occurs if the content size does not match
|
|
ErrSizeMismatch = errors.New("Content size does not match")
|
|
)
|
|
|
|
// ContentStore provides a simple file system based storage.
|
|
type ContentStore struct {
|
|
storage.ObjectStorage
|
|
}
|
|
|
|
// NewContentStore creates the default ContentStore
|
|
func NewContentStore() *ContentStore {
|
|
contentStore := &ContentStore{ObjectStorage: storage.LFS}
|
|
return contentStore
|
|
}
|
|
|
|
// Get takes a Meta object and retrieves the content from the store, returning
|
|
// it as an io.ReadSeekCloser.
|
|
func (s *ContentStore) Get(pointer Pointer) (storage.Object, error) {
|
|
f, err := s.Open(pointer.RelativePath())
|
|
if err != nil {
|
|
log.Error("Whilst trying to read LFS OID[%s]: Unable to open Error: %v", pointer.Oid, err)
|
|
return nil, err
|
|
}
|
|
return f, err
|
|
}
|
|
|
|
// Put takes a Meta object and an io.Reader and writes the content to the store.
|
|
func (s *ContentStore) Put(pointer Pointer, r io.Reader) error {
|
|
p := pointer.RelativePath()
|
|
|
|
// Wrap the provided reader with an inline hashing and size checker
|
|
wrappedRd := newHashingReader(pointer.Size, pointer.Oid, r)
|
|
|
|
// now pass the wrapped reader to Save - if there is a size mismatch or hash mismatch then
|
|
// the errors returned by the newHashingReader should percolate up to here
|
|
written, err := s.Save(p, wrappedRd, pointer.Size)
|
|
if err != nil {
|
|
log.Error("Whilst putting LFS OID[%s]: Failed to copy to tmpPath: %s Error: %v", pointer.Oid, p, err)
|
|
return err
|
|
}
|
|
|
|
// check again whether there is any error during the Save operation
|
|
// because some errors might be ignored by the Reader's caller
|
|
if wrappedRd.lastError != nil && !errors.Is(wrappedRd.lastError, io.EOF) {
|
|
err = wrappedRd.lastError
|
|
} else if written != pointer.Size {
|
|
err = ErrSizeMismatch
|
|
}
|
|
|
|
// if the upload failed, try to delete the file
|
|
if err != nil {
|
|
if errDel := s.Delete(p); errDel != nil {
|
|
log.Error("Cleaning the LFS OID[%s] failed: %v", pointer.Oid, errDel)
|
|
}
|
|
}
|
|
|
|
return err
|
|
}
|
|
|
|
// Exists returns true if the object exists in the content store.
|
|
func (s *ContentStore) Exists(pointer Pointer) (bool, error) {
|
|
_, err := s.ObjectStorage.Stat(pointer.RelativePath())
|
|
if err != nil {
|
|
if os.IsNotExist(err) {
|
|
return false, nil
|
|
}
|
|
return false, err
|
|
}
|
|
return true, nil
|
|
}
|
|
|
|
// Verify returns true if the object exists in the content store and size is correct.
|
|
func (s *ContentStore) Verify(pointer Pointer) (bool, error) {
|
|
p := pointer.RelativePath()
|
|
fi, err := s.ObjectStorage.Stat(p)
|
|
if os.IsNotExist(err) || (err == nil && fi.Size() != pointer.Size) {
|
|
return false, nil
|
|
} else if err != nil {
|
|
log.Error("Unable stat file: %s for LFS OID[%s] Error: %v", p, pointer.Oid, err)
|
|
return false, err
|
|
}
|
|
|
|
return true, nil
|
|
}
|
|
|
|
// ReadMetaObject will read a git_model.LFSMetaObject and return a reader
|
|
func ReadMetaObject(pointer Pointer) (io.ReadCloser, error) {
|
|
contentStore := NewContentStore()
|
|
return contentStore.Get(pointer)
|
|
}
|
|
|
|
type hashingReader struct {
|
|
internal io.Reader
|
|
currentSize int64
|
|
expectedSize int64
|
|
hash hash.Hash
|
|
expectedHash string
|
|
lastError error
|
|
}
|
|
|
|
// recordError records the last error during the Save operation
|
|
// Some callers of the Reader doesn't respect the returned "err"
|
|
// For example, MinIO's Put will ignore errors if the written size could equal to expected size
|
|
// So we must remember the error by ourselves,
|
|
// and later check again whether ErrSizeMismatch or ErrHashMismatch occurs during the Save operation
|
|
func (r *hashingReader) recordError(err error) error {
|
|
r.lastError = err
|
|
return err
|
|
}
|
|
|
|
func (r *hashingReader) Read(b []byte) (int, error) {
|
|
n, err := r.internal.Read(b)
|
|
|
|
if n > 0 {
|
|
r.currentSize += int64(n)
|
|
wn, werr := r.hash.Write(b[:n])
|
|
if wn != n || werr != nil {
|
|
return n, r.recordError(werr)
|
|
}
|
|
}
|
|
|
|
if errors.Is(err, io.EOF) || r.currentSize >= r.expectedSize {
|
|
if r.currentSize != r.expectedSize {
|
|
return n, r.recordError(ErrSizeMismatch)
|
|
}
|
|
|
|
shaStr := hex.EncodeToString(r.hash.Sum(nil))
|
|
if shaStr != r.expectedHash {
|
|
return n, r.recordError(ErrHashMismatch)
|
|
}
|
|
}
|
|
|
|
return n, r.recordError(err)
|
|
}
|
|
|
|
func newHashingReader(expectedSize int64, expectedHash string, reader io.Reader) *hashingReader {
|
|
return &hashingReader{
|
|
internal: reader,
|
|
expectedSize: expectedSize,
|
|
expectedHash: expectedHash,
|
|
hash: sha256.New(),
|
|
}
|
|
}
|