mirror of
https://github.com/superseriousbusiness/gotosocial.git
synced 2024-12-29 18:46:31 +00:00
9d0df426da
* feat: vendor minio client * feat: introduce storage package with s3 support * feat: serve s3 files directly this saves a lot of bandwith as the files are fetched from the object store directly * fix: use explicit local storage in tests * feat: integrate s3 storage with the main server * fix: add s3 config to cli tests * docs: explicitly set values in example config also adds license header to the storage package * fix: use better http status code on s3 redirect HTTP 302 Found is the best fit, as it signifies that the resource requested was found but not under its presumed URL 307/TemporaryRedirect would mean that this resource is usually located here, not in this case 303/SeeOther indicates that the redirection does not link to the requested resource but to another page * refactor: use context in storage driver interface
200 lines
5.7 KiB
Go
200 lines
5.7 KiB
Go
//+build !noasm,!appengine,gc
|
|
|
|
// Copyright (c) 2020 MinIO Inc. All rights reserved.
|
|
// Use of this source code is governed by a license that can be
|
|
// found in the LICENSE file.
|
|
|
|
package md5simd
|
|
|
|
import (
|
|
"fmt"
|
|
"math"
|
|
"sync"
|
|
"unsafe"
|
|
|
|
"github.com/klauspost/cpuid"
|
|
)
|
|
|
|
var hasAVX512 bool
|
|
|
|
//go:noescape
|
|
func block8(state *uint32, base uintptr, bufs *int32, cache *byte, n int)
|
|
|
|
//go:noescape
|
|
func block16(state *uint32, base uintptr, ptrs *int32, mask uint64, n int)
|
|
|
|
// 8-way 4x uint32 digests in 4 ymm registers
|
|
// (ymm0, ymm1, ymm2, ymm3)
|
|
type digest8 struct {
|
|
v0, v1, v2, v3 [8]uint32
|
|
}
|
|
|
|
// Stack cache for 8x64 byte md5.BlockSize bytes.
|
|
// Must be 32-byte aligned, so allocate 512+32 and
|
|
// align upwards at runtime.
|
|
type cache8 [512 + 32]byte
|
|
|
|
// MD5 magic numbers for one lane of hashing; inflated
|
|
// 8x below at init time.
|
|
var md5consts = [64]uint32{
|
|
0xd76aa478, 0xe8c7b756, 0x242070db, 0xc1bdceee,
|
|
0xf57c0faf, 0x4787c62a, 0xa8304613, 0xfd469501,
|
|
0x698098d8, 0x8b44f7af, 0xffff5bb1, 0x895cd7be,
|
|
0x6b901122, 0xfd987193, 0xa679438e, 0x49b40821,
|
|
0xf61e2562, 0xc040b340, 0x265e5a51, 0xe9b6c7aa,
|
|
0xd62f105d, 0x02441453, 0xd8a1e681, 0xe7d3fbc8,
|
|
0x21e1cde6, 0xc33707d6, 0xf4d50d87, 0x455a14ed,
|
|
0xa9e3e905, 0xfcefa3f8, 0x676f02d9, 0x8d2a4c8a,
|
|
0xfffa3942, 0x8771f681, 0x6d9d6122, 0xfde5380c,
|
|
0xa4beea44, 0x4bdecfa9, 0xf6bb4b60, 0xbebfbc70,
|
|
0x289b7ec6, 0xeaa127fa, 0xd4ef3085, 0x04881d05,
|
|
0xd9d4d039, 0xe6db99e5, 0x1fa27cf8, 0xc4ac5665,
|
|
0xf4292244, 0x432aff97, 0xab9423a7, 0xfc93a039,
|
|
0x655b59c3, 0x8f0ccc92, 0xffeff47d, 0x85845dd1,
|
|
0x6fa87e4f, 0xfe2ce6e0, 0xa3014314, 0x4e0811a1,
|
|
0xf7537e82, 0xbd3af235, 0x2ad7d2bb, 0xeb86d391,
|
|
}
|
|
|
|
// inflate the consts 8-way for 8x md5 (256 bit ymm registers)
|
|
var avx256md5consts = func(c []uint32) []uint32 {
|
|
inf := make([]uint32, 8*len(c))
|
|
for i := range c {
|
|
for j := 0; j < 8; j++ {
|
|
inf[(i*8)+j] = c[i]
|
|
}
|
|
}
|
|
return inf
|
|
}(md5consts[:])
|
|
|
|
// 16-way 4x uint32 digests in 4 zmm registers
|
|
type digest16 struct {
|
|
v0, v1, v2, v3 [16]uint32
|
|
}
|
|
|
|
// inflate the consts 16-way for 16x md5 (512 bit zmm registers)
|
|
var avx512md5consts = func(c []uint32) []uint32 {
|
|
inf := make([]uint32, 16*len(c))
|
|
for i := range c {
|
|
for j := 0; j < 16; j++ {
|
|
inf[(i*16)+j] = c[i]
|
|
}
|
|
}
|
|
return inf
|
|
}(md5consts[:])
|
|
|
|
func init() {
|
|
hasAVX512 = cpuid.CPU.AVX512F()
|
|
}
|
|
|
|
// Interface function to assembly code
|
|
func (s *md5Server) blockMd5_x16(d *digest16, input [16][]byte, half bool) {
|
|
if hasAVX512 {
|
|
blockMd5_avx512(d, input, s.allBufs, &s.maskRounds16)
|
|
} else {
|
|
d8a, d8b := digest8{}, digest8{}
|
|
for i := range d8a.v0 {
|
|
j := i + 8
|
|
d8a.v0[i], d8a.v1[i], d8a.v2[i], d8a.v3[i] = d.v0[i], d.v1[i], d.v2[i], d.v3[i]
|
|
if !half {
|
|
d8b.v0[i], d8b.v1[i], d8b.v2[i], d8b.v3[i] = d.v0[j], d.v1[j], d.v2[j], d.v3[j]
|
|
}
|
|
}
|
|
|
|
i8 := [2][8][]byte{}
|
|
for i := range i8[0] {
|
|
i8[0][i], i8[1][i] = input[i], input[8+i]
|
|
}
|
|
if half {
|
|
blockMd5_avx2(&d8a, i8[0], s.allBufs, &s.maskRounds8a)
|
|
} else {
|
|
wg := sync.WaitGroup{}
|
|
wg.Add(2)
|
|
go func() { blockMd5_avx2(&d8a, i8[0], s.allBufs, &s.maskRounds8a); wg.Done() }()
|
|
go func() { blockMd5_avx2(&d8b, i8[1], s.allBufs, &s.maskRounds8b); wg.Done() }()
|
|
wg.Wait()
|
|
}
|
|
|
|
for i := range d8a.v0 {
|
|
j := i + 8
|
|
d.v0[i], d.v1[i], d.v2[i], d.v3[i] = d8a.v0[i], d8a.v1[i], d8a.v2[i], d8a.v3[i]
|
|
if !half {
|
|
d.v0[j], d.v1[j], d.v2[j], d.v3[j] = d8b.v0[i], d8b.v1[i], d8b.v2[i], d8b.v3[i]
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Interface function to AVX512 assembly code
|
|
func blockMd5_avx512(s *digest16, input [16][]byte, base []byte, maskRounds *[16]maskRounds) {
|
|
baseMin := uint64(uintptr(unsafe.Pointer(&(base[0]))))
|
|
ptrs := [16]int32{}
|
|
|
|
for i := range ptrs {
|
|
if len(input[i]) > 0 {
|
|
if len(input[i]) > internalBlockSize {
|
|
panic(fmt.Sprintf("Sanity check fails for lane %d: maximum input length cannot exceed internalBlockSize", i))
|
|
}
|
|
|
|
off := uint64(uintptr(unsafe.Pointer(&(input[i][0])))) - baseMin
|
|
if off > math.MaxUint32 {
|
|
panic(fmt.Sprintf("invalid buffer sent with offset %x", off))
|
|
}
|
|
ptrs[i] = int32(off)
|
|
}
|
|
}
|
|
|
|
sdup := *s // create copy of initial states to receive intermediate updates
|
|
|
|
rounds := generateMaskAndRounds16(input, maskRounds)
|
|
|
|
for r := 0; r < rounds; r++ {
|
|
m := maskRounds[r]
|
|
|
|
block16(&sdup.v0[0], uintptr(baseMin), &ptrs[0], m.mask, int(64*m.rounds))
|
|
|
|
for j := 0; j < len(ptrs); j++ {
|
|
ptrs[j] += int32(64 * m.rounds) // update pointers for next round
|
|
if m.mask&(1<<j) != 0 { // update digest if still masked as active
|
|
(*s).v0[j], (*s).v1[j], (*s).v2[j], (*s).v3[j] = sdup.v0[j], sdup.v1[j], sdup.v2[j], sdup.v3[j]
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Interface function to AVX2 assembly code
|
|
func blockMd5_avx2(s *digest8, input [8][]byte, base []byte, maskRounds *[8]maskRounds) {
|
|
baseMin := uint64(uintptr(unsafe.Pointer(&(base[0])))) - 4
|
|
ptrs := [8]int32{}
|
|
|
|
for i := range ptrs {
|
|
if len(input[i]) > 0 {
|
|
if len(input[i]) > internalBlockSize {
|
|
panic(fmt.Sprintf("Sanity check fails for lane %d: maximum input length cannot exceed internalBlockSize", i))
|
|
}
|
|
|
|
off := uint64(uintptr(unsafe.Pointer(&(input[i][0])))) - baseMin
|
|
if off > math.MaxUint32 {
|
|
panic(fmt.Sprintf("invalid buffer sent with offset %x", off))
|
|
}
|
|
ptrs[i] = int32(off)
|
|
}
|
|
}
|
|
|
|
sdup := *s // create copy of initial states to receive intermediate updates
|
|
|
|
rounds := generateMaskAndRounds8(input, maskRounds)
|
|
|
|
for r := 0; r < rounds; r++ {
|
|
m := maskRounds[r]
|
|
var cache cache8 // stack storage for block8 tmp state
|
|
block8(&sdup.v0[0], uintptr(baseMin), &ptrs[0], &cache[0], int(64*m.rounds))
|
|
|
|
for j := 0; j < len(ptrs); j++ {
|
|
ptrs[j] += int32(64 * m.rounds) // update pointers for next round
|
|
if m.mask&(1<<j) != 0 { // update digest if still masked as active
|
|
(*s).v0[j], (*s).v1[j], (*s).v2[j], (*s).v3[j] = sdup.v0[j], sdup.v1[j], sdup.v2[j], sdup.v3[j]
|
|
}
|
|
}
|
|
}
|
|
}
|