mirror of
https://github.com/optim-enterprises-bv/vault.git
synced 2025-11-02 19:47:54 +00:00
Mongodb driver switch to mongo-driver (#8140)
* Switch mongodb driver to mongo-driver * Tidy mod * Make writeConcern private * Implement review feedback * Add retry functionality * Added backoff time * go mod vendor * Fix failing test * goimport
This commit is contained in:
@@ -14,8 +14,9 @@ import (
|
|||||||
"github.com/hashicorp/vault/sdk/framework"
|
"github.com/hashicorp/vault/sdk/framework"
|
||||||
"github.com/hashicorp/vault/sdk/helper/dbtxn"
|
"github.com/hashicorp/vault/sdk/helper/dbtxn"
|
||||||
"github.com/hashicorp/vault/sdk/logical"
|
"github.com/hashicorp/vault/sdk/logical"
|
||||||
|
|
||||||
"github.com/lib/pq"
|
"github.com/lib/pq"
|
||||||
|
"go.mongodb.org/mongo-driver/mongo"
|
||||||
|
"go.mongodb.org/mongo-driver/mongo/options"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@@ -843,6 +844,11 @@ func TestBackend_StaticRole_Rotations_MongoDB(t *testing.T) {
|
|||||||
// configure backend, add item and confirm length
|
// configure backend, add item and confirm length
|
||||||
cleanup, connURL := mongodb.PrepareTestContainerWithDatabase(t, "latest", "vaulttestdb")
|
cleanup, connURL := mongodb.PrepareTestContainerWithDatabase(t, "latest", "vaulttestdb")
|
||||||
defer cleanup()
|
defer cleanup()
|
||||||
|
testCases := []string{"65", "130", "5400"}
|
||||||
|
// Create database users ahead
|
||||||
|
for _, tc := range testCases {
|
||||||
|
testCreateDBUser(t, connURL, "vaulttestdb", "statictestMongo"+tc, "test")
|
||||||
|
}
|
||||||
|
|
||||||
// Configure a connection
|
// Configure a connection
|
||||||
data := map[string]interface{}{
|
data := map[string]interface{}{
|
||||||
@@ -865,7 +871,6 @@ func TestBackend_StaticRole_Rotations_MongoDB(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// create three static roles with different rotation periods
|
// create three static roles with different rotation periods
|
||||||
testCases := []string{"65", "130", "5400"}
|
|
||||||
for _, tc := range testCases {
|
for _, tc := range testCases {
|
||||||
roleName := "plugin-static-role-" + tc
|
roleName := "plugin-static-role-" + tc
|
||||||
data = map[string]interface{}{
|
data = map[string]interface{}{
|
||||||
@@ -956,6 +961,30 @@ func TestBackend_StaticRole_Rotations_MongoDB(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func testCreateDBUser(t testing.TB, connURL, db, username, password string) {
|
||||||
|
ctx, _ := context.WithTimeout(context.Background(), 10*time.Second)
|
||||||
|
client, err := mongo.Connect(ctx, options.Client().ApplyURI(connURL))
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
createUserCmd := &createUserCommand{
|
||||||
|
Username: username,
|
||||||
|
Password: password,
|
||||||
|
Roles: []interface{}{},
|
||||||
|
}
|
||||||
|
result := client.Database(db).RunCommand(ctx, createUserCmd, nil)
|
||||||
|
if result.Err() != nil {
|
||||||
|
t.Fatal(result.Err())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type createUserCommand struct {
|
||||||
|
Username string `bson:"createUser"`
|
||||||
|
Password string `bson:"pwd"`
|
||||||
|
Roles []interface{} `bson:"roles"`
|
||||||
|
}
|
||||||
|
|
||||||
// Demonstrates a bug fix for the credential rotation not releasing locks
|
// Demonstrates a bug fix for the credential rotation not releasing locks
|
||||||
func TestBackend_StaticRole_LockRegression(t *testing.T) {
|
func TestBackend_StaticRole_LockRegression(t *testing.T) {
|
||||||
cluster, sys := getCluster(t)
|
cluster, sys := getCluster(t)
|
||||||
|
|||||||
5
go.mod
5
go.mod
@@ -10,6 +10,7 @@ require (
|
|||||||
cloud.google.com/go v0.39.0
|
cloud.google.com/go v0.39.0
|
||||||
github.com/Azure/azure-sdk-for-go v36.2.0+incompatible
|
github.com/Azure/azure-sdk-for-go v36.2.0+incompatible
|
||||||
github.com/Azure/go-autorest/autorest v0.9.2
|
github.com/Azure/go-autorest/autorest v0.9.2
|
||||||
|
github.com/DataDog/zstd v1.4.4 // indirect
|
||||||
github.com/NYTimes/gziphandler v1.1.1
|
github.com/NYTimes/gziphandler v1.1.1
|
||||||
github.com/SAP/go-hdb v0.14.1
|
github.com/SAP/go-hdb v0.14.1
|
||||||
github.com/StackExchange/wmi v0.0.0-20180116203802-5d049714c4a6 // indirect
|
github.com/StackExchange/wmi v0.0.0-20180116203802-5d049714c4a6 // indirect
|
||||||
@@ -124,9 +125,13 @@ require (
|
|||||||
github.com/shirou/w32 v0.0.0-20160930032740-bb4de0191aa4 // indirect
|
github.com/shirou/w32 v0.0.0-20160930032740-bb4de0191aa4 // indirect
|
||||||
github.com/shopspring/decimal v0.0.0-20180709203117-cd690d0c9e24 // indirect
|
github.com/shopspring/decimal v0.0.0-20180709203117-cd690d0c9e24 // indirect
|
||||||
github.com/stretchr/testify v1.4.0
|
github.com/stretchr/testify v1.4.0
|
||||||
|
github.com/tidwall/pretty v1.0.0 // indirect
|
||||||
|
github.com/xdg/scram v0.0.0-20180814205039-7eeb5667e42c // indirect
|
||||||
|
github.com/xdg/stringprep v1.0.0 // indirect
|
||||||
github.com/xi2/xz v0.0.0-20171230120015-48954b6210f8 // indirect
|
github.com/xi2/xz v0.0.0-20171230120015-48954b6210f8 // indirect
|
||||||
go.etcd.io/bbolt v1.3.2
|
go.etcd.io/bbolt v1.3.2
|
||||||
go.etcd.io/etcd v0.0.0-20190412021913-f29b1ada1971
|
go.etcd.io/etcd v0.0.0-20190412021913-f29b1ada1971
|
||||||
|
go.mongodb.org/mongo-driver v1.2.1
|
||||||
go.uber.org/atomic v1.4.0
|
go.uber.org/atomic v1.4.0
|
||||||
golang.org/x/crypto v0.0.0-20191106202628-ed6320f186d4
|
golang.org/x/crypto v0.0.0-20191106202628-ed6320f186d4
|
||||||
golang.org/x/net v0.0.0-20190813141303-74dc4d7220e7
|
golang.org/x/net v0.0.0-20190813141303-74dc4d7220e7
|
||||||
|
|||||||
11
go.sum
11
go.sum
@@ -45,6 +45,8 @@ github.com/BurntSushi/toml v0.3.1/go.mod h1:xHWCNGjB5oqiDr8zfno3MHue2Ht5sIBksp03
|
|||||||
github.com/DataDog/datadog-go v2.2.0+incompatible/go.mod h1:LButxg5PwREeZtORoXG3tL4fMGNddJ+vMq1mwgfaqoQ=
|
github.com/DataDog/datadog-go v2.2.0+incompatible/go.mod h1:LButxg5PwREeZtORoXG3tL4fMGNddJ+vMq1mwgfaqoQ=
|
||||||
github.com/DataDog/datadog-go v3.2.0+incompatible h1:qSG2N4FghB1He/r2mFrWKCaL7dXCilEuNEeAn20fdD4=
|
github.com/DataDog/datadog-go v3.2.0+incompatible h1:qSG2N4FghB1He/r2mFrWKCaL7dXCilEuNEeAn20fdD4=
|
||||||
github.com/DataDog/datadog-go v3.2.0+incompatible/go.mod h1:LButxg5PwREeZtORoXG3tL4fMGNddJ+vMq1mwgfaqoQ=
|
github.com/DataDog/datadog-go v3.2.0+incompatible/go.mod h1:LButxg5PwREeZtORoXG3tL4fMGNddJ+vMq1mwgfaqoQ=
|
||||||
|
github.com/DataDog/zstd v1.4.4 h1:+IawcoXhCBylN7ccwdwf8LOH2jKq7NavGpEPanrlTzE=
|
||||||
|
github.com/DataDog/zstd v1.4.4/go.mod h1:1jcaCB/ufaK+sKp1NBhlGmpz41jOoPQ35bpF36t7BBo=
|
||||||
github.com/Jeffail/gabs v1.1.1 h1:V0uzR08Hj22EX8+8QMhyI9sX2hwRu+/RJhJUmnwda/E=
|
github.com/Jeffail/gabs v1.1.1 h1:V0uzR08Hj22EX8+8QMhyI9sX2hwRu+/RJhJUmnwda/E=
|
||||||
github.com/Jeffail/gabs v1.1.1/go.mod h1:6xMvQMK4k33lb7GUUpaAPh6nKMmemQeg5d4gn7/bOXc=
|
github.com/Jeffail/gabs v1.1.1/go.mod h1:6xMvQMK4k33lb7GUUpaAPh6nKMmemQeg5d4gn7/bOXc=
|
||||||
github.com/Masterminds/semver v1.4.2 h1:WBLTQ37jOCzSLtXNdoo8bNM8876KhNqOKvrlGITgsTc=
|
github.com/Masterminds/semver v1.4.2 h1:WBLTQ37jOCzSLtXNdoo8bNM8876KhNqOKvrlGITgsTc=
|
||||||
@@ -202,6 +204,7 @@ github.com/go-ole/go-ole v1.2.1 h1:2lOsA72HgjxAuMlKpFiCbHTvu44PIVkZ5hqm3RSdI/E=
|
|||||||
github.com/go-ole/go-ole v1.2.1/go.mod h1:7FAglXiTm7HKlQRDeOQ6ZNUHidzCWXuZWq/1dTyBNF8=
|
github.com/go-ole/go-ole v1.2.1/go.mod h1:7FAglXiTm7HKlQRDeOQ6ZNUHidzCWXuZWq/1dTyBNF8=
|
||||||
github.com/go-sql-driver/mysql v1.4.1 h1:g24URVg0OFbNUTx9qqY1IRZ9D9z3iPyi5zKhQZpNwpA=
|
github.com/go-sql-driver/mysql v1.4.1 h1:g24URVg0OFbNUTx9qqY1IRZ9D9z3iPyi5zKhQZpNwpA=
|
||||||
github.com/go-sql-driver/mysql v1.4.1/go.mod h1:zAC/RDZ24gD3HViQzih4MyKcchzm+sOG5ZlKdlhCg5w=
|
github.com/go-sql-driver/mysql v1.4.1/go.mod h1:zAC/RDZ24gD3HViQzih4MyKcchzm+sOG5ZlKdlhCg5w=
|
||||||
|
github.com/go-stack/stack v1.8.0 h1:5SgMzNM5HxrEjV0ww2lTmX6E2Izsfxas4+YHWRs3Lsk=
|
||||||
github.com/go-stack/stack v1.8.0/go.mod h1:v0f6uXyyMGvRgIKkXu+yp6POWl0qKG85gN/melR3HDY=
|
github.com/go-stack/stack v1.8.0/go.mod h1:v0f6uXyyMGvRgIKkXu+yp6POWl0qKG85gN/melR3HDY=
|
||||||
github.com/go-test/deep v1.0.2-0.20181118220953-042da051cf31/go.mod h1:wGDj63lr65AM2AQyKZd/NYHGb0R+1RLqB8NKt3aSFNA=
|
github.com/go-test/deep v1.0.2-0.20181118220953-042da051cf31/go.mod h1:wGDj63lr65AM2AQyKZd/NYHGb0R+1RLqB8NKt3aSFNA=
|
||||||
github.com/go-test/deep v1.0.2 h1:onZX1rnHT3Wv6cqNgYyFOOlgVKJrksuCMCRvJStbMYw=
|
github.com/go-test/deep v1.0.2 h1:onZX1rnHT3Wv6cqNgYyFOOlgVKJrksuCMCRvJStbMYw=
|
||||||
@@ -631,6 +634,8 @@ github.com/stretchr/testify v1.3.0/go.mod h1:M5WIy9Dh21IEIfnGCwXGc5bZfKNJtfHm1UV
|
|||||||
github.com/stretchr/testify v1.4.0 h1:2E4SXV/wtOkTonXsotYi4li6zVWxYlZuYNCXe9XRJyk=
|
github.com/stretchr/testify v1.4.0 h1:2E4SXV/wtOkTonXsotYi4li6zVWxYlZuYNCXe9XRJyk=
|
||||||
github.com/stretchr/testify v1.4.0/go.mod h1:j7eGeouHqKxXV5pUuKE4zz7dFj8WfuZ+81PSLYec5m4=
|
github.com/stretchr/testify v1.4.0/go.mod h1:j7eGeouHqKxXV5pUuKE4zz7dFj8WfuZ+81PSLYec5m4=
|
||||||
github.com/tarm/serial v0.0.0-20180830185346-98f6abe2eb07/go.mod h1:kDXzergiv9cbyO7IOYJZWg1U88JhDg3PB6klq9Hg2pA=
|
github.com/tarm/serial v0.0.0-20180830185346-98f6abe2eb07/go.mod h1:kDXzergiv9cbyO7IOYJZWg1U88JhDg3PB6klq9Hg2pA=
|
||||||
|
github.com/tidwall/pretty v1.0.0 h1:HsD+QiTn7sK6flMKIvNmpqz1qrpP3Ps6jOKIKMooyg4=
|
||||||
|
github.com/tidwall/pretty v1.0.0/go.mod h1:XNkn88O1ChpSDQmQeStsy+sBenx6DDtFZJxhVysOjyk=
|
||||||
github.com/tmc/grpc-websocket-proxy v0.0.0-20170815181823-89b8d40f7ca8 h1:ndzgwNDnKIqyCvHTXaCqh9KlOWKvBry6nuXMJmonVsE=
|
github.com/tmc/grpc-websocket-proxy v0.0.0-20170815181823-89b8d40f7ca8 h1:ndzgwNDnKIqyCvHTXaCqh9KlOWKvBry6nuXMJmonVsE=
|
||||||
github.com/tmc/grpc-websocket-proxy v0.0.0-20170815181823-89b8d40f7ca8/go.mod h1:ncp9v5uamzpCO7NfCPTXjqaC+bZgJeR0sMTm6dMHP7U=
|
github.com/tmc/grpc-websocket-proxy v0.0.0-20170815181823-89b8d40f7ca8/go.mod h1:ncp9v5uamzpCO7NfCPTXjqaC+bZgJeR0sMTm6dMHP7U=
|
||||||
github.com/tv42/httpunix v0.0.0-20150427012821-b75d8614f926 h1:G3dpKMzFDjgEh2q1Z7zUUtKa8ViPtH+ocF0bE0g00O8=
|
github.com/tv42/httpunix v0.0.0-20150427012821-b75d8614f926 h1:G3dpKMzFDjgEh2q1Z7zUUtKa8ViPtH+ocF0bE0g00O8=
|
||||||
@@ -642,6 +647,10 @@ github.com/ugorji/go/codec v0.0.0-20190204201341-e444a5086c43/go.mod h1:iT03XoTw
|
|||||||
github.com/ulikunitz/xz v0.5.6 h1:jGHAfXawEGZQ3blwU5wnWKQJvAraT7Ftq9EXjnXYgt8=
|
github.com/ulikunitz/xz v0.5.6 h1:jGHAfXawEGZQ3blwU5wnWKQJvAraT7Ftq9EXjnXYgt8=
|
||||||
github.com/ulikunitz/xz v0.5.6/go.mod h1:2bypXElzHzzJZwzH67Y6wb67pO62Rzfn7BSiF4ABRW8=
|
github.com/ulikunitz/xz v0.5.6/go.mod h1:2bypXElzHzzJZwzH67Y6wb67pO62Rzfn7BSiF4ABRW8=
|
||||||
github.com/urfave/cli v1.20.0/go.mod h1:70zkFmudgCuE/ngEzBv17Jvp/497gISqfk5gWijbERA=
|
github.com/urfave/cli v1.20.0/go.mod h1:70zkFmudgCuE/ngEzBv17Jvp/497gISqfk5gWijbERA=
|
||||||
|
github.com/xdg/scram v0.0.0-20180814205039-7eeb5667e42c h1:u40Z8hqBAAQyv+vATcGgV0YCnDjqSL7/q/JyPhhJSPk=
|
||||||
|
github.com/xdg/scram v0.0.0-20180814205039-7eeb5667e42c/go.mod h1:lB8K/P019DLNhemzwFU4jHLhdvlE6uDZjXFejJXr49I=
|
||||||
|
github.com/xdg/stringprep v1.0.0 h1:d9X0esnoa3dFsV0FG35rAT0RIhYFlPq7MiP+DW89La0=
|
||||||
|
github.com/xdg/stringprep v1.0.0/go.mod h1:Jhud4/sHMO4oL310DaZAKk9ZaJ08SJfe+sJh0HrGL1Y=
|
||||||
github.com/xi2/xz v0.0.0-20171230120015-48954b6210f8 h1:nIPpBwaJSVYIxUFsDv3M8ofmx9yWTog9BfvIu0q41lo=
|
github.com/xi2/xz v0.0.0-20171230120015-48954b6210f8 h1:nIPpBwaJSVYIxUFsDv3M8ofmx9yWTog9BfvIu0q41lo=
|
||||||
github.com/xi2/xz v0.0.0-20171230120015-48954b6210f8/go.mod h1:HUYIGzjTL3rfEspMxjDjgmT5uz5wzYJKVo23qUhYTos=
|
github.com/xi2/xz v0.0.0-20171230120015-48954b6210f8/go.mod h1:HUYIGzjTL3rfEspMxjDjgmT5uz5wzYJKVo23qUhYTos=
|
||||||
github.com/xiang90/probing v0.0.0-20190116061207-43a291ad63a2 h1:eY9dn8+vbi4tKz5Qo6v2eYzo7kUS51QINcR5jNpbZS8=
|
github.com/xiang90/probing v0.0.0-20190116061207-43a291ad63a2 h1:eY9dn8+vbi4tKz5Qo6v2eYzo7kUS51QINcR5jNpbZS8=
|
||||||
@@ -650,6 +659,8 @@ go.etcd.io/bbolt v1.3.2 h1:Z/90sZLPOeCy2PwprqkFa25PdkusRzaj9P8zm/KNyvk=
|
|||||||
go.etcd.io/bbolt v1.3.2/go.mod h1:IbVyRI1SCnLcuJnV2u8VeU0CEYM7e686BmAb1XKL+uU=
|
go.etcd.io/bbolt v1.3.2/go.mod h1:IbVyRI1SCnLcuJnV2u8VeU0CEYM7e686BmAb1XKL+uU=
|
||||||
go.etcd.io/etcd v0.0.0-20190412021913-f29b1ada1971 h1:C+ye4QyWT3rbVj8As5DUc+Dsp067xJxCC6aa9+UnCmU=
|
go.etcd.io/etcd v0.0.0-20190412021913-f29b1ada1971 h1:C+ye4QyWT3rbVj8As5DUc+Dsp067xJxCC6aa9+UnCmU=
|
||||||
go.etcd.io/etcd v0.0.0-20190412021913-f29b1ada1971/go.mod h1:KSGwdbiFchh5KIC9My2+ZVl5/3ANcwohw50dpPwa2cw=
|
go.etcd.io/etcd v0.0.0-20190412021913-f29b1ada1971/go.mod h1:KSGwdbiFchh5KIC9My2+ZVl5/3ANcwohw50dpPwa2cw=
|
||||||
|
go.mongodb.org/mongo-driver v1.2.1 h1:ANAlYXXM5XmOdW/Nc38jOr+wS5nlk7YihT24U1imiWM=
|
||||||
|
go.mongodb.org/mongo-driver v1.2.1/go.mod h1:u7ryQJ+DOzQmeO7zB6MHyr8jkEQvC8vH7qLUO4lqsUM=
|
||||||
go.opencensus.io v0.18.0/go.mod h1:vKdFvxhtzZ9onBp9VKHK8z/sRpBMnKAsufL7wlDrCOA=
|
go.opencensus.io v0.18.0/go.mod h1:vKdFvxhtzZ9onBp9VKHK8z/sRpBMnKAsufL7wlDrCOA=
|
||||||
go.opencensus.io v0.19.1/go.mod h1:gug0GbSHa8Pafr0d2urOSgoXHZ6x/RUlaiT0d9pqb4A=
|
go.opencensus.io v0.19.1/go.mod h1:gug0GbSHa8Pafr0d2urOSgoXHZ6x/RUlaiT0d9pqb4A=
|
||||||
go.opencensus.io v0.19.2/go.mod h1:NO/8qkisMZLZ1FCsKNqtJPwc8/TaclWyY0B6wcYNg9M=
|
go.opencensus.io v0.19.2/go.mod h1:NO/8qkisMZLZ1FCsKNqtJPwc8/TaclWyY0B6wcYNg9M=
|
||||||
|
|||||||
@@ -2,15 +2,9 @@ package mongodb
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"crypto/tls"
|
|
||||||
"encoding/base64"
|
"encoding/base64"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
|
||||||
"fmt"
|
"fmt"
|
||||||
"net"
|
|
||||||
"net/url"
|
|
||||||
"strconv"
|
|
||||||
"strings"
|
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
@@ -18,7 +12,10 @@ import (
|
|||||||
"github.com/hashicorp/vault/sdk/database/helper/connutil"
|
"github.com/hashicorp/vault/sdk/database/helper/connutil"
|
||||||
"github.com/hashicorp/vault/sdk/database/helper/dbutil"
|
"github.com/hashicorp/vault/sdk/database/helper/dbutil"
|
||||||
"github.com/mitchellh/mapstructure"
|
"github.com/mitchellh/mapstructure"
|
||||||
mgo "gopkg.in/mgo.v2"
|
"go.mongodb.org/mongo-driver/mongo"
|
||||||
|
"go.mongodb.org/mongo-driver/mongo/options"
|
||||||
|
"go.mongodb.org/mongo-driver/mongo/readpref"
|
||||||
|
"go.mongodb.org/mongo-driver/mongo/writeconcern"
|
||||||
)
|
)
|
||||||
|
|
||||||
// mongoDBConnectionProducer implements ConnectionProducer and provides an
|
// mongoDBConnectionProducer implements ConnectionProducer and provides an
|
||||||
@@ -32,11 +29,20 @@ type mongoDBConnectionProducer struct {
|
|||||||
Initialized bool
|
Initialized bool
|
||||||
RawConfig map[string]interface{}
|
RawConfig map[string]interface{}
|
||||||
Type string
|
Type string
|
||||||
session *mgo.Session
|
clientOptions *options.ClientOptions
|
||||||
safe *mgo.Safe
|
client *mongo.Client
|
||||||
sync.Mutex
|
sync.Mutex
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// writeConcern defines the write concern options
|
||||||
|
type writeConcern struct {
|
||||||
|
W int // Min # of servers to ack before success
|
||||||
|
WMode string // Write mode for MongoDB 2.0+ (e.g. "majority")
|
||||||
|
WTimeout int // Milliseconds to wait for W before timing out
|
||||||
|
FSync bool // DEPRECATED: Is now handled by J. See: https://jira.mongodb.org/browse/CXX-910
|
||||||
|
J bool // Sync via the journal if present
|
||||||
|
}
|
||||||
|
|
||||||
func (c *mongoDBConnectionProducer) Initialize(ctx context.Context, conf map[string]interface{}, verifyConnection bool) error {
|
func (c *mongoDBConnectionProducer) Initialize(ctx context.Context, conf map[string]interface{}, verifyConnection bool) error {
|
||||||
_, err := c.Init(ctx, conf, verifyConnection)
|
_, err := c.Init(ctx, conf, verifyConnection)
|
||||||
return err
|
return err
|
||||||
@@ -73,18 +79,41 @@ func (c *mongoDBConnectionProducer) Init(ctx context.Context, conf map[string]in
|
|||||||
input = string(inputBytes)
|
input = string(inputBytes)
|
||||||
}
|
}
|
||||||
|
|
||||||
concern := &mgo.Safe{}
|
concern := &writeConcern{}
|
||||||
err = json.Unmarshal([]byte(input), concern)
|
err = json.Unmarshal([]byte(input), concern)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, errwrap.Wrapf("error mashalling write_concern: {{err}}", err)
|
return nil, errwrap.Wrapf("error unmarshalling write_concern: {{err}}", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Guard against empty, non-nil mgo.Safe object; we don't want to pass that
|
// Translate write concern to mongo options
|
||||||
// into mgo.SetSafe in Connection().
|
var w writeconcern.Option
|
||||||
if (mgo.Safe{} == *concern) {
|
switch {
|
||||||
return nil, fmt.Errorf("provided write_concern values did not map to any mgo.Safe fields")
|
case concern.W != 0:
|
||||||
|
w = writeconcern.W(concern.W)
|
||||||
|
case concern.WMode != "":
|
||||||
|
w = writeconcern.WTagSet(concern.WMode)
|
||||||
|
default:
|
||||||
|
w = writeconcern.WMajority()
|
||||||
|
}
|
||||||
|
|
||||||
|
var j writeconcern.Option
|
||||||
|
switch {
|
||||||
|
case concern.FSync:
|
||||||
|
j = writeconcern.J(concern.FSync)
|
||||||
|
case concern.J:
|
||||||
|
j = writeconcern.J(concern.J)
|
||||||
|
default:
|
||||||
|
j = writeconcern.J(false)
|
||||||
|
}
|
||||||
|
|
||||||
|
writeConcern := writeconcern.New(
|
||||||
|
w,
|
||||||
|
j,
|
||||||
|
writeconcern.WTimeout(time.Duration(concern.WTimeout)*time.Millisecond))
|
||||||
|
|
||||||
|
c.clientOptions = &options.ClientOptions{
|
||||||
|
WriteConcern: writeConcern,
|
||||||
}
|
}
|
||||||
c.safe = concern
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Set initialized to true at this point since all fields are set,
|
// Set initialized to true at this point since all fields are set,
|
||||||
@@ -96,7 +125,7 @@ func (c *mongoDBConnectionProducer) Init(ctx context.Context, conf map[string]in
|
|||||||
return nil, errwrap.Wrapf("error verifying connection: {{err}}", err)
|
return nil, errwrap.Wrapf("error verifying connection: {{err}}", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := c.session.Ping(); err != nil {
|
if err := c.client.Ping(ctx, readpref.Primary()); err != nil {
|
||||||
return nil, errwrap.Wrapf("error verifying connection: {{err}}", err)
|
return nil, errwrap.Wrapf("error verifying connection: {{err}}", err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -106,36 +135,33 @@ func (c *mongoDBConnectionProducer) Init(ctx context.Context, conf map[string]in
|
|||||||
|
|
||||||
// Connection creates or returns an existing a database connection. If the session fails
|
// Connection creates or returns an existing a database connection. If the session fails
|
||||||
// on a ping check, the session will be closed and then re-created.
|
// on a ping check, the session will be closed and then re-created.
|
||||||
func (c *mongoDBConnectionProducer) Connection(_ context.Context) (interface{}, error) {
|
// This method does not lock the mutex and it is intended that this is the callers
|
||||||
|
// responsibility.
|
||||||
|
func (c *mongoDBConnectionProducer) Connection(ctx context.Context) (interface{}, error) {
|
||||||
if !c.Initialized {
|
if !c.Initialized {
|
||||||
return nil, connutil.ErrNotInitialized
|
return nil, connutil.ErrNotInitialized
|
||||||
}
|
}
|
||||||
|
|
||||||
if c.session != nil {
|
if c.client != nil {
|
||||||
if err := c.session.Ping(); err == nil {
|
if err := c.client.Ping(ctx, readpref.Primary()); err == nil {
|
||||||
return c.session, nil
|
return c.client, nil
|
||||||
}
|
}
|
||||||
c.session.Close()
|
// Ignore error on purpose since we want to re-create a session
|
||||||
|
_ = c.client.Disconnect(ctx)
|
||||||
}
|
}
|
||||||
|
|
||||||
dialInfo, err := parseMongoURL(c.ConnectionURL)
|
if c.clientOptions == nil {
|
||||||
|
c.clientOptions = options.Client()
|
||||||
|
}
|
||||||
|
c.clientOptions.SetSocketTimeout(1 * time.Minute)
|
||||||
|
c.clientOptions.SetConnectTimeout(1 * time.Minute)
|
||||||
|
|
||||||
|
var err error
|
||||||
|
c.client, err = mongo.Connect(ctx, c.clientOptions.ApplyURI(c.ConnectionURL))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
return c.client, nil
|
||||||
c.session, err = mgo.DialWithInfo(dialInfo)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
if c.safe != nil {
|
|
||||||
c.session.SetSafe(c.safe)
|
|
||||||
}
|
|
||||||
|
|
||||||
c.session.SetSyncTimeout(1 * time.Minute)
|
|
||||||
c.session.SetSocketTimeout(1 * time.Minute)
|
|
||||||
|
|
||||||
return c.session, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Close terminates the database connection.
|
// Close terminates the database connection.
|
||||||
@@ -143,83 +169,19 @@ func (c *mongoDBConnectionProducer) Close() error {
|
|||||||
c.Lock()
|
c.Lock()
|
||||||
defer c.Unlock()
|
defer c.Unlock()
|
||||||
|
|
||||||
if c.session != nil {
|
if c.client != nil {
|
||||||
c.session.Close()
|
ctx, cancel := context.WithTimeout(context.Background(), 1 * time.Minute)
|
||||||
|
defer cancel()
|
||||||
|
if err := c.client.Disconnect(ctx); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
c.session = nil
|
c.client = nil
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func parseMongoURL(rawURL string) (*mgo.DialInfo, error) {
|
|
||||||
url, err := url.Parse(rawURL)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
info := mgo.DialInfo{
|
|
||||||
Addrs: strings.Split(url.Host, ","),
|
|
||||||
Database: strings.TrimPrefix(url.Path, "/"),
|
|
||||||
Timeout: 10 * time.Second,
|
|
||||||
}
|
|
||||||
|
|
||||||
if url.User != nil {
|
|
||||||
info.Username = url.User.Username()
|
|
||||||
info.Password, _ = url.User.Password()
|
|
||||||
}
|
|
||||||
|
|
||||||
query := url.Query()
|
|
||||||
for key, values := range query {
|
|
||||||
var value string
|
|
||||||
if len(values) > 0 {
|
|
||||||
value = values[0]
|
|
||||||
}
|
|
||||||
|
|
||||||
switch key {
|
|
||||||
case "authSource":
|
|
||||||
info.Source = value
|
|
||||||
case "authMechanism":
|
|
||||||
info.Mechanism = value
|
|
||||||
case "gssapiServiceName":
|
|
||||||
info.Service = value
|
|
||||||
case "replicaSet":
|
|
||||||
info.ReplicaSetName = value
|
|
||||||
case "maxPoolSize":
|
|
||||||
poolLimit, err := strconv.Atoi(value)
|
|
||||||
if err != nil {
|
|
||||||
return nil, errors.New("bad value for maxPoolSize: " + value)
|
|
||||||
}
|
|
||||||
info.PoolLimit = poolLimit
|
|
||||||
case "ssl":
|
|
||||||
// Unfortunately, mgo doesn't support the ssl parameter in its MongoDB URI parsing logic, so we have to handle that
|
|
||||||
// ourselves. See https://github.com/go-mgo/mgo/issues/84
|
|
||||||
ssl, err := strconv.ParseBool(value)
|
|
||||||
if err != nil {
|
|
||||||
return nil, errors.New("bad value for ssl: " + value)
|
|
||||||
}
|
|
||||||
if ssl {
|
|
||||||
info.DialServer = func(addr *mgo.ServerAddr) (net.Conn, error) {
|
|
||||||
return tls.Dial("tcp", addr.String(), &tls.Config{})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
case "connect":
|
|
||||||
if value == "direct" {
|
|
||||||
info.Direct = true
|
|
||||||
break
|
|
||||||
}
|
|
||||||
if value == "replicaSet" {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
fallthrough
|
|
||||||
default:
|
|
||||||
return nil, errors.New("unsupported connection URL option: " + key + "=" + value)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return &info, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c *mongoDBConnectionProducer) secretValues() map[string]interface{} {
|
func (c *mongoDBConnectionProducer) secretValues() map[string]interface{} {
|
||||||
return map[string]interface{}{
|
return map[string]interface{}{
|
||||||
c.Password: "[password]",
|
c.Password: "[password]",
|
||||||
|
|||||||
@@ -2,21 +2,18 @@ package mongodb
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
"io"
|
"fmt"
|
||||||
"strings"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"encoding/json"
|
|
||||||
|
|
||||||
"fmt"
|
|
||||||
|
|
||||||
"github.com/hashicorp/errwrap"
|
|
||||||
"github.com/hashicorp/vault/api"
|
"github.com/hashicorp/vault/api"
|
||||||
"github.com/hashicorp/vault/sdk/database/dbplugin"
|
"github.com/hashicorp/vault/sdk/database/dbplugin"
|
||||||
"github.com/hashicorp/vault/sdk/database/helper/credsutil"
|
"github.com/hashicorp/vault/sdk/database/helper/credsutil"
|
||||||
"github.com/hashicorp/vault/sdk/database/helper/dbutil"
|
"github.com/hashicorp/vault/sdk/database/helper/dbutil"
|
||||||
mgo "gopkg.in/mgo.v2"
|
"go.mongodb.org/mongo-driver/mongo"
|
||||||
|
"go.mongodb.org/mongo-driver/mongo/writeconcern"
|
||||||
|
"go.mongodb.org/mongo-driver/x/mongo/driver/connstring"
|
||||||
)
|
)
|
||||||
|
|
||||||
const mongoDBTypeName = "mongodb"
|
const mongoDBTypeName = "mongodb"
|
||||||
@@ -70,13 +67,13 @@ func (m *MongoDB) Type() (string, error) {
|
|||||||
return mongoDBTypeName, nil
|
return mongoDBTypeName, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *MongoDB) getConnection(ctx context.Context) (*mgo.Session, error) {
|
func (m *MongoDB) getConnection(ctx context.Context) (*mongo.Client, error) {
|
||||||
session, err := m.Connection(ctx)
|
client, err := m.Connection(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
return session.(*mgo.Session), nil
|
return client.(*mongo.Client), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// CreateUser generates the username/password on the underlying secret backend as instructed by
|
// CreateUser generates the username/password on the underlying secret backend as instructed by
|
||||||
@@ -98,7 +95,7 @@ func (m *MongoDB) CreateUser(ctx context.Context, statements dbplugin.Statements
|
|||||||
return "", "", dbutil.ErrEmptyCreationStatement
|
return "", "", dbutil.ErrEmptyCreationStatement
|
||||||
}
|
}
|
||||||
|
|
||||||
session, err := m.getConnection(ctx)
|
client, err := m.getConnection(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", "", err
|
return "", "", err
|
||||||
}
|
}
|
||||||
@@ -135,20 +132,7 @@ func (m *MongoDB) CreateUser(ctx context.Context, statements dbplugin.Statements
|
|||||||
Roles: mongoCS.Roles.toStandardRolesArray(),
|
Roles: mongoCS.Roles.toStandardRolesArray(),
|
||||||
}
|
}
|
||||||
|
|
||||||
err = session.DB(mongoCS.DB).Run(createUserCmd, nil)
|
if err := runCommandWithRetry(ctx, client, mongoCS.DB, createUserCmd); err != nil {
|
||||||
switch {
|
|
||||||
case err == nil:
|
|
||||||
case err == io.EOF, strings.Contains(err.Error(), "EOF"):
|
|
||||||
// Call getConnection to reset and retry query if we get an EOF error on first attempt.
|
|
||||||
session, err := m.getConnection(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return "", "", err
|
|
||||||
}
|
|
||||||
err = session.DB(mongoCS.DB).Run(createUserCmd, nil)
|
|
||||||
if err != nil {
|
|
||||||
return "", "", err
|
|
||||||
}
|
|
||||||
default:
|
|
||||||
return "", "", err
|
return "", "", err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -166,7 +150,7 @@ func (m *MongoDB) SetCredentials(ctx context.Context, statements dbplugin.Statem
|
|||||||
m.Lock()
|
m.Lock()
|
||||||
defer m.Unlock()
|
defer m.Unlock()
|
||||||
|
|
||||||
session, err := m.getConnection(ctx)
|
client, err := m.getConnection(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", "", err
|
return "", "", err
|
||||||
}
|
}
|
||||||
@@ -174,31 +158,16 @@ func (m *MongoDB) SetCredentials(ctx context.Context, statements dbplugin.Statem
|
|||||||
username = staticUser.Username
|
username = staticUser.Username
|
||||||
password = staticUser.Password
|
password = staticUser.Password
|
||||||
|
|
||||||
dialInfo, err := parseMongoURL(m.ConnectionURL)
|
changeUserCmd := &updateUserCommand{
|
||||||
if err != nil {
|
|
||||||
return "", "", err
|
|
||||||
}
|
|
||||||
|
|
||||||
mongoUser := mgo.User{
|
|
||||||
Username: username,
|
Username: username,
|
||||||
Password: password,
|
Password: password,
|
||||||
}
|
}
|
||||||
|
|
||||||
err = session.DB(dialInfo.Database).UpsertUser(&mongoUser)
|
cs, err := connstring.Parse(m.ConnectionURL)
|
||||||
|
|
||||||
switch {
|
|
||||||
case err == nil:
|
|
||||||
case err == io.EOF, strings.Contains(err.Error(), "EOF"):
|
|
||||||
// Call getConnection to reset and retry query if we get an EOF error on first attempt.
|
|
||||||
session, err := m.getConnection(ctx)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", "", err
|
return "", "", err
|
||||||
}
|
}
|
||||||
err = session.DB(dialInfo.Database).UpsertUser(&mongoUser)
|
if err := runCommandWithRetry(ctx, client, cs.Database, changeUserCmd); err != nil {
|
||||||
if err != nil {
|
|
||||||
return "", "", err
|
|
||||||
}
|
|
||||||
default:
|
|
||||||
return "", "", err
|
return "", "", err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -219,7 +188,7 @@ func (m *MongoDB) RevokeUser(ctx context.Context, statements dbplugin.Statements
|
|||||||
|
|
||||||
statements = dbutil.StatementCompatibilityHelper(statements)
|
statements = dbutil.StatementCompatibilityHelper(statements)
|
||||||
|
|
||||||
session, err := m.getConnection(ctx)
|
client, err := m.getConnection(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@@ -248,29 +217,35 @@ func (m *MongoDB) RevokeUser(ctx context.Context, statements dbplugin.Statements
|
|||||||
db = "admin"
|
db = "admin"
|
||||||
}
|
}
|
||||||
|
|
||||||
err = session.DB(db).RemoveUser(username)
|
dropUserCmd := &dropUserCommand{
|
||||||
switch {
|
Username: username,
|
||||||
case err == nil, err == mgo.ErrNotFound:
|
WriteConcern: writeconcern.New(writeconcern.WMajority()),
|
||||||
case err == io.EOF, strings.Contains(err.Error(), "EOF"):
|
|
||||||
if err := m.Close(); err != nil {
|
|
||||||
return errwrap.Wrapf("error closing EOF'd mongo connection: {{err}}", err)
|
|
||||||
}
|
|
||||||
session, err := m.getConnection(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
err = session.DB(db).RemoveUser(username)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
default:
|
|
||||||
return err
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil
|
return runCommandWithRetry(ctx, client, db, dropUserCmd)
|
||||||
}
|
}
|
||||||
|
|
||||||
// RotateRootCredentials is not currently supported on MongoDB
|
// RotateRootCredentials is not currently supported on MongoDB
|
||||||
func (m *MongoDB) RotateRootCredentials(ctx context.Context, statements []string) (map[string]interface{}, error) {
|
func (m *MongoDB) RotateRootCredentials(ctx context.Context, statements []string) (map[string]interface{}, error) {
|
||||||
return nil, errors.New("root credential rotation is not currently implemented in this database secrets engine")
|
return nil, errors.New("root credential rotation is not currently implemented in this database secrets engine")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// runCommandWithRetry runs a command with retry.
|
||||||
|
func runCommandWithRetry(ctx context.Context, client *mongo.Client, db string, cmd interface{}) error {
|
||||||
|
timeout := time.Now().Add(1 * time.Minute)
|
||||||
|
backoffTime := 3
|
||||||
|
for {
|
||||||
|
// Run command
|
||||||
|
result := client.Database(db).RunCommand(ctx, cmd, nil)
|
||||||
|
if result.Err() == nil {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
if time.Now().After(timeout) {
|
||||||
|
return result.Err()
|
||||||
|
}
|
||||||
|
time.Sleep(time.Duration(backoffTime) * time.Second)
|
||||||
|
backoffTime += backoffTime
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|||||||
@@ -9,7 +9,9 @@ import (
|
|||||||
|
|
||||||
"github.com/hashicorp/vault/helper/testhelpers/mongodb"
|
"github.com/hashicorp/vault/helper/testhelpers/mongodb"
|
||||||
"github.com/hashicorp/vault/sdk/database/dbplugin"
|
"github.com/hashicorp/vault/sdk/database/dbplugin"
|
||||||
"gopkg.in/mgo.v2"
|
"go.mongodb.org/mongo-driver/mongo"
|
||||||
|
"go.mongodb.org/mongo-driver/mongo/options"
|
||||||
|
"go.mongodb.org/mongo-driver/mongo/readpref"
|
||||||
)
|
)
|
||||||
|
|
||||||
const testMongoDBRole = `{ "db": "admin", "roles": [ { "role": "readWrite" } ] }`
|
const testMongoDBRole = `{ "db": "admin", "roles": [ { "role": "readWrite" } ] }`
|
||||||
@@ -152,18 +154,13 @@ func TestMongoDB_RevokeUser(t *testing.T) {
|
|||||||
|
|
||||||
func testCredsExist(t testing.TB, connURL, username, password string) error {
|
func testCredsExist(t testing.TB, connURL, username, password string) error {
|
||||||
connURL = strings.Replace(connURL, "mongodb://", fmt.Sprintf("mongodb://%s:%s@", username, password), 1)
|
connURL = strings.Replace(connURL, "mongodb://", fmt.Sprintf("mongodb://%s:%s@", username, password), 1)
|
||||||
dialInfo, err := parseMongoURL(connURL)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
session, err := mgo.DialWithInfo(dialInfo)
|
ctx, _ := context.WithTimeout(context.Background(), 10*time.Second)
|
||||||
|
client, err := mongo.Connect(ctx, options.Client().ApplyURI(connURL))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
session.SetSyncTimeout(1 * time.Minute)
|
return client.Ping(ctx, readpref.Primary())
|
||||||
session.SetSocketTimeout(1 * time.Minute)
|
|
||||||
return session.Ping()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestMongoDB_SetCredentials(t *testing.T) {
|
func TestMongoDB_SetCredentials(t *testing.T) {
|
||||||
@@ -186,7 +183,7 @@ func TestMongoDB_SetCredentials(t *testing.T) {
|
|||||||
// create the database user in advance, and test the connection
|
// create the database user in advance, and test the connection
|
||||||
dbUser := "testmongouser"
|
dbUser := "testmongouser"
|
||||||
startingPassword := "password"
|
startingPassword := "password"
|
||||||
testCreateDBUser(t, connURL, dbUser, startingPassword)
|
testCreateDBUser(t, connURL, "test", dbUser, startingPassword)
|
||||||
if err := testCredsExist(t, connURL, dbUser, startingPassword); err != nil {
|
if err := testCredsExist(t, connURL, dbUser, startingPassword); err != nil {
|
||||||
t.Fatalf("Could not connect with new credentials: %s", err)
|
t.Fatalf("Could not connect with new credentials: %s", err)
|
||||||
}
|
}
|
||||||
@@ -219,24 +216,20 @@ func TestMongoDB_SetCredentials(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func testCreateDBUser(t testing.TB, connURL, username, password string) {
|
func testCreateDBUser(t testing.TB, connURL, db, username, password string) {
|
||||||
dialInfo, err := parseMongoURL(connURL)
|
ctx, _ := context.WithTimeout(context.Background(), 10*time.Second)
|
||||||
|
client, err := mongo.Connect(ctx, options.Client().ApplyURI(connURL))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
session, err := mgo.DialWithInfo(dialInfo)
|
createUserCmd := &createUserCommand{
|
||||||
if err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
session.SetSyncTimeout(1 * time.Minute)
|
|
||||||
session.SetSocketTimeout(1 * time.Minute)
|
|
||||||
mUser := mgo.User{
|
|
||||||
Username: username,
|
Username: username,
|
||||||
Password: password,
|
Password: password,
|
||||||
|
Roles: []interface{}{},
|
||||||
}
|
}
|
||||||
|
result := client.Database(db).RunCommand(ctx, createUserCmd, nil)
|
||||||
if err := session.DB(dialInfo.Database).UpsertUser(&mUser); err != nil {
|
if result.Err() != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(result.Err())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,11 +1,23 @@
|
|||||||
package mongodb
|
package mongodb
|
||||||
|
|
||||||
|
import "go.mongodb.org/mongo-driver/mongo/writeconcern"
|
||||||
|
|
||||||
type createUserCommand struct {
|
type createUserCommand struct {
|
||||||
Username string `bson:"createUser"`
|
Username string `bson:"createUser"`
|
||||||
Password string `bson:"pwd"`
|
Password string `bson:"pwd"`
|
||||||
Roles []interface{} `bson:"roles"`
|
Roles []interface{} `bson:"roles"`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type updateUserCommand struct {
|
||||||
|
Username string `bson:"updateUser"`
|
||||||
|
Password string `bson:"pwd"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type dropUserCommand struct {
|
||||||
|
Username string `bson:"dropUser"`
|
||||||
|
WriteConcern *writeconcern.WriteConcern `bson:"writeConcern"`
|
||||||
|
}
|
||||||
|
|
||||||
type mongodbRole struct {
|
type mongodbRole struct {
|
||||||
Role string `json:"role" bson:"role"`
|
Role string `json:"role" bson:"role"`
|
||||||
DB string `json:"db" bson:"db"`
|
DB string `json:"db" bson:"db"`
|
||||||
|
|||||||
27
vendor/github.com/DataDog/zstd/LICENSE
generated
vendored
Normal file
27
vendor/github.com/DataDog/zstd/LICENSE
generated
vendored
Normal file
@@ -0,0 +1,27 @@
|
|||||||
|
Simplified BSD License
|
||||||
|
|
||||||
|
Copyright (c) 2016, Datadog <info@datadoghq.com>
|
||||||
|
All rights reserved.
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright notice,
|
||||||
|
this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above copyright notice,
|
||||||
|
this list of conditions and the following disclaimer in the documentation
|
||||||
|
and/or other materials provided with the distribution.
|
||||||
|
* Neither the name of the copyright holder nor the names of its contributors
|
||||||
|
may be used to endorse or promote products derived from this software
|
||||||
|
without specific prior written permission.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||||
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||||
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||||
|
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE
|
||||||
|
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
||||||
|
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
||||||
|
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
||||||
|
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
||||||
|
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
124
vendor/github.com/DataDog/zstd/README.md
generated
vendored
Normal file
124
vendor/github.com/DataDog/zstd/README.md
generated
vendored
Normal file
@@ -0,0 +1,124 @@
|
|||||||
|
# Zstd Go Wrapper
|
||||||
|
|
||||||
|
[](https://circleci.com/gh/DataDog/zstd/tree/1.x)
|
||||||
|
[](https://godoc.org/github.com/DataDog/zstd)
|
||||||
|
|
||||||
|
|
||||||
|
[C Zstd Homepage](https://github.com/facebook/zstd)
|
||||||
|
|
||||||
|
The current headers and C files are from *v1.4.4* (Commit
|
||||||
|
[10f0e699](https://github.com/facebook/zstd/releases/tag/v1.4.4)).
|
||||||
|
|
||||||
|
## Usage
|
||||||
|
|
||||||
|
There are two main APIs:
|
||||||
|
|
||||||
|
* simple Compress/Decompress
|
||||||
|
* streaming API (io.Reader/io.Writer)
|
||||||
|
|
||||||
|
The compress/decompress APIs mirror that of lz4, while the streaming API was
|
||||||
|
designed to be a drop-in replacement for zlib.
|
||||||
|
|
||||||
|
### Simple `Compress/Decompress`
|
||||||
|
|
||||||
|
|
||||||
|
```go
|
||||||
|
// Compress compresses the byte array given in src and writes it to dst.
|
||||||
|
// If you already have a buffer allocated, you can pass it to prevent allocation
|
||||||
|
// If not, you can pass nil as dst.
|
||||||
|
// If the buffer is too small, it will be reallocated, resized, and returned bu the function
|
||||||
|
// If dst is nil, this will allocate the worst case size (CompressBound(src))
|
||||||
|
Compress(dst, src []byte) ([]byte, error)
|
||||||
|
```
|
||||||
|
|
||||||
|
```go
|
||||||
|
// CompressLevel is the same as Compress but you can pass another compression level
|
||||||
|
CompressLevel(dst, src []byte, level int) ([]byte, error)
|
||||||
|
```
|
||||||
|
|
||||||
|
```go
|
||||||
|
// Decompress will decompress your payload into dst.
|
||||||
|
// If you already have a buffer allocated, you can pass it to prevent allocation
|
||||||
|
// If not, you can pass nil as dst (allocates a 4*src size as default).
|
||||||
|
// If the buffer is too small, it will retry 3 times by doubling the dst size
|
||||||
|
// After max retries, it will switch to the slower stream API to be sure to be able
|
||||||
|
// to decompress. Currently switches if compression ratio > 4*2**3=32.
|
||||||
|
Decompress(dst, src []byte) ([]byte, error)
|
||||||
|
```
|
||||||
|
|
||||||
|
### Stream API
|
||||||
|
|
||||||
|
```go
|
||||||
|
// NewWriter creates a new object that can optionally be initialized with
|
||||||
|
// a precomputed dictionary. If dict is nil, compress without a dictionary.
|
||||||
|
// The dictionary array should not be changed during the use of this object.
|
||||||
|
// You MUST CALL Close() to write the last bytes of a zstd stream and free C objects.
|
||||||
|
NewWriter(w io.Writer) *Writer
|
||||||
|
NewWriterLevel(w io.Writer, level int) *Writer
|
||||||
|
NewWriterLevelDict(w io.Writer, level int, dict []byte) *Writer
|
||||||
|
|
||||||
|
// Write compresses the input data and write it to the underlying writer
|
||||||
|
(w *Writer) Write(p []byte) (int, error)
|
||||||
|
|
||||||
|
// Close flushes the buffer and frees C zstd objects
|
||||||
|
(w *Writer) Close() error
|
||||||
|
```
|
||||||
|
|
||||||
|
```go
|
||||||
|
// NewReader returns a new io.ReadCloser that will decompress data from the
|
||||||
|
// underlying reader. If a dictionary is provided to NewReaderDict, it must
|
||||||
|
// not be modified until Close is called. It is the caller's responsibility
|
||||||
|
// to call Close, which frees up C objects.
|
||||||
|
NewReader(r io.Reader) io.ReadCloser
|
||||||
|
NewReaderDict(r io.Reader, dict []byte) io.ReadCloser
|
||||||
|
```
|
||||||
|
|
||||||
|
### Benchmarks (benchmarked with v0.5.0)
|
||||||
|
|
||||||
|
The author of Zstd also wrote lz4. Zstd is intended to occupy a speed/ratio
|
||||||
|
level similar to what zlib currently provides. In our tests, the can always
|
||||||
|
be made to be better than zlib by chosing an appropriate level while still
|
||||||
|
keeping compression and decompression time faster than zlib.
|
||||||
|
|
||||||
|
You can run the benchmarks against your own payloads by using the Go benchmarks tool.
|
||||||
|
Just export your payload filepath as the `PAYLOAD` environment variable and run the benchmarks:
|
||||||
|
|
||||||
|
```go
|
||||||
|
go test -bench .
|
||||||
|
```
|
||||||
|
|
||||||
|
Compression of a 7Mb pdf zstd (this wrapper) vs [czlib](https://github.com/DataDog/czlib):
|
||||||
|
```
|
||||||
|
BenchmarkCompression 5 221056624 ns/op 67.34 MB/s
|
||||||
|
BenchmarkDecompression 100 18370416 ns/op 810.32 MB/s
|
||||||
|
|
||||||
|
BenchmarkFzlibCompress 2 610156603 ns/op 24.40 MB/s
|
||||||
|
BenchmarkFzlibDecompress 20 81195246 ns/op 183.33 MB/s
|
||||||
|
```
|
||||||
|
|
||||||
|
Ratio is also better by a margin of ~20%.
|
||||||
|
Compression speed is always better than zlib on all the payloads we tested;
|
||||||
|
However, [czlib](https://github.com/DataDog/czlib) has optimisations that make it
|
||||||
|
faster at decompressiong small payloads:
|
||||||
|
|
||||||
|
```
|
||||||
|
Testing with size: 11... czlib: 8.97 MB/s, zstd: 3.26 MB/s
|
||||||
|
Testing with size: 27... czlib: 23.3 MB/s, zstd: 8.22 MB/s
|
||||||
|
Testing with size: 62... czlib: 31.6 MB/s, zstd: 19.49 MB/s
|
||||||
|
Testing with size: 141... czlib: 74.54 MB/s, zstd: 42.55 MB/s
|
||||||
|
Testing with size: 323... czlib: 155.14 MB/s, zstd: 99.39 MB/s
|
||||||
|
Testing with size: 739... czlib: 235.9 MB/s, zstd: 216.45 MB/s
|
||||||
|
Testing with size: 1689... czlib: 116.45 MB/s, zstd: 345.64 MB/s
|
||||||
|
Testing with size: 3858... czlib: 176.39 MB/s, zstd: 617.56 MB/s
|
||||||
|
Testing with size: 8811... czlib: 254.11 MB/s, zstd: 824.34 MB/s
|
||||||
|
Testing with size: 20121... czlib: 197.43 MB/s, zstd: 1339.11 MB/s
|
||||||
|
Testing with size: 45951... czlib: 201.62 MB/s, zstd: 1951.57 MB/s
|
||||||
|
```
|
||||||
|
|
||||||
|
zstd starts to shine with payloads > 1KB
|
||||||
|
|
||||||
|
### Stability - Current state: STABLE
|
||||||
|
|
||||||
|
The C library seems to be pretty stable and according to the author has been tested and fuzzed.
|
||||||
|
|
||||||
|
For the Go wrapper, the test cover most usual cases and we have succesfully tested it on all staging and prod data.
|
||||||
30
vendor/github.com/DataDog/zstd/ZSTD_LICENSE
generated
vendored
Normal file
30
vendor/github.com/DataDog/zstd/ZSTD_LICENSE
generated
vendored
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
BSD License
|
||||||
|
|
||||||
|
For Zstandard software
|
||||||
|
|
||||||
|
Copyright (c) 2016-present, Facebook, Inc. All rights reserved.
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without modification,
|
||||||
|
are permitted provided that the following conditions are met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright notice, this
|
||||||
|
list of conditions and the following disclaimer.
|
||||||
|
|
||||||
|
* Redistributions in binary form must reproduce the above copyright notice,
|
||||||
|
this list of conditions and the following disclaimer in the documentation
|
||||||
|
and/or other materials provided with the distribution.
|
||||||
|
|
||||||
|
* Neither the name Facebook nor the names of its contributors may be used to
|
||||||
|
endorse or promote products derived from this software without specific
|
||||||
|
prior written permission.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
|
||||||
|
ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
||||||
|
WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||||
|
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR
|
||||||
|
ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
|
||||||
|
(INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
|
||||||
|
LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON
|
||||||
|
ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
|
||||||
|
SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
460
vendor/github.com/DataDog/zstd/bitstream.h
generated
vendored
Normal file
460
vendor/github.com/DataDog/zstd/bitstream.h
generated
vendored
Normal file
@@ -0,0 +1,460 @@
|
|||||||
|
/* ******************************************************************
|
||||||
|
bitstream
|
||||||
|
Part of FSE library
|
||||||
|
Copyright (C) 2013-present, Yann Collet.
|
||||||
|
|
||||||
|
BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
You can contact the author at :
|
||||||
|
- Source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
||||||
|
****************************************************************** */
|
||||||
|
#ifndef BITSTREAM_H_MODULE
|
||||||
|
#define BITSTREAM_H_MODULE
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*
|
||||||
|
* This API consists of small unitary functions, which must be inlined for best performance.
|
||||||
|
* Since link-time-optimization is not available for all compilers,
|
||||||
|
* these functions are defined into a .h to be included.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/*-****************************************
|
||||||
|
* Dependencies
|
||||||
|
******************************************/
|
||||||
|
#include "mem.h" /* unaligned access routines */
|
||||||
|
#include "debug.h" /* assert(), DEBUGLOG(), RAWLOG() */
|
||||||
|
#include "error_private.h" /* error codes and messages */
|
||||||
|
|
||||||
|
|
||||||
|
/*=========================================
|
||||||
|
* Target specific
|
||||||
|
=========================================*/
|
||||||
|
#if defined(__BMI__) && defined(__GNUC__)
|
||||||
|
# include <immintrin.h> /* support for bextr (experimental) */
|
||||||
|
#elif defined(__ICCARM__)
|
||||||
|
# include <intrinsics.h>
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#define STREAM_ACCUMULATOR_MIN_32 25
|
||||||
|
#define STREAM_ACCUMULATOR_MIN_64 57
|
||||||
|
#define STREAM_ACCUMULATOR_MIN ((U32)(MEM_32bits() ? STREAM_ACCUMULATOR_MIN_32 : STREAM_ACCUMULATOR_MIN_64))
|
||||||
|
|
||||||
|
|
||||||
|
/*-******************************************
|
||||||
|
* bitStream encoding API (write forward)
|
||||||
|
********************************************/
|
||||||
|
/* bitStream can mix input from multiple sources.
|
||||||
|
* A critical property of these streams is that they encode and decode in **reverse** direction.
|
||||||
|
* So the first bit sequence you add will be the last to be read, like a LIFO stack.
|
||||||
|
*/
|
||||||
|
typedef struct {
|
||||||
|
size_t bitContainer;
|
||||||
|
unsigned bitPos;
|
||||||
|
char* startPtr;
|
||||||
|
char* ptr;
|
||||||
|
char* endPtr;
|
||||||
|
} BIT_CStream_t;
|
||||||
|
|
||||||
|
MEM_STATIC size_t BIT_initCStream(BIT_CStream_t* bitC, void* dstBuffer, size_t dstCapacity);
|
||||||
|
MEM_STATIC void BIT_addBits(BIT_CStream_t* bitC, size_t value, unsigned nbBits);
|
||||||
|
MEM_STATIC void BIT_flushBits(BIT_CStream_t* bitC);
|
||||||
|
MEM_STATIC size_t BIT_closeCStream(BIT_CStream_t* bitC);
|
||||||
|
|
||||||
|
/* Start with initCStream, providing the size of buffer to write into.
|
||||||
|
* bitStream will never write outside of this buffer.
|
||||||
|
* `dstCapacity` must be >= sizeof(bitD->bitContainer), otherwise @return will be an error code.
|
||||||
|
*
|
||||||
|
* bits are first added to a local register.
|
||||||
|
* Local register is size_t, hence 64-bits on 64-bits systems, or 32-bits on 32-bits systems.
|
||||||
|
* Writing data into memory is an explicit operation, performed by the flushBits function.
|
||||||
|
* Hence keep track how many bits are potentially stored into local register to avoid register overflow.
|
||||||
|
* After a flushBits, a maximum of 7 bits might still be stored into local register.
|
||||||
|
*
|
||||||
|
* Avoid storing elements of more than 24 bits if you want compatibility with 32-bits bitstream readers.
|
||||||
|
*
|
||||||
|
* Last operation is to close the bitStream.
|
||||||
|
* The function returns the final size of CStream in bytes.
|
||||||
|
* If data couldn't fit into `dstBuffer`, it will return a 0 ( == not storable)
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/*-********************************************
|
||||||
|
* bitStream decoding API (read backward)
|
||||||
|
**********************************************/
|
||||||
|
typedef struct {
|
||||||
|
size_t bitContainer;
|
||||||
|
unsigned bitsConsumed;
|
||||||
|
const char* ptr;
|
||||||
|
const char* start;
|
||||||
|
const char* limitPtr;
|
||||||
|
} BIT_DStream_t;
|
||||||
|
|
||||||
|
typedef enum { BIT_DStream_unfinished = 0,
|
||||||
|
BIT_DStream_endOfBuffer = 1,
|
||||||
|
BIT_DStream_completed = 2,
|
||||||
|
BIT_DStream_overflow = 3 } BIT_DStream_status; /* result of BIT_reloadDStream() */
|
||||||
|
/* 1,2,4,8 would be better for bitmap combinations, but slows down performance a bit ... :( */
|
||||||
|
|
||||||
|
MEM_STATIC size_t BIT_initDStream(BIT_DStream_t* bitD, const void* srcBuffer, size_t srcSize);
|
||||||
|
MEM_STATIC size_t BIT_readBits(BIT_DStream_t* bitD, unsigned nbBits);
|
||||||
|
MEM_STATIC BIT_DStream_status BIT_reloadDStream(BIT_DStream_t* bitD);
|
||||||
|
MEM_STATIC unsigned BIT_endOfDStream(const BIT_DStream_t* bitD);
|
||||||
|
|
||||||
|
|
||||||
|
/* Start by invoking BIT_initDStream().
|
||||||
|
* A chunk of the bitStream is then stored into a local register.
|
||||||
|
* Local register size is 64-bits on 64-bits systems, 32-bits on 32-bits systems (size_t).
|
||||||
|
* You can then retrieve bitFields stored into the local register, **in reverse order**.
|
||||||
|
* Local register is explicitly reloaded from memory by the BIT_reloadDStream() method.
|
||||||
|
* A reload guarantee a minimum of ((8*sizeof(bitD->bitContainer))-7) bits when its result is BIT_DStream_unfinished.
|
||||||
|
* Otherwise, it can be less than that, so proceed accordingly.
|
||||||
|
* Checking if DStream has reached its end can be performed with BIT_endOfDStream().
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/*-****************************************
|
||||||
|
* unsafe API
|
||||||
|
******************************************/
|
||||||
|
MEM_STATIC void BIT_addBitsFast(BIT_CStream_t* bitC, size_t value, unsigned nbBits);
|
||||||
|
/* faster, but works only if value is "clean", meaning all high bits above nbBits are 0 */
|
||||||
|
|
||||||
|
MEM_STATIC void BIT_flushBitsFast(BIT_CStream_t* bitC);
|
||||||
|
/* unsafe version; does not check buffer overflow */
|
||||||
|
|
||||||
|
MEM_STATIC size_t BIT_readBitsFast(BIT_DStream_t* bitD, unsigned nbBits);
|
||||||
|
/* faster, but works only if nbBits >= 1 */
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/*-**************************************************************
|
||||||
|
* Internal functions
|
||||||
|
****************************************************************/
|
||||||
|
MEM_STATIC unsigned BIT_highbit32 (U32 val)
|
||||||
|
{
|
||||||
|
assert(val != 0);
|
||||||
|
{
|
||||||
|
# if defined(_MSC_VER) /* Visual */
|
||||||
|
unsigned long r=0;
|
||||||
|
_BitScanReverse ( &r, val );
|
||||||
|
return (unsigned) r;
|
||||||
|
# elif defined(__GNUC__) && (__GNUC__ >= 3) /* Use GCC Intrinsic */
|
||||||
|
return __builtin_clz (val) ^ 31;
|
||||||
|
# elif defined(__ICCARM__) /* IAR Intrinsic */
|
||||||
|
return 31 - __CLZ(val);
|
||||||
|
# else /* Software version */
|
||||||
|
static const unsigned DeBruijnClz[32] = { 0, 9, 1, 10, 13, 21, 2, 29,
|
||||||
|
11, 14, 16, 18, 22, 25, 3, 30,
|
||||||
|
8, 12, 20, 28, 15, 17, 24, 7,
|
||||||
|
19, 27, 23, 6, 26, 5, 4, 31 };
|
||||||
|
U32 v = val;
|
||||||
|
v |= v >> 1;
|
||||||
|
v |= v >> 2;
|
||||||
|
v |= v >> 4;
|
||||||
|
v |= v >> 8;
|
||||||
|
v |= v >> 16;
|
||||||
|
return DeBruijnClz[ (U32) (v * 0x07C4ACDDU) >> 27];
|
||||||
|
# endif
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/*===== Local Constants =====*/
|
||||||
|
static const unsigned BIT_mask[] = {
|
||||||
|
0, 1, 3, 7, 0xF, 0x1F,
|
||||||
|
0x3F, 0x7F, 0xFF, 0x1FF, 0x3FF, 0x7FF,
|
||||||
|
0xFFF, 0x1FFF, 0x3FFF, 0x7FFF, 0xFFFF, 0x1FFFF,
|
||||||
|
0x3FFFF, 0x7FFFF, 0xFFFFF, 0x1FFFFF, 0x3FFFFF, 0x7FFFFF,
|
||||||
|
0xFFFFFF, 0x1FFFFFF, 0x3FFFFFF, 0x7FFFFFF, 0xFFFFFFF, 0x1FFFFFFF,
|
||||||
|
0x3FFFFFFF, 0x7FFFFFFF}; /* up to 31 bits */
|
||||||
|
#define BIT_MASK_SIZE (sizeof(BIT_mask) / sizeof(BIT_mask[0]))
|
||||||
|
|
||||||
|
/*-**************************************************************
|
||||||
|
* bitStream encoding
|
||||||
|
****************************************************************/
|
||||||
|
/*! BIT_initCStream() :
|
||||||
|
* `dstCapacity` must be > sizeof(size_t)
|
||||||
|
* @return : 0 if success,
|
||||||
|
* otherwise an error code (can be tested using ERR_isError()) */
|
||||||
|
MEM_STATIC size_t BIT_initCStream(BIT_CStream_t* bitC,
|
||||||
|
void* startPtr, size_t dstCapacity)
|
||||||
|
{
|
||||||
|
bitC->bitContainer = 0;
|
||||||
|
bitC->bitPos = 0;
|
||||||
|
bitC->startPtr = (char*)startPtr;
|
||||||
|
bitC->ptr = bitC->startPtr;
|
||||||
|
bitC->endPtr = bitC->startPtr + dstCapacity - sizeof(bitC->bitContainer);
|
||||||
|
if (dstCapacity <= sizeof(bitC->bitContainer)) return ERROR(dstSize_tooSmall);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! BIT_addBits() :
|
||||||
|
* can add up to 31 bits into `bitC`.
|
||||||
|
* Note : does not check for register overflow ! */
|
||||||
|
MEM_STATIC void BIT_addBits(BIT_CStream_t* bitC,
|
||||||
|
size_t value, unsigned nbBits)
|
||||||
|
{
|
||||||
|
MEM_STATIC_ASSERT(BIT_MASK_SIZE == 32);
|
||||||
|
assert(nbBits < BIT_MASK_SIZE);
|
||||||
|
assert(nbBits + bitC->bitPos < sizeof(bitC->bitContainer) * 8);
|
||||||
|
bitC->bitContainer |= (value & BIT_mask[nbBits]) << bitC->bitPos;
|
||||||
|
bitC->bitPos += nbBits;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! BIT_addBitsFast() :
|
||||||
|
* works only if `value` is _clean_,
|
||||||
|
* meaning all high bits above nbBits are 0 */
|
||||||
|
MEM_STATIC void BIT_addBitsFast(BIT_CStream_t* bitC,
|
||||||
|
size_t value, unsigned nbBits)
|
||||||
|
{
|
||||||
|
assert((value>>nbBits) == 0);
|
||||||
|
assert(nbBits + bitC->bitPos < sizeof(bitC->bitContainer) * 8);
|
||||||
|
bitC->bitContainer |= value << bitC->bitPos;
|
||||||
|
bitC->bitPos += nbBits;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! BIT_flushBitsFast() :
|
||||||
|
* assumption : bitContainer has not overflowed
|
||||||
|
* unsafe version; does not check buffer overflow */
|
||||||
|
MEM_STATIC void BIT_flushBitsFast(BIT_CStream_t* bitC)
|
||||||
|
{
|
||||||
|
size_t const nbBytes = bitC->bitPos >> 3;
|
||||||
|
assert(bitC->bitPos < sizeof(bitC->bitContainer) * 8);
|
||||||
|
assert(bitC->ptr <= bitC->endPtr);
|
||||||
|
MEM_writeLEST(bitC->ptr, bitC->bitContainer);
|
||||||
|
bitC->ptr += nbBytes;
|
||||||
|
bitC->bitPos &= 7;
|
||||||
|
bitC->bitContainer >>= nbBytes*8;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! BIT_flushBits() :
|
||||||
|
* assumption : bitContainer has not overflowed
|
||||||
|
* safe version; check for buffer overflow, and prevents it.
|
||||||
|
* note : does not signal buffer overflow.
|
||||||
|
* overflow will be revealed later on using BIT_closeCStream() */
|
||||||
|
MEM_STATIC void BIT_flushBits(BIT_CStream_t* bitC)
|
||||||
|
{
|
||||||
|
size_t const nbBytes = bitC->bitPos >> 3;
|
||||||
|
assert(bitC->bitPos < sizeof(bitC->bitContainer) * 8);
|
||||||
|
assert(bitC->ptr <= bitC->endPtr);
|
||||||
|
MEM_writeLEST(bitC->ptr, bitC->bitContainer);
|
||||||
|
bitC->ptr += nbBytes;
|
||||||
|
if (bitC->ptr > bitC->endPtr) bitC->ptr = bitC->endPtr;
|
||||||
|
bitC->bitPos &= 7;
|
||||||
|
bitC->bitContainer >>= nbBytes*8;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! BIT_closeCStream() :
|
||||||
|
* @return : size of CStream, in bytes,
|
||||||
|
* or 0 if it could not fit into dstBuffer */
|
||||||
|
MEM_STATIC size_t BIT_closeCStream(BIT_CStream_t* bitC)
|
||||||
|
{
|
||||||
|
BIT_addBitsFast(bitC, 1, 1); /* endMark */
|
||||||
|
BIT_flushBits(bitC);
|
||||||
|
if (bitC->ptr >= bitC->endPtr) return 0; /* overflow detected */
|
||||||
|
return (bitC->ptr - bitC->startPtr) + (bitC->bitPos > 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/*-********************************************************
|
||||||
|
* bitStream decoding
|
||||||
|
**********************************************************/
|
||||||
|
/*! BIT_initDStream() :
|
||||||
|
* Initialize a BIT_DStream_t.
|
||||||
|
* `bitD` : a pointer to an already allocated BIT_DStream_t structure.
|
||||||
|
* `srcSize` must be the *exact* size of the bitStream, in bytes.
|
||||||
|
* @return : size of stream (== srcSize), or an errorCode if a problem is detected
|
||||||
|
*/
|
||||||
|
MEM_STATIC size_t BIT_initDStream(BIT_DStream_t* bitD, const void* srcBuffer, size_t srcSize)
|
||||||
|
{
|
||||||
|
if (srcSize < 1) { memset(bitD, 0, sizeof(*bitD)); return ERROR(srcSize_wrong); }
|
||||||
|
|
||||||
|
bitD->start = (const char*)srcBuffer;
|
||||||
|
bitD->limitPtr = bitD->start + sizeof(bitD->bitContainer);
|
||||||
|
|
||||||
|
if (srcSize >= sizeof(bitD->bitContainer)) { /* normal case */
|
||||||
|
bitD->ptr = (const char*)srcBuffer + srcSize - sizeof(bitD->bitContainer);
|
||||||
|
bitD->bitContainer = MEM_readLEST(bitD->ptr);
|
||||||
|
{ BYTE const lastByte = ((const BYTE*)srcBuffer)[srcSize-1];
|
||||||
|
bitD->bitsConsumed = lastByte ? 8 - BIT_highbit32(lastByte) : 0; /* ensures bitsConsumed is always set */
|
||||||
|
if (lastByte == 0) return ERROR(GENERIC); /* endMark not present */ }
|
||||||
|
} else {
|
||||||
|
bitD->ptr = bitD->start;
|
||||||
|
bitD->bitContainer = *(const BYTE*)(bitD->start);
|
||||||
|
switch(srcSize)
|
||||||
|
{
|
||||||
|
case 7: bitD->bitContainer += (size_t)(((const BYTE*)(srcBuffer))[6]) << (sizeof(bitD->bitContainer)*8 - 16);
|
||||||
|
/* fall-through */
|
||||||
|
|
||||||
|
case 6: bitD->bitContainer += (size_t)(((const BYTE*)(srcBuffer))[5]) << (sizeof(bitD->bitContainer)*8 - 24);
|
||||||
|
/* fall-through */
|
||||||
|
|
||||||
|
case 5: bitD->bitContainer += (size_t)(((const BYTE*)(srcBuffer))[4]) << (sizeof(bitD->bitContainer)*8 - 32);
|
||||||
|
/* fall-through */
|
||||||
|
|
||||||
|
case 4: bitD->bitContainer += (size_t)(((const BYTE*)(srcBuffer))[3]) << 24;
|
||||||
|
/* fall-through */
|
||||||
|
|
||||||
|
case 3: bitD->bitContainer += (size_t)(((const BYTE*)(srcBuffer))[2]) << 16;
|
||||||
|
/* fall-through */
|
||||||
|
|
||||||
|
case 2: bitD->bitContainer += (size_t)(((const BYTE*)(srcBuffer))[1]) << 8;
|
||||||
|
/* fall-through */
|
||||||
|
|
||||||
|
default: break;
|
||||||
|
}
|
||||||
|
{ BYTE const lastByte = ((const BYTE*)srcBuffer)[srcSize-1];
|
||||||
|
bitD->bitsConsumed = lastByte ? 8 - BIT_highbit32(lastByte) : 0;
|
||||||
|
if (lastByte == 0) return ERROR(corruption_detected); /* endMark not present */
|
||||||
|
}
|
||||||
|
bitD->bitsConsumed += (U32)(sizeof(bitD->bitContainer) - srcSize)*8;
|
||||||
|
}
|
||||||
|
|
||||||
|
return srcSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC size_t BIT_getUpperBits(size_t bitContainer, U32 const start)
|
||||||
|
{
|
||||||
|
return bitContainer >> start;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC size_t BIT_getMiddleBits(size_t bitContainer, U32 const start, U32 const nbBits)
|
||||||
|
{
|
||||||
|
U32 const regMask = sizeof(bitContainer)*8 - 1;
|
||||||
|
/* if start > regMask, bitstream is corrupted, and result is undefined */
|
||||||
|
assert(nbBits < BIT_MASK_SIZE);
|
||||||
|
return (bitContainer >> (start & regMask)) & BIT_mask[nbBits];
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC size_t BIT_getLowerBits(size_t bitContainer, U32 const nbBits)
|
||||||
|
{
|
||||||
|
assert(nbBits < BIT_MASK_SIZE);
|
||||||
|
return bitContainer & BIT_mask[nbBits];
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! BIT_lookBits() :
|
||||||
|
* Provides next n bits from local register.
|
||||||
|
* local register is not modified.
|
||||||
|
* On 32-bits, maxNbBits==24.
|
||||||
|
* On 64-bits, maxNbBits==56.
|
||||||
|
* @return : value extracted */
|
||||||
|
MEM_STATIC size_t BIT_lookBits(const BIT_DStream_t* bitD, U32 nbBits)
|
||||||
|
{
|
||||||
|
/* arbitrate between double-shift and shift+mask */
|
||||||
|
#if 1
|
||||||
|
/* if bitD->bitsConsumed + nbBits > sizeof(bitD->bitContainer)*8,
|
||||||
|
* bitstream is likely corrupted, and result is undefined */
|
||||||
|
return BIT_getMiddleBits(bitD->bitContainer, (sizeof(bitD->bitContainer)*8) - bitD->bitsConsumed - nbBits, nbBits);
|
||||||
|
#else
|
||||||
|
/* this code path is slower on my os-x laptop */
|
||||||
|
U32 const regMask = sizeof(bitD->bitContainer)*8 - 1;
|
||||||
|
return ((bitD->bitContainer << (bitD->bitsConsumed & regMask)) >> 1) >> ((regMask-nbBits) & regMask);
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! BIT_lookBitsFast() :
|
||||||
|
* unsafe version; only works if nbBits >= 1 */
|
||||||
|
MEM_STATIC size_t BIT_lookBitsFast(const BIT_DStream_t* bitD, U32 nbBits)
|
||||||
|
{
|
||||||
|
U32 const regMask = sizeof(bitD->bitContainer)*8 - 1;
|
||||||
|
assert(nbBits >= 1);
|
||||||
|
return (bitD->bitContainer << (bitD->bitsConsumed & regMask)) >> (((regMask+1)-nbBits) & regMask);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void BIT_skipBits(BIT_DStream_t* bitD, U32 nbBits)
|
||||||
|
{
|
||||||
|
bitD->bitsConsumed += nbBits;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! BIT_readBits() :
|
||||||
|
* Read (consume) next n bits from local register and update.
|
||||||
|
* Pay attention to not read more than nbBits contained into local register.
|
||||||
|
* @return : extracted value. */
|
||||||
|
MEM_STATIC size_t BIT_readBits(BIT_DStream_t* bitD, unsigned nbBits)
|
||||||
|
{
|
||||||
|
size_t const value = BIT_lookBits(bitD, nbBits);
|
||||||
|
BIT_skipBits(bitD, nbBits);
|
||||||
|
return value;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! BIT_readBitsFast() :
|
||||||
|
* unsafe version; only works only if nbBits >= 1 */
|
||||||
|
MEM_STATIC size_t BIT_readBitsFast(BIT_DStream_t* bitD, unsigned nbBits)
|
||||||
|
{
|
||||||
|
size_t const value = BIT_lookBitsFast(bitD, nbBits);
|
||||||
|
assert(nbBits >= 1);
|
||||||
|
BIT_skipBits(bitD, nbBits);
|
||||||
|
return value;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! BIT_reloadDStream() :
|
||||||
|
* Refill `bitD` from buffer previously set in BIT_initDStream() .
|
||||||
|
* This function is safe, it guarantees it will not read beyond src buffer.
|
||||||
|
* @return : status of `BIT_DStream_t` internal register.
|
||||||
|
* when status == BIT_DStream_unfinished, internal register is filled with at least 25 or 57 bits */
|
||||||
|
MEM_STATIC BIT_DStream_status BIT_reloadDStream(BIT_DStream_t* bitD)
|
||||||
|
{
|
||||||
|
if (bitD->bitsConsumed > (sizeof(bitD->bitContainer)*8)) /* overflow detected, like end of stream */
|
||||||
|
return BIT_DStream_overflow;
|
||||||
|
|
||||||
|
if (bitD->ptr >= bitD->limitPtr) {
|
||||||
|
bitD->ptr -= bitD->bitsConsumed >> 3;
|
||||||
|
bitD->bitsConsumed &= 7;
|
||||||
|
bitD->bitContainer = MEM_readLEST(bitD->ptr);
|
||||||
|
return BIT_DStream_unfinished;
|
||||||
|
}
|
||||||
|
if (bitD->ptr == bitD->start) {
|
||||||
|
if (bitD->bitsConsumed < sizeof(bitD->bitContainer)*8) return BIT_DStream_endOfBuffer;
|
||||||
|
return BIT_DStream_completed;
|
||||||
|
}
|
||||||
|
/* start < ptr < limitPtr */
|
||||||
|
{ U32 nbBytes = bitD->bitsConsumed >> 3;
|
||||||
|
BIT_DStream_status result = BIT_DStream_unfinished;
|
||||||
|
if (bitD->ptr - nbBytes < bitD->start) {
|
||||||
|
nbBytes = (U32)(bitD->ptr - bitD->start); /* ptr > start */
|
||||||
|
result = BIT_DStream_endOfBuffer;
|
||||||
|
}
|
||||||
|
bitD->ptr -= nbBytes;
|
||||||
|
bitD->bitsConsumed -= nbBytes*8;
|
||||||
|
bitD->bitContainer = MEM_readLEST(bitD->ptr); /* reminder : srcSize > sizeof(bitD->bitContainer), otherwise bitD->ptr == bitD->start */
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! BIT_endOfDStream() :
|
||||||
|
* @return : 1 if DStream has _exactly_ reached its end (all bits consumed).
|
||||||
|
*/
|
||||||
|
MEM_STATIC unsigned BIT_endOfDStream(const BIT_DStream_t* DStream)
|
||||||
|
{
|
||||||
|
return ((DStream->ptr == DStream->start) && (DStream->bitsConsumed == sizeof(DStream->bitContainer)*8));
|
||||||
|
}
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* BITSTREAM_H_MODULE */
|
||||||
159
vendor/github.com/DataDog/zstd/compiler.h
generated
vendored
Normal file
159
vendor/github.com/DataDog/zstd/compiler.h
generated
vendored
Normal file
@@ -0,0 +1,159 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_COMPILER_H
|
||||||
|
#define ZSTD_COMPILER_H
|
||||||
|
|
||||||
|
/*-*******************************************************
|
||||||
|
* Compiler specifics
|
||||||
|
*********************************************************/
|
||||||
|
/* force inlining */
|
||||||
|
|
||||||
|
#if !defined(ZSTD_NO_INLINE)
|
||||||
|
#if defined (__GNUC__) || defined(__cplusplus) || defined(__STDC_VERSION__) && __STDC_VERSION__ >= 199901L /* C99 */
|
||||||
|
# define INLINE_KEYWORD inline
|
||||||
|
#else
|
||||||
|
# define INLINE_KEYWORD
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#if defined(__GNUC__) || defined(__ICCARM__)
|
||||||
|
# define FORCE_INLINE_ATTR __attribute__((always_inline))
|
||||||
|
#elif defined(_MSC_VER)
|
||||||
|
# define FORCE_INLINE_ATTR __forceinline
|
||||||
|
#else
|
||||||
|
# define FORCE_INLINE_ATTR
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#else
|
||||||
|
|
||||||
|
#define INLINE_KEYWORD
|
||||||
|
#define FORCE_INLINE_ATTR
|
||||||
|
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/**
|
||||||
|
* FORCE_INLINE_TEMPLATE is used to define C "templates", which take constant
|
||||||
|
* parameters. They must be inlined for the compiler to eliminate the constant
|
||||||
|
* branches.
|
||||||
|
*/
|
||||||
|
#define FORCE_INLINE_TEMPLATE static INLINE_KEYWORD FORCE_INLINE_ATTR
|
||||||
|
/**
|
||||||
|
* HINT_INLINE is used to help the compiler generate better code. It is *not*
|
||||||
|
* used for "templates", so it can be tweaked based on the compilers
|
||||||
|
* performance.
|
||||||
|
*
|
||||||
|
* gcc-4.8 and gcc-4.9 have been shown to benefit from leaving off the
|
||||||
|
* always_inline attribute.
|
||||||
|
*
|
||||||
|
* clang up to 5.0.0 (trunk) benefit tremendously from the always_inline
|
||||||
|
* attribute.
|
||||||
|
*/
|
||||||
|
#if !defined(__clang__) && defined(__GNUC__) && __GNUC__ >= 4 && __GNUC_MINOR__ >= 8 && __GNUC__ < 5
|
||||||
|
# define HINT_INLINE static INLINE_KEYWORD
|
||||||
|
#else
|
||||||
|
# define HINT_INLINE static INLINE_KEYWORD FORCE_INLINE_ATTR
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* UNUSED_ATTR tells the compiler it is okay if the function is unused. */
|
||||||
|
#if defined(__GNUC__)
|
||||||
|
# define UNUSED_ATTR __attribute__((unused))
|
||||||
|
#else
|
||||||
|
# define UNUSED_ATTR
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* force no inlining */
|
||||||
|
#ifdef _MSC_VER
|
||||||
|
# define FORCE_NOINLINE static __declspec(noinline)
|
||||||
|
#else
|
||||||
|
# if defined(__GNUC__) || defined(__ICCARM__)
|
||||||
|
# define FORCE_NOINLINE static __attribute__((__noinline__))
|
||||||
|
# else
|
||||||
|
# define FORCE_NOINLINE static
|
||||||
|
# endif
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* target attribute */
|
||||||
|
#ifndef __has_attribute
|
||||||
|
#define __has_attribute(x) 0 /* Compatibility with non-clang compilers. */
|
||||||
|
#endif
|
||||||
|
#if defined(__GNUC__) || defined(__ICCARM__)
|
||||||
|
# define TARGET_ATTRIBUTE(target) __attribute__((__target__(target)))
|
||||||
|
#else
|
||||||
|
# define TARGET_ATTRIBUTE(target)
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* Enable runtime BMI2 dispatch based on the CPU.
|
||||||
|
* Enabled for clang & gcc >=4.8 on x86 when BMI2 isn't enabled by default.
|
||||||
|
*/
|
||||||
|
#ifndef DYNAMIC_BMI2
|
||||||
|
#if ((defined(__clang__) && __has_attribute(__target__)) \
|
||||||
|
|| (defined(__GNUC__) \
|
||||||
|
&& (__GNUC__ >= 5 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8)))) \
|
||||||
|
&& (defined(__x86_64__) || defined(_M_X86)) \
|
||||||
|
&& !defined(__BMI2__)
|
||||||
|
# define DYNAMIC_BMI2 1
|
||||||
|
#else
|
||||||
|
# define DYNAMIC_BMI2 0
|
||||||
|
#endif
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* prefetch
|
||||||
|
* can be disabled, by declaring NO_PREFETCH build macro */
|
||||||
|
#if defined(NO_PREFETCH)
|
||||||
|
# define PREFETCH_L1(ptr) (void)(ptr) /* disabled */
|
||||||
|
# define PREFETCH_L2(ptr) (void)(ptr) /* disabled */
|
||||||
|
#else
|
||||||
|
# if defined(_MSC_VER) && (defined(_M_X64) || defined(_M_I86)) /* _mm_prefetch() is not defined outside of x86/x64 */
|
||||||
|
# include <mmintrin.h> /* https://msdn.microsoft.com/fr-fr/library/84szxsww(v=vs.90).aspx */
|
||||||
|
# define PREFETCH_L1(ptr) _mm_prefetch((const char*)(ptr), _MM_HINT_T0)
|
||||||
|
# define PREFETCH_L2(ptr) _mm_prefetch((const char*)(ptr), _MM_HINT_T1)
|
||||||
|
# elif defined(__GNUC__) && ( (__GNUC__ >= 4) || ( (__GNUC__ == 3) && (__GNUC_MINOR__ >= 1) ) )
|
||||||
|
# define PREFETCH_L1(ptr) __builtin_prefetch((ptr), 0 /* rw==read */, 3 /* locality */)
|
||||||
|
# define PREFETCH_L2(ptr) __builtin_prefetch((ptr), 0 /* rw==read */, 2 /* locality */)
|
||||||
|
# else
|
||||||
|
# define PREFETCH_L1(ptr) (void)(ptr) /* disabled */
|
||||||
|
# define PREFETCH_L2(ptr) (void)(ptr) /* disabled */
|
||||||
|
# endif
|
||||||
|
#endif /* NO_PREFETCH */
|
||||||
|
|
||||||
|
#define CACHELINE_SIZE 64
|
||||||
|
|
||||||
|
#define PREFETCH_AREA(p, s) { \
|
||||||
|
const char* const _ptr = (const char*)(p); \
|
||||||
|
size_t const _size = (size_t)(s); \
|
||||||
|
size_t _pos; \
|
||||||
|
for (_pos=0; _pos<_size; _pos+=CACHELINE_SIZE) { \
|
||||||
|
PREFETCH_L2(_ptr + _pos); \
|
||||||
|
} \
|
||||||
|
}
|
||||||
|
|
||||||
|
/* vectorization
|
||||||
|
* older GCC (pre gcc-4.3 picked as the cutoff) uses a different syntax */
|
||||||
|
#if !defined(__clang__) && defined(__GNUC__)
|
||||||
|
# if (__GNUC__ == 4 && __GNUC_MINOR__ > 3) || (__GNUC__ >= 5)
|
||||||
|
# define DONT_VECTORIZE __attribute__((optimize("no-tree-vectorize")))
|
||||||
|
# else
|
||||||
|
# define DONT_VECTORIZE _Pragma("GCC optimize(\"no-tree-vectorize\")")
|
||||||
|
# endif
|
||||||
|
#else
|
||||||
|
# define DONT_VECTORIZE
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* disable warnings */
|
||||||
|
#ifdef _MSC_VER /* Visual Studio */
|
||||||
|
# include <intrin.h> /* For Visual 2005 */
|
||||||
|
# pragma warning(disable : 4100) /* disable: C4100: unreferenced formal parameter */
|
||||||
|
# pragma warning(disable : 4127) /* disable: C4127: conditional expression is constant */
|
||||||
|
# pragma warning(disable : 4204) /* disable: C4204: non-constant aggregate initializer */
|
||||||
|
# pragma warning(disable : 4214) /* disable: C4214: non-int bitfields */
|
||||||
|
# pragma warning(disable : 4324) /* disable: C4324: padded structure */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_COMPILER_H */
|
||||||
1236
vendor/github.com/DataDog/zstd/cover.c
generated
vendored
Normal file
1236
vendor/github.com/DataDog/zstd/cover.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
147
vendor/github.com/DataDog/zstd/cover.h
generated
vendored
Normal file
147
vendor/github.com/DataDog/zstd/cover.h
generated
vendored
Normal file
@@ -0,0 +1,147 @@
|
|||||||
|
#include <stdio.h> /* fprintf */
|
||||||
|
#include <stdlib.h> /* malloc, free, qsort */
|
||||||
|
#include <string.h> /* memset */
|
||||||
|
#include <time.h> /* clock */
|
||||||
|
#include "mem.h" /* read */
|
||||||
|
#include "pool.h"
|
||||||
|
#include "threading.h"
|
||||||
|
#include "zstd_internal.h" /* includes zstd.h */
|
||||||
|
#ifndef ZDICT_STATIC_LINKING_ONLY
|
||||||
|
#define ZDICT_STATIC_LINKING_ONLY
|
||||||
|
#endif
|
||||||
|
#include "zdict.h"
|
||||||
|
|
||||||
|
/**
|
||||||
|
* COVER_best_t is used for two purposes:
|
||||||
|
* 1. Synchronizing threads.
|
||||||
|
* 2. Saving the best parameters and dictionary.
|
||||||
|
*
|
||||||
|
* All of the methods except COVER_best_init() are thread safe if zstd is
|
||||||
|
* compiled with multithreaded support.
|
||||||
|
*/
|
||||||
|
typedef struct COVER_best_s {
|
||||||
|
ZSTD_pthread_mutex_t mutex;
|
||||||
|
ZSTD_pthread_cond_t cond;
|
||||||
|
size_t liveJobs;
|
||||||
|
void *dict;
|
||||||
|
size_t dictSize;
|
||||||
|
ZDICT_cover_params_t parameters;
|
||||||
|
size_t compressedSize;
|
||||||
|
} COVER_best_t;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A segment is a range in the source as well as the score of the segment.
|
||||||
|
*/
|
||||||
|
typedef struct {
|
||||||
|
U32 begin;
|
||||||
|
U32 end;
|
||||||
|
U32 score;
|
||||||
|
} COVER_segment_t;
|
||||||
|
|
||||||
|
/**
|
||||||
|
*Number of epochs and size of each epoch.
|
||||||
|
*/
|
||||||
|
typedef struct {
|
||||||
|
U32 num;
|
||||||
|
U32 size;
|
||||||
|
} COVER_epoch_info_t;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Struct used for the dictionary selection function.
|
||||||
|
*/
|
||||||
|
typedef struct COVER_dictSelection {
|
||||||
|
BYTE* dictContent;
|
||||||
|
size_t dictSize;
|
||||||
|
size_t totalCompressedSize;
|
||||||
|
} COVER_dictSelection_t;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Computes the number of epochs and the size of each epoch.
|
||||||
|
* We will make sure that each epoch gets at least 10 * k bytes.
|
||||||
|
*
|
||||||
|
* The COVER algorithms divide the data up into epochs of equal size and
|
||||||
|
* select one segment from each epoch.
|
||||||
|
*
|
||||||
|
* @param maxDictSize The maximum allowed dictionary size.
|
||||||
|
* @param nbDmers The number of dmers we are training on.
|
||||||
|
* @param k The parameter k (segment size).
|
||||||
|
* @param passes The target number of passes over the dmer corpus.
|
||||||
|
* More passes means a better dictionary.
|
||||||
|
*/
|
||||||
|
COVER_epoch_info_t COVER_computeEpochs(U32 maxDictSize, U32 nbDmers,
|
||||||
|
U32 k, U32 passes);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Warns the user when their corpus is too small.
|
||||||
|
*/
|
||||||
|
void COVER_warnOnSmallCorpus(size_t maxDictSize, size_t nbDmers, int displayLevel);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Checks total compressed size of a dictionary
|
||||||
|
*/
|
||||||
|
size_t COVER_checkTotalCompressedSize(const ZDICT_cover_params_t parameters,
|
||||||
|
const size_t *samplesSizes, const BYTE *samples,
|
||||||
|
size_t *offsets,
|
||||||
|
size_t nbTrainSamples, size_t nbSamples,
|
||||||
|
BYTE *const dict, size_t dictBufferCapacity);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns the sum of the sample sizes.
|
||||||
|
*/
|
||||||
|
size_t COVER_sum(const size_t *samplesSizes, unsigned nbSamples) ;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Initialize the `COVER_best_t`.
|
||||||
|
*/
|
||||||
|
void COVER_best_init(COVER_best_t *best);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Wait until liveJobs == 0.
|
||||||
|
*/
|
||||||
|
void COVER_best_wait(COVER_best_t *best);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Call COVER_best_wait() and then destroy the COVER_best_t.
|
||||||
|
*/
|
||||||
|
void COVER_best_destroy(COVER_best_t *best);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Called when a thread is about to be launched.
|
||||||
|
* Increments liveJobs.
|
||||||
|
*/
|
||||||
|
void COVER_best_start(COVER_best_t *best);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Called when a thread finishes executing, both on error or success.
|
||||||
|
* Decrements liveJobs and signals any waiting threads if liveJobs == 0.
|
||||||
|
* If this dictionary is the best so far save it and its parameters.
|
||||||
|
*/
|
||||||
|
void COVER_best_finish(COVER_best_t *best, ZDICT_cover_params_t parameters,
|
||||||
|
COVER_dictSelection_t selection);
|
||||||
|
/**
|
||||||
|
* Error function for COVER_selectDict function. Checks if the return
|
||||||
|
* value is an error.
|
||||||
|
*/
|
||||||
|
unsigned COVER_dictSelectionIsError(COVER_dictSelection_t selection);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Error function for COVER_selectDict function. Returns a struct where
|
||||||
|
* return.totalCompressedSize is a ZSTD error.
|
||||||
|
*/
|
||||||
|
COVER_dictSelection_t COVER_dictSelectionError(size_t error);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Always call after selectDict is called to free up used memory from
|
||||||
|
* newly created dictionary.
|
||||||
|
*/
|
||||||
|
void COVER_dictSelectionFree(COVER_dictSelection_t selection);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Called to finalize the dictionary and select one based on whether or not
|
||||||
|
* the shrink-dict flag was enabled. If enabled the dictionary used is the
|
||||||
|
* smallest dictionary within a specified regression of the compressed size
|
||||||
|
* from the largest dictionary.
|
||||||
|
*/
|
||||||
|
COVER_dictSelection_t COVER_selectDict(BYTE* customDictContent,
|
||||||
|
size_t dictContentSize, const BYTE* samplesBuffer, const size_t* samplesSizes, unsigned nbFinalizeSamples,
|
||||||
|
size_t nbCheckSamples, size_t nbSamples, ZDICT_cover_params_t params, size_t* offsets, size_t totalCompressedSize);
|
||||||
215
vendor/github.com/DataDog/zstd/cpu.h
generated
vendored
Normal file
215
vendor/github.com/DataDog/zstd/cpu.h
generated
vendored
Normal file
@@ -0,0 +1,215 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2018-present, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_COMMON_CPU_H
|
||||||
|
#define ZSTD_COMMON_CPU_H
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Implementation taken from folly/CpuId.h
|
||||||
|
* https://github.com/facebook/folly/blob/master/folly/CpuId.h
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <string.h>
|
||||||
|
|
||||||
|
#include "mem.h"
|
||||||
|
|
||||||
|
#ifdef _MSC_VER
|
||||||
|
#include <intrin.h>
|
||||||
|
#endif
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
U32 f1c;
|
||||||
|
U32 f1d;
|
||||||
|
U32 f7b;
|
||||||
|
U32 f7c;
|
||||||
|
} ZSTD_cpuid_t;
|
||||||
|
|
||||||
|
MEM_STATIC ZSTD_cpuid_t ZSTD_cpuid(void) {
|
||||||
|
U32 f1c = 0;
|
||||||
|
U32 f1d = 0;
|
||||||
|
U32 f7b = 0;
|
||||||
|
U32 f7c = 0;
|
||||||
|
#if defined(_MSC_VER) && (defined(_M_X64) || defined(_M_IX86))
|
||||||
|
int reg[4];
|
||||||
|
__cpuid((int*)reg, 0);
|
||||||
|
{
|
||||||
|
int const n = reg[0];
|
||||||
|
if (n >= 1) {
|
||||||
|
__cpuid((int*)reg, 1);
|
||||||
|
f1c = (U32)reg[2];
|
||||||
|
f1d = (U32)reg[3];
|
||||||
|
}
|
||||||
|
if (n >= 7) {
|
||||||
|
__cpuidex((int*)reg, 7, 0);
|
||||||
|
f7b = (U32)reg[1];
|
||||||
|
f7c = (U32)reg[2];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#elif defined(__i386__) && defined(__PIC__) && !defined(__clang__) && defined(__GNUC__)
|
||||||
|
/* The following block like the normal cpuid branch below, but gcc
|
||||||
|
* reserves ebx for use of its pic register so we must specially
|
||||||
|
* handle the save and restore to avoid clobbering the register
|
||||||
|
*/
|
||||||
|
U32 n;
|
||||||
|
__asm__(
|
||||||
|
"pushl %%ebx\n\t"
|
||||||
|
"cpuid\n\t"
|
||||||
|
"popl %%ebx\n\t"
|
||||||
|
: "=a"(n)
|
||||||
|
: "a"(0)
|
||||||
|
: "ecx", "edx");
|
||||||
|
if (n >= 1) {
|
||||||
|
U32 f1a;
|
||||||
|
__asm__(
|
||||||
|
"pushl %%ebx\n\t"
|
||||||
|
"cpuid\n\t"
|
||||||
|
"popl %%ebx\n\t"
|
||||||
|
: "=a"(f1a), "=c"(f1c), "=d"(f1d)
|
||||||
|
: "a"(1));
|
||||||
|
}
|
||||||
|
if (n >= 7) {
|
||||||
|
__asm__(
|
||||||
|
"pushl %%ebx\n\t"
|
||||||
|
"cpuid\n\t"
|
||||||
|
"movl %%ebx, %%eax\n\t"
|
||||||
|
"popl %%ebx"
|
||||||
|
: "=a"(f7b), "=c"(f7c)
|
||||||
|
: "a"(7), "c"(0)
|
||||||
|
: "edx");
|
||||||
|
}
|
||||||
|
#elif defined(__x86_64__) || defined(_M_X64) || defined(__i386__)
|
||||||
|
U32 n;
|
||||||
|
__asm__("cpuid" : "=a"(n) : "a"(0) : "ebx", "ecx", "edx");
|
||||||
|
if (n >= 1) {
|
||||||
|
U32 f1a;
|
||||||
|
__asm__("cpuid" : "=a"(f1a), "=c"(f1c), "=d"(f1d) : "a"(1) : "ebx");
|
||||||
|
}
|
||||||
|
if (n >= 7) {
|
||||||
|
U32 f7a;
|
||||||
|
__asm__("cpuid"
|
||||||
|
: "=a"(f7a), "=b"(f7b), "=c"(f7c)
|
||||||
|
: "a"(7), "c"(0)
|
||||||
|
: "edx");
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
{
|
||||||
|
ZSTD_cpuid_t cpuid;
|
||||||
|
cpuid.f1c = f1c;
|
||||||
|
cpuid.f1d = f1d;
|
||||||
|
cpuid.f7b = f7b;
|
||||||
|
cpuid.f7c = f7c;
|
||||||
|
return cpuid;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#define X(name, r, bit) \
|
||||||
|
MEM_STATIC int ZSTD_cpuid_##name(ZSTD_cpuid_t const cpuid) { \
|
||||||
|
return ((cpuid.r) & (1U << bit)) != 0; \
|
||||||
|
}
|
||||||
|
|
||||||
|
/* cpuid(1): Processor Info and Feature Bits. */
|
||||||
|
#define C(name, bit) X(name, f1c, bit)
|
||||||
|
C(sse3, 0)
|
||||||
|
C(pclmuldq, 1)
|
||||||
|
C(dtes64, 2)
|
||||||
|
C(monitor, 3)
|
||||||
|
C(dscpl, 4)
|
||||||
|
C(vmx, 5)
|
||||||
|
C(smx, 6)
|
||||||
|
C(eist, 7)
|
||||||
|
C(tm2, 8)
|
||||||
|
C(ssse3, 9)
|
||||||
|
C(cnxtid, 10)
|
||||||
|
C(fma, 12)
|
||||||
|
C(cx16, 13)
|
||||||
|
C(xtpr, 14)
|
||||||
|
C(pdcm, 15)
|
||||||
|
C(pcid, 17)
|
||||||
|
C(dca, 18)
|
||||||
|
C(sse41, 19)
|
||||||
|
C(sse42, 20)
|
||||||
|
C(x2apic, 21)
|
||||||
|
C(movbe, 22)
|
||||||
|
C(popcnt, 23)
|
||||||
|
C(tscdeadline, 24)
|
||||||
|
C(aes, 25)
|
||||||
|
C(xsave, 26)
|
||||||
|
C(osxsave, 27)
|
||||||
|
C(avx, 28)
|
||||||
|
C(f16c, 29)
|
||||||
|
C(rdrand, 30)
|
||||||
|
#undef C
|
||||||
|
#define D(name, bit) X(name, f1d, bit)
|
||||||
|
D(fpu, 0)
|
||||||
|
D(vme, 1)
|
||||||
|
D(de, 2)
|
||||||
|
D(pse, 3)
|
||||||
|
D(tsc, 4)
|
||||||
|
D(msr, 5)
|
||||||
|
D(pae, 6)
|
||||||
|
D(mce, 7)
|
||||||
|
D(cx8, 8)
|
||||||
|
D(apic, 9)
|
||||||
|
D(sep, 11)
|
||||||
|
D(mtrr, 12)
|
||||||
|
D(pge, 13)
|
||||||
|
D(mca, 14)
|
||||||
|
D(cmov, 15)
|
||||||
|
D(pat, 16)
|
||||||
|
D(pse36, 17)
|
||||||
|
D(psn, 18)
|
||||||
|
D(clfsh, 19)
|
||||||
|
D(ds, 21)
|
||||||
|
D(acpi, 22)
|
||||||
|
D(mmx, 23)
|
||||||
|
D(fxsr, 24)
|
||||||
|
D(sse, 25)
|
||||||
|
D(sse2, 26)
|
||||||
|
D(ss, 27)
|
||||||
|
D(htt, 28)
|
||||||
|
D(tm, 29)
|
||||||
|
D(pbe, 31)
|
||||||
|
#undef D
|
||||||
|
|
||||||
|
/* cpuid(7): Extended Features. */
|
||||||
|
#define B(name, bit) X(name, f7b, bit)
|
||||||
|
B(bmi1, 3)
|
||||||
|
B(hle, 4)
|
||||||
|
B(avx2, 5)
|
||||||
|
B(smep, 7)
|
||||||
|
B(bmi2, 8)
|
||||||
|
B(erms, 9)
|
||||||
|
B(invpcid, 10)
|
||||||
|
B(rtm, 11)
|
||||||
|
B(mpx, 14)
|
||||||
|
B(avx512f, 16)
|
||||||
|
B(avx512dq, 17)
|
||||||
|
B(rdseed, 18)
|
||||||
|
B(adx, 19)
|
||||||
|
B(smap, 20)
|
||||||
|
B(avx512ifma, 21)
|
||||||
|
B(pcommit, 22)
|
||||||
|
B(clflushopt, 23)
|
||||||
|
B(clwb, 24)
|
||||||
|
B(avx512pf, 26)
|
||||||
|
B(avx512er, 27)
|
||||||
|
B(avx512cd, 28)
|
||||||
|
B(sha, 29)
|
||||||
|
B(avx512bw, 30)
|
||||||
|
B(avx512vl, 31)
|
||||||
|
#undef B
|
||||||
|
#define C(name, bit) X(name, f7c, bit)
|
||||||
|
C(prefetchwt1, 0)
|
||||||
|
C(avx512vbmi, 1)
|
||||||
|
#undef C
|
||||||
|
|
||||||
|
#undef X
|
||||||
|
|
||||||
|
#endif /* ZSTD_COMMON_CPU_H */
|
||||||
44
vendor/github.com/DataDog/zstd/debug.c
generated
vendored
Normal file
44
vendor/github.com/DataDog/zstd/debug.c
generated
vendored
Normal file
@@ -0,0 +1,44 @@
|
|||||||
|
/* ******************************************************************
|
||||||
|
debug
|
||||||
|
Part of FSE library
|
||||||
|
Copyright (C) 2013-present, Yann Collet.
|
||||||
|
|
||||||
|
BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
You can contact the author at :
|
||||||
|
- Source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
||||||
|
****************************************************************** */
|
||||||
|
|
||||||
|
|
||||||
|
/*
|
||||||
|
* This module only hosts one global variable
|
||||||
|
* which can be used to dynamically influence the verbosity of traces,
|
||||||
|
* such as DEBUGLOG and RAWLOG
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include "debug.h"
|
||||||
|
|
||||||
|
int g_debuglevel = DEBUGLEVEL;
|
||||||
134
vendor/github.com/DataDog/zstd/debug.h
generated
vendored
Normal file
134
vendor/github.com/DataDog/zstd/debug.h
generated
vendored
Normal file
@@ -0,0 +1,134 @@
|
|||||||
|
/* ******************************************************************
|
||||||
|
debug
|
||||||
|
Part of FSE library
|
||||||
|
Copyright (C) 2013-present, Yann Collet.
|
||||||
|
|
||||||
|
BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
You can contact the author at :
|
||||||
|
- Source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
||||||
|
****************************************************************** */
|
||||||
|
|
||||||
|
|
||||||
|
/*
|
||||||
|
* The purpose of this header is to enable debug functions.
|
||||||
|
* They regroup assert(), DEBUGLOG() and RAWLOG() for run-time,
|
||||||
|
* and DEBUG_STATIC_ASSERT() for compile-time.
|
||||||
|
*
|
||||||
|
* By default, DEBUGLEVEL==0, which means run-time debug is disabled.
|
||||||
|
*
|
||||||
|
* Level 1 enables assert() only.
|
||||||
|
* Starting level 2, traces can be generated and pushed to stderr.
|
||||||
|
* The higher the level, the more verbose the traces.
|
||||||
|
*
|
||||||
|
* It's possible to dynamically adjust level using variable g_debug_level,
|
||||||
|
* which is only declared if DEBUGLEVEL>=2,
|
||||||
|
* and is a global variable, not multi-thread protected (use with care)
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef DEBUG_H_12987983217
|
||||||
|
#define DEBUG_H_12987983217
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* static assert is triggered at compile time, leaving no runtime artefact.
|
||||||
|
* static assert only works with compile-time constants.
|
||||||
|
* Also, this variant can only be used inside a function. */
|
||||||
|
#define DEBUG_STATIC_ASSERT(c) (void)sizeof(char[(c) ? 1 : -1])
|
||||||
|
|
||||||
|
|
||||||
|
/* DEBUGLEVEL is expected to be defined externally,
|
||||||
|
* typically through compiler command line.
|
||||||
|
* Value must be a number. */
|
||||||
|
#ifndef DEBUGLEVEL
|
||||||
|
# define DEBUGLEVEL 0
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* DEBUGFILE can be defined externally,
|
||||||
|
* typically through compiler command line.
|
||||||
|
* note : currently useless.
|
||||||
|
* Value must be stderr or stdout */
|
||||||
|
#ifndef DEBUGFILE
|
||||||
|
# define DEBUGFILE stderr
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* recommended values for DEBUGLEVEL :
|
||||||
|
* 0 : release mode, no debug, all run-time checks disabled
|
||||||
|
* 1 : enables assert() only, no display
|
||||||
|
* 2 : reserved, for currently active debug path
|
||||||
|
* 3 : events once per object lifetime (CCtx, CDict, etc.)
|
||||||
|
* 4 : events once per frame
|
||||||
|
* 5 : events once per block
|
||||||
|
* 6 : events once per sequence (verbose)
|
||||||
|
* 7+: events at every position (*very* verbose)
|
||||||
|
*
|
||||||
|
* It's generally inconvenient to output traces > 5.
|
||||||
|
* In which case, it's possible to selectively trigger high verbosity levels
|
||||||
|
* by modifying g_debug_level.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#if (DEBUGLEVEL>=1)
|
||||||
|
# include <assert.h>
|
||||||
|
#else
|
||||||
|
# ifndef assert /* assert may be already defined, due to prior #include <assert.h> */
|
||||||
|
# define assert(condition) ((void)0) /* disable assert (default) */
|
||||||
|
# endif
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#if (DEBUGLEVEL>=2)
|
||||||
|
# include <stdio.h>
|
||||||
|
extern int g_debuglevel; /* the variable is only declared,
|
||||||
|
it actually lives in debug.c,
|
||||||
|
and is shared by the whole process.
|
||||||
|
It's not thread-safe.
|
||||||
|
It's useful when enabling very verbose levels
|
||||||
|
on selective conditions (such as position in src) */
|
||||||
|
|
||||||
|
# define RAWLOG(l, ...) { \
|
||||||
|
if (l<=g_debuglevel) { \
|
||||||
|
fprintf(stderr, __VA_ARGS__); \
|
||||||
|
} }
|
||||||
|
# define DEBUGLOG(l, ...) { \
|
||||||
|
if (l<=g_debuglevel) { \
|
||||||
|
fprintf(stderr, __FILE__ ": " __VA_ARGS__); \
|
||||||
|
fprintf(stderr, " \n"); \
|
||||||
|
} }
|
||||||
|
#else
|
||||||
|
# define RAWLOG(l, ...) {} /* disabled */
|
||||||
|
# define DEBUGLOG(l, ...) {} /* disabled */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* DEBUG_H_12987983217 */
|
||||||
1913
vendor/github.com/DataDog/zstd/divsufsort.c
generated
vendored
Normal file
1913
vendor/github.com/DataDog/zstd/divsufsort.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
67
vendor/github.com/DataDog/zstd/divsufsort.h
generated
vendored
Normal file
67
vendor/github.com/DataDog/zstd/divsufsort.h
generated
vendored
Normal file
@@ -0,0 +1,67 @@
|
|||||||
|
/*
|
||||||
|
* divsufsort.h for libdivsufsort-lite
|
||||||
|
* Copyright (c) 2003-2008 Yuta Mori All Rights Reserved.
|
||||||
|
*
|
||||||
|
* Permission is hereby granted, free of charge, to any person
|
||||||
|
* obtaining a copy of this software and associated documentation
|
||||||
|
* files (the "Software"), to deal in the Software without
|
||||||
|
* restriction, including without limitation the rights to use,
|
||||||
|
* copy, modify, merge, publish, distribute, sublicense, and/or sell
|
||||||
|
* copies of the Software, and to permit persons to whom the
|
||||||
|
* Software is furnished to do so, subject to the following
|
||||||
|
* conditions:
|
||||||
|
*
|
||||||
|
* The above copyright notice and this permission notice shall be
|
||||||
|
* included in all copies or substantial portions of the Software.
|
||||||
|
*
|
||||||
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
||||||
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
|
||||||
|
* OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
||||||
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
|
||||||
|
* HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
|
||||||
|
* WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||||
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
|
||||||
|
* OTHER DEALINGS IN THE SOFTWARE.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef _DIVSUFSORT_H
|
||||||
|
#define _DIVSUFSORT_H 1
|
||||||
|
|
||||||
|
#ifdef __cplusplus
|
||||||
|
extern "C" {
|
||||||
|
#endif /* __cplusplus */
|
||||||
|
|
||||||
|
|
||||||
|
/*- Prototypes -*/
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructs the suffix array of a given string.
|
||||||
|
* @param T [0..n-1] The input string.
|
||||||
|
* @param SA [0..n-1] The output array of suffixes.
|
||||||
|
* @param n The length of the given string.
|
||||||
|
* @param openMP enables OpenMP optimization.
|
||||||
|
* @return 0 if no error occurred, -1 or -2 otherwise.
|
||||||
|
*/
|
||||||
|
int
|
||||||
|
divsufsort(const unsigned char *T, int *SA, int n, int openMP);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructs the burrows-wheeler transformed string of a given string.
|
||||||
|
* @param T [0..n-1] The input string.
|
||||||
|
* @param U [0..n-1] The output string. (can be T)
|
||||||
|
* @param A [0..n-1] The temporary array. (can be NULL)
|
||||||
|
* @param n The length of the given string.
|
||||||
|
* @param num_indexes The length of secondary indexes array. (can be NULL)
|
||||||
|
* @param indexes The secondary indexes array. (can be NULL)
|
||||||
|
* @param openMP enables OpenMP optimization.
|
||||||
|
* @return The primary index if no error occurred, -1 or -2 otherwise.
|
||||||
|
*/
|
||||||
|
int
|
||||||
|
divbwt(const unsigned char *T, unsigned char *U, int *A, int n, unsigned char * num_indexes, int * indexes, int openMP);
|
||||||
|
|
||||||
|
|
||||||
|
#ifdef __cplusplus
|
||||||
|
} /* extern "C" */
|
||||||
|
#endif /* __cplusplus */
|
||||||
|
|
||||||
|
#endif /* _DIVSUFSORT_H */
|
||||||
236
vendor/github.com/DataDog/zstd/entropy_common.c
generated
vendored
Normal file
236
vendor/github.com/DataDog/zstd/entropy_common.c
generated
vendored
Normal file
@@ -0,0 +1,236 @@
|
|||||||
|
/*
|
||||||
|
Common functions of New Generation Entropy library
|
||||||
|
Copyright (C) 2016, Yann Collet.
|
||||||
|
|
||||||
|
BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
You can contact the author at :
|
||||||
|
- FSE+HUF source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
||||||
|
- Public forum : https://groups.google.com/forum/#!forum/lz4c
|
||||||
|
*************************************************************************** */
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Dependencies
|
||||||
|
***************************************/
|
||||||
|
#include "mem.h"
|
||||||
|
#include "error_private.h" /* ERR_*, ERROR */
|
||||||
|
#define FSE_STATIC_LINKING_ONLY /* FSE_MIN_TABLELOG */
|
||||||
|
#include "fse.h"
|
||||||
|
#define HUF_STATIC_LINKING_ONLY /* HUF_TABLELOG_ABSOLUTEMAX */
|
||||||
|
#include "huf.h"
|
||||||
|
|
||||||
|
|
||||||
|
/*=== Version ===*/
|
||||||
|
unsigned FSE_versionNumber(void) { return FSE_VERSION_NUMBER; }
|
||||||
|
|
||||||
|
|
||||||
|
/*=== Error Management ===*/
|
||||||
|
unsigned FSE_isError(size_t code) { return ERR_isError(code); }
|
||||||
|
const char* FSE_getErrorName(size_t code) { return ERR_getErrorName(code); }
|
||||||
|
|
||||||
|
unsigned HUF_isError(size_t code) { return ERR_isError(code); }
|
||||||
|
const char* HUF_getErrorName(size_t code) { return ERR_getErrorName(code); }
|
||||||
|
|
||||||
|
|
||||||
|
/*-**************************************************************
|
||||||
|
* FSE NCount encoding-decoding
|
||||||
|
****************************************************************/
|
||||||
|
size_t FSE_readNCount (short* normalizedCounter, unsigned* maxSVPtr, unsigned* tableLogPtr,
|
||||||
|
const void* headerBuffer, size_t hbSize)
|
||||||
|
{
|
||||||
|
const BYTE* const istart = (const BYTE*) headerBuffer;
|
||||||
|
const BYTE* const iend = istart + hbSize;
|
||||||
|
const BYTE* ip = istart;
|
||||||
|
int nbBits;
|
||||||
|
int remaining;
|
||||||
|
int threshold;
|
||||||
|
U32 bitStream;
|
||||||
|
int bitCount;
|
||||||
|
unsigned charnum = 0;
|
||||||
|
int previous0 = 0;
|
||||||
|
|
||||||
|
if (hbSize < 4) {
|
||||||
|
/* This function only works when hbSize >= 4 */
|
||||||
|
char buffer[4];
|
||||||
|
memset(buffer, 0, sizeof(buffer));
|
||||||
|
memcpy(buffer, headerBuffer, hbSize);
|
||||||
|
{ size_t const countSize = FSE_readNCount(normalizedCounter, maxSVPtr, tableLogPtr,
|
||||||
|
buffer, sizeof(buffer));
|
||||||
|
if (FSE_isError(countSize)) return countSize;
|
||||||
|
if (countSize > hbSize) return ERROR(corruption_detected);
|
||||||
|
return countSize;
|
||||||
|
} }
|
||||||
|
assert(hbSize >= 4);
|
||||||
|
|
||||||
|
/* init */
|
||||||
|
memset(normalizedCounter, 0, (*maxSVPtr+1) * sizeof(normalizedCounter[0])); /* all symbols not present in NCount have a frequency of 0 */
|
||||||
|
bitStream = MEM_readLE32(ip);
|
||||||
|
nbBits = (bitStream & 0xF) + FSE_MIN_TABLELOG; /* extract tableLog */
|
||||||
|
if (nbBits > FSE_TABLELOG_ABSOLUTE_MAX) return ERROR(tableLog_tooLarge);
|
||||||
|
bitStream >>= 4;
|
||||||
|
bitCount = 4;
|
||||||
|
*tableLogPtr = nbBits;
|
||||||
|
remaining = (1<<nbBits)+1;
|
||||||
|
threshold = 1<<nbBits;
|
||||||
|
nbBits++;
|
||||||
|
|
||||||
|
while ((remaining>1) & (charnum<=*maxSVPtr)) {
|
||||||
|
if (previous0) {
|
||||||
|
unsigned n0 = charnum;
|
||||||
|
while ((bitStream & 0xFFFF) == 0xFFFF) {
|
||||||
|
n0 += 24;
|
||||||
|
if (ip < iend-5) {
|
||||||
|
ip += 2;
|
||||||
|
bitStream = MEM_readLE32(ip) >> bitCount;
|
||||||
|
} else {
|
||||||
|
bitStream >>= 16;
|
||||||
|
bitCount += 16;
|
||||||
|
} }
|
||||||
|
while ((bitStream & 3) == 3) {
|
||||||
|
n0 += 3;
|
||||||
|
bitStream >>= 2;
|
||||||
|
bitCount += 2;
|
||||||
|
}
|
||||||
|
n0 += bitStream & 3;
|
||||||
|
bitCount += 2;
|
||||||
|
if (n0 > *maxSVPtr) return ERROR(maxSymbolValue_tooSmall);
|
||||||
|
while (charnum < n0) normalizedCounter[charnum++] = 0;
|
||||||
|
if ((ip <= iend-7) || (ip + (bitCount>>3) <= iend-4)) {
|
||||||
|
assert((bitCount >> 3) <= 3); /* For first condition to work */
|
||||||
|
ip += bitCount>>3;
|
||||||
|
bitCount &= 7;
|
||||||
|
bitStream = MEM_readLE32(ip) >> bitCount;
|
||||||
|
} else {
|
||||||
|
bitStream >>= 2;
|
||||||
|
} }
|
||||||
|
{ int const max = (2*threshold-1) - remaining;
|
||||||
|
int count;
|
||||||
|
|
||||||
|
if ((bitStream & (threshold-1)) < (U32)max) {
|
||||||
|
count = bitStream & (threshold-1);
|
||||||
|
bitCount += nbBits-1;
|
||||||
|
} else {
|
||||||
|
count = bitStream & (2*threshold-1);
|
||||||
|
if (count >= threshold) count -= max;
|
||||||
|
bitCount += nbBits;
|
||||||
|
}
|
||||||
|
|
||||||
|
count--; /* extra accuracy */
|
||||||
|
remaining -= count < 0 ? -count : count; /* -1 means +1 */
|
||||||
|
normalizedCounter[charnum++] = (short)count;
|
||||||
|
previous0 = !count;
|
||||||
|
while (remaining < threshold) {
|
||||||
|
nbBits--;
|
||||||
|
threshold >>= 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
if ((ip <= iend-7) || (ip + (bitCount>>3) <= iend-4)) {
|
||||||
|
ip += bitCount>>3;
|
||||||
|
bitCount &= 7;
|
||||||
|
} else {
|
||||||
|
bitCount -= (int)(8 * (iend - 4 - ip));
|
||||||
|
ip = iend - 4;
|
||||||
|
}
|
||||||
|
bitStream = MEM_readLE32(ip) >> (bitCount & 31);
|
||||||
|
} } /* while ((remaining>1) & (charnum<=*maxSVPtr)) */
|
||||||
|
if (remaining != 1) return ERROR(corruption_detected);
|
||||||
|
if (bitCount > 32) return ERROR(corruption_detected);
|
||||||
|
*maxSVPtr = charnum-1;
|
||||||
|
|
||||||
|
ip += (bitCount+7)>>3;
|
||||||
|
return ip-istart;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/*! HUF_readStats() :
|
||||||
|
Read compact Huffman tree, saved by HUF_writeCTable().
|
||||||
|
`huffWeight` is destination buffer.
|
||||||
|
`rankStats` is assumed to be a table of at least HUF_TABLELOG_MAX U32.
|
||||||
|
@return : size read from `src` , or an error Code .
|
||||||
|
Note : Needed by HUF_readCTable() and HUF_readDTableX?() .
|
||||||
|
*/
|
||||||
|
size_t HUF_readStats(BYTE* huffWeight, size_t hwSize, U32* rankStats,
|
||||||
|
U32* nbSymbolsPtr, U32* tableLogPtr,
|
||||||
|
const void* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
U32 weightTotal;
|
||||||
|
const BYTE* ip = (const BYTE*) src;
|
||||||
|
size_t iSize;
|
||||||
|
size_t oSize;
|
||||||
|
|
||||||
|
if (!srcSize) return ERROR(srcSize_wrong);
|
||||||
|
iSize = ip[0];
|
||||||
|
/* memset(huffWeight, 0, hwSize); *//* is not necessary, even though some analyzer complain ... */
|
||||||
|
|
||||||
|
if (iSize >= 128) { /* special header */
|
||||||
|
oSize = iSize - 127;
|
||||||
|
iSize = ((oSize+1)/2);
|
||||||
|
if (iSize+1 > srcSize) return ERROR(srcSize_wrong);
|
||||||
|
if (oSize >= hwSize) return ERROR(corruption_detected);
|
||||||
|
ip += 1;
|
||||||
|
{ U32 n;
|
||||||
|
for (n=0; n<oSize; n+=2) {
|
||||||
|
huffWeight[n] = ip[n/2] >> 4;
|
||||||
|
huffWeight[n+1] = ip[n/2] & 15;
|
||||||
|
} } }
|
||||||
|
else { /* header compressed with FSE (normal case) */
|
||||||
|
FSE_DTable fseWorkspace[FSE_DTABLE_SIZE_U32(6)]; /* 6 is max possible tableLog for HUF header (maybe even 5, to be tested) */
|
||||||
|
if (iSize+1 > srcSize) return ERROR(srcSize_wrong);
|
||||||
|
oSize = FSE_decompress_wksp(huffWeight, hwSize-1, ip+1, iSize, fseWorkspace, 6); /* max (hwSize-1) values decoded, as last one is implied */
|
||||||
|
if (FSE_isError(oSize)) return oSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* collect weight stats */
|
||||||
|
memset(rankStats, 0, (HUF_TABLELOG_MAX + 1) * sizeof(U32));
|
||||||
|
weightTotal = 0;
|
||||||
|
{ U32 n; for (n=0; n<oSize; n++) {
|
||||||
|
if (huffWeight[n] >= HUF_TABLELOG_MAX) return ERROR(corruption_detected);
|
||||||
|
rankStats[huffWeight[n]]++;
|
||||||
|
weightTotal += (1 << huffWeight[n]) >> 1;
|
||||||
|
} }
|
||||||
|
if (weightTotal == 0) return ERROR(corruption_detected);
|
||||||
|
|
||||||
|
/* get last non-null symbol weight (implied, total must be 2^n) */
|
||||||
|
{ U32 const tableLog = BIT_highbit32(weightTotal) + 1;
|
||||||
|
if (tableLog > HUF_TABLELOG_MAX) return ERROR(corruption_detected);
|
||||||
|
*tableLogPtr = tableLog;
|
||||||
|
/* determine last weight */
|
||||||
|
{ U32 const total = 1 << tableLog;
|
||||||
|
U32 const rest = total - weightTotal;
|
||||||
|
U32 const verif = 1 << BIT_highbit32(rest);
|
||||||
|
U32 const lastWeight = BIT_highbit32(rest) + 1;
|
||||||
|
if (verif != rest) return ERROR(corruption_detected); /* last value must be a clean power of 2 */
|
||||||
|
huffWeight[oSize] = (BYTE)lastWeight;
|
||||||
|
rankStats[lastWeight]++;
|
||||||
|
} }
|
||||||
|
|
||||||
|
/* check tree construction validity */
|
||||||
|
if ((rankStats[1] < 2) || (rankStats[1] & 1)) return ERROR(corruption_detected); /* by construction : at least 2 elts of rank 1, must be even */
|
||||||
|
|
||||||
|
/* results */
|
||||||
|
*nbSymbolsPtr = (U32)(oSize+1);
|
||||||
|
return iSize+1;
|
||||||
|
}
|
||||||
54
vendor/github.com/DataDog/zstd/error_private.c
generated
vendored
Normal file
54
vendor/github.com/DataDog/zstd/error_private.c
generated
vendored
Normal file
@@ -0,0 +1,54 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* The purpose of this file is to have a single list of error strings embedded in binary */
|
||||||
|
|
||||||
|
#include "error_private.h"
|
||||||
|
|
||||||
|
const char* ERR_getErrorString(ERR_enum code)
|
||||||
|
{
|
||||||
|
#ifdef ZSTD_STRIP_ERROR_STRINGS
|
||||||
|
(void)code;
|
||||||
|
return "Error strings stripped";
|
||||||
|
#else
|
||||||
|
static const char* const notErrorCode = "Unspecified error code";
|
||||||
|
switch( code )
|
||||||
|
{
|
||||||
|
case PREFIX(no_error): return "No error detected";
|
||||||
|
case PREFIX(GENERIC): return "Error (generic)";
|
||||||
|
case PREFIX(prefix_unknown): return "Unknown frame descriptor";
|
||||||
|
case PREFIX(version_unsupported): return "Version not supported";
|
||||||
|
case PREFIX(frameParameter_unsupported): return "Unsupported frame parameter";
|
||||||
|
case PREFIX(frameParameter_windowTooLarge): return "Frame requires too much memory for decoding";
|
||||||
|
case PREFIX(corruption_detected): return "Corrupted block detected";
|
||||||
|
case PREFIX(checksum_wrong): return "Restored data doesn't match checksum";
|
||||||
|
case PREFIX(parameter_unsupported): return "Unsupported parameter";
|
||||||
|
case PREFIX(parameter_outOfBound): return "Parameter is out of bound";
|
||||||
|
case PREFIX(init_missing): return "Context should be init first";
|
||||||
|
case PREFIX(memory_allocation): return "Allocation error : not enough memory";
|
||||||
|
case PREFIX(workSpace_tooSmall): return "workSpace buffer is not large enough";
|
||||||
|
case PREFIX(stage_wrong): return "Operation not authorized at current processing stage";
|
||||||
|
case PREFIX(tableLog_tooLarge): return "tableLog requires too much memory : unsupported";
|
||||||
|
case PREFIX(maxSymbolValue_tooLarge): return "Unsupported max Symbol Value : too large";
|
||||||
|
case PREFIX(maxSymbolValue_tooSmall): return "Specified maxSymbolValue is too small";
|
||||||
|
case PREFIX(dictionary_corrupted): return "Dictionary is corrupted";
|
||||||
|
case PREFIX(dictionary_wrong): return "Dictionary mismatch";
|
||||||
|
case PREFIX(dictionaryCreation_failed): return "Cannot create Dictionary from provided samples";
|
||||||
|
case PREFIX(dstSize_tooSmall): return "Destination buffer is too small";
|
||||||
|
case PREFIX(srcSize_wrong): return "Src size is incorrect";
|
||||||
|
case PREFIX(dstBuffer_null): return "Operation on NULL destination buffer";
|
||||||
|
/* following error codes are not stable and may be removed or changed in a future version */
|
||||||
|
case PREFIX(frameIndex_tooLarge): return "Frame index is too large";
|
||||||
|
case PREFIX(seekableIO): return "An I/O error occurred when reading/seeking";
|
||||||
|
case PREFIX(maxCode):
|
||||||
|
default: return notErrorCode;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
}
|
||||||
76
vendor/github.com/DataDog/zstd/error_private.h
generated
vendored
Normal file
76
vendor/github.com/DataDog/zstd/error_private.h
generated
vendored
Normal file
@@ -0,0 +1,76 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* Note : this module is expected to remain private, do not expose it */
|
||||||
|
|
||||||
|
#ifndef ERROR_H_MODULE
|
||||||
|
#define ERROR_H_MODULE
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* ****************************************
|
||||||
|
* Dependencies
|
||||||
|
******************************************/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
#include "zstd_errors.h" /* enum list */
|
||||||
|
|
||||||
|
|
||||||
|
/* ****************************************
|
||||||
|
* Compiler-specific
|
||||||
|
******************************************/
|
||||||
|
#if defined(__GNUC__)
|
||||||
|
# define ERR_STATIC static __attribute__((unused))
|
||||||
|
#elif defined (__cplusplus) || (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L) /* C99 */)
|
||||||
|
# define ERR_STATIC static inline
|
||||||
|
#elif defined(_MSC_VER)
|
||||||
|
# define ERR_STATIC static __inline
|
||||||
|
#else
|
||||||
|
# define ERR_STATIC static /* this version may generate warnings for unused static functions; disable the relevant warning */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/*-****************************************
|
||||||
|
* Customization (error_public.h)
|
||||||
|
******************************************/
|
||||||
|
typedef ZSTD_ErrorCode ERR_enum;
|
||||||
|
#define PREFIX(name) ZSTD_error_##name
|
||||||
|
|
||||||
|
|
||||||
|
/*-****************************************
|
||||||
|
* Error codes handling
|
||||||
|
******************************************/
|
||||||
|
#undef ERROR /* reported already defined on VS 2015 (Rich Geldreich) */
|
||||||
|
#define ERROR(name) ZSTD_ERROR(name)
|
||||||
|
#define ZSTD_ERROR(name) ((size_t)-PREFIX(name))
|
||||||
|
|
||||||
|
ERR_STATIC unsigned ERR_isError(size_t code) { return (code > ERROR(maxCode)); }
|
||||||
|
|
||||||
|
ERR_STATIC ERR_enum ERR_getErrorCode(size_t code) { if (!ERR_isError(code)) return (ERR_enum)0; return (ERR_enum) (0-code); }
|
||||||
|
|
||||||
|
|
||||||
|
/*-****************************************
|
||||||
|
* Error Strings
|
||||||
|
******************************************/
|
||||||
|
|
||||||
|
const char* ERR_getErrorString(ERR_enum code); /* error_private.c */
|
||||||
|
|
||||||
|
ERR_STATIC const char* ERR_getErrorName(size_t code)
|
||||||
|
{
|
||||||
|
return ERR_getErrorString(ERR_getErrorCode(code));
|
||||||
|
}
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ERROR_H_MODULE */
|
||||||
35
vendor/github.com/DataDog/zstd/errors.go
generated
vendored
Normal file
35
vendor/github.com/DataDog/zstd/errors.go
generated
vendored
Normal file
@@ -0,0 +1,35 @@
|
|||||||
|
package zstd
|
||||||
|
|
||||||
|
/*
|
||||||
|
#define ZSTD_STATIC_LINKING_ONLY
|
||||||
|
#include "zstd.h"
|
||||||
|
*/
|
||||||
|
import "C"
|
||||||
|
|
||||||
|
// ErrorCode is an error returned by the zstd library.
|
||||||
|
type ErrorCode int
|
||||||
|
|
||||||
|
// Error returns the error string given by zstd
|
||||||
|
func (e ErrorCode) Error() string {
|
||||||
|
return C.GoString(C.ZSTD_getErrorName(C.size_t(e)))
|
||||||
|
}
|
||||||
|
|
||||||
|
func cIsError(code int) bool {
|
||||||
|
return int(C.ZSTD_isError(C.size_t(code))) != 0
|
||||||
|
}
|
||||||
|
|
||||||
|
// getError returns an error for the return code, or nil if it's not an error
|
||||||
|
func getError(code int) error {
|
||||||
|
if code < 0 && cIsError(code) {
|
||||||
|
return ErrorCode(code)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsDstSizeTooSmallError returns whether the error correspond to zstd standard sDstSizeTooSmall error
|
||||||
|
func IsDstSizeTooSmallError(e error) bool {
|
||||||
|
if e != nil && e.Error() == "Destination buffer is too small" {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
747
vendor/github.com/DataDog/zstd/fastcover.c
generated
vendored
Normal file
747
vendor/github.com/DataDog/zstd/fastcover.c
generated
vendored
Normal file
@@ -0,0 +1,747 @@
|
|||||||
|
/*-*************************************
|
||||||
|
* Dependencies
|
||||||
|
***************************************/
|
||||||
|
#include <stdio.h> /* fprintf */
|
||||||
|
#include <stdlib.h> /* malloc, free, qsort */
|
||||||
|
#include <string.h> /* memset */
|
||||||
|
#include <time.h> /* clock */
|
||||||
|
|
||||||
|
#include "mem.h" /* read */
|
||||||
|
#include "pool.h"
|
||||||
|
#include "threading.h"
|
||||||
|
#include "cover.h"
|
||||||
|
#include "zstd_internal.h" /* includes zstd.h */
|
||||||
|
#ifndef ZDICT_STATIC_LINKING_ONLY
|
||||||
|
#define ZDICT_STATIC_LINKING_ONLY
|
||||||
|
#endif
|
||||||
|
#include "zdict.h"
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Constants
|
||||||
|
***************************************/
|
||||||
|
#define FASTCOVER_MAX_SAMPLES_SIZE (sizeof(size_t) == 8 ? ((unsigned)-1) : ((unsigned)1 GB))
|
||||||
|
#define FASTCOVER_MAX_F 31
|
||||||
|
#define FASTCOVER_MAX_ACCEL 10
|
||||||
|
#define DEFAULT_SPLITPOINT 0.75
|
||||||
|
#define DEFAULT_F 20
|
||||||
|
#define DEFAULT_ACCEL 1
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Console display
|
||||||
|
***************************************/
|
||||||
|
static int g_displayLevel = 2;
|
||||||
|
#define DISPLAY(...) \
|
||||||
|
{ \
|
||||||
|
fprintf(stderr, __VA_ARGS__); \
|
||||||
|
fflush(stderr); \
|
||||||
|
}
|
||||||
|
#define LOCALDISPLAYLEVEL(displayLevel, l, ...) \
|
||||||
|
if (displayLevel >= l) { \
|
||||||
|
DISPLAY(__VA_ARGS__); \
|
||||||
|
} /* 0 : no display; 1: errors; 2: default; 3: details; 4: debug */
|
||||||
|
#define DISPLAYLEVEL(l, ...) LOCALDISPLAYLEVEL(g_displayLevel, l, __VA_ARGS__)
|
||||||
|
|
||||||
|
#define LOCALDISPLAYUPDATE(displayLevel, l, ...) \
|
||||||
|
if (displayLevel >= l) { \
|
||||||
|
if ((clock() - g_time > refreshRate) || (displayLevel >= 4)) { \
|
||||||
|
g_time = clock(); \
|
||||||
|
DISPLAY(__VA_ARGS__); \
|
||||||
|
} \
|
||||||
|
}
|
||||||
|
#define DISPLAYUPDATE(l, ...) LOCALDISPLAYUPDATE(g_displayLevel, l, __VA_ARGS__)
|
||||||
|
static const clock_t refreshRate = CLOCKS_PER_SEC * 15 / 100;
|
||||||
|
static clock_t g_time = 0;
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Hash Functions
|
||||||
|
***************************************/
|
||||||
|
static const U64 prime6bytes = 227718039650203ULL;
|
||||||
|
static size_t ZSTD_hash6(U64 u, U32 h) { return (size_t)(((u << (64-48)) * prime6bytes) >> (64-h)) ; }
|
||||||
|
static size_t ZSTD_hash6Ptr(const void* p, U32 h) { return ZSTD_hash6(MEM_readLE64(p), h); }
|
||||||
|
|
||||||
|
static const U64 prime8bytes = 0xCF1BBCDCB7A56463ULL;
|
||||||
|
static size_t ZSTD_hash8(U64 u, U32 h) { return (size_t)(((u) * prime8bytes) >> (64-h)) ; }
|
||||||
|
static size_t ZSTD_hash8Ptr(const void* p, U32 h) { return ZSTD_hash8(MEM_readLE64(p), h); }
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Hash the d-byte value pointed to by p and mod 2^f
|
||||||
|
*/
|
||||||
|
static size_t FASTCOVER_hashPtrToIndex(const void* p, U32 h, unsigned d) {
|
||||||
|
if (d == 6) {
|
||||||
|
return ZSTD_hash6Ptr(p, h) & ((1 << h) - 1);
|
||||||
|
}
|
||||||
|
return ZSTD_hash8Ptr(p, h) & ((1 << h) - 1);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Acceleration
|
||||||
|
***************************************/
|
||||||
|
typedef struct {
|
||||||
|
unsigned finalize; /* Percentage of training samples used for ZDICT_finalizeDictionary */
|
||||||
|
unsigned skip; /* Number of dmer skipped between each dmer counted in computeFrequency */
|
||||||
|
} FASTCOVER_accel_t;
|
||||||
|
|
||||||
|
|
||||||
|
static const FASTCOVER_accel_t FASTCOVER_defaultAccelParameters[FASTCOVER_MAX_ACCEL+1] = {
|
||||||
|
{ 100, 0 }, /* accel = 0, should not happen because accel = 0 defaults to accel = 1 */
|
||||||
|
{ 100, 0 }, /* accel = 1 */
|
||||||
|
{ 50, 1 }, /* accel = 2 */
|
||||||
|
{ 34, 2 }, /* accel = 3 */
|
||||||
|
{ 25, 3 }, /* accel = 4 */
|
||||||
|
{ 20, 4 }, /* accel = 5 */
|
||||||
|
{ 17, 5 }, /* accel = 6 */
|
||||||
|
{ 14, 6 }, /* accel = 7 */
|
||||||
|
{ 13, 7 }, /* accel = 8 */
|
||||||
|
{ 11, 8 }, /* accel = 9 */
|
||||||
|
{ 10, 9 }, /* accel = 10 */
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Context
|
||||||
|
***************************************/
|
||||||
|
typedef struct {
|
||||||
|
const BYTE *samples;
|
||||||
|
size_t *offsets;
|
||||||
|
const size_t *samplesSizes;
|
||||||
|
size_t nbSamples;
|
||||||
|
size_t nbTrainSamples;
|
||||||
|
size_t nbTestSamples;
|
||||||
|
size_t nbDmers;
|
||||||
|
U32 *freqs;
|
||||||
|
unsigned d;
|
||||||
|
unsigned f;
|
||||||
|
FASTCOVER_accel_t accelParams;
|
||||||
|
} FASTCOVER_ctx_t;
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Helper functions
|
||||||
|
***************************************/
|
||||||
|
/**
|
||||||
|
* Selects the best segment in an epoch.
|
||||||
|
* Segments of are scored according to the function:
|
||||||
|
*
|
||||||
|
* Let F(d) be the frequency of all dmers with hash value d.
|
||||||
|
* Let S_i be hash value of the dmer at position i of segment S which has length k.
|
||||||
|
*
|
||||||
|
* Score(S) = F(S_1) + F(S_2) + ... + F(S_{k-d+1})
|
||||||
|
*
|
||||||
|
* Once the dmer with hash value d is in the dictionary we set F(d) = 0.
|
||||||
|
*/
|
||||||
|
static COVER_segment_t FASTCOVER_selectSegment(const FASTCOVER_ctx_t *ctx,
|
||||||
|
U32 *freqs, U32 begin, U32 end,
|
||||||
|
ZDICT_cover_params_t parameters,
|
||||||
|
U16* segmentFreqs) {
|
||||||
|
/* Constants */
|
||||||
|
const U32 k = parameters.k;
|
||||||
|
const U32 d = parameters.d;
|
||||||
|
const U32 f = ctx->f;
|
||||||
|
const U32 dmersInK = k - d + 1;
|
||||||
|
|
||||||
|
/* Try each segment (activeSegment) and save the best (bestSegment) */
|
||||||
|
COVER_segment_t bestSegment = {0, 0, 0};
|
||||||
|
COVER_segment_t activeSegment;
|
||||||
|
|
||||||
|
/* Reset the activeDmers in the segment */
|
||||||
|
/* The activeSegment starts at the beginning of the epoch. */
|
||||||
|
activeSegment.begin = begin;
|
||||||
|
activeSegment.end = begin;
|
||||||
|
activeSegment.score = 0;
|
||||||
|
|
||||||
|
/* Slide the activeSegment through the whole epoch.
|
||||||
|
* Save the best segment in bestSegment.
|
||||||
|
*/
|
||||||
|
while (activeSegment.end < end) {
|
||||||
|
/* Get hash value of current dmer */
|
||||||
|
const size_t idx = FASTCOVER_hashPtrToIndex(ctx->samples + activeSegment.end, f, d);
|
||||||
|
|
||||||
|
/* Add frequency of this index to score if this is the first occurrence of index in active segment */
|
||||||
|
if (segmentFreqs[idx] == 0) {
|
||||||
|
activeSegment.score += freqs[idx];
|
||||||
|
}
|
||||||
|
/* Increment end of segment and segmentFreqs*/
|
||||||
|
activeSegment.end += 1;
|
||||||
|
segmentFreqs[idx] += 1;
|
||||||
|
/* If the window is now too large, drop the first position */
|
||||||
|
if (activeSegment.end - activeSegment.begin == dmersInK + 1) {
|
||||||
|
/* Get hash value of the dmer to be eliminated from active segment */
|
||||||
|
const size_t delIndex = FASTCOVER_hashPtrToIndex(ctx->samples + activeSegment.begin, f, d);
|
||||||
|
segmentFreqs[delIndex] -= 1;
|
||||||
|
/* Subtract frequency of this index from score if this is the last occurrence of this index in active segment */
|
||||||
|
if (segmentFreqs[delIndex] == 0) {
|
||||||
|
activeSegment.score -= freqs[delIndex];
|
||||||
|
}
|
||||||
|
/* Increment start of segment */
|
||||||
|
activeSegment.begin += 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* If this segment is the best so far save it */
|
||||||
|
if (activeSegment.score > bestSegment.score) {
|
||||||
|
bestSegment = activeSegment;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Zero out rest of segmentFreqs array */
|
||||||
|
while (activeSegment.begin < end) {
|
||||||
|
const size_t delIndex = FASTCOVER_hashPtrToIndex(ctx->samples + activeSegment.begin, f, d);
|
||||||
|
segmentFreqs[delIndex] -= 1;
|
||||||
|
activeSegment.begin += 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
{
|
||||||
|
/* Zero the frequency of hash value of each dmer covered by the chosen segment. */
|
||||||
|
U32 pos;
|
||||||
|
for (pos = bestSegment.begin; pos != bestSegment.end; ++pos) {
|
||||||
|
const size_t i = FASTCOVER_hashPtrToIndex(ctx->samples + pos, f, d);
|
||||||
|
freqs[i] = 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return bestSegment;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static int FASTCOVER_checkParameters(ZDICT_cover_params_t parameters,
|
||||||
|
size_t maxDictSize, unsigned f,
|
||||||
|
unsigned accel) {
|
||||||
|
/* k, d, and f are required parameters */
|
||||||
|
if (parameters.d == 0 || parameters.k == 0) {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
/* d has to be 6 or 8 */
|
||||||
|
if (parameters.d != 6 && parameters.d != 8) {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
/* k <= maxDictSize */
|
||||||
|
if (parameters.k > maxDictSize) {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
/* d <= k */
|
||||||
|
if (parameters.d > parameters.k) {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
/* 0 < f <= FASTCOVER_MAX_F*/
|
||||||
|
if (f > FASTCOVER_MAX_F || f == 0) {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
/* 0 < splitPoint <= 1 */
|
||||||
|
if (parameters.splitPoint <= 0 || parameters.splitPoint > 1) {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
/* 0 < accel <= 10 */
|
||||||
|
if (accel > 10 || accel == 0) {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Clean up a context initialized with `FASTCOVER_ctx_init()`.
|
||||||
|
*/
|
||||||
|
static void
|
||||||
|
FASTCOVER_ctx_destroy(FASTCOVER_ctx_t* ctx)
|
||||||
|
{
|
||||||
|
if (!ctx) return;
|
||||||
|
|
||||||
|
free(ctx->freqs);
|
||||||
|
ctx->freqs = NULL;
|
||||||
|
|
||||||
|
free(ctx->offsets);
|
||||||
|
ctx->offsets = NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Calculate for frequency of hash value of each dmer in ctx->samples
|
||||||
|
*/
|
||||||
|
static void
|
||||||
|
FASTCOVER_computeFrequency(U32* freqs, const FASTCOVER_ctx_t* ctx)
|
||||||
|
{
|
||||||
|
const unsigned f = ctx->f;
|
||||||
|
const unsigned d = ctx->d;
|
||||||
|
const unsigned skip = ctx->accelParams.skip;
|
||||||
|
const unsigned readLength = MAX(d, 8);
|
||||||
|
size_t i;
|
||||||
|
assert(ctx->nbTrainSamples >= 5);
|
||||||
|
assert(ctx->nbTrainSamples <= ctx->nbSamples);
|
||||||
|
for (i = 0; i < ctx->nbTrainSamples; i++) {
|
||||||
|
size_t start = ctx->offsets[i]; /* start of current dmer */
|
||||||
|
size_t const currSampleEnd = ctx->offsets[i+1];
|
||||||
|
while (start + readLength <= currSampleEnd) {
|
||||||
|
const size_t dmerIndex = FASTCOVER_hashPtrToIndex(ctx->samples + start, f, d);
|
||||||
|
freqs[dmerIndex]++;
|
||||||
|
start = start + skip + 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Prepare a context for dictionary building.
|
||||||
|
* The context is only dependent on the parameter `d` and can used multiple
|
||||||
|
* times.
|
||||||
|
* Returns 0 on success or error code on error.
|
||||||
|
* The context must be destroyed with `FASTCOVER_ctx_destroy()`.
|
||||||
|
*/
|
||||||
|
static size_t
|
||||||
|
FASTCOVER_ctx_init(FASTCOVER_ctx_t* ctx,
|
||||||
|
const void* samplesBuffer,
|
||||||
|
const size_t* samplesSizes, unsigned nbSamples,
|
||||||
|
unsigned d, double splitPoint, unsigned f,
|
||||||
|
FASTCOVER_accel_t accelParams)
|
||||||
|
{
|
||||||
|
const BYTE* const samples = (const BYTE*)samplesBuffer;
|
||||||
|
const size_t totalSamplesSize = COVER_sum(samplesSizes, nbSamples);
|
||||||
|
/* Split samples into testing and training sets */
|
||||||
|
const unsigned nbTrainSamples = splitPoint < 1.0 ? (unsigned)((double)nbSamples * splitPoint) : nbSamples;
|
||||||
|
const unsigned nbTestSamples = splitPoint < 1.0 ? nbSamples - nbTrainSamples : nbSamples;
|
||||||
|
const size_t trainingSamplesSize = splitPoint < 1.0 ? COVER_sum(samplesSizes, nbTrainSamples) : totalSamplesSize;
|
||||||
|
const size_t testSamplesSize = splitPoint < 1.0 ? COVER_sum(samplesSizes + nbTrainSamples, nbTestSamples) : totalSamplesSize;
|
||||||
|
|
||||||
|
/* Checks */
|
||||||
|
if (totalSamplesSize < MAX(d, sizeof(U64)) ||
|
||||||
|
totalSamplesSize >= (size_t)FASTCOVER_MAX_SAMPLES_SIZE) {
|
||||||
|
DISPLAYLEVEL(1, "Total samples size is too large (%u MB), maximum size is %u MB\n",
|
||||||
|
(unsigned)(totalSamplesSize >> 20), (FASTCOVER_MAX_SAMPLES_SIZE >> 20));
|
||||||
|
return ERROR(srcSize_wrong);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Check if there are at least 5 training samples */
|
||||||
|
if (nbTrainSamples < 5) {
|
||||||
|
DISPLAYLEVEL(1, "Total number of training samples is %u and is invalid\n", nbTrainSamples);
|
||||||
|
return ERROR(srcSize_wrong);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Check if there's testing sample */
|
||||||
|
if (nbTestSamples < 1) {
|
||||||
|
DISPLAYLEVEL(1, "Total number of testing samples is %u and is invalid.\n", nbTestSamples);
|
||||||
|
return ERROR(srcSize_wrong);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Zero the context */
|
||||||
|
memset(ctx, 0, sizeof(*ctx));
|
||||||
|
DISPLAYLEVEL(2, "Training on %u samples of total size %u\n", nbTrainSamples,
|
||||||
|
(unsigned)trainingSamplesSize);
|
||||||
|
DISPLAYLEVEL(2, "Testing on %u samples of total size %u\n", nbTestSamples,
|
||||||
|
(unsigned)testSamplesSize);
|
||||||
|
|
||||||
|
ctx->samples = samples;
|
||||||
|
ctx->samplesSizes = samplesSizes;
|
||||||
|
ctx->nbSamples = nbSamples;
|
||||||
|
ctx->nbTrainSamples = nbTrainSamples;
|
||||||
|
ctx->nbTestSamples = nbTestSamples;
|
||||||
|
ctx->nbDmers = trainingSamplesSize - MAX(d, sizeof(U64)) + 1;
|
||||||
|
ctx->d = d;
|
||||||
|
ctx->f = f;
|
||||||
|
ctx->accelParams = accelParams;
|
||||||
|
|
||||||
|
/* The offsets of each file */
|
||||||
|
ctx->offsets = (size_t*)calloc((nbSamples + 1), sizeof(size_t));
|
||||||
|
if (ctx->offsets == NULL) {
|
||||||
|
DISPLAYLEVEL(1, "Failed to allocate scratch buffers \n");
|
||||||
|
FASTCOVER_ctx_destroy(ctx);
|
||||||
|
return ERROR(memory_allocation);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Fill offsets from the samplesSizes */
|
||||||
|
{ U32 i;
|
||||||
|
ctx->offsets[0] = 0;
|
||||||
|
assert(nbSamples >= 5);
|
||||||
|
for (i = 1; i <= nbSamples; ++i) {
|
||||||
|
ctx->offsets[i] = ctx->offsets[i - 1] + samplesSizes[i - 1];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Initialize frequency array of size 2^f */
|
||||||
|
ctx->freqs = (U32*)calloc(((U64)1 << f), sizeof(U32));
|
||||||
|
if (ctx->freqs == NULL) {
|
||||||
|
DISPLAYLEVEL(1, "Failed to allocate frequency table \n");
|
||||||
|
FASTCOVER_ctx_destroy(ctx);
|
||||||
|
return ERROR(memory_allocation);
|
||||||
|
}
|
||||||
|
|
||||||
|
DISPLAYLEVEL(2, "Computing frequencies\n");
|
||||||
|
FASTCOVER_computeFrequency(ctx->freqs, ctx);
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Given the prepared context build the dictionary.
|
||||||
|
*/
|
||||||
|
static size_t
|
||||||
|
FASTCOVER_buildDictionary(const FASTCOVER_ctx_t* ctx,
|
||||||
|
U32* freqs,
|
||||||
|
void* dictBuffer, size_t dictBufferCapacity,
|
||||||
|
ZDICT_cover_params_t parameters,
|
||||||
|
U16* segmentFreqs)
|
||||||
|
{
|
||||||
|
BYTE *const dict = (BYTE *)dictBuffer;
|
||||||
|
size_t tail = dictBufferCapacity;
|
||||||
|
/* Divide the data into epochs. We will select one segment from each epoch. */
|
||||||
|
const COVER_epoch_info_t epochs = COVER_computeEpochs(
|
||||||
|
(U32)dictBufferCapacity, (U32)ctx->nbDmers, parameters.k, 1);
|
||||||
|
const size_t maxZeroScoreRun = 10;
|
||||||
|
size_t zeroScoreRun = 0;
|
||||||
|
size_t epoch;
|
||||||
|
DISPLAYLEVEL(2, "Breaking content into %u epochs of size %u\n",
|
||||||
|
(U32)epochs.num, (U32)epochs.size);
|
||||||
|
/* Loop through the epochs until there are no more segments or the dictionary
|
||||||
|
* is full.
|
||||||
|
*/
|
||||||
|
for (epoch = 0; tail > 0; epoch = (epoch + 1) % epochs.num) {
|
||||||
|
const U32 epochBegin = (U32)(epoch * epochs.size);
|
||||||
|
const U32 epochEnd = epochBegin + epochs.size;
|
||||||
|
size_t segmentSize;
|
||||||
|
/* Select a segment */
|
||||||
|
COVER_segment_t segment = FASTCOVER_selectSegment(
|
||||||
|
ctx, freqs, epochBegin, epochEnd, parameters, segmentFreqs);
|
||||||
|
|
||||||
|
/* If the segment covers no dmers, then we are out of content.
|
||||||
|
* There may be new content in other epochs, for continue for some time.
|
||||||
|
*/
|
||||||
|
if (segment.score == 0) {
|
||||||
|
if (++zeroScoreRun >= maxZeroScoreRun) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
zeroScoreRun = 0;
|
||||||
|
|
||||||
|
/* Trim the segment if necessary and if it is too small then we are done */
|
||||||
|
segmentSize = MIN(segment.end - segment.begin + parameters.d - 1, tail);
|
||||||
|
if (segmentSize < parameters.d) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* We fill the dictionary from the back to allow the best segments to be
|
||||||
|
* referenced with the smallest offsets.
|
||||||
|
*/
|
||||||
|
tail -= segmentSize;
|
||||||
|
memcpy(dict + tail, ctx->samples + segment.begin, segmentSize);
|
||||||
|
DISPLAYUPDATE(
|
||||||
|
2, "\r%u%% ",
|
||||||
|
(unsigned)(((dictBufferCapacity - tail) * 100) / dictBufferCapacity));
|
||||||
|
}
|
||||||
|
DISPLAYLEVEL(2, "\r%79s\r", "");
|
||||||
|
return tail;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Parameters for FASTCOVER_tryParameters().
|
||||||
|
*/
|
||||||
|
typedef struct FASTCOVER_tryParameters_data_s {
|
||||||
|
const FASTCOVER_ctx_t* ctx;
|
||||||
|
COVER_best_t* best;
|
||||||
|
size_t dictBufferCapacity;
|
||||||
|
ZDICT_cover_params_t parameters;
|
||||||
|
} FASTCOVER_tryParameters_data_t;
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Tries a set of parameters and updates the COVER_best_t with the results.
|
||||||
|
* This function is thread safe if zstd is compiled with multithreaded support.
|
||||||
|
* It takes its parameters as an *OWNING* opaque pointer to support threading.
|
||||||
|
*/
|
||||||
|
static void FASTCOVER_tryParameters(void *opaque)
|
||||||
|
{
|
||||||
|
/* Save parameters as local variables */
|
||||||
|
FASTCOVER_tryParameters_data_t *const data = (FASTCOVER_tryParameters_data_t *)opaque;
|
||||||
|
const FASTCOVER_ctx_t *const ctx = data->ctx;
|
||||||
|
const ZDICT_cover_params_t parameters = data->parameters;
|
||||||
|
size_t dictBufferCapacity = data->dictBufferCapacity;
|
||||||
|
size_t totalCompressedSize = ERROR(GENERIC);
|
||||||
|
/* Initialize array to keep track of frequency of dmer within activeSegment */
|
||||||
|
U16* segmentFreqs = (U16 *)calloc(((U64)1 << ctx->f), sizeof(U16));
|
||||||
|
/* Allocate space for hash table, dict, and freqs */
|
||||||
|
BYTE *const dict = (BYTE * const)malloc(dictBufferCapacity);
|
||||||
|
COVER_dictSelection_t selection = COVER_dictSelectionError(ERROR(GENERIC));
|
||||||
|
U32 *freqs = (U32*) malloc(((U64)1 << ctx->f) * sizeof(U32));
|
||||||
|
if (!segmentFreqs || !dict || !freqs) {
|
||||||
|
DISPLAYLEVEL(1, "Failed to allocate buffers: out of memory\n");
|
||||||
|
goto _cleanup;
|
||||||
|
}
|
||||||
|
/* Copy the frequencies because we need to modify them */
|
||||||
|
memcpy(freqs, ctx->freqs, ((U64)1 << ctx->f) * sizeof(U32));
|
||||||
|
/* Build the dictionary */
|
||||||
|
{ const size_t tail = FASTCOVER_buildDictionary(ctx, freqs, dict, dictBufferCapacity,
|
||||||
|
parameters, segmentFreqs);
|
||||||
|
|
||||||
|
const unsigned nbFinalizeSamples = (unsigned)(ctx->nbTrainSamples * ctx->accelParams.finalize / 100);
|
||||||
|
selection = COVER_selectDict(dict + tail, dictBufferCapacity - tail,
|
||||||
|
ctx->samples, ctx->samplesSizes, nbFinalizeSamples, ctx->nbTrainSamples, ctx->nbSamples, parameters, ctx->offsets,
|
||||||
|
totalCompressedSize);
|
||||||
|
|
||||||
|
if (COVER_dictSelectionIsError(selection)) {
|
||||||
|
DISPLAYLEVEL(1, "Failed to select dictionary\n");
|
||||||
|
goto _cleanup;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
_cleanup:
|
||||||
|
free(dict);
|
||||||
|
COVER_best_finish(data->best, parameters, selection);
|
||||||
|
free(data);
|
||||||
|
free(segmentFreqs);
|
||||||
|
COVER_dictSelectionFree(selection);
|
||||||
|
free(freqs);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static void
|
||||||
|
FASTCOVER_convertToCoverParams(ZDICT_fastCover_params_t fastCoverParams,
|
||||||
|
ZDICT_cover_params_t* coverParams)
|
||||||
|
{
|
||||||
|
coverParams->k = fastCoverParams.k;
|
||||||
|
coverParams->d = fastCoverParams.d;
|
||||||
|
coverParams->steps = fastCoverParams.steps;
|
||||||
|
coverParams->nbThreads = fastCoverParams.nbThreads;
|
||||||
|
coverParams->splitPoint = fastCoverParams.splitPoint;
|
||||||
|
coverParams->zParams = fastCoverParams.zParams;
|
||||||
|
coverParams->shrinkDict = fastCoverParams.shrinkDict;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static void
|
||||||
|
FASTCOVER_convertToFastCoverParams(ZDICT_cover_params_t coverParams,
|
||||||
|
ZDICT_fastCover_params_t* fastCoverParams,
|
||||||
|
unsigned f, unsigned accel)
|
||||||
|
{
|
||||||
|
fastCoverParams->k = coverParams.k;
|
||||||
|
fastCoverParams->d = coverParams.d;
|
||||||
|
fastCoverParams->steps = coverParams.steps;
|
||||||
|
fastCoverParams->nbThreads = coverParams.nbThreads;
|
||||||
|
fastCoverParams->splitPoint = coverParams.splitPoint;
|
||||||
|
fastCoverParams->f = f;
|
||||||
|
fastCoverParams->accel = accel;
|
||||||
|
fastCoverParams->zParams = coverParams.zParams;
|
||||||
|
fastCoverParams->shrinkDict = coverParams.shrinkDict;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
ZDICTLIB_API size_t
|
||||||
|
ZDICT_trainFromBuffer_fastCover(void* dictBuffer, size_t dictBufferCapacity,
|
||||||
|
const void* samplesBuffer,
|
||||||
|
const size_t* samplesSizes, unsigned nbSamples,
|
||||||
|
ZDICT_fastCover_params_t parameters)
|
||||||
|
{
|
||||||
|
BYTE* const dict = (BYTE*)dictBuffer;
|
||||||
|
FASTCOVER_ctx_t ctx;
|
||||||
|
ZDICT_cover_params_t coverParams;
|
||||||
|
FASTCOVER_accel_t accelParams;
|
||||||
|
/* Initialize global data */
|
||||||
|
g_displayLevel = parameters.zParams.notificationLevel;
|
||||||
|
/* Assign splitPoint and f if not provided */
|
||||||
|
parameters.splitPoint = 1.0;
|
||||||
|
parameters.f = parameters.f == 0 ? DEFAULT_F : parameters.f;
|
||||||
|
parameters.accel = parameters.accel == 0 ? DEFAULT_ACCEL : parameters.accel;
|
||||||
|
/* Convert to cover parameter */
|
||||||
|
memset(&coverParams, 0 , sizeof(coverParams));
|
||||||
|
FASTCOVER_convertToCoverParams(parameters, &coverParams);
|
||||||
|
/* Checks */
|
||||||
|
if (!FASTCOVER_checkParameters(coverParams, dictBufferCapacity, parameters.f,
|
||||||
|
parameters.accel)) {
|
||||||
|
DISPLAYLEVEL(1, "FASTCOVER parameters incorrect\n");
|
||||||
|
return ERROR(parameter_outOfBound);
|
||||||
|
}
|
||||||
|
if (nbSamples == 0) {
|
||||||
|
DISPLAYLEVEL(1, "FASTCOVER must have at least one input file\n");
|
||||||
|
return ERROR(srcSize_wrong);
|
||||||
|
}
|
||||||
|
if (dictBufferCapacity < ZDICT_DICTSIZE_MIN) {
|
||||||
|
DISPLAYLEVEL(1, "dictBufferCapacity must be at least %u\n",
|
||||||
|
ZDICT_DICTSIZE_MIN);
|
||||||
|
return ERROR(dstSize_tooSmall);
|
||||||
|
}
|
||||||
|
/* Assign corresponding FASTCOVER_accel_t to accelParams*/
|
||||||
|
accelParams = FASTCOVER_defaultAccelParameters[parameters.accel];
|
||||||
|
/* Initialize context */
|
||||||
|
{
|
||||||
|
size_t const initVal = FASTCOVER_ctx_init(&ctx, samplesBuffer, samplesSizes, nbSamples,
|
||||||
|
coverParams.d, parameters.splitPoint, parameters.f,
|
||||||
|
accelParams);
|
||||||
|
if (ZSTD_isError(initVal)) {
|
||||||
|
DISPLAYLEVEL(1, "Failed to initialize context\n");
|
||||||
|
return initVal;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
COVER_warnOnSmallCorpus(dictBufferCapacity, ctx.nbDmers, g_displayLevel);
|
||||||
|
/* Build the dictionary */
|
||||||
|
DISPLAYLEVEL(2, "Building dictionary\n");
|
||||||
|
{
|
||||||
|
/* Initialize array to keep track of frequency of dmer within activeSegment */
|
||||||
|
U16* segmentFreqs = (U16 *)calloc(((U64)1 << parameters.f), sizeof(U16));
|
||||||
|
const size_t tail = FASTCOVER_buildDictionary(&ctx, ctx.freqs, dictBuffer,
|
||||||
|
dictBufferCapacity, coverParams, segmentFreqs);
|
||||||
|
const unsigned nbFinalizeSamples = (unsigned)(ctx.nbTrainSamples * ctx.accelParams.finalize / 100);
|
||||||
|
const size_t dictionarySize = ZDICT_finalizeDictionary(
|
||||||
|
dict, dictBufferCapacity, dict + tail, dictBufferCapacity - tail,
|
||||||
|
samplesBuffer, samplesSizes, nbFinalizeSamples, coverParams.zParams);
|
||||||
|
if (!ZSTD_isError(dictionarySize)) {
|
||||||
|
DISPLAYLEVEL(2, "Constructed dictionary of size %u\n",
|
||||||
|
(unsigned)dictionarySize);
|
||||||
|
}
|
||||||
|
FASTCOVER_ctx_destroy(&ctx);
|
||||||
|
free(segmentFreqs);
|
||||||
|
return dictionarySize;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
ZDICTLIB_API size_t
|
||||||
|
ZDICT_optimizeTrainFromBuffer_fastCover(
|
||||||
|
void* dictBuffer, size_t dictBufferCapacity,
|
||||||
|
const void* samplesBuffer,
|
||||||
|
const size_t* samplesSizes, unsigned nbSamples,
|
||||||
|
ZDICT_fastCover_params_t* parameters)
|
||||||
|
{
|
||||||
|
ZDICT_cover_params_t coverParams;
|
||||||
|
FASTCOVER_accel_t accelParams;
|
||||||
|
/* constants */
|
||||||
|
const unsigned nbThreads = parameters->nbThreads;
|
||||||
|
const double splitPoint =
|
||||||
|
parameters->splitPoint <= 0.0 ? DEFAULT_SPLITPOINT : parameters->splitPoint;
|
||||||
|
const unsigned kMinD = parameters->d == 0 ? 6 : parameters->d;
|
||||||
|
const unsigned kMaxD = parameters->d == 0 ? 8 : parameters->d;
|
||||||
|
const unsigned kMinK = parameters->k == 0 ? 50 : parameters->k;
|
||||||
|
const unsigned kMaxK = parameters->k == 0 ? 2000 : parameters->k;
|
||||||
|
const unsigned kSteps = parameters->steps == 0 ? 40 : parameters->steps;
|
||||||
|
const unsigned kStepSize = MAX((kMaxK - kMinK) / kSteps, 1);
|
||||||
|
const unsigned kIterations =
|
||||||
|
(1 + (kMaxD - kMinD) / 2) * (1 + (kMaxK - kMinK) / kStepSize);
|
||||||
|
const unsigned f = parameters->f == 0 ? DEFAULT_F : parameters->f;
|
||||||
|
const unsigned accel = parameters->accel == 0 ? DEFAULT_ACCEL : parameters->accel;
|
||||||
|
const unsigned shrinkDict = 0;
|
||||||
|
/* Local variables */
|
||||||
|
const int displayLevel = parameters->zParams.notificationLevel;
|
||||||
|
unsigned iteration = 1;
|
||||||
|
unsigned d;
|
||||||
|
unsigned k;
|
||||||
|
COVER_best_t best;
|
||||||
|
POOL_ctx *pool = NULL;
|
||||||
|
int warned = 0;
|
||||||
|
/* Checks */
|
||||||
|
if (splitPoint <= 0 || splitPoint > 1) {
|
||||||
|
LOCALDISPLAYLEVEL(displayLevel, 1, "Incorrect splitPoint\n");
|
||||||
|
return ERROR(parameter_outOfBound);
|
||||||
|
}
|
||||||
|
if (accel == 0 || accel > FASTCOVER_MAX_ACCEL) {
|
||||||
|
LOCALDISPLAYLEVEL(displayLevel, 1, "Incorrect accel\n");
|
||||||
|
return ERROR(parameter_outOfBound);
|
||||||
|
}
|
||||||
|
if (kMinK < kMaxD || kMaxK < kMinK) {
|
||||||
|
LOCALDISPLAYLEVEL(displayLevel, 1, "Incorrect k\n");
|
||||||
|
return ERROR(parameter_outOfBound);
|
||||||
|
}
|
||||||
|
if (nbSamples == 0) {
|
||||||
|
LOCALDISPLAYLEVEL(displayLevel, 1, "FASTCOVER must have at least one input file\n");
|
||||||
|
return ERROR(srcSize_wrong);
|
||||||
|
}
|
||||||
|
if (dictBufferCapacity < ZDICT_DICTSIZE_MIN) {
|
||||||
|
LOCALDISPLAYLEVEL(displayLevel, 1, "dictBufferCapacity must be at least %u\n",
|
||||||
|
ZDICT_DICTSIZE_MIN);
|
||||||
|
return ERROR(dstSize_tooSmall);
|
||||||
|
}
|
||||||
|
if (nbThreads > 1) {
|
||||||
|
pool = POOL_create(nbThreads, 1);
|
||||||
|
if (!pool) {
|
||||||
|
return ERROR(memory_allocation);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
/* Initialization */
|
||||||
|
COVER_best_init(&best);
|
||||||
|
memset(&coverParams, 0 , sizeof(coverParams));
|
||||||
|
FASTCOVER_convertToCoverParams(*parameters, &coverParams);
|
||||||
|
accelParams = FASTCOVER_defaultAccelParameters[accel];
|
||||||
|
/* Turn down global display level to clean up display at level 2 and below */
|
||||||
|
g_displayLevel = displayLevel == 0 ? 0 : displayLevel - 1;
|
||||||
|
/* Loop through d first because each new value needs a new context */
|
||||||
|
LOCALDISPLAYLEVEL(displayLevel, 2, "Trying %u different sets of parameters\n",
|
||||||
|
kIterations);
|
||||||
|
for (d = kMinD; d <= kMaxD; d += 2) {
|
||||||
|
/* Initialize the context for this value of d */
|
||||||
|
FASTCOVER_ctx_t ctx;
|
||||||
|
LOCALDISPLAYLEVEL(displayLevel, 3, "d=%u\n", d);
|
||||||
|
{
|
||||||
|
size_t const initVal = FASTCOVER_ctx_init(&ctx, samplesBuffer, samplesSizes, nbSamples, d, splitPoint, f, accelParams);
|
||||||
|
if (ZSTD_isError(initVal)) {
|
||||||
|
LOCALDISPLAYLEVEL(displayLevel, 1, "Failed to initialize context\n");
|
||||||
|
COVER_best_destroy(&best);
|
||||||
|
POOL_free(pool);
|
||||||
|
return initVal;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (!warned) {
|
||||||
|
COVER_warnOnSmallCorpus(dictBufferCapacity, ctx.nbDmers, displayLevel);
|
||||||
|
warned = 1;
|
||||||
|
}
|
||||||
|
/* Loop through k reusing the same context */
|
||||||
|
for (k = kMinK; k <= kMaxK; k += kStepSize) {
|
||||||
|
/* Prepare the arguments */
|
||||||
|
FASTCOVER_tryParameters_data_t *data = (FASTCOVER_tryParameters_data_t *)malloc(
|
||||||
|
sizeof(FASTCOVER_tryParameters_data_t));
|
||||||
|
LOCALDISPLAYLEVEL(displayLevel, 3, "k=%u\n", k);
|
||||||
|
if (!data) {
|
||||||
|
LOCALDISPLAYLEVEL(displayLevel, 1, "Failed to allocate parameters\n");
|
||||||
|
COVER_best_destroy(&best);
|
||||||
|
FASTCOVER_ctx_destroy(&ctx);
|
||||||
|
POOL_free(pool);
|
||||||
|
return ERROR(memory_allocation);
|
||||||
|
}
|
||||||
|
data->ctx = &ctx;
|
||||||
|
data->best = &best;
|
||||||
|
data->dictBufferCapacity = dictBufferCapacity;
|
||||||
|
data->parameters = coverParams;
|
||||||
|
data->parameters.k = k;
|
||||||
|
data->parameters.d = d;
|
||||||
|
data->parameters.splitPoint = splitPoint;
|
||||||
|
data->parameters.steps = kSteps;
|
||||||
|
data->parameters.shrinkDict = shrinkDict;
|
||||||
|
data->parameters.zParams.notificationLevel = g_displayLevel;
|
||||||
|
/* Check the parameters */
|
||||||
|
if (!FASTCOVER_checkParameters(data->parameters, dictBufferCapacity,
|
||||||
|
data->ctx->f, accel)) {
|
||||||
|
DISPLAYLEVEL(1, "FASTCOVER parameters incorrect\n");
|
||||||
|
free(data);
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
/* Call the function and pass ownership of data to it */
|
||||||
|
COVER_best_start(&best);
|
||||||
|
if (pool) {
|
||||||
|
POOL_add(pool, &FASTCOVER_tryParameters, data);
|
||||||
|
} else {
|
||||||
|
FASTCOVER_tryParameters(data);
|
||||||
|
}
|
||||||
|
/* Print status */
|
||||||
|
LOCALDISPLAYUPDATE(displayLevel, 2, "\r%u%% ",
|
||||||
|
(unsigned)((iteration * 100) / kIterations));
|
||||||
|
++iteration;
|
||||||
|
}
|
||||||
|
COVER_best_wait(&best);
|
||||||
|
FASTCOVER_ctx_destroy(&ctx);
|
||||||
|
}
|
||||||
|
LOCALDISPLAYLEVEL(displayLevel, 2, "\r%79s\r", "");
|
||||||
|
/* Fill the output buffer and parameters with output of the best parameters */
|
||||||
|
{
|
||||||
|
const size_t dictSize = best.dictSize;
|
||||||
|
if (ZSTD_isError(best.compressedSize)) {
|
||||||
|
const size_t compressedSize = best.compressedSize;
|
||||||
|
COVER_best_destroy(&best);
|
||||||
|
POOL_free(pool);
|
||||||
|
return compressedSize;
|
||||||
|
}
|
||||||
|
FASTCOVER_convertToFastCoverParams(best.parameters, parameters, f, accel);
|
||||||
|
memcpy(dictBuffer, best.dict, dictSize);
|
||||||
|
COVER_best_destroy(&best);
|
||||||
|
POOL_free(pool);
|
||||||
|
return dictSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
708
vendor/github.com/DataDog/zstd/fse.h
generated
vendored
Normal file
708
vendor/github.com/DataDog/zstd/fse.h
generated
vendored
Normal file
@@ -0,0 +1,708 @@
|
|||||||
|
/* ******************************************************************
|
||||||
|
FSE : Finite State Entropy codec
|
||||||
|
Public Prototypes declaration
|
||||||
|
Copyright (C) 2013-2016, Yann Collet.
|
||||||
|
|
||||||
|
BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
You can contact the author at :
|
||||||
|
- Source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
||||||
|
****************************************************************** */
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#ifndef FSE_H
|
||||||
|
#define FSE_H
|
||||||
|
|
||||||
|
|
||||||
|
/*-*****************************************
|
||||||
|
* Dependencies
|
||||||
|
******************************************/
|
||||||
|
#include <stddef.h> /* size_t, ptrdiff_t */
|
||||||
|
|
||||||
|
|
||||||
|
/*-*****************************************
|
||||||
|
* FSE_PUBLIC_API : control library symbols visibility
|
||||||
|
******************************************/
|
||||||
|
#if defined(FSE_DLL_EXPORT) && (FSE_DLL_EXPORT==1) && defined(__GNUC__) && (__GNUC__ >= 4)
|
||||||
|
# define FSE_PUBLIC_API __attribute__ ((visibility ("default")))
|
||||||
|
#elif defined(FSE_DLL_EXPORT) && (FSE_DLL_EXPORT==1) /* Visual expected */
|
||||||
|
# define FSE_PUBLIC_API __declspec(dllexport)
|
||||||
|
#elif defined(FSE_DLL_IMPORT) && (FSE_DLL_IMPORT==1)
|
||||||
|
# define FSE_PUBLIC_API __declspec(dllimport) /* It isn't required but allows to generate better code, saving a function pointer load from the IAT and an indirect jump.*/
|
||||||
|
#else
|
||||||
|
# define FSE_PUBLIC_API
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*------ Version ------*/
|
||||||
|
#define FSE_VERSION_MAJOR 0
|
||||||
|
#define FSE_VERSION_MINOR 9
|
||||||
|
#define FSE_VERSION_RELEASE 0
|
||||||
|
|
||||||
|
#define FSE_LIB_VERSION FSE_VERSION_MAJOR.FSE_VERSION_MINOR.FSE_VERSION_RELEASE
|
||||||
|
#define FSE_QUOTE(str) #str
|
||||||
|
#define FSE_EXPAND_AND_QUOTE(str) FSE_QUOTE(str)
|
||||||
|
#define FSE_VERSION_STRING FSE_EXPAND_AND_QUOTE(FSE_LIB_VERSION)
|
||||||
|
|
||||||
|
#define FSE_VERSION_NUMBER (FSE_VERSION_MAJOR *100*100 + FSE_VERSION_MINOR *100 + FSE_VERSION_RELEASE)
|
||||||
|
FSE_PUBLIC_API unsigned FSE_versionNumber(void); /**< library version number; to be used when checking dll version */
|
||||||
|
|
||||||
|
|
||||||
|
/*-****************************************
|
||||||
|
* FSE simple functions
|
||||||
|
******************************************/
|
||||||
|
/*! FSE_compress() :
|
||||||
|
Compress content of buffer 'src', of size 'srcSize', into destination buffer 'dst'.
|
||||||
|
'dst' buffer must be already allocated. Compression runs faster is dstCapacity >= FSE_compressBound(srcSize).
|
||||||
|
@return : size of compressed data (<= dstCapacity).
|
||||||
|
Special values : if return == 0, srcData is not compressible => Nothing is stored within dst !!!
|
||||||
|
if return == 1, srcData is a single byte symbol * srcSize times. Use RLE compression instead.
|
||||||
|
if FSE_isError(return), compression failed (more details using FSE_getErrorName())
|
||||||
|
*/
|
||||||
|
FSE_PUBLIC_API size_t FSE_compress(void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
/*! FSE_decompress():
|
||||||
|
Decompress FSE data from buffer 'cSrc', of size 'cSrcSize',
|
||||||
|
into already allocated destination buffer 'dst', of size 'dstCapacity'.
|
||||||
|
@return : size of regenerated data (<= maxDstSize),
|
||||||
|
or an error code, which can be tested using FSE_isError() .
|
||||||
|
|
||||||
|
** Important ** : FSE_decompress() does not decompress non-compressible nor RLE data !!!
|
||||||
|
Why ? : making this distinction requires a header.
|
||||||
|
Header management is intentionally delegated to the user layer, which can better manage special cases.
|
||||||
|
*/
|
||||||
|
FSE_PUBLIC_API size_t FSE_decompress(void* dst, size_t dstCapacity,
|
||||||
|
const void* cSrc, size_t cSrcSize);
|
||||||
|
|
||||||
|
|
||||||
|
/*-*****************************************
|
||||||
|
* Tool functions
|
||||||
|
******************************************/
|
||||||
|
FSE_PUBLIC_API size_t FSE_compressBound(size_t size); /* maximum compressed size */
|
||||||
|
|
||||||
|
/* Error Management */
|
||||||
|
FSE_PUBLIC_API unsigned FSE_isError(size_t code); /* tells if a return value is an error code */
|
||||||
|
FSE_PUBLIC_API const char* FSE_getErrorName(size_t code); /* provides error code string (useful for debugging) */
|
||||||
|
|
||||||
|
|
||||||
|
/*-*****************************************
|
||||||
|
* FSE advanced functions
|
||||||
|
******************************************/
|
||||||
|
/*! FSE_compress2() :
|
||||||
|
Same as FSE_compress(), but allows the selection of 'maxSymbolValue' and 'tableLog'
|
||||||
|
Both parameters can be defined as '0' to mean : use default value
|
||||||
|
@return : size of compressed data
|
||||||
|
Special values : if return == 0, srcData is not compressible => Nothing is stored within cSrc !!!
|
||||||
|
if return == 1, srcData is a single byte symbol * srcSize times. Use RLE compression.
|
||||||
|
if FSE_isError(return), it's an error code.
|
||||||
|
*/
|
||||||
|
FSE_PUBLIC_API size_t FSE_compress2 (void* dst, size_t dstSize, const void* src, size_t srcSize, unsigned maxSymbolValue, unsigned tableLog);
|
||||||
|
|
||||||
|
|
||||||
|
/*-*****************************************
|
||||||
|
* FSE detailed API
|
||||||
|
******************************************/
|
||||||
|
/*!
|
||||||
|
FSE_compress() does the following:
|
||||||
|
1. count symbol occurrence from source[] into table count[] (see hist.h)
|
||||||
|
2. normalize counters so that sum(count[]) == Power_of_2 (2^tableLog)
|
||||||
|
3. save normalized counters to memory buffer using writeNCount()
|
||||||
|
4. build encoding table 'CTable' from normalized counters
|
||||||
|
5. encode the data stream using encoding table 'CTable'
|
||||||
|
|
||||||
|
FSE_decompress() does the following:
|
||||||
|
1. read normalized counters with readNCount()
|
||||||
|
2. build decoding table 'DTable' from normalized counters
|
||||||
|
3. decode the data stream using decoding table 'DTable'
|
||||||
|
|
||||||
|
The following API allows targeting specific sub-functions for advanced tasks.
|
||||||
|
For example, it's possible to compress several blocks using the same 'CTable',
|
||||||
|
or to save and provide normalized distribution using external method.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* *** COMPRESSION *** */
|
||||||
|
|
||||||
|
/*! FSE_optimalTableLog():
|
||||||
|
dynamically downsize 'tableLog' when conditions are met.
|
||||||
|
It saves CPU time, by using smaller tables, while preserving or even improving compression ratio.
|
||||||
|
@return : recommended tableLog (necessarily <= 'maxTableLog') */
|
||||||
|
FSE_PUBLIC_API unsigned FSE_optimalTableLog(unsigned maxTableLog, size_t srcSize, unsigned maxSymbolValue);
|
||||||
|
|
||||||
|
/*! FSE_normalizeCount():
|
||||||
|
normalize counts so that sum(count[]) == Power_of_2 (2^tableLog)
|
||||||
|
'normalizedCounter' is a table of short, of minimum size (maxSymbolValue+1).
|
||||||
|
@return : tableLog,
|
||||||
|
or an errorCode, which can be tested using FSE_isError() */
|
||||||
|
FSE_PUBLIC_API size_t FSE_normalizeCount(short* normalizedCounter, unsigned tableLog,
|
||||||
|
const unsigned* count, size_t srcSize, unsigned maxSymbolValue);
|
||||||
|
|
||||||
|
/*! FSE_NCountWriteBound():
|
||||||
|
Provides the maximum possible size of an FSE normalized table, given 'maxSymbolValue' and 'tableLog'.
|
||||||
|
Typically useful for allocation purpose. */
|
||||||
|
FSE_PUBLIC_API size_t FSE_NCountWriteBound(unsigned maxSymbolValue, unsigned tableLog);
|
||||||
|
|
||||||
|
/*! FSE_writeNCount():
|
||||||
|
Compactly save 'normalizedCounter' into 'buffer'.
|
||||||
|
@return : size of the compressed table,
|
||||||
|
or an errorCode, which can be tested using FSE_isError(). */
|
||||||
|
FSE_PUBLIC_API size_t FSE_writeNCount (void* buffer, size_t bufferSize,
|
||||||
|
const short* normalizedCounter,
|
||||||
|
unsigned maxSymbolValue, unsigned tableLog);
|
||||||
|
|
||||||
|
/*! Constructor and Destructor of FSE_CTable.
|
||||||
|
Note that FSE_CTable size depends on 'tableLog' and 'maxSymbolValue' */
|
||||||
|
typedef unsigned FSE_CTable; /* don't allocate that. It's only meant to be more restrictive than void* */
|
||||||
|
FSE_PUBLIC_API FSE_CTable* FSE_createCTable (unsigned maxSymbolValue, unsigned tableLog);
|
||||||
|
FSE_PUBLIC_API void FSE_freeCTable (FSE_CTable* ct);
|
||||||
|
|
||||||
|
/*! FSE_buildCTable():
|
||||||
|
Builds `ct`, which must be already allocated, using FSE_createCTable().
|
||||||
|
@return : 0, or an errorCode, which can be tested using FSE_isError() */
|
||||||
|
FSE_PUBLIC_API size_t FSE_buildCTable(FSE_CTable* ct, const short* normalizedCounter, unsigned maxSymbolValue, unsigned tableLog);
|
||||||
|
|
||||||
|
/*! FSE_compress_usingCTable():
|
||||||
|
Compress `src` using `ct` into `dst` which must be already allocated.
|
||||||
|
@return : size of compressed data (<= `dstCapacity`),
|
||||||
|
or 0 if compressed data could not fit into `dst`,
|
||||||
|
or an errorCode, which can be tested using FSE_isError() */
|
||||||
|
FSE_PUBLIC_API size_t FSE_compress_usingCTable (void* dst, size_t dstCapacity, const void* src, size_t srcSize, const FSE_CTable* ct);
|
||||||
|
|
||||||
|
/*!
|
||||||
|
Tutorial :
|
||||||
|
----------
|
||||||
|
The first step is to count all symbols. FSE_count() does this job very fast.
|
||||||
|
Result will be saved into 'count', a table of unsigned int, which must be already allocated, and have 'maxSymbolValuePtr[0]+1' cells.
|
||||||
|
'src' is a table of bytes of size 'srcSize'. All values within 'src' MUST be <= maxSymbolValuePtr[0]
|
||||||
|
maxSymbolValuePtr[0] will be updated, with its real value (necessarily <= original value)
|
||||||
|
FSE_count() will return the number of occurrence of the most frequent symbol.
|
||||||
|
This can be used to know if there is a single symbol within 'src', and to quickly evaluate its compressibility.
|
||||||
|
If there is an error, the function will return an ErrorCode (which can be tested using FSE_isError()).
|
||||||
|
|
||||||
|
The next step is to normalize the frequencies.
|
||||||
|
FSE_normalizeCount() will ensure that sum of frequencies is == 2 ^'tableLog'.
|
||||||
|
It also guarantees a minimum of 1 to any Symbol with frequency >= 1.
|
||||||
|
You can use 'tableLog'==0 to mean "use default tableLog value".
|
||||||
|
If you are unsure of which tableLog value to use, you can ask FSE_optimalTableLog(),
|
||||||
|
which will provide the optimal valid tableLog given sourceSize, maxSymbolValue, and a user-defined maximum (0 means "default").
|
||||||
|
|
||||||
|
The result of FSE_normalizeCount() will be saved into a table,
|
||||||
|
called 'normalizedCounter', which is a table of signed short.
|
||||||
|
'normalizedCounter' must be already allocated, and have at least 'maxSymbolValue+1' cells.
|
||||||
|
The return value is tableLog if everything proceeded as expected.
|
||||||
|
It is 0 if there is a single symbol within distribution.
|
||||||
|
If there is an error (ex: invalid tableLog value), the function will return an ErrorCode (which can be tested using FSE_isError()).
|
||||||
|
|
||||||
|
'normalizedCounter' can be saved in a compact manner to a memory area using FSE_writeNCount().
|
||||||
|
'buffer' must be already allocated.
|
||||||
|
For guaranteed success, buffer size must be at least FSE_headerBound().
|
||||||
|
The result of the function is the number of bytes written into 'buffer'.
|
||||||
|
If there is an error, the function will return an ErrorCode (which can be tested using FSE_isError(); ex : buffer size too small).
|
||||||
|
|
||||||
|
'normalizedCounter' can then be used to create the compression table 'CTable'.
|
||||||
|
The space required by 'CTable' must be already allocated, using FSE_createCTable().
|
||||||
|
You can then use FSE_buildCTable() to fill 'CTable'.
|
||||||
|
If there is an error, both functions will return an ErrorCode (which can be tested using FSE_isError()).
|
||||||
|
|
||||||
|
'CTable' can then be used to compress 'src', with FSE_compress_usingCTable().
|
||||||
|
Similar to FSE_count(), the convention is that 'src' is assumed to be a table of char of size 'srcSize'
|
||||||
|
The function returns the size of compressed data (without header), necessarily <= `dstCapacity`.
|
||||||
|
If it returns '0', compressed data could not fit into 'dst'.
|
||||||
|
If there is an error, the function will return an ErrorCode (which can be tested using FSE_isError()).
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/* *** DECOMPRESSION *** */
|
||||||
|
|
||||||
|
/*! FSE_readNCount():
|
||||||
|
Read compactly saved 'normalizedCounter' from 'rBuffer'.
|
||||||
|
@return : size read from 'rBuffer',
|
||||||
|
or an errorCode, which can be tested using FSE_isError().
|
||||||
|
maxSymbolValuePtr[0] and tableLogPtr[0] will also be updated with their respective values */
|
||||||
|
FSE_PUBLIC_API size_t FSE_readNCount (short* normalizedCounter,
|
||||||
|
unsigned* maxSymbolValuePtr, unsigned* tableLogPtr,
|
||||||
|
const void* rBuffer, size_t rBuffSize);
|
||||||
|
|
||||||
|
/*! Constructor and Destructor of FSE_DTable.
|
||||||
|
Note that its size depends on 'tableLog' */
|
||||||
|
typedef unsigned FSE_DTable; /* don't allocate that. It's just a way to be more restrictive than void* */
|
||||||
|
FSE_PUBLIC_API FSE_DTable* FSE_createDTable(unsigned tableLog);
|
||||||
|
FSE_PUBLIC_API void FSE_freeDTable(FSE_DTable* dt);
|
||||||
|
|
||||||
|
/*! FSE_buildDTable():
|
||||||
|
Builds 'dt', which must be already allocated, using FSE_createDTable().
|
||||||
|
return : 0, or an errorCode, which can be tested using FSE_isError() */
|
||||||
|
FSE_PUBLIC_API size_t FSE_buildDTable (FSE_DTable* dt, const short* normalizedCounter, unsigned maxSymbolValue, unsigned tableLog);
|
||||||
|
|
||||||
|
/*! FSE_decompress_usingDTable():
|
||||||
|
Decompress compressed source `cSrc` of size `cSrcSize` using `dt`
|
||||||
|
into `dst` which must be already allocated.
|
||||||
|
@return : size of regenerated data (necessarily <= `dstCapacity`),
|
||||||
|
or an errorCode, which can be tested using FSE_isError() */
|
||||||
|
FSE_PUBLIC_API size_t FSE_decompress_usingDTable(void* dst, size_t dstCapacity, const void* cSrc, size_t cSrcSize, const FSE_DTable* dt);
|
||||||
|
|
||||||
|
/*!
|
||||||
|
Tutorial :
|
||||||
|
----------
|
||||||
|
(Note : these functions only decompress FSE-compressed blocks.
|
||||||
|
If block is uncompressed, use memcpy() instead
|
||||||
|
If block is a single repeated byte, use memset() instead )
|
||||||
|
|
||||||
|
The first step is to obtain the normalized frequencies of symbols.
|
||||||
|
This can be performed by FSE_readNCount() if it was saved using FSE_writeNCount().
|
||||||
|
'normalizedCounter' must be already allocated, and have at least 'maxSymbolValuePtr[0]+1' cells of signed short.
|
||||||
|
In practice, that means it's necessary to know 'maxSymbolValue' beforehand,
|
||||||
|
or size the table to handle worst case situations (typically 256).
|
||||||
|
FSE_readNCount() will provide 'tableLog' and 'maxSymbolValue'.
|
||||||
|
The result of FSE_readNCount() is the number of bytes read from 'rBuffer'.
|
||||||
|
Note that 'rBufferSize' must be at least 4 bytes, even if useful information is less than that.
|
||||||
|
If there is an error, the function will return an error code, which can be tested using FSE_isError().
|
||||||
|
|
||||||
|
The next step is to build the decompression tables 'FSE_DTable' from 'normalizedCounter'.
|
||||||
|
This is performed by the function FSE_buildDTable().
|
||||||
|
The space required by 'FSE_DTable' must be already allocated using FSE_createDTable().
|
||||||
|
If there is an error, the function will return an error code, which can be tested using FSE_isError().
|
||||||
|
|
||||||
|
`FSE_DTable` can then be used to decompress `cSrc`, with FSE_decompress_usingDTable().
|
||||||
|
`cSrcSize` must be strictly correct, otherwise decompression will fail.
|
||||||
|
FSE_decompress_usingDTable() result will tell how many bytes were regenerated (<=`dstCapacity`).
|
||||||
|
If there is an error, the function will return an error code, which can be tested using FSE_isError(). (ex: dst buffer too small)
|
||||||
|
*/
|
||||||
|
|
||||||
|
#endif /* FSE_H */
|
||||||
|
|
||||||
|
#if defined(FSE_STATIC_LINKING_ONLY) && !defined(FSE_H_FSE_STATIC_LINKING_ONLY)
|
||||||
|
#define FSE_H_FSE_STATIC_LINKING_ONLY
|
||||||
|
|
||||||
|
/* *** Dependency *** */
|
||||||
|
#include "bitstream.h"
|
||||||
|
|
||||||
|
|
||||||
|
/* *****************************************
|
||||||
|
* Static allocation
|
||||||
|
*******************************************/
|
||||||
|
/* FSE buffer bounds */
|
||||||
|
#define FSE_NCOUNTBOUND 512
|
||||||
|
#define FSE_BLOCKBOUND(size) (size + (size>>7) + 4 /* fse states */ + sizeof(size_t) /* bitContainer */)
|
||||||
|
#define FSE_COMPRESSBOUND(size) (FSE_NCOUNTBOUND + FSE_BLOCKBOUND(size)) /* Macro version, useful for static allocation */
|
||||||
|
|
||||||
|
/* It is possible to statically allocate FSE CTable/DTable as a table of FSE_CTable/FSE_DTable using below macros */
|
||||||
|
#define FSE_CTABLE_SIZE_U32(maxTableLog, maxSymbolValue) (1 + (1<<(maxTableLog-1)) + ((maxSymbolValue+1)*2))
|
||||||
|
#define FSE_DTABLE_SIZE_U32(maxTableLog) (1 + (1<<maxTableLog))
|
||||||
|
|
||||||
|
/* or use the size to malloc() space directly. Pay attention to alignment restrictions though */
|
||||||
|
#define FSE_CTABLE_SIZE(maxTableLog, maxSymbolValue) (FSE_CTABLE_SIZE_U32(maxTableLog, maxSymbolValue) * sizeof(FSE_CTable))
|
||||||
|
#define FSE_DTABLE_SIZE(maxTableLog) (FSE_DTABLE_SIZE_U32(maxTableLog) * sizeof(FSE_DTable))
|
||||||
|
|
||||||
|
|
||||||
|
/* *****************************************
|
||||||
|
* FSE advanced API
|
||||||
|
***************************************** */
|
||||||
|
|
||||||
|
unsigned FSE_optimalTableLog_internal(unsigned maxTableLog, size_t srcSize, unsigned maxSymbolValue, unsigned minus);
|
||||||
|
/**< same as FSE_optimalTableLog(), which used `minus==2` */
|
||||||
|
|
||||||
|
/* FSE_compress_wksp() :
|
||||||
|
* Same as FSE_compress2(), but using an externally allocated scratch buffer (`workSpace`).
|
||||||
|
* FSE_WKSP_SIZE_U32() provides the minimum size required for `workSpace` as a table of FSE_CTable.
|
||||||
|
*/
|
||||||
|
#define FSE_WKSP_SIZE_U32(maxTableLog, maxSymbolValue) ( FSE_CTABLE_SIZE_U32(maxTableLog, maxSymbolValue) + ((maxTableLog > 12) ? (1 << (maxTableLog - 2)) : 1024) )
|
||||||
|
size_t FSE_compress_wksp (void* dst, size_t dstSize, const void* src, size_t srcSize, unsigned maxSymbolValue, unsigned tableLog, void* workSpace, size_t wkspSize);
|
||||||
|
|
||||||
|
size_t FSE_buildCTable_raw (FSE_CTable* ct, unsigned nbBits);
|
||||||
|
/**< build a fake FSE_CTable, designed for a flat distribution, where each symbol uses nbBits */
|
||||||
|
|
||||||
|
size_t FSE_buildCTable_rle (FSE_CTable* ct, unsigned char symbolValue);
|
||||||
|
/**< build a fake FSE_CTable, designed to compress always the same symbolValue */
|
||||||
|
|
||||||
|
/* FSE_buildCTable_wksp() :
|
||||||
|
* Same as FSE_buildCTable(), but using an externally allocated scratch buffer (`workSpace`).
|
||||||
|
* `wkspSize` must be >= `(1<<tableLog)`.
|
||||||
|
*/
|
||||||
|
size_t FSE_buildCTable_wksp(FSE_CTable* ct, const short* normalizedCounter, unsigned maxSymbolValue, unsigned tableLog, void* workSpace, size_t wkspSize);
|
||||||
|
|
||||||
|
size_t FSE_buildDTable_raw (FSE_DTable* dt, unsigned nbBits);
|
||||||
|
/**< build a fake FSE_DTable, designed to read a flat distribution where each symbol uses nbBits */
|
||||||
|
|
||||||
|
size_t FSE_buildDTable_rle (FSE_DTable* dt, unsigned char symbolValue);
|
||||||
|
/**< build a fake FSE_DTable, designed to always generate the same symbolValue */
|
||||||
|
|
||||||
|
size_t FSE_decompress_wksp(void* dst, size_t dstCapacity, const void* cSrc, size_t cSrcSize, FSE_DTable* workSpace, unsigned maxLog);
|
||||||
|
/**< same as FSE_decompress(), using an externally allocated `workSpace` produced with `FSE_DTABLE_SIZE_U32(maxLog)` */
|
||||||
|
|
||||||
|
typedef enum {
|
||||||
|
FSE_repeat_none, /**< Cannot use the previous table */
|
||||||
|
FSE_repeat_check, /**< Can use the previous table but it must be checked */
|
||||||
|
FSE_repeat_valid /**< Can use the previous table and it is assumed to be valid */
|
||||||
|
} FSE_repeat;
|
||||||
|
|
||||||
|
/* *****************************************
|
||||||
|
* FSE symbol compression API
|
||||||
|
*******************************************/
|
||||||
|
/*!
|
||||||
|
This API consists of small unitary functions, which highly benefit from being inlined.
|
||||||
|
Hence their body are included in next section.
|
||||||
|
*/
|
||||||
|
typedef struct {
|
||||||
|
ptrdiff_t value;
|
||||||
|
const void* stateTable;
|
||||||
|
const void* symbolTT;
|
||||||
|
unsigned stateLog;
|
||||||
|
} FSE_CState_t;
|
||||||
|
|
||||||
|
static void FSE_initCState(FSE_CState_t* CStatePtr, const FSE_CTable* ct);
|
||||||
|
|
||||||
|
static void FSE_encodeSymbol(BIT_CStream_t* bitC, FSE_CState_t* CStatePtr, unsigned symbol);
|
||||||
|
|
||||||
|
static void FSE_flushCState(BIT_CStream_t* bitC, const FSE_CState_t* CStatePtr);
|
||||||
|
|
||||||
|
/**<
|
||||||
|
These functions are inner components of FSE_compress_usingCTable().
|
||||||
|
They allow the creation of custom streams, mixing multiple tables and bit sources.
|
||||||
|
|
||||||
|
A key property to keep in mind is that encoding and decoding are done **in reverse direction**.
|
||||||
|
So the first symbol you will encode is the last you will decode, like a LIFO stack.
|
||||||
|
|
||||||
|
You will need a few variables to track your CStream. They are :
|
||||||
|
|
||||||
|
FSE_CTable ct; // Provided by FSE_buildCTable()
|
||||||
|
BIT_CStream_t bitStream; // bitStream tracking structure
|
||||||
|
FSE_CState_t state; // State tracking structure (can have several)
|
||||||
|
|
||||||
|
|
||||||
|
The first thing to do is to init bitStream and state.
|
||||||
|
size_t errorCode = BIT_initCStream(&bitStream, dstBuffer, maxDstSize);
|
||||||
|
FSE_initCState(&state, ct);
|
||||||
|
|
||||||
|
Note that BIT_initCStream() can produce an error code, so its result should be tested, using FSE_isError();
|
||||||
|
You can then encode your input data, byte after byte.
|
||||||
|
FSE_encodeSymbol() outputs a maximum of 'tableLog' bits at a time.
|
||||||
|
Remember decoding will be done in reverse direction.
|
||||||
|
FSE_encodeByte(&bitStream, &state, symbol);
|
||||||
|
|
||||||
|
At any time, you can also add any bit sequence.
|
||||||
|
Note : maximum allowed nbBits is 25, for compatibility with 32-bits decoders
|
||||||
|
BIT_addBits(&bitStream, bitField, nbBits);
|
||||||
|
|
||||||
|
The above methods don't commit data to memory, they just store it into local register, for speed.
|
||||||
|
Local register size is 64-bits on 64-bits systems, 32-bits on 32-bits systems (size_t).
|
||||||
|
Writing data to memory is a manual operation, performed by the flushBits function.
|
||||||
|
BIT_flushBits(&bitStream);
|
||||||
|
|
||||||
|
Your last FSE encoding operation shall be to flush your last state value(s).
|
||||||
|
FSE_flushState(&bitStream, &state);
|
||||||
|
|
||||||
|
Finally, you must close the bitStream.
|
||||||
|
The function returns the size of CStream in bytes.
|
||||||
|
If data couldn't fit into dstBuffer, it will return a 0 ( == not compressible)
|
||||||
|
If there is an error, it returns an errorCode (which can be tested using FSE_isError()).
|
||||||
|
size_t size = BIT_closeCStream(&bitStream);
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/* *****************************************
|
||||||
|
* FSE symbol decompression API
|
||||||
|
*******************************************/
|
||||||
|
typedef struct {
|
||||||
|
size_t state;
|
||||||
|
const void* table; /* precise table may vary, depending on U16 */
|
||||||
|
} FSE_DState_t;
|
||||||
|
|
||||||
|
|
||||||
|
static void FSE_initDState(FSE_DState_t* DStatePtr, BIT_DStream_t* bitD, const FSE_DTable* dt);
|
||||||
|
|
||||||
|
static unsigned char FSE_decodeSymbol(FSE_DState_t* DStatePtr, BIT_DStream_t* bitD);
|
||||||
|
|
||||||
|
static unsigned FSE_endOfDState(const FSE_DState_t* DStatePtr);
|
||||||
|
|
||||||
|
/**<
|
||||||
|
Let's now decompose FSE_decompress_usingDTable() into its unitary components.
|
||||||
|
You will decode FSE-encoded symbols from the bitStream,
|
||||||
|
and also any other bitFields you put in, **in reverse order**.
|
||||||
|
|
||||||
|
You will need a few variables to track your bitStream. They are :
|
||||||
|
|
||||||
|
BIT_DStream_t DStream; // Stream context
|
||||||
|
FSE_DState_t DState; // State context. Multiple ones are possible
|
||||||
|
FSE_DTable* DTablePtr; // Decoding table, provided by FSE_buildDTable()
|
||||||
|
|
||||||
|
The first thing to do is to init the bitStream.
|
||||||
|
errorCode = BIT_initDStream(&DStream, srcBuffer, srcSize);
|
||||||
|
|
||||||
|
You should then retrieve your initial state(s)
|
||||||
|
(in reverse flushing order if you have several ones) :
|
||||||
|
errorCode = FSE_initDState(&DState, &DStream, DTablePtr);
|
||||||
|
|
||||||
|
You can then decode your data, symbol after symbol.
|
||||||
|
For information the maximum number of bits read by FSE_decodeSymbol() is 'tableLog'.
|
||||||
|
Keep in mind that symbols are decoded in reverse order, like a LIFO stack (last in, first out).
|
||||||
|
unsigned char symbol = FSE_decodeSymbol(&DState, &DStream);
|
||||||
|
|
||||||
|
You can retrieve any bitfield you eventually stored into the bitStream (in reverse order)
|
||||||
|
Note : maximum allowed nbBits is 25, for 32-bits compatibility
|
||||||
|
size_t bitField = BIT_readBits(&DStream, nbBits);
|
||||||
|
|
||||||
|
All above operations only read from local register (which size depends on size_t).
|
||||||
|
Refueling the register from memory is manually performed by the reload method.
|
||||||
|
endSignal = FSE_reloadDStream(&DStream);
|
||||||
|
|
||||||
|
BIT_reloadDStream() result tells if there is still some more data to read from DStream.
|
||||||
|
BIT_DStream_unfinished : there is still some data left into the DStream.
|
||||||
|
BIT_DStream_endOfBuffer : Dstream reached end of buffer. Its container may no longer be completely filled.
|
||||||
|
BIT_DStream_completed : Dstream reached its exact end, corresponding in general to decompression completed.
|
||||||
|
BIT_DStream_tooFar : Dstream went too far. Decompression result is corrupted.
|
||||||
|
|
||||||
|
When reaching end of buffer (BIT_DStream_endOfBuffer), progress slowly, notably if you decode multiple symbols per loop,
|
||||||
|
to properly detect the exact end of stream.
|
||||||
|
After each decoded symbol, check if DStream is fully consumed using this simple test :
|
||||||
|
BIT_reloadDStream(&DStream) >= BIT_DStream_completed
|
||||||
|
|
||||||
|
When it's done, verify decompression is fully completed, by checking both DStream and the relevant states.
|
||||||
|
Checking if DStream has reached its end is performed by :
|
||||||
|
BIT_endOfDStream(&DStream);
|
||||||
|
Check also the states. There might be some symbols left there, if some high probability ones (>50%) are possible.
|
||||||
|
FSE_endOfDState(&DState);
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/* *****************************************
|
||||||
|
* FSE unsafe API
|
||||||
|
*******************************************/
|
||||||
|
static unsigned char FSE_decodeSymbolFast(FSE_DState_t* DStatePtr, BIT_DStream_t* bitD);
|
||||||
|
/* faster, but works only if nbBits is always >= 1 (otherwise, result will be corrupted) */
|
||||||
|
|
||||||
|
|
||||||
|
/* *****************************************
|
||||||
|
* Implementation of inlined functions
|
||||||
|
*******************************************/
|
||||||
|
typedef struct {
|
||||||
|
int deltaFindState;
|
||||||
|
U32 deltaNbBits;
|
||||||
|
} FSE_symbolCompressionTransform; /* total 8 bytes */
|
||||||
|
|
||||||
|
MEM_STATIC void FSE_initCState(FSE_CState_t* statePtr, const FSE_CTable* ct)
|
||||||
|
{
|
||||||
|
const void* ptr = ct;
|
||||||
|
const U16* u16ptr = (const U16*) ptr;
|
||||||
|
const U32 tableLog = MEM_read16(ptr);
|
||||||
|
statePtr->value = (ptrdiff_t)1<<tableLog;
|
||||||
|
statePtr->stateTable = u16ptr+2;
|
||||||
|
statePtr->symbolTT = ct + 1 + (tableLog ? (1<<(tableLog-1)) : 1);
|
||||||
|
statePtr->stateLog = tableLog;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/*! FSE_initCState2() :
|
||||||
|
* Same as FSE_initCState(), but the first symbol to include (which will be the last to be read)
|
||||||
|
* uses the smallest state value possible, saving the cost of this symbol */
|
||||||
|
MEM_STATIC void FSE_initCState2(FSE_CState_t* statePtr, const FSE_CTable* ct, U32 symbol)
|
||||||
|
{
|
||||||
|
FSE_initCState(statePtr, ct);
|
||||||
|
{ const FSE_symbolCompressionTransform symbolTT = ((const FSE_symbolCompressionTransform*)(statePtr->symbolTT))[symbol];
|
||||||
|
const U16* stateTable = (const U16*)(statePtr->stateTable);
|
||||||
|
U32 nbBitsOut = (U32)((symbolTT.deltaNbBits + (1<<15)) >> 16);
|
||||||
|
statePtr->value = (nbBitsOut << 16) - symbolTT.deltaNbBits;
|
||||||
|
statePtr->value = stateTable[(statePtr->value >> nbBitsOut) + symbolTT.deltaFindState];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void FSE_encodeSymbol(BIT_CStream_t* bitC, FSE_CState_t* statePtr, unsigned symbol)
|
||||||
|
{
|
||||||
|
FSE_symbolCompressionTransform const symbolTT = ((const FSE_symbolCompressionTransform*)(statePtr->symbolTT))[symbol];
|
||||||
|
const U16* const stateTable = (const U16*)(statePtr->stateTable);
|
||||||
|
U32 const nbBitsOut = (U32)((statePtr->value + symbolTT.deltaNbBits) >> 16);
|
||||||
|
BIT_addBits(bitC, statePtr->value, nbBitsOut);
|
||||||
|
statePtr->value = stateTable[ (statePtr->value >> nbBitsOut) + symbolTT.deltaFindState];
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void FSE_flushCState(BIT_CStream_t* bitC, const FSE_CState_t* statePtr)
|
||||||
|
{
|
||||||
|
BIT_addBits(bitC, statePtr->value, statePtr->stateLog);
|
||||||
|
BIT_flushBits(bitC);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* FSE_getMaxNbBits() :
|
||||||
|
* Approximate maximum cost of a symbol, in bits.
|
||||||
|
* Fractional get rounded up (i.e : a symbol with a normalized frequency of 3 gives the same result as a frequency of 2)
|
||||||
|
* note 1 : assume symbolValue is valid (<= maxSymbolValue)
|
||||||
|
* note 2 : if freq[symbolValue]==0, @return a fake cost of tableLog+1 bits */
|
||||||
|
MEM_STATIC U32 FSE_getMaxNbBits(const void* symbolTTPtr, U32 symbolValue)
|
||||||
|
{
|
||||||
|
const FSE_symbolCompressionTransform* symbolTT = (const FSE_symbolCompressionTransform*) symbolTTPtr;
|
||||||
|
return (symbolTT[symbolValue].deltaNbBits + ((1<<16)-1)) >> 16;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* FSE_bitCost() :
|
||||||
|
* Approximate symbol cost, as fractional value, using fixed-point format (accuracyLog fractional bits)
|
||||||
|
* note 1 : assume symbolValue is valid (<= maxSymbolValue)
|
||||||
|
* note 2 : if freq[symbolValue]==0, @return a fake cost of tableLog+1 bits */
|
||||||
|
MEM_STATIC U32 FSE_bitCost(const void* symbolTTPtr, U32 tableLog, U32 symbolValue, U32 accuracyLog)
|
||||||
|
{
|
||||||
|
const FSE_symbolCompressionTransform* symbolTT = (const FSE_symbolCompressionTransform*) symbolTTPtr;
|
||||||
|
U32 const minNbBits = symbolTT[symbolValue].deltaNbBits >> 16;
|
||||||
|
U32 const threshold = (minNbBits+1) << 16;
|
||||||
|
assert(tableLog < 16);
|
||||||
|
assert(accuracyLog < 31-tableLog); /* ensure enough room for renormalization double shift */
|
||||||
|
{ U32 const tableSize = 1 << tableLog;
|
||||||
|
U32 const deltaFromThreshold = threshold - (symbolTT[symbolValue].deltaNbBits + tableSize);
|
||||||
|
U32 const normalizedDeltaFromThreshold = (deltaFromThreshold << accuracyLog) >> tableLog; /* linear interpolation (very approximate) */
|
||||||
|
U32 const bitMultiplier = 1 << accuracyLog;
|
||||||
|
assert(symbolTT[symbolValue].deltaNbBits + tableSize <= threshold);
|
||||||
|
assert(normalizedDeltaFromThreshold <= bitMultiplier);
|
||||||
|
return (minNbBits+1)*bitMultiplier - normalizedDeltaFromThreshold;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* ====== Decompression ====== */
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
U16 tableLog;
|
||||||
|
U16 fastMode;
|
||||||
|
} FSE_DTableHeader; /* sizeof U32 */
|
||||||
|
|
||||||
|
typedef struct
|
||||||
|
{
|
||||||
|
unsigned short newState;
|
||||||
|
unsigned char symbol;
|
||||||
|
unsigned char nbBits;
|
||||||
|
} FSE_decode_t; /* size == U32 */
|
||||||
|
|
||||||
|
MEM_STATIC void FSE_initDState(FSE_DState_t* DStatePtr, BIT_DStream_t* bitD, const FSE_DTable* dt)
|
||||||
|
{
|
||||||
|
const void* ptr = dt;
|
||||||
|
const FSE_DTableHeader* const DTableH = (const FSE_DTableHeader*)ptr;
|
||||||
|
DStatePtr->state = BIT_readBits(bitD, DTableH->tableLog);
|
||||||
|
BIT_reloadDStream(bitD);
|
||||||
|
DStatePtr->table = dt + 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC BYTE FSE_peekSymbol(const FSE_DState_t* DStatePtr)
|
||||||
|
{
|
||||||
|
FSE_decode_t const DInfo = ((const FSE_decode_t*)(DStatePtr->table))[DStatePtr->state];
|
||||||
|
return DInfo.symbol;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void FSE_updateState(FSE_DState_t* DStatePtr, BIT_DStream_t* bitD)
|
||||||
|
{
|
||||||
|
FSE_decode_t const DInfo = ((const FSE_decode_t*)(DStatePtr->table))[DStatePtr->state];
|
||||||
|
U32 const nbBits = DInfo.nbBits;
|
||||||
|
size_t const lowBits = BIT_readBits(bitD, nbBits);
|
||||||
|
DStatePtr->state = DInfo.newState + lowBits;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC BYTE FSE_decodeSymbol(FSE_DState_t* DStatePtr, BIT_DStream_t* bitD)
|
||||||
|
{
|
||||||
|
FSE_decode_t const DInfo = ((const FSE_decode_t*)(DStatePtr->table))[DStatePtr->state];
|
||||||
|
U32 const nbBits = DInfo.nbBits;
|
||||||
|
BYTE const symbol = DInfo.symbol;
|
||||||
|
size_t const lowBits = BIT_readBits(bitD, nbBits);
|
||||||
|
|
||||||
|
DStatePtr->state = DInfo.newState + lowBits;
|
||||||
|
return symbol;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! FSE_decodeSymbolFast() :
|
||||||
|
unsafe, only works if no symbol has a probability > 50% */
|
||||||
|
MEM_STATIC BYTE FSE_decodeSymbolFast(FSE_DState_t* DStatePtr, BIT_DStream_t* bitD)
|
||||||
|
{
|
||||||
|
FSE_decode_t const DInfo = ((const FSE_decode_t*)(DStatePtr->table))[DStatePtr->state];
|
||||||
|
U32 const nbBits = DInfo.nbBits;
|
||||||
|
BYTE const symbol = DInfo.symbol;
|
||||||
|
size_t const lowBits = BIT_readBitsFast(bitD, nbBits);
|
||||||
|
|
||||||
|
DStatePtr->state = DInfo.newState + lowBits;
|
||||||
|
return symbol;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC unsigned FSE_endOfDState(const FSE_DState_t* DStatePtr)
|
||||||
|
{
|
||||||
|
return DStatePtr->state == 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
#ifndef FSE_COMMONDEFS_ONLY
|
||||||
|
|
||||||
|
/* **************************************************************
|
||||||
|
* Tuning parameters
|
||||||
|
****************************************************************/
|
||||||
|
/*!MEMORY_USAGE :
|
||||||
|
* Memory usage formula : N->2^N Bytes (examples : 10 -> 1KB; 12 -> 4KB ; 16 -> 64KB; 20 -> 1MB; etc.)
|
||||||
|
* Increasing memory usage improves compression ratio
|
||||||
|
* Reduced memory usage can improve speed, due to cache effect
|
||||||
|
* Recommended max value is 14, for 16KB, which nicely fits into Intel x86 L1 cache */
|
||||||
|
#ifndef FSE_MAX_MEMORY_USAGE
|
||||||
|
# define FSE_MAX_MEMORY_USAGE 14
|
||||||
|
#endif
|
||||||
|
#ifndef FSE_DEFAULT_MEMORY_USAGE
|
||||||
|
# define FSE_DEFAULT_MEMORY_USAGE 13
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*!FSE_MAX_SYMBOL_VALUE :
|
||||||
|
* Maximum symbol value authorized.
|
||||||
|
* Required for proper stack allocation */
|
||||||
|
#ifndef FSE_MAX_SYMBOL_VALUE
|
||||||
|
# define FSE_MAX_SYMBOL_VALUE 255
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* **************************************************************
|
||||||
|
* template functions type & suffix
|
||||||
|
****************************************************************/
|
||||||
|
#define FSE_FUNCTION_TYPE BYTE
|
||||||
|
#define FSE_FUNCTION_EXTENSION
|
||||||
|
#define FSE_DECODE_TYPE FSE_decode_t
|
||||||
|
|
||||||
|
|
||||||
|
#endif /* !FSE_COMMONDEFS_ONLY */
|
||||||
|
|
||||||
|
|
||||||
|
/* ***************************************************************
|
||||||
|
* Constants
|
||||||
|
*****************************************************************/
|
||||||
|
#define FSE_MAX_TABLELOG (FSE_MAX_MEMORY_USAGE-2)
|
||||||
|
#define FSE_MAX_TABLESIZE (1U<<FSE_MAX_TABLELOG)
|
||||||
|
#define FSE_MAXTABLESIZE_MASK (FSE_MAX_TABLESIZE-1)
|
||||||
|
#define FSE_DEFAULT_TABLELOG (FSE_DEFAULT_MEMORY_USAGE-2)
|
||||||
|
#define FSE_MIN_TABLELOG 5
|
||||||
|
|
||||||
|
#define FSE_TABLELOG_ABSOLUTE_MAX 15
|
||||||
|
#if FSE_MAX_TABLELOG > FSE_TABLELOG_ABSOLUTE_MAX
|
||||||
|
# error "FSE_MAX_TABLELOG > FSE_TABLELOG_ABSOLUTE_MAX is not supported"
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#define FSE_TABLESTEP(tableSize) ((tableSize>>1) + (tableSize>>3) + 3)
|
||||||
|
|
||||||
|
|
||||||
|
#endif /* FSE_STATIC_LINKING_ONLY */
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
721
vendor/github.com/DataDog/zstd/fse_compress.c
generated
vendored
Normal file
721
vendor/github.com/DataDog/zstd/fse_compress.c
generated
vendored
Normal file
@@ -0,0 +1,721 @@
|
|||||||
|
/* ******************************************************************
|
||||||
|
FSE : Finite State Entropy encoder
|
||||||
|
Copyright (C) 2013-present, Yann Collet.
|
||||||
|
|
||||||
|
BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
You can contact the author at :
|
||||||
|
- FSE source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
||||||
|
- Public forum : https://groups.google.com/forum/#!forum/lz4c
|
||||||
|
****************************************************************** */
|
||||||
|
|
||||||
|
/* **************************************************************
|
||||||
|
* Includes
|
||||||
|
****************************************************************/
|
||||||
|
#include <stdlib.h> /* malloc, free, qsort */
|
||||||
|
#include <string.h> /* memcpy, memset */
|
||||||
|
#include "compiler.h"
|
||||||
|
#include "mem.h" /* U32, U16, etc. */
|
||||||
|
#include "debug.h" /* assert, DEBUGLOG */
|
||||||
|
#include "hist.h" /* HIST_count_wksp */
|
||||||
|
#include "bitstream.h"
|
||||||
|
#define FSE_STATIC_LINKING_ONLY
|
||||||
|
#include "fse.h"
|
||||||
|
#include "error_private.h"
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************************************************
|
||||||
|
* Error Management
|
||||||
|
****************************************************************/
|
||||||
|
#define FSE_isError ERR_isError
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************************************************
|
||||||
|
* Templates
|
||||||
|
****************************************************************/
|
||||||
|
/*
|
||||||
|
designed to be included
|
||||||
|
for type-specific functions (template emulation in C)
|
||||||
|
Objective is to write these functions only once, for improved maintenance
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* safety checks */
|
||||||
|
#ifndef FSE_FUNCTION_EXTENSION
|
||||||
|
# error "FSE_FUNCTION_EXTENSION must be defined"
|
||||||
|
#endif
|
||||||
|
#ifndef FSE_FUNCTION_TYPE
|
||||||
|
# error "FSE_FUNCTION_TYPE must be defined"
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* Function names */
|
||||||
|
#define FSE_CAT(X,Y) X##Y
|
||||||
|
#define FSE_FUNCTION_NAME(X,Y) FSE_CAT(X,Y)
|
||||||
|
#define FSE_TYPE_NAME(X,Y) FSE_CAT(X,Y)
|
||||||
|
|
||||||
|
|
||||||
|
/* Function templates */
|
||||||
|
|
||||||
|
/* FSE_buildCTable_wksp() :
|
||||||
|
* Same as FSE_buildCTable(), but using an externally allocated scratch buffer (`workSpace`).
|
||||||
|
* wkspSize should be sized to handle worst case situation, which is `1<<max_tableLog * sizeof(FSE_FUNCTION_TYPE)`
|
||||||
|
* workSpace must also be properly aligned with FSE_FUNCTION_TYPE requirements
|
||||||
|
*/
|
||||||
|
size_t FSE_buildCTable_wksp(FSE_CTable* ct,
|
||||||
|
const short* normalizedCounter, unsigned maxSymbolValue, unsigned tableLog,
|
||||||
|
void* workSpace, size_t wkspSize)
|
||||||
|
{
|
||||||
|
U32 const tableSize = 1 << tableLog;
|
||||||
|
U32 const tableMask = tableSize - 1;
|
||||||
|
void* const ptr = ct;
|
||||||
|
U16* const tableU16 = ( (U16*) ptr) + 2;
|
||||||
|
void* const FSCT = ((U32*)ptr) + 1 /* header */ + (tableLog ? tableSize>>1 : 1) ;
|
||||||
|
FSE_symbolCompressionTransform* const symbolTT = (FSE_symbolCompressionTransform*) (FSCT);
|
||||||
|
U32 const step = FSE_TABLESTEP(tableSize);
|
||||||
|
U32 cumul[FSE_MAX_SYMBOL_VALUE+2];
|
||||||
|
|
||||||
|
FSE_FUNCTION_TYPE* const tableSymbol = (FSE_FUNCTION_TYPE*)workSpace;
|
||||||
|
U32 highThreshold = tableSize-1;
|
||||||
|
|
||||||
|
/* CTable header */
|
||||||
|
if (((size_t)1 << tableLog) * sizeof(FSE_FUNCTION_TYPE) > wkspSize) return ERROR(tableLog_tooLarge);
|
||||||
|
tableU16[-2] = (U16) tableLog;
|
||||||
|
tableU16[-1] = (U16) maxSymbolValue;
|
||||||
|
assert(tableLog < 16); /* required for threshold strategy to work */
|
||||||
|
|
||||||
|
/* For explanations on how to distribute symbol values over the table :
|
||||||
|
* http://fastcompression.blogspot.fr/2014/02/fse-distributing-symbol-values.html */
|
||||||
|
|
||||||
|
#ifdef __clang_analyzer__
|
||||||
|
memset(tableSymbol, 0, sizeof(*tableSymbol) * tableSize); /* useless initialization, just to keep scan-build happy */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* symbol start positions */
|
||||||
|
{ U32 u;
|
||||||
|
cumul[0] = 0;
|
||||||
|
for (u=1; u <= maxSymbolValue+1; u++) {
|
||||||
|
if (normalizedCounter[u-1]==-1) { /* Low proba symbol */
|
||||||
|
cumul[u] = cumul[u-1] + 1;
|
||||||
|
tableSymbol[highThreshold--] = (FSE_FUNCTION_TYPE)(u-1);
|
||||||
|
} else {
|
||||||
|
cumul[u] = cumul[u-1] + normalizedCounter[u-1];
|
||||||
|
} }
|
||||||
|
cumul[maxSymbolValue+1] = tableSize+1;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Spread symbols */
|
||||||
|
{ U32 position = 0;
|
||||||
|
U32 symbol;
|
||||||
|
for (symbol=0; symbol<=maxSymbolValue; symbol++) {
|
||||||
|
int nbOccurrences;
|
||||||
|
int const freq = normalizedCounter[symbol];
|
||||||
|
for (nbOccurrences=0; nbOccurrences<freq; nbOccurrences++) {
|
||||||
|
tableSymbol[position] = (FSE_FUNCTION_TYPE)symbol;
|
||||||
|
position = (position + step) & tableMask;
|
||||||
|
while (position > highThreshold)
|
||||||
|
position = (position + step) & tableMask; /* Low proba area */
|
||||||
|
} }
|
||||||
|
|
||||||
|
assert(position==0); /* Must have initialized all positions */
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Build table */
|
||||||
|
{ U32 u; for (u=0; u<tableSize; u++) {
|
||||||
|
FSE_FUNCTION_TYPE s = tableSymbol[u]; /* note : static analyzer may not understand tableSymbol is properly initialized */
|
||||||
|
tableU16[cumul[s]++] = (U16) (tableSize+u); /* TableU16 : sorted by symbol order; gives next state value */
|
||||||
|
} }
|
||||||
|
|
||||||
|
/* Build Symbol Transformation Table */
|
||||||
|
{ unsigned total = 0;
|
||||||
|
unsigned s;
|
||||||
|
for (s=0; s<=maxSymbolValue; s++) {
|
||||||
|
switch (normalizedCounter[s])
|
||||||
|
{
|
||||||
|
case 0:
|
||||||
|
/* filling nonetheless, for compatibility with FSE_getMaxNbBits() */
|
||||||
|
symbolTT[s].deltaNbBits = ((tableLog+1) << 16) - (1<<tableLog);
|
||||||
|
break;
|
||||||
|
|
||||||
|
case -1:
|
||||||
|
case 1:
|
||||||
|
symbolTT[s].deltaNbBits = (tableLog << 16) - (1<<tableLog);
|
||||||
|
symbolTT[s].deltaFindState = total - 1;
|
||||||
|
total ++;
|
||||||
|
break;
|
||||||
|
default :
|
||||||
|
{
|
||||||
|
U32 const maxBitsOut = tableLog - BIT_highbit32 (normalizedCounter[s]-1);
|
||||||
|
U32 const minStatePlus = normalizedCounter[s] << maxBitsOut;
|
||||||
|
symbolTT[s].deltaNbBits = (maxBitsOut << 16) - minStatePlus;
|
||||||
|
symbolTT[s].deltaFindState = total - normalizedCounter[s];
|
||||||
|
total += normalizedCounter[s];
|
||||||
|
} } } }
|
||||||
|
|
||||||
|
#if 0 /* debug : symbol costs */
|
||||||
|
DEBUGLOG(5, "\n --- table statistics : ");
|
||||||
|
{ U32 symbol;
|
||||||
|
for (symbol=0; symbol<=maxSymbolValue; symbol++) {
|
||||||
|
DEBUGLOG(5, "%3u: w=%3i, maxBits=%u, fracBits=%.2f",
|
||||||
|
symbol, normalizedCounter[symbol],
|
||||||
|
FSE_getMaxNbBits(symbolTT, symbol),
|
||||||
|
(double)FSE_bitCost(symbolTT, tableLog, symbol, 8) / 256);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t FSE_buildCTable(FSE_CTable* ct, const short* normalizedCounter, unsigned maxSymbolValue, unsigned tableLog)
|
||||||
|
{
|
||||||
|
FSE_FUNCTION_TYPE tableSymbol[FSE_MAX_TABLESIZE]; /* memset() is not necessary, even if static analyzer complain about it */
|
||||||
|
return FSE_buildCTable_wksp(ct, normalizedCounter, maxSymbolValue, tableLog, tableSymbol, sizeof(tableSymbol));
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
#ifndef FSE_COMMONDEFS_ONLY
|
||||||
|
|
||||||
|
|
||||||
|
/*-**************************************************************
|
||||||
|
* FSE NCount encoding
|
||||||
|
****************************************************************/
|
||||||
|
size_t FSE_NCountWriteBound(unsigned maxSymbolValue, unsigned tableLog)
|
||||||
|
{
|
||||||
|
size_t const maxHeaderSize = (((maxSymbolValue+1) * tableLog) >> 3) + 3;
|
||||||
|
return maxSymbolValue ? maxHeaderSize : FSE_NCOUNTBOUND; /* maxSymbolValue==0 ? use default */
|
||||||
|
}
|
||||||
|
|
||||||
|
static size_t
|
||||||
|
FSE_writeNCount_generic (void* header, size_t headerBufferSize,
|
||||||
|
const short* normalizedCounter, unsigned maxSymbolValue, unsigned tableLog,
|
||||||
|
unsigned writeIsSafe)
|
||||||
|
{
|
||||||
|
BYTE* const ostart = (BYTE*) header;
|
||||||
|
BYTE* out = ostart;
|
||||||
|
BYTE* const oend = ostart + headerBufferSize;
|
||||||
|
int nbBits;
|
||||||
|
const int tableSize = 1 << tableLog;
|
||||||
|
int remaining;
|
||||||
|
int threshold;
|
||||||
|
U32 bitStream = 0;
|
||||||
|
int bitCount = 0;
|
||||||
|
unsigned symbol = 0;
|
||||||
|
unsigned const alphabetSize = maxSymbolValue + 1;
|
||||||
|
int previousIs0 = 0;
|
||||||
|
|
||||||
|
/* Table Size */
|
||||||
|
bitStream += (tableLog-FSE_MIN_TABLELOG) << bitCount;
|
||||||
|
bitCount += 4;
|
||||||
|
|
||||||
|
/* Init */
|
||||||
|
remaining = tableSize+1; /* +1 for extra accuracy */
|
||||||
|
threshold = tableSize;
|
||||||
|
nbBits = tableLog+1;
|
||||||
|
|
||||||
|
while ((symbol < alphabetSize) && (remaining>1)) { /* stops at 1 */
|
||||||
|
if (previousIs0) {
|
||||||
|
unsigned start = symbol;
|
||||||
|
while ((symbol < alphabetSize) && !normalizedCounter[symbol]) symbol++;
|
||||||
|
if (symbol == alphabetSize) break; /* incorrect distribution */
|
||||||
|
while (symbol >= start+24) {
|
||||||
|
start+=24;
|
||||||
|
bitStream += 0xFFFFU << bitCount;
|
||||||
|
if ((!writeIsSafe) && (out > oend-2))
|
||||||
|
return ERROR(dstSize_tooSmall); /* Buffer overflow */
|
||||||
|
out[0] = (BYTE) bitStream;
|
||||||
|
out[1] = (BYTE)(bitStream>>8);
|
||||||
|
out+=2;
|
||||||
|
bitStream>>=16;
|
||||||
|
}
|
||||||
|
while (symbol >= start+3) {
|
||||||
|
start+=3;
|
||||||
|
bitStream += 3 << bitCount;
|
||||||
|
bitCount += 2;
|
||||||
|
}
|
||||||
|
bitStream += (symbol-start) << bitCount;
|
||||||
|
bitCount += 2;
|
||||||
|
if (bitCount>16) {
|
||||||
|
if ((!writeIsSafe) && (out > oend - 2))
|
||||||
|
return ERROR(dstSize_tooSmall); /* Buffer overflow */
|
||||||
|
out[0] = (BYTE)bitStream;
|
||||||
|
out[1] = (BYTE)(bitStream>>8);
|
||||||
|
out += 2;
|
||||||
|
bitStream >>= 16;
|
||||||
|
bitCount -= 16;
|
||||||
|
} }
|
||||||
|
{ int count = normalizedCounter[symbol++];
|
||||||
|
int const max = (2*threshold-1) - remaining;
|
||||||
|
remaining -= count < 0 ? -count : count;
|
||||||
|
count++; /* +1 for extra accuracy */
|
||||||
|
if (count>=threshold)
|
||||||
|
count += max; /* [0..max[ [max..threshold[ (...) [threshold+max 2*threshold[ */
|
||||||
|
bitStream += count << bitCount;
|
||||||
|
bitCount += nbBits;
|
||||||
|
bitCount -= (count<max);
|
||||||
|
previousIs0 = (count==1);
|
||||||
|
if (remaining<1) return ERROR(GENERIC);
|
||||||
|
while (remaining<threshold) { nbBits--; threshold>>=1; }
|
||||||
|
}
|
||||||
|
if (bitCount>16) {
|
||||||
|
if ((!writeIsSafe) && (out > oend - 2))
|
||||||
|
return ERROR(dstSize_tooSmall); /* Buffer overflow */
|
||||||
|
out[0] = (BYTE)bitStream;
|
||||||
|
out[1] = (BYTE)(bitStream>>8);
|
||||||
|
out += 2;
|
||||||
|
bitStream >>= 16;
|
||||||
|
bitCount -= 16;
|
||||||
|
} }
|
||||||
|
|
||||||
|
if (remaining != 1)
|
||||||
|
return ERROR(GENERIC); /* incorrect normalized distribution */
|
||||||
|
assert(symbol <= alphabetSize);
|
||||||
|
|
||||||
|
/* flush remaining bitStream */
|
||||||
|
if ((!writeIsSafe) && (out > oend - 2))
|
||||||
|
return ERROR(dstSize_tooSmall); /* Buffer overflow */
|
||||||
|
out[0] = (BYTE)bitStream;
|
||||||
|
out[1] = (BYTE)(bitStream>>8);
|
||||||
|
out+= (bitCount+7) /8;
|
||||||
|
|
||||||
|
return (out-ostart);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t FSE_writeNCount (void* buffer, size_t bufferSize,
|
||||||
|
const short* normalizedCounter, unsigned maxSymbolValue, unsigned tableLog)
|
||||||
|
{
|
||||||
|
if (tableLog > FSE_MAX_TABLELOG) return ERROR(tableLog_tooLarge); /* Unsupported */
|
||||||
|
if (tableLog < FSE_MIN_TABLELOG) return ERROR(GENERIC); /* Unsupported */
|
||||||
|
|
||||||
|
if (bufferSize < FSE_NCountWriteBound(maxSymbolValue, tableLog))
|
||||||
|
return FSE_writeNCount_generic(buffer, bufferSize, normalizedCounter, maxSymbolValue, tableLog, 0);
|
||||||
|
|
||||||
|
return FSE_writeNCount_generic(buffer, bufferSize, normalizedCounter, maxSymbolValue, tableLog, 1 /* write in buffer is safe */);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/*-**************************************************************
|
||||||
|
* FSE Compression Code
|
||||||
|
****************************************************************/
|
||||||
|
|
||||||
|
FSE_CTable* FSE_createCTable (unsigned maxSymbolValue, unsigned tableLog)
|
||||||
|
{
|
||||||
|
size_t size;
|
||||||
|
if (tableLog > FSE_TABLELOG_ABSOLUTE_MAX) tableLog = FSE_TABLELOG_ABSOLUTE_MAX;
|
||||||
|
size = FSE_CTABLE_SIZE_U32 (tableLog, maxSymbolValue) * sizeof(U32);
|
||||||
|
return (FSE_CTable*)malloc(size);
|
||||||
|
}
|
||||||
|
|
||||||
|
void FSE_freeCTable (FSE_CTable* ct) { free(ct); }
|
||||||
|
|
||||||
|
/* provides the minimum logSize to safely represent a distribution */
|
||||||
|
static unsigned FSE_minTableLog(size_t srcSize, unsigned maxSymbolValue)
|
||||||
|
{
|
||||||
|
U32 minBitsSrc = BIT_highbit32((U32)(srcSize)) + 1;
|
||||||
|
U32 minBitsSymbols = BIT_highbit32(maxSymbolValue) + 2;
|
||||||
|
U32 minBits = minBitsSrc < minBitsSymbols ? minBitsSrc : minBitsSymbols;
|
||||||
|
assert(srcSize > 1); /* Not supported, RLE should be used instead */
|
||||||
|
return minBits;
|
||||||
|
}
|
||||||
|
|
||||||
|
unsigned FSE_optimalTableLog_internal(unsigned maxTableLog, size_t srcSize, unsigned maxSymbolValue, unsigned minus)
|
||||||
|
{
|
||||||
|
U32 maxBitsSrc = BIT_highbit32((U32)(srcSize - 1)) - minus;
|
||||||
|
U32 tableLog = maxTableLog;
|
||||||
|
U32 minBits = FSE_minTableLog(srcSize, maxSymbolValue);
|
||||||
|
assert(srcSize > 1); /* Not supported, RLE should be used instead */
|
||||||
|
if (tableLog==0) tableLog = FSE_DEFAULT_TABLELOG;
|
||||||
|
if (maxBitsSrc < tableLog) tableLog = maxBitsSrc; /* Accuracy can be reduced */
|
||||||
|
if (minBits > tableLog) tableLog = minBits; /* Need a minimum to safely represent all symbol values */
|
||||||
|
if (tableLog < FSE_MIN_TABLELOG) tableLog = FSE_MIN_TABLELOG;
|
||||||
|
if (tableLog > FSE_MAX_TABLELOG) tableLog = FSE_MAX_TABLELOG;
|
||||||
|
return tableLog;
|
||||||
|
}
|
||||||
|
|
||||||
|
unsigned FSE_optimalTableLog(unsigned maxTableLog, size_t srcSize, unsigned maxSymbolValue)
|
||||||
|
{
|
||||||
|
return FSE_optimalTableLog_internal(maxTableLog, srcSize, maxSymbolValue, 2);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* Secondary normalization method.
|
||||||
|
To be used when primary method fails. */
|
||||||
|
|
||||||
|
static size_t FSE_normalizeM2(short* norm, U32 tableLog, const unsigned* count, size_t total, U32 maxSymbolValue)
|
||||||
|
{
|
||||||
|
short const NOT_YET_ASSIGNED = -2;
|
||||||
|
U32 s;
|
||||||
|
U32 distributed = 0;
|
||||||
|
U32 ToDistribute;
|
||||||
|
|
||||||
|
/* Init */
|
||||||
|
U32 const lowThreshold = (U32)(total >> tableLog);
|
||||||
|
U32 lowOne = (U32)((total * 3) >> (tableLog + 1));
|
||||||
|
|
||||||
|
for (s=0; s<=maxSymbolValue; s++) {
|
||||||
|
if (count[s] == 0) {
|
||||||
|
norm[s]=0;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
if (count[s] <= lowThreshold) {
|
||||||
|
norm[s] = -1;
|
||||||
|
distributed++;
|
||||||
|
total -= count[s];
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
if (count[s] <= lowOne) {
|
||||||
|
norm[s] = 1;
|
||||||
|
distributed++;
|
||||||
|
total -= count[s];
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
norm[s]=NOT_YET_ASSIGNED;
|
||||||
|
}
|
||||||
|
ToDistribute = (1 << tableLog) - distributed;
|
||||||
|
|
||||||
|
if (ToDistribute == 0)
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
if ((total / ToDistribute) > lowOne) {
|
||||||
|
/* risk of rounding to zero */
|
||||||
|
lowOne = (U32)((total * 3) / (ToDistribute * 2));
|
||||||
|
for (s=0; s<=maxSymbolValue; s++) {
|
||||||
|
if ((norm[s] == NOT_YET_ASSIGNED) && (count[s] <= lowOne)) {
|
||||||
|
norm[s] = 1;
|
||||||
|
distributed++;
|
||||||
|
total -= count[s];
|
||||||
|
continue;
|
||||||
|
} }
|
||||||
|
ToDistribute = (1 << tableLog) - distributed;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (distributed == maxSymbolValue+1) {
|
||||||
|
/* all values are pretty poor;
|
||||||
|
probably incompressible data (should have already been detected);
|
||||||
|
find max, then give all remaining points to max */
|
||||||
|
U32 maxV = 0, maxC = 0;
|
||||||
|
for (s=0; s<=maxSymbolValue; s++)
|
||||||
|
if (count[s] > maxC) { maxV=s; maxC=count[s]; }
|
||||||
|
norm[maxV] += (short)ToDistribute;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (total == 0) {
|
||||||
|
/* all of the symbols were low enough for the lowOne or lowThreshold */
|
||||||
|
for (s=0; ToDistribute > 0; s = (s+1)%(maxSymbolValue+1))
|
||||||
|
if (norm[s] > 0) { ToDistribute--; norm[s]++; }
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
{ U64 const vStepLog = 62 - tableLog;
|
||||||
|
U64 const mid = (1ULL << (vStepLog-1)) - 1;
|
||||||
|
U64 const rStep = ((((U64)1<<vStepLog) * ToDistribute) + mid) / total; /* scale on remaining */
|
||||||
|
U64 tmpTotal = mid;
|
||||||
|
for (s=0; s<=maxSymbolValue; s++) {
|
||||||
|
if (norm[s]==NOT_YET_ASSIGNED) {
|
||||||
|
U64 const end = tmpTotal + (count[s] * rStep);
|
||||||
|
U32 const sStart = (U32)(tmpTotal >> vStepLog);
|
||||||
|
U32 const sEnd = (U32)(end >> vStepLog);
|
||||||
|
U32 const weight = sEnd - sStart;
|
||||||
|
if (weight < 1)
|
||||||
|
return ERROR(GENERIC);
|
||||||
|
norm[s] = (short)weight;
|
||||||
|
tmpTotal = end;
|
||||||
|
} } }
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t FSE_normalizeCount (short* normalizedCounter, unsigned tableLog,
|
||||||
|
const unsigned* count, size_t total,
|
||||||
|
unsigned maxSymbolValue)
|
||||||
|
{
|
||||||
|
/* Sanity checks */
|
||||||
|
if (tableLog==0) tableLog = FSE_DEFAULT_TABLELOG;
|
||||||
|
if (tableLog < FSE_MIN_TABLELOG) return ERROR(GENERIC); /* Unsupported size */
|
||||||
|
if (tableLog > FSE_MAX_TABLELOG) return ERROR(tableLog_tooLarge); /* Unsupported size */
|
||||||
|
if (tableLog < FSE_minTableLog(total, maxSymbolValue)) return ERROR(GENERIC); /* Too small tableLog, compression potentially impossible */
|
||||||
|
|
||||||
|
{ static U32 const rtbTable[] = { 0, 473195, 504333, 520860, 550000, 700000, 750000, 830000 };
|
||||||
|
U64 const scale = 62 - tableLog;
|
||||||
|
U64 const step = ((U64)1<<62) / total; /* <== here, one division ! */
|
||||||
|
U64 const vStep = 1ULL<<(scale-20);
|
||||||
|
int stillToDistribute = 1<<tableLog;
|
||||||
|
unsigned s;
|
||||||
|
unsigned largest=0;
|
||||||
|
short largestP=0;
|
||||||
|
U32 lowThreshold = (U32)(total >> tableLog);
|
||||||
|
|
||||||
|
for (s=0; s<=maxSymbolValue; s++) {
|
||||||
|
if (count[s] == total) return 0; /* rle special case */
|
||||||
|
if (count[s] == 0) { normalizedCounter[s]=0; continue; }
|
||||||
|
if (count[s] <= lowThreshold) {
|
||||||
|
normalizedCounter[s] = -1;
|
||||||
|
stillToDistribute--;
|
||||||
|
} else {
|
||||||
|
short proba = (short)((count[s]*step) >> scale);
|
||||||
|
if (proba<8) {
|
||||||
|
U64 restToBeat = vStep * rtbTable[proba];
|
||||||
|
proba += (count[s]*step) - ((U64)proba<<scale) > restToBeat;
|
||||||
|
}
|
||||||
|
if (proba > largestP) { largestP=proba; largest=s; }
|
||||||
|
normalizedCounter[s] = proba;
|
||||||
|
stillToDistribute -= proba;
|
||||||
|
} }
|
||||||
|
if (-stillToDistribute >= (normalizedCounter[largest] >> 1)) {
|
||||||
|
/* corner case, need another normalization method */
|
||||||
|
size_t const errorCode = FSE_normalizeM2(normalizedCounter, tableLog, count, total, maxSymbolValue);
|
||||||
|
if (FSE_isError(errorCode)) return errorCode;
|
||||||
|
}
|
||||||
|
else normalizedCounter[largest] += (short)stillToDistribute;
|
||||||
|
}
|
||||||
|
|
||||||
|
#if 0
|
||||||
|
{ /* Print Table (debug) */
|
||||||
|
U32 s;
|
||||||
|
U32 nTotal = 0;
|
||||||
|
for (s=0; s<=maxSymbolValue; s++)
|
||||||
|
RAWLOG(2, "%3i: %4i \n", s, normalizedCounter[s]);
|
||||||
|
for (s=0; s<=maxSymbolValue; s++)
|
||||||
|
nTotal += abs(normalizedCounter[s]);
|
||||||
|
if (nTotal != (1U<<tableLog))
|
||||||
|
RAWLOG(2, "Warning !!! Total == %u != %u !!!", nTotal, 1U<<tableLog);
|
||||||
|
getchar();
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
return tableLog;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* fake FSE_CTable, for raw (uncompressed) input */
|
||||||
|
size_t FSE_buildCTable_raw (FSE_CTable* ct, unsigned nbBits)
|
||||||
|
{
|
||||||
|
const unsigned tableSize = 1 << nbBits;
|
||||||
|
const unsigned tableMask = tableSize - 1;
|
||||||
|
const unsigned maxSymbolValue = tableMask;
|
||||||
|
void* const ptr = ct;
|
||||||
|
U16* const tableU16 = ( (U16*) ptr) + 2;
|
||||||
|
void* const FSCT = ((U32*)ptr) + 1 /* header */ + (tableSize>>1); /* assumption : tableLog >= 1 */
|
||||||
|
FSE_symbolCompressionTransform* const symbolTT = (FSE_symbolCompressionTransform*) (FSCT);
|
||||||
|
unsigned s;
|
||||||
|
|
||||||
|
/* Sanity checks */
|
||||||
|
if (nbBits < 1) return ERROR(GENERIC); /* min size */
|
||||||
|
|
||||||
|
/* header */
|
||||||
|
tableU16[-2] = (U16) nbBits;
|
||||||
|
tableU16[-1] = (U16) maxSymbolValue;
|
||||||
|
|
||||||
|
/* Build table */
|
||||||
|
for (s=0; s<tableSize; s++)
|
||||||
|
tableU16[s] = (U16)(tableSize + s);
|
||||||
|
|
||||||
|
/* Build Symbol Transformation Table */
|
||||||
|
{ const U32 deltaNbBits = (nbBits << 16) - (1 << nbBits);
|
||||||
|
for (s=0; s<=maxSymbolValue; s++) {
|
||||||
|
symbolTT[s].deltaNbBits = deltaNbBits;
|
||||||
|
symbolTT[s].deltaFindState = s-1;
|
||||||
|
} }
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* fake FSE_CTable, for rle input (always same symbol) */
|
||||||
|
size_t FSE_buildCTable_rle (FSE_CTable* ct, BYTE symbolValue)
|
||||||
|
{
|
||||||
|
void* ptr = ct;
|
||||||
|
U16* tableU16 = ( (U16*) ptr) + 2;
|
||||||
|
void* FSCTptr = (U32*)ptr + 2;
|
||||||
|
FSE_symbolCompressionTransform* symbolTT = (FSE_symbolCompressionTransform*) FSCTptr;
|
||||||
|
|
||||||
|
/* header */
|
||||||
|
tableU16[-2] = (U16) 0;
|
||||||
|
tableU16[-1] = (U16) symbolValue;
|
||||||
|
|
||||||
|
/* Build table */
|
||||||
|
tableU16[0] = 0;
|
||||||
|
tableU16[1] = 0; /* just in case */
|
||||||
|
|
||||||
|
/* Build Symbol Transformation Table */
|
||||||
|
symbolTT[symbolValue].deltaNbBits = 0;
|
||||||
|
symbolTT[symbolValue].deltaFindState = 0;
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static size_t FSE_compress_usingCTable_generic (void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const FSE_CTable* ct, const unsigned fast)
|
||||||
|
{
|
||||||
|
const BYTE* const istart = (const BYTE*) src;
|
||||||
|
const BYTE* const iend = istart + srcSize;
|
||||||
|
const BYTE* ip=iend;
|
||||||
|
|
||||||
|
BIT_CStream_t bitC;
|
||||||
|
FSE_CState_t CState1, CState2;
|
||||||
|
|
||||||
|
/* init */
|
||||||
|
if (srcSize <= 2) return 0;
|
||||||
|
{ size_t const initError = BIT_initCStream(&bitC, dst, dstSize);
|
||||||
|
if (FSE_isError(initError)) return 0; /* not enough space available to write a bitstream */ }
|
||||||
|
|
||||||
|
#define FSE_FLUSHBITS(s) (fast ? BIT_flushBitsFast(s) : BIT_flushBits(s))
|
||||||
|
|
||||||
|
if (srcSize & 1) {
|
||||||
|
FSE_initCState2(&CState1, ct, *--ip);
|
||||||
|
FSE_initCState2(&CState2, ct, *--ip);
|
||||||
|
FSE_encodeSymbol(&bitC, &CState1, *--ip);
|
||||||
|
FSE_FLUSHBITS(&bitC);
|
||||||
|
} else {
|
||||||
|
FSE_initCState2(&CState2, ct, *--ip);
|
||||||
|
FSE_initCState2(&CState1, ct, *--ip);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* join to mod 4 */
|
||||||
|
srcSize -= 2;
|
||||||
|
if ((sizeof(bitC.bitContainer)*8 > FSE_MAX_TABLELOG*4+7 ) && (srcSize & 2)) { /* test bit 2 */
|
||||||
|
FSE_encodeSymbol(&bitC, &CState2, *--ip);
|
||||||
|
FSE_encodeSymbol(&bitC, &CState1, *--ip);
|
||||||
|
FSE_FLUSHBITS(&bitC);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* 2 or 4 encoding per loop */
|
||||||
|
while ( ip>istart ) {
|
||||||
|
|
||||||
|
FSE_encodeSymbol(&bitC, &CState2, *--ip);
|
||||||
|
|
||||||
|
if (sizeof(bitC.bitContainer)*8 < FSE_MAX_TABLELOG*2+7 ) /* this test must be static */
|
||||||
|
FSE_FLUSHBITS(&bitC);
|
||||||
|
|
||||||
|
FSE_encodeSymbol(&bitC, &CState1, *--ip);
|
||||||
|
|
||||||
|
if (sizeof(bitC.bitContainer)*8 > FSE_MAX_TABLELOG*4+7 ) { /* this test must be static */
|
||||||
|
FSE_encodeSymbol(&bitC, &CState2, *--ip);
|
||||||
|
FSE_encodeSymbol(&bitC, &CState1, *--ip);
|
||||||
|
}
|
||||||
|
|
||||||
|
FSE_FLUSHBITS(&bitC);
|
||||||
|
}
|
||||||
|
|
||||||
|
FSE_flushCState(&bitC, &CState2);
|
||||||
|
FSE_flushCState(&bitC, &CState1);
|
||||||
|
return BIT_closeCStream(&bitC);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t FSE_compress_usingCTable (void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const FSE_CTable* ct)
|
||||||
|
{
|
||||||
|
unsigned const fast = (dstSize >= FSE_BLOCKBOUND(srcSize));
|
||||||
|
|
||||||
|
if (fast)
|
||||||
|
return FSE_compress_usingCTable_generic(dst, dstSize, src, srcSize, ct, 1);
|
||||||
|
else
|
||||||
|
return FSE_compress_usingCTable_generic(dst, dstSize, src, srcSize, ct, 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t FSE_compressBound(size_t size) { return FSE_COMPRESSBOUND(size); }
|
||||||
|
|
||||||
|
#define CHECK_V_F(e, f) size_t const e = f; if (ERR_isError(e)) return e
|
||||||
|
#define CHECK_F(f) { CHECK_V_F(_var_err__, f); }
|
||||||
|
|
||||||
|
/* FSE_compress_wksp() :
|
||||||
|
* Same as FSE_compress2(), but using an externally allocated scratch buffer (`workSpace`).
|
||||||
|
* `wkspSize` size must be `(1<<tableLog)`.
|
||||||
|
*/
|
||||||
|
size_t FSE_compress_wksp (void* dst, size_t dstSize, const void* src, size_t srcSize, unsigned maxSymbolValue, unsigned tableLog, void* workSpace, size_t wkspSize)
|
||||||
|
{
|
||||||
|
BYTE* const ostart = (BYTE*) dst;
|
||||||
|
BYTE* op = ostart;
|
||||||
|
BYTE* const oend = ostart + dstSize;
|
||||||
|
|
||||||
|
unsigned count[FSE_MAX_SYMBOL_VALUE+1];
|
||||||
|
S16 norm[FSE_MAX_SYMBOL_VALUE+1];
|
||||||
|
FSE_CTable* CTable = (FSE_CTable*)workSpace;
|
||||||
|
size_t const CTableSize = FSE_CTABLE_SIZE_U32(tableLog, maxSymbolValue);
|
||||||
|
void* scratchBuffer = (void*)(CTable + CTableSize);
|
||||||
|
size_t const scratchBufferSize = wkspSize - (CTableSize * sizeof(FSE_CTable));
|
||||||
|
|
||||||
|
/* init conditions */
|
||||||
|
if (wkspSize < FSE_WKSP_SIZE_U32(tableLog, maxSymbolValue)) return ERROR(tableLog_tooLarge);
|
||||||
|
if (srcSize <= 1) return 0; /* Not compressible */
|
||||||
|
if (!maxSymbolValue) maxSymbolValue = FSE_MAX_SYMBOL_VALUE;
|
||||||
|
if (!tableLog) tableLog = FSE_DEFAULT_TABLELOG;
|
||||||
|
|
||||||
|
/* Scan input and build symbol stats */
|
||||||
|
{ CHECK_V_F(maxCount, HIST_count_wksp(count, &maxSymbolValue, src, srcSize, scratchBuffer, scratchBufferSize) );
|
||||||
|
if (maxCount == srcSize) return 1; /* only a single symbol in src : rle */
|
||||||
|
if (maxCount == 1) return 0; /* each symbol present maximum once => not compressible */
|
||||||
|
if (maxCount < (srcSize >> 7)) return 0; /* Heuristic : not compressible enough */
|
||||||
|
}
|
||||||
|
|
||||||
|
tableLog = FSE_optimalTableLog(tableLog, srcSize, maxSymbolValue);
|
||||||
|
CHECK_F( FSE_normalizeCount(norm, tableLog, count, srcSize, maxSymbolValue) );
|
||||||
|
|
||||||
|
/* Write table description header */
|
||||||
|
{ CHECK_V_F(nc_err, FSE_writeNCount(op, oend-op, norm, maxSymbolValue, tableLog) );
|
||||||
|
op += nc_err;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Compress */
|
||||||
|
CHECK_F( FSE_buildCTable_wksp(CTable, norm, maxSymbolValue, tableLog, scratchBuffer, scratchBufferSize) );
|
||||||
|
{ CHECK_V_F(cSize, FSE_compress_usingCTable(op, oend - op, src, srcSize, CTable) );
|
||||||
|
if (cSize == 0) return 0; /* not enough space for compressed data */
|
||||||
|
op += cSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* check compressibility */
|
||||||
|
if ( (size_t)(op-ostart) >= srcSize-1 ) return 0;
|
||||||
|
|
||||||
|
return op-ostart;
|
||||||
|
}
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
FSE_CTable CTable_max[FSE_CTABLE_SIZE_U32(FSE_MAX_TABLELOG, FSE_MAX_SYMBOL_VALUE)];
|
||||||
|
BYTE scratchBuffer[1 << FSE_MAX_TABLELOG];
|
||||||
|
} fseWkspMax_t;
|
||||||
|
|
||||||
|
size_t FSE_compress2 (void* dst, size_t dstCapacity, const void* src, size_t srcSize, unsigned maxSymbolValue, unsigned tableLog)
|
||||||
|
{
|
||||||
|
fseWkspMax_t scratchBuffer;
|
||||||
|
DEBUG_STATIC_ASSERT(sizeof(scratchBuffer) >= FSE_WKSP_SIZE_U32(FSE_MAX_TABLELOG, FSE_MAX_SYMBOL_VALUE)); /* compilation failures here means scratchBuffer is not large enough */
|
||||||
|
if (tableLog > FSE_MAX_TABLELOG) return ERROR(tableLog_tooLarge);
|
||||||
|
return FSE_compress_wksp(dst, dstCapacity, src, srcSize, maxSymbolValue, tableLog, &scratchBuffer, sizeof(scratchBuffer));
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t FSE_compress (void* dst, size_t dstCapacity, const void* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
return FSE_compress2(dst, dstCapacity, src, srcSize, FSE_MAX_SYMBOL_VALUE, FSE_DEFAULT_TABLELOG);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#endif /* FSE_COMMONDEFS_ONLY */
|
||||||
311
vendor/github.com/DataDog/zstd/fse_decompress.c
generated
vendored
Normal file
311
vendor/github.com/DataDog/zstd/fse_decompress.c
generated
vendored
Normal file
@@ -0,0 +1,311 @@
|
|||||||
|
/* ******************************************************************
|
||||||
|
FSE : Finite State Entropy decoder
|
||||||
|
Copyright (C) 2013-2015, Yann Collet.
|
||||||
|
|
||||||
|
BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
You can contact the author at :
|
||||||
|
- FSE source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
||||||
|
- Public forum : https://groups.google.com/forum/#!forum/lz4c
|
||||||
|
****************************************************************** */
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************************************************
|
||||||
|
* Includes
|
||||||
|
****************************************************************/
|
||||||
|
#include <stdlib.h> /* malloc, free, qsort */
|
||||||
|
#include <string.h> /* memcpy, memset */
|
||||||
|
#include "bitstream.h"
|
||||||
|
#include "compiler.h"
|
||||||
|
#define FSE_STATIC_LINKING_ONLY
|
||||||
|
#include "fse.h"
|
||||||
|
#include "error_private.h"
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************************************************
|
||||||
|
* Error Management
|
||||||
|
****************************************************************/
|
||||||
|
#define FSE_isError ERR_isError
|
||||||
|
#define FSE_STATIC_ASSERT(c) DEBUG_STATIC_ASSERT(c) /* use only *after* variable declarations */
|
||||||
|
|
||||||
|
/* check and forward error code */
|
||||||
|
#ifndef CHECK_F
|
||||||
|
#define CHECK_F(f) { size_t const e = f; if (FSE_isError(e)) return e; }
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************************************************
|
||||||
|
* Templates
|
||||||
|
****************************************************************/
|
||||||
|
/*
|
||||||
|
designed to be included
|
||||||
|
for type-specific functions (template emulation in C)
|
||||||
|
Objective is to write these functions only once, for improved maintenance
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* safety checks */
|
||||||
|
#ifndef FSE_FUNCTION_EXTENSION
|
||||||
|
# error "FSE_FUNCTION_EXTENSION must be defined"
|
||||||
|
#endif
|
||||||
|
#ifndef FSE_FUNCTION_TYPE
|
||||||
|
# error "FSE_FUNCTION_TYPE must be defined"
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* Function names */
|
||||||
|
#define FSE_CAT(X,Y) X##Y
|
||||||
|
#define FSE_FUNCTION_NAME(X,Y) FSE_CAT(X,Y)
|
||||||
|
#define FSE_TYPE_NAME(X,Y) FSE_CAT(X,Y)
|
||||||
|
|
||||||
|
|
||||||
|
/* Function templates */
|
||||||
|
FSE_DTable* FSE_createDTable (unsigned tableLog)
|
||||||
|
{
|
||||||
|
if (tableLog > FSE_TABLELOG_ABSOLUTE_MAX) tableLog = FSE_TABLELOG_ABSOLUTE_MAX;
|
||||||
|
return (FSE_DTable*)malloc( FSE_DTABLE_SIZE_U32(tableLog) * sizeof (U32) );
|
||||||
|
}
|
||||||
|
|
||||||
|
void FSE_freeDTable (FSE_DTable* dt)
|
||||||
|
{
|
||||||
|
free(dt);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t FSE_buildDTable(FSE_DTable* dt, const short* normalizedCounter, unsigned maxSymbolValue, unsigned tableLog)
|
||||||
|
{
|
||||||
|
void* const tdPtr = dt+1; /* because *dt is unsigned, 32-bits aligned on 32-bits */
|
||||||
|
FSE_DECODE_TYPE* const tableDecode = (FSE_DECODE_TYPE*) (tdPtr);
|
||||||
|
U16 symbolNext[FSE_MAX_SYMBOL_VALUE+1];
|
||||||
|
|
||||||
|
U32 const maxSV1 = maxSymbolValue + 1;
|
||||||
|
U32 const tableSize = 1 << tableLog;
|
||||||
|
U32 highThreshold = tableSize-1;
|
||||||
|
|
||||||
|
/* Sanity Checks */
|
||||||
|
if (maxSymbolValue > FSE_MAX_SYMBOL_VALUE) return ERROR(maxSymbolValue_tooLarge);
|
||||||
|
if (tableLog > FSE_MAX_TABLELOG) return ERROR(tableLog_tooLarge);
|
||||||
|
|
||||||
|
/* Init, lay down lowprob symbols */
|
||||||
|
{ FSE_DTableHeader DTableH;
|
||||||
|
DTableH.tableLog = (U16)tableLog;
|
||||||
|
DTableH.fastMode = 1;
|
||||||
|
{ S16 const largeLimit= (S16)(1 << (tableLog-1));
|
||||||
|
U32 s;
|
||||||
|
for (s=0; s<maxSV1; s++) {
|
||||||
|
if (normalizedCounter[s]==-1) {
|
||||||
|
tableDecode[highThreshold--].symbol = (FSE_FUNCTION_TYPE)s;
|
||||||
|
symbolNext[s] = 1;
|
||||||
|
} else {
|
||||||
|
if (normalizedCounter[s] >= largeLimit) DTableH.fastMode=0;
|
||||||
|
symbolNext[s] = normalizedCounter[s];
|
||||||
|
} } }
|
||||||
|
memcpy(dt, &DTableH, sizeof(DTableH));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Spread symbols */
|
||||||
|
{ U32 const tableMask = tableSize-1;
|
||||||
|
U32 const step = FSE_TABLESTEP(tableSize);
|
||||||
|
U32 s, position = 0;
|
||||||
|
for (s=0; s<maxSV1; s++) {
|
||||||
|
int i;
|
||||||
|
for (i=0; i<normalizedCounter[s]; i++) {
|
||||||
|
tableDecode[position].symbol = (FSE_FUNCTION_TYPE)s;
|
||||||
|
position = (position + step) & tableMask;
|
||||||
|
while (position > highThreshold) position = (position + step) & tableMask; /* lowprob area */
|
||||||
|
} }
|
||||||
|
if (position!=0) return ERROR(GENERIC); /* position must reach all cells once, otherwise normalizedCounter is incorrect */
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Build Decoding table */
|
||||||
|
{ U32 u;
|
||||||
|
for (u=0; u<tableSize; u++) {
|
||||||
|
FSE_FUNCTION_TYPE const symbol = (FSE_FUNCTION_TYPE)(tableDecode[u].symbol);
|
||||||
|
U32 const nextState = symbolNext[symbol]++;
|
||||||
|
tableDecode[u].nbBits = (BYTE) (tableLog - BIT_highbit32(nextState) );
|
||||||
|
tableDecode[u].newState = (U16) ( (nextState << tableDecode[u].nbBits) - tableSize);
|
||||||
|
} }
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#ifndef FSE_COMMONDEFS_ONLY
|
||||||
|
|
||||||
|
/*-*******************************************************
|
||||||
|
* Decompression (Byte symbols)
|
||||||
|
*********************************************************/
|
||||||
|
size_t FSE_buildDTable_rle (FSE_DTable* dt, BYTE symbolValue)
|
||||||
|
{
|
||||||
|
void* ptr = dt;
|
||||||
|
FSE_DTableHeader* const DTableH = (FSE_DTableHeader*)ptr;
|
||||||
|
void* dPtr = dt + 1;
|
||||||
|
FSE_decode_t* const cell = (FSE_decode_t*)dPtr;
|
||||||
|
|
||||||
|
DTableH->tableLog = 0;
|
||||||
|
DTableH->fastMode = 0;
|
||||||
|
|
||||||
|
cell->newState = 0;
|
||||||
|
cell->symbol = symbolValue;
|
||||||
|
cell->nbBits = 0;
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t FSE_buildDTable_raw (FSE_DTable* dt, unsigned nbBits)
|
||||||
|
{
|
||||||
|
void* ptr = dt;
|
||||||
|
FSE_DTableHeader* const DTableH = (FSE_DTableHeader*)ptr;
|
||||||
|
void* dPtr = dt + 1;
|
||||||
|
FSE_decode_t* const dinfo = (FSE_decode_t*)dPtr;
|
||||||
|
const unsigned tableSize = 1 << nbBits;
|
||||||
|
const unsigned tableMask = tableSize - 1;
|
||||||
|
const unsigned maxSV1 = tableMask+1;
|
||||||
|
unsigned s;
|
||||||
|
|
||||||
|
/* Sanity checks */
|
||||||
|
if (nbBits < 1) return ERROR(GENERIC); /* min size */
|
||||||
|
|
||||||
|
/* Build Decoding Table */
|
||||||
|
DTableH->tableLog = (U16)nbBits;
|
||||||
|
DTableH->fastMode = 1;
|
||||||
|
for (s=0; s<maxSV1; s++) {
|
||||||
|
dinfo[s].newState = 0;
|
||||||
|
dinfo[s].symbol = (BYTE)s;
|
||||||
|
dinfo[s].nbBits = (BYTE)nbBits;
|
||||||
|
}
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE size_t FSE_decompress_usingDTable_generic(
|
||||||
|
void* dst, size_t maxDstSize,
|
||||||
|
const void* cSrc, size_t cSrcSize,
|
||||||
|
const FSE_DTable* dt, const unsigned fast)
|
||||||
|
{
|
||||||
|
BYTE* const ostart = (BYTE*) dst;
|
||||||
|
BYTE* op = ostart;
|
||||||
|
BYTE* const omax = op + maxDstSize;
|
||||||
|
BYTE* const olimit = omax-3;
|
||||||
|
|
||||||
|
BIT_DStream_t bitD;
|
||||||
|
FSE_DState_t state1;
|
||||||
|
FSE_DState_t state2;
|
||||||
|
|
||||||
|
/* Init */
|
||||||
|
CHECK_F(BIT_initDStream(&bitD, cSrc, cSrcSize));
|
||||||
|
|
||||||
|
FSE_initDState(&state1, &bitD, dt);
|
||||||
|
FSE_initDState(&state2, &bitD, dt);
|
||||||
|
|
||||||
|
#define FSE_GETSYMBOL(statePtr) fast ? FSE_decodeSymbolFast(statePtr, &bitD) : FSE_decodeSymbol(statePtr, &bitD)
|
||||||
|
|
||||||
|
/* 4 symbols per loop */
|
||||||
|
for ( ; (BIT_reloadDStream(&bitD)==BIT_DStream_unfinished) & (op<olimit) ; op+=4) {
|
||||||
|
op[0] = FSE_GETSYMBOL(&state1);
|
||||||
|
|
||||||
|
if (FSE_MAX_TABLELOG*2+7 > sizeof(bitD.bitContainer)*8) /* This test must be static */
|
||||||
|
BIT_reloadDStream(&bitD);
|
||||||
|
|
||||||
|
op[1] = FSE_GETSYMBOL(&state2);
|
||||||
|
|
||||||
|
if (FSE_MAX_TABLELOG*4+7 > sizeof(bitD.bitContainer)*8) /* This test must be static */
|
||||||
|
{ if (BIT_reloadDStream(&bitD) > BIT_DStream_unfinished) { op+=2; break; } }
|
||||||
|
|
||||||
|
op[2] = FSE_GETSYMBOL(&state1);
|
||||||
|
|
||||||
|
if (FSE_MAX_TABLELOG*2+7 > sizeof(bitD.bitContainer)*8) /* This test must be static */
|
||||||
|
BIT_reloadDStream(&bitD);
|
||||||
|
|
||||||
|
op[3] = FSE_GETSYMBOL(&state2);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* tail */
|
||||||
|
/* note : BIT_reloadDStream(&bitD) >= FSE_DStream_partiallyFilled; Ends at exactly BIT_DStream_completed */
|
||||||
|
while (1) {
|
||||||
|
if (op>(omax-2)) return ERROR(dstSize_tooSmall);
|
||||||
|
*op++ = FSE_GETSYMBOL(&state1);
|
||||||
|
if (BIT_reloadDStream(&bitD)==BIT_DStream_overflow) {
|
||||||
|
*op++ = FSE_GETSYMBOL(&state2);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (op>(omax-2)) return ERROR(dstSize_tooSmall);
|
||||||
|
*op++ = FSE_GETSYMBOL(&state2);
|
||||||
|
if (BIT_reloadDStream(&bitD)==BIT_DStream_overflow) {
|
||||||
|
*op++ = FSE_GETSYMBOL(&state1);
|
||||||
|
break;
|
||||||
|
} }
|
||||||
|
|
||||||
|
return op-ostart;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t FSE_decompress_usingDTable(void* dst, size_t originalSize,
|
||||||
|
const void* cSrc, size_t cSrcSize,
|
||||||
|
const FSE_DTable* dt)
|
||||||
|
{
|
||||||
|
const void* ptr = dt;
|
||||||
|
const FSE_DTableHeader* DTableH = (const FSE_DTableHeader*)ptr;
|
||||||
|
const U32 fastMode = DTableH->fastMode;
|
||||||
|
|
||||||
|
/* select fast mode (static) */
|
||||||
|
if (fastMode) return FSE_decompress_usingDTable_generic(dst, originalSize, cSrc, cSrcSize, dt, 1);
|
||||||
|
return FSE_decompress_usingDTable_generic(dst, originalSize, cSrc, cSrcSize, dt, 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t FSE_decompress_wksp(void* dst, size_t dstCapacity, const void* cSrc, size_t cSrcSize, FSE_DTable* workSpace, unsigned maxLog)
|
||||||
|
{
|
||||||
|
const BYTE* const istart = (const BYTE*)cSrc;
|
||||||
|
const BYTE* ip = istart;
|
||||||
|
short counting[FSE_MAX_SYMBOL_VALUE+1];
|
||||||
|
unsigned tableLog;
|
||||||
|
unsigned maxSymbolValue = FSE_MAX_SYMBOL_VALUE;
|
||||||
|
|
||||||
|
/* normal FSE decoding mode */
|
||||||
|
size_t const NCountLength = FSE_readNCount (counting, &maxSymbolValue, &tableLog, istart, cSrcSize);
|
||||||
|
if (FSE_isError(NCountLength)) return NCountLength;
|
||||||
|
//if (NCountLength >= cSrcSize) return ERROR(srcSize_wrong); /* too small input size; supposed to be already checked in NCountLength, only remaining case : NCountLength==cSrcSize */
|
||||||
|
if (tableLog > maxLog) return ERROR(tableLog_tooLarge);
|
||||||
|
ip += NCountLength;
|
||||||
|
cSrcSize -= NCountLength;
|
||||||
|
|
||||||
|
CHECK_F( FSE_buildDTable (workSpace, counting, maxSymbolValue, tableLog) );
|
||||||
|
|
||||||
|
return FSE_decompress_usingDTable (dst, dstCapacity, ip, cSrcSize, workSpace); /* always return, even if it is an error code */
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
typedef FSE_DTable DTable_max_t[FSE_DTABLE_SIZE_U32(FSE_MAX_TABLELOG)];
|
||||||
|
|
||||||
|
size_t FSE_decompress(void* dst, size_t dstCapacity, const void* cSrc, size_t cSrcSize)
|
||||||
|
{
|
||||||
|
DTable_max_t dt; /* Static analyzer seems unable to understand this table will be properly initialized later */
|
||||||
|
return FSE_decompress_wksp(dst, dstCapacity, cSrc, cSrcSize, dt, FSE_MAX_TABLELOG);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
#endif /* FSE_COMMONDEFS_ONLY */
|
||||||
203
vendor/github.com/DataDog/zstd/hist.c
generated
vendored
Normal file
203
vendor/github.com/DataDog/zstd/hist.c
generated
vendored
Normal file
@@ -0,0 +1,203 @@
|
|||||||
|
/* ******************************************************************
|
||||||
|
hist : Histogram functions
|
||||||
|
part of Finite State Entropy project
|
||||||
|
Copyright (C) 2013-present, Yann Collet.
|
||||||
|
|
||||||
|
BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
You can contact the author at :
|
||||||
|
- FSE source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
||||||
|
- Public forum : https://groups.google.com/forum/#!forum/lz4c
|
||||||
|
****************************************************************** */
|
||||||
|
|
||||||
|
/* --- dependencies --- */
|
||||||
|
#include "mem.h" /* U32, BYTE, etc. */
|
||||||
|
#include "debug.h" /* assert, DEBUGLOG */
|
||||||
|
#include "error_private.h" /* ERROR */
|
||||||
|
#include "hist.h"
|
||||||
|
|
||||||
|
|
||||||
|
/* --- Error management --- */
|
||||||
|
unsigned HIST_isError(size_t code) { return ERR_isError(code); }
|
||||||
|
|
||||||
|
/*-**************************************************************
|
||||||
|
* Histogram functions
|
||||||
|
****************************************************************/
|
||||||
|
unsigned HIST_count_simple(unsigned* count, unsigned* maxSymbolValuePtr,
|
||||||
|
const void* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
const BYTE* ip = (const BYTE*)src;
|
||||||
|
const BYTE* const end = ip + srcSize;
|
||||||
|
unsigned maxSymbolValue = *maxSymbolValuePtr;
|
||||||
|
unsigned largestCount=0;
|
||||||
|
|
||||||
|
memset(count, 0, (maxSymbolValue+1) * sizeof(*count));
|
||||||
|
if (srcSize==0) { *maxSymbolValuePtr = 0; return 0; }
|
||||||
|
|
||||||
|
while (ip<end) {
|
||||||
|
assert(*ip <= maxSymbolValue);
|
||||||
|
count[*ip++]++;
|
||||||
|
}
|
||||||
|
|
||||||
|
while (!count[maxSymbolValue]) maxSymbolValue--;
|
||||||
|
*maxSymbolValuePtr = maxSymbolValue;
|
||||||
|
|
||||||
|
{ U32 s;
|
||||||
|
for (s=0; s<=maxSymbolValue; s++)
|
||||||
|
if (count[s] > largestCount) largestCount = count[s];
|
||||||
|
}
|
||||||
|
|
||||||
|
return largestCount;
|
||||||
|
}
|
||||||
|
|
||||||
|
typedef enum { trustInput, checkMaxSymbolValue } HIST_checkInput_e;
|
||||||
|
|
||||||
|
/* HIST_count_parallel_wksp() :
|
||||||
|
* store histogram into 4 intermediate tables, recombined at the end.
|
||||||
|
* this design makes better use of OoO cpus,
|
||||||
|
* and is noticeably faster when some values are heavily repeated.
|
||||||
|
* But it needs some additional workspace for intermediate tables.
|
||||||
|
* `workSpace` size must be a table of size >= HIST_WKSP_SIZE_U32.
|
||||||
|
* @return : largest histogram frequency,
|
||||||
|
* or an error code (notably when histogram would be larger than *maxSymbolValuePtr). */
|
||||||
|
static size_t HIST_count_parallel_wksp(
|
||||||
|
unsigned* count, unsigned* maxSymbolValuePtr,
|
||||||
|
const void* source, size_t sourceSize,
|
||||||
|
HIST_checkInput_e check,
|
||||||
|
U32* const workSpace)
|
||||||
|
{
|
||||||
|
const BYTE* ip = (const BYTE*)source;
|
||||||
|
const BYTE* const iend = ip+sourceSize;
|
||||||
|
unsigned maxSymbolValue = *maxSymbolValuePtr;
|
||||||
|
unsigned max=0;
|
||||||
|
U32* const Counting1 = workSpace;
|
||||||
|
U32* const Counting2 = Counting1 + 256;
|
||||||
|
U32* const Counting3 = Counting2 + 256;
|
||||||
|
U32* const Counting4 = Counting3 + 256;
|
||||||
|
|
||||||
|
memset(workSpace, 0, 4*256*sizeof(unsigned));
|
||||||
|
|
||||||
|
/* safety checks */
|
||||||
|
if (!sourceSize) {
|
||||||
|
memset(count, 0, maxSymbolValue + 1);
|
||||||
|
*maxSymbolValuePtr = 0;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
if (!maxSymbolValue) maxSymbolValue = 255; /* 0 == default */
|
||||||
|
|
||||||
|
/* by stripes of 16 bytes */
|
||||||
|
{ U32 cached = MEM_read32(ip); ip += 4;
|
||||||
|
while (ip < iend-15) {
|
||||||
|
U32 c = cached; cached = MEM_read32(ip); ip += 4;
|
||||||
|
Counting1[(BYTE) c ]++;
|
||||||
|
Counting2[(BYTE)(c>>8) ]++;
|
||||||
|
Counting3[(BYTE)(c>>16)]++;
|
||||||
|
Counting4[ c>>24 ]++;
|
||||||
|
c = cached; cached = MEM_read32(ip); ip += 4;
|
||||||
|
Counting1[(BYTE) c ]++;
|
||||||
|
Counting2[(BYTE)(c>>8) ]++;
|
||||||
|
Counting3[(BYTE)(c>>16)]++;
|
||||||
|
Counting4[ c>>24 ]++;
|
||||||
|
c = cached; cached = MEM_read32(ip); ip += 4;
|
||||||
|
Counting1[(BYTE) c ]++;
|
||||||
|
Counting2[(BYTE)(c>>8) ]++;
|
||||||
|
Counting3[(BYTE)(c>>16)]++;
|
||||||
|
Counting4[ c>>24 ]++;
|
||||||
|
c = cached; cached = MEM_read32(ip); ip += 4;
|
||||||
|
Counting1[(BYTE) c ]++;
|
||||||
|
Counting2[(BYTE)(c>>8) ]++;
|
||||||
|
Counting3[(BYTE)(c>>16)]++;
|
||||||
|
Counting4[ c>>24 ]++;
|
||||||
|
}
|
||||||
|
ip-=4;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* finish last symbols */
|
||||||
|
while (ip<iend) Counting1[*ip++]++;
|
||||||
|
|
||||||
|
if (check) { /* verify stats will fit into destination table */
|
||||||
|
U32 s; for (s=255; s>maxSymbolValue; s--) {
|
||||||
|
Counting1[s] += Counting2[s] + Counting3[s] + Counting4[s];
|
||||||
|
if (Counting1[s]) return ERROR(maxSymbolValue_tooSmall);
|
||||||
|
} }
|
||||||
|
|
||||||
|
{ U32 s;
|
||||||
|
if (maxSymbolValue > 255) maxSymbolValue = 255;
|
||||||
|
for (s=0; s<=maxSymbolValue; s++) {
|
||||||
|
count[s] = Counting1[s] + Counting2[s] + Counting3[s] + Counting4[s];
|
||||||
|
if (count[s] > max) max = count[s];
|
||||||
|
} }
|
||||||
|
|
||||||
|
while (!count[maxSymbolValue]) maxSymbolValue--;
|
||||||
|
*maxSymbolValuePtr = maxSymbolValue;
|
||||||
|
return (size_t)max;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* HIST_countFast_wksp() :
|
||||||
|
* Same as HIST_countFast(), but using an externally provided scratch buffer.
|
||||||
|
* `workSpace` is a writable buffer which must be 4-bytes aligned,
|
||||||
|
* `workSpaceSize` must be >= HIST_WKSP_SIZE
|
||||||
|
*/
|
||||||
|
size_t HIST_countFast_wksp(unsigned* count, unsigned* maxSymbolValuePtr,
|
||||||
|
const void* source, size_t sourceSize,
|
||||||
|
void* workSpace, size_t workSpaceSize)
|
||||||
|
{
|
||||||
|
if (sourceSize < 1500) /* heuristic threshold */
|
||||||
|
return HIST_count_simple(count, maxSymbolValuePtr, source, sourceSize);
|
||||||
|
if ((size_t)workSpace & 3) return ERROR(GENERIC); /* must be aligned on 4-bytes boundaries */
|
||||||
|
if (workSpaceSize < HIST_WKSP_SIZE) return ERROR(workSpace_tooSmall);
|
||||||
|
return HIST_count_parallel_wksp(count, maxSymbolValuePtr, source, sourceSize, trustInput, (U32*)workSpace);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* fast variant (unsafe : won't check if src contains values beyond count[] limit) */
|
||||||
|
size_t HIST_countFast(unsigned* count, unsigned* maxSymbolValuePtr,
|
||||||
|
const void* source, size_t sourceSize)
|
||||||
|
{
|
||||||
|
unsigned tmpCounters[HIST_WKSP_SIZE_U32];
|
||||||
|
return HIST_countFast_wksp(count, maxSymbolValuePtr, source, sourceSize, tmpCounters, sizeof(tmpCounters));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* HIST_count_wksp() :
|
||||||
|
* Same as HIST_count(), but using an externally provided scratch buffer.
|
||||||
|
* `workSpace` size must be table of >= HIST_WKSP_SIZE_U32 unsigned */
|
||||||
|
size_t HIST_count_wksp(unsigned* count, unsigned* maxSymbolValuePtr,
|
||||||
|
const void* source, size_t sourceSize,
|
||||||
|
void* workSpace, size_t workSpaceSize)
|
||||||
|
{
|
||||||
|
if ((size_t)workSpace & 3) return ERROR(GENERIC); /* must be aligned on 4-bytes boundaries */
|
||||||
|
if (workSpaceSize < HIST_WKSP_SIZE) return ERROR(workSpace_tooSmall);
|
||||||
|
if (*maxSymbolValuePtr < 255)
|
||||||
|
return HIST_count_parallel_wksp(count, maxSymbolValuePtr, source, sourceSize, checkMaxSymbolValue, (U32*)workSpace);
|
||||||
|
*maxSymbolValuePtr = 255;
|
||||||
|
return HIST_countFast_wksp(count, maxSymbolValuePtr, source, sourceSize, workSpace, workSpaceSize);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t HIST_count(unsigned* count, unsigned* maxSymbolValuePtr,
|
||||||
|
const void* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
unsigned tmpCounters[HIST_WKSP_SIZE_U32];
|
||||||
|
return HIST_count_wksp(count, maxSymbolValuePtr, src, srcSize, tmpCounters, sizeof(tmpCounters));
|
||||||
|
}
|
||||||
95
vendor/github.com/DataDog/zstd/hist.h
generated
vendored
Normal file
95
vendor/github.com/DataDog/zstd/hist.h
generated
vendored
Normal file
@@ -0,0 +1,95 @@
|
|||||||
|
/* ******************************************************************
|
||||||
|
hist : Histogram functions
|
||||||
|
part of Finite State Entropy project
|
||||||
|
Copyright (C) 2013-present, Yann Collet.
|
||||||
|
|
||||||
|
BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
You can contact the author at :
|
||||||
|
- FSE source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
||||||
|
- Public forum : https://groups.google.com/forum/#!forum/lz4c
|
||||||
|
****************************************************************** */
|
||||||
|
|
||||||
|
/* --- dependencies --- */
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
|
||||||
|
|
||||||
|
/* --- simple histogram functions --- */
|
||||||
|
|
||||||
|
/*! HIST_count():
|
||||||
|
* Provides the precise count of each byte within a table 'count'.
|
||||||
|
* 'count' is a table of unsigned int, of minimum size (*maxSymbolValuePtr+1).
|
||||||
|
* Updates *maxSymbolValuePtr with actual largest symbol value detected.
|
||||||
|
* @return : count of the most frequent symbol (which isn't identified).
|
||||||
|
* or an error code, which can be tested using HIST_isError().
|
||||||
|
* note : if return == srcSize, there is only one symbol.
|
||||||
|
*/
|
||||||
|
size_t HIST_count(unsigned* count, unsigned* maxSymbolValuePtr,
|
||||||
|
const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
unsigned HIST_isError(size_t code); /**< tells if a return value is an error code */
|
||||||
|
|
||||||
|
|
||||||
|
/* --- advanced histogram functions --- */
|
||||||
|
|
||||||
|
#define HIST_WKSP_SIZE_U32 1024
|
||||||
|
#define HIST_WKSP_SIZE (HIST_WKSP_SIZE_U32 * sizeof(unsigned))
|
||||||
|
/** HIST_count_wksp() :
|
||||||
|
* Same as HIST_count(), but using an externally provided scratch buffer.
|
||||||
|
* Benefit is this function will use very little stack space.
|
||||||
|
* `workSpace` is a writable buffer which must be 4-bytes aligned,
|
||||||
|
* `workSpaceSize` must be >= HIST_WKSP_SIZE
|
||||||
|
*/
|
||||||
|
size_t HIST_count_wksp(unsigned* count, unsigned* maxSymbolValuePtr,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
void* workSpace, size_t workSpaceSize);
|
||||||
|
|
||||||
|
/** HIST_countFast() :
|
||||||
|
* same as HIST_count(), but blindly trusts that all byte values within src are <= *maxSymbolValuePtr.
|
||||||
|
* This function is unsafe, and will segfault if any value within `src` is `> *maxSymbolValuePtr`
|
||||||
|
*/
|
||||||
|
size_t HIST_countFast(unsigned* count, unsigned* maxSymbolValuePtr,
|
||||||
|
const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
/** HIST_countFast_wksp() :
|
||||||
|
* Same as HIST_countFast(), but using an externally provided scratch buffer.
|
||||||
|
* `workSpace` is a writable buffer which must be 4-bytes aligned,
|
||||||
|
* `workSpaceSize` must be >= HIST_WKSP_SIZE
|
||||||
|
*/
|
||||||
|
size_t HIST_countFast_wksp(unsigned* count, unsigned* maxSymbolValuePtr,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
void* workSpace, size_t workSpaceSize);
|
||||||
|
|
||||||
|
/*! HIST_count_simple() :
|
||||||
|
* Same as HIST_countFast(), this function is unsafe,
|
||||||
|
* and will segfault if any value within `src` is `> *maxSymbolValuePtr`.
|
||||||
|
* It is also a bit slower for large inputs.
|
||||||
|
* However, it does not need any additional memory (not even on stack).
|
||||||
|
* @return : count of the most frequent symbol.
|
||||||
|
* Note this function doesn't produce any error (i.e. it must succeed).
|
||||||
|
*/
|
||||||
|
unsigned HIST_count_simple(unsigned* count, unsigned* maxSymbolValuePtr,
|
||||||
|
const void* src, size_t srcSize);
|
||||||
358
vendor/github.com/DataDog/zstd/huf.h
generated
vendored
Normal file
358
vendor/github.com/DataDog/zstd/huf.h
generated
vendored
Normal file
@@ -0,0 +1,358 @@
|
|||||||
|
/* ******************************************************************
|
||||||
|
huff0 huffman codec,
|
||||||
|
part of Finite State Entropy library
|
||||||
|
Copyright (C) 2013-present, Yann Collet.
|
||||||
|
|
||||||
|
BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
You can contact the author at :
|
||||||
|
- Source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
||||||
|
****************************************************************** */
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#ifndef HUF_H_298734234
|
||||||
|
#define HUF_H_298734234
|
||||||
|
|
||||||
|
/* *** Dependencies *** */
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
|
||||||
|
|
||||||
|
/* *** library symbols visibility *** */
|
||||||
|
/* Note : when linking with -fvisibility=hidden on gcc, or by default on Visual,
|
||||||
|
* HUF symbols remain "private" (internal symbols for library only).
|
||||||
|
* Set macro FSE_DLL_EXPORT to 1 if you want HUF symbols visible on DLL interface */
|
||||||
|
#if defined(FSE_DLL_EXPORT) && (FSE_DLL_EXPORT==1) && defined(__GNUC__) && (__GNUC__ >= 4)
|
||||||
|
# define HUF_PUBLIC_API __attribute__ ((visibility ("default")))
|
||||||
|
#elif defined(FSE_DLL_EXPORT) && (FSE_DLL_EXPORT==1) /* Visual expected */
|
||||||
|
# define HUF_PUBLIC_API __declspec(dllexport)
|
||||||
|
#elif defined(FSE_DLL_IMPORT) && (FSE_DLL_IMPORT==1)
|
||||||
|
# define HUF_PUBLIC_API __declspec(dllimport) /* not required, just to generate faster code (saves a function pointer load from IAT and an indirect jump) */
|
||||||
|
#else
|
||||||
|
# define HUF_PUBLIC_API
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* ========================== */
|
||||||
|
/* *** simple functions *** */
|
||||||
|
/* ========================== */
|
||||||
|
|
||||||
|
/** HUF_compress() :
|
||||||
|
* Compress content from buffer 'src', of size 'srcSize', into buffer 'dst'.
|
||||||
|
* 'dst' buffer must be already allocated.
|
||||||
|
* Compression runs faster if `dstCapacity` >= HUF_compressBound(srcSize).
|
||||||
|
* `srcSize` must be <= `HUF_BLOCKSIZE_MAX` == 128 KB.
|
||||||
|
* @return : size of compressed data (<= `dstCapacity`).
|
||||||
|
* Special values : if return == 0, srcData is not compressible => Nothing is stored within dst !!!
|
||||||
|
* if HUF_isError(return), compression failed (more details using HUF_getErrorName())
|
||||||
|
*/
|
||||||
|
HUF_PUBLIC_API size_t HUF_compress(void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
/** HUF_decompress() :
|
||||||
|
* Decompress HUF data from buffer 'cSrc', of size 'cSrcSize',
|
||||||
|
* into already allocated buffer 'dst', of minimum size 'dstSize'.
|
||||||
|
* `originalSize` : **must** be the ***exact*** size of original (uncompressed) data.
|
||||||
|
* Note : in contrast with FSE, HUF_decompress can regenerate
|
||||||
|
* RLE (cSrcSize==1) and uncompressed (cSrcSize==dstSize) data,
|
||||||
|
* because it knows size to regenerate (originalSize).
|
||||||
|
* @return : size of regenerated data (== originalSize),
|
||||||
|
* or an error code, which can be tested using HUF_isError()
|
||||||
|
*/
|
||||||
|
HUF_PUBLIC_API size_t HUF_decompress(void* dst, size_t originalSize,
|
||||||
|
const void* cSrc, size_t cSrcSize);
|
||||||
|
|
||||||
|
|
||||||
|
/* *** Tool functions *** */
|
||||||
|
#define HUF_BLOCKSIZE_MAX (128 * 1024) /**< maximum input size for a single block compressed with HUF_compress */
|
||||||
|
HUF_PUBLIC_API size_t HUF_compressBound(size_t size); /**< maximum compressed size (worst case) */
|
||||||
|
|
||||||
|
/* Error Management */
|
||||||
|
HUF_PUBLIC_API unsigned HUF_isError(size_t code); /**< tells if a return value is an error code */
|
||||||
|
HUF_PUBLIC_API const char* HUF_getErrorName(size_t code); /**< provides error code string (useful for debugging) */
|
||||||
|
|
||||||
|
|
||||||
|
/* *** Advanced function *** */
|
||||||
|
|
||||||
|
/** HUF_compress2() :
|
||||||
|
* Same as HUF_compress(), but offers control over `maxSymbolValue` and `tableLog`.
|
||||||
|
* `maxSymbolValue` must be <= HUF_SYMBOLVALUE_MAX .
|
||||||
|
* `tableLog` must be `<= HUF_TABLELOG_MAX` . */
|
||||||
|
HUF_PUBLIC_API size_t HUF_compress2 (void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
unsigned maxSymbolValue, unsigned tableLog);
|
||||||
|
|
||||||
|
/** HUF_compress4X_wksp() :
|
||||||
|
* Same as HUF_compress2(), but uses externally allocated `workSpace`.
|
||||||
|
* `workspace` must have minimum alignment of 4, and be at least as large as HUF_WORKSPACE_SIZE */
|
||||||
|
#define HUF_WORKSPACE_SIZE (6 << 10)
|
||||||
|
#define HUF_WORKSPACE_SIZE_U32 (HUF_WORKSPACE_SIZE / sizeof(U32))
|
||||||
|
HUF_PUBLIC_API size_t HUF_compress4X_wksp (void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
unsigned maxSymbolValue, unsigned tableLog,
|
||||||
|
void* workSpace, size_t wkspSize);
|
||||||
|
|
||||||
|
#endif /* HUF_H_298734234 */
|
||||||
|
|
||||||
|
/* ******************************************************************
|
||||||
|
* WARNING !!
|
||||||
|
* The following section contains advanced and experimental definitions
|
||||||
|
* which shall never be used in the context of a dynamic library,
|
||||||
|
* because they are not guaranteed to remain stable in the future.
|
||||||
|
* Only consider them in association with static linking.
|
||||||
|
* *****************************************************************/
|
||||||
|
#if defined(HUF_STATIC_LINKING_ONLY) && !defined(HUF_H_HUF_STATIC_LINKING_ONLY)
|
||||||
|
#define HUF_H_HUF_STATIC_LINKING_ONLY
|
||||||
|
|
||||||
|
/* *** Dependencies *** */
|
||||||
|
#include "mem.h" /* U32 */
|
||||||
|
|
||||||
|
|
||||||
|
/* *** Constants *** */
|
||||||
|
#define HUF_TABLELOG_MAX 12 /* max runtime value of tableLog (due to static allocation); can be modified up to HUF_ABSOLUTEMAX_TABLELOG */
|
||||||
|
#define HUF_TABLELOG_DEFAULT 11 /* default tableLog value when none specified */
|
||||||
|
#define HUF_SYMBOLVALUE_MAX 255
|
||||||
|
|
||||||
|
#define HUF_TABLELOG_ABSOLUTEMAX 15 /* absolute limit of HUF_MAX_TABLELOG. Beyond that value, code does not work */
|
||||||
|
#if (HUF_TABLELOG_MAX > HUF_TABLELOG_ABSOLUTEMAX)
|
||||||
|
# error "HUF_TABLELOG_MAX is too large !"
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* ****************************************
|
||||||
|
* Static allocation
|
||||||
|
******************************************/
|
||||||
|
/* HUF buffer bounds */
|
||||||
|
#define HUF_CTABLEBOUND 129
|
||||||
|
#define HUF_BLOCKBOUND(size) (size + (size>>8) + 8) /* only true when incompressible is pre-filtered with fast heuristic */
|
||||||
|
#define HUF_COMPRESSBOUND(size) (HUF_CTABLEBOUND + HUF_BLOCKBOUND(size)) /* Macro version, useful for static allocation */
|
||||||
|
|
||||||
|
/* static allocation of HUF's Compression Table */
|
||||||
|
#define HUF_CTABLE_SIZE_U32(maxSymbolValue) ((maxSymbolValue)+1) /* Use tables of U32, for proper alignment */
|
||||||
|
#define HUF_CTABLE_SIZE(maxSymbolValue) (HUF_CTABLE_SIZE_U32(maxSymbolValue) * sizeof(U32))
|
||||||
|
#define HUF_CREATE_STATIC_CTABLE(name, maxSymbolValue) \
|
||||||
|
U32 name##hb[HUF_CTABLE_SIZE_U32(maxSymbolValue)]; \
|
||||||
|
void* name##hv = &(name##hb); \
|
||||||
|
HUF_CElt* name = (HUF_CElt*)(name##hv) /* no final ; */
|
||||||
|
|
||||||
|
/* static allocation of HUF's DTable */
|
||||||
|
typedef U32 HUF_DTable;
|
||||||
|
#define HUF_DTABLE_SIZE(maxTableLog) (1 + (1<<(maxTableLog)))
|
||||||
|
#define HUF_CREATE_STATIC_DTABLEX1(DTable, maxTableLog) \
|
||||||
|
HUF_DTable DTable[HUF_DTABLE_SIZE((maxTableLog)-1)] = { ((U32)((maxTableLog)-1) * 0x01000001) }
|
||||||
|
#define HUF_CREATE_STATIC_DTABLEX2(DTable, maxTableLog) \
|
||||||
|
HUF_DTable DTable[HUF_DTABLE_SIZE(maxTableLog)] = { ((U32)(maxTableLog) * 0x01000001) }
|
||||||
|
|
||||||
|
|
||||||
|
/* ****************************************
|
||||||
|
* Advanced decompression functions
|
||||||
|
******************************************/
|
||||||
|
size_t HUF_decompress4X1 (void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize); /**< single-symbol decoder */
|
||||||
|
#ifndef HUF_FORCE_DECOMPRESS_X1
|
||||||
|
size_t HUF_decompress4X2 (void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize); /**< double-symbols decoder */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
size_t HUF_decompress4X_DCtx (HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize); /**< decodes RLE and uncompressed */
|
||||||
|
size_t HUF_decompress4X_hufOnly(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize); /**< considers RLE and uncompressed as errors */
|
||||||
|
size_t HUF_decompress4X_hufOnly_wksp(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize, void* workSpace, size_t wkspSize); /**< considers RLE and uncompressed as errors */
|
||||||
|
size_t HUF_decompress4X1_DCtx(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize); /**< single-symbol decoder */
|
||||||
|
size_t HUF_decompress4X1_DCtx_wksp(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize, void* workSpace, size_t wkspSize); /**< single-symbol decoder */
|
||||||
|
#ifndef HUF_FORCE_DECOMPRESS_X1
|
||||||
|
size_t HUF_decompress4X2_DCtx(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize); /**< double-symbols decoder */
|
||||||
|
size_t HUF_decompress4X2_DCtx_wksp(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize, void* workSpace, size_t wkspSize); /**< double-symbols decoder */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* ****************************************
|
||||||
|
* HUF detailed API
|
||||||
|
* ****************************************/
|
||||||
|
|
||||||
|
/*! HUF_compress() does the following:
|
||||||
|
* 1. count symbol occurrence from source[] into table count[] using FSE_count() (exposed within "fse.h")
|
||||||
|
* 2. (optional) refine tableLog using HUF_optimalTableLog()
|
||||||
|
* 3. build Huffman table from count using HUF_buildCTable()
|
||||||
|
* 4. save Huffman table to memory buffer using HUF_writeCTable()
|
||||||
|
* 5. encode the data stream using HUF_compress4X_usingCTable()
|
||||||
|
*
|
||||||
|
* The following API allows targeting specific sub-functions for advanced tasks.
|
||||||
|
* For example, it's possible to compress several blocks using the same 'CTable',
|
||||||
|
* or to save and regenerate 'CTable' using external methods.
|
||||||
|
*/
|
||||||
|
unsigned HUF_optimalTableLog(unsigned maxTableLog, size_t srcSize, unsigned maxSymbolValue);
|
||||||
|
typedef struct HUF_CElt_s HUF_CElt; /* incomplete type */
|
||||||
|
size_t HUF_buildCTable (HUF_CElt* CTable, const unsigned* count, unsigned maxSymbolValue, unsigned maxNbBits); /* @return : maxNbBits; CTable and count can overlap. In which case, CTable will overwrite count content */
|
||||||
|
size_t HUF_writeCTable (void* dst, size_t maxDstSize, const HUF_CElt* CTable, unsigned maxSymbolValue, unsigned huffLog);
|
||||||
|
size_t HUF_compress4X_usingCTable(void* dst, size_t dstSize, const void* src, size_t srcSize, const HUF_CElt* CTable);
|
||||||
|
|
||||||
|
typedef enum {
|
||||||
|
HUF_repeat_none, /**< Cannot use the previous table */
|
||||||
|
HUF_repeat_check, /**< Can use the previous table but it must be checked. Note : The previous table must have been constructed by HUF_compress{1, 4}X_repeat */
|
||||||
|
HUF_repeat_valid /**< Can use the previous table and it is assumed to be valid */
|
||||||
|
} HUF_repeat;
|
||||||
|
/** HUF_compress4X_repeat() :
|
||||||
|
* Same as HUF_compress4X_wksp(), but considers using hufTable if *repeat != HUF_repeat_none.
|
||||||
|
* If it uses hufTable it does not modify hufTable or repeat.
|
||||||
|
* If it doesn't, it sets *repeat = HUF_repeat_none, and it sets hufTable to the table used.
|
||||||
|
* If preferRepeat then the old table will always be used if valid. */
|
||||||
|
size_t HUF_compress4X_repeat(void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
unsigned maxSymbolValue, unsigned tableLog,
|
||||||
|
void* workSpace, size_t wkspSize, /**< `workSpace` must be aligned on 4-bytes boundaries, `wkspSize` must be >= HUF_WORKSPACE_SIZE */
|
||||||
|
HUF_CElt* hufTable, HUF_repeat* repeat, int preferRepeat, int bmi2);
|
||||||
|
|
||||||
|
/** HUF_buildCTable_wksp() :
|
||||||
|
* Same as HUF_buildCTable(), but using externally allocated scratch buffer.
|
||||||
|
* `workSpace` must be aligned on 4-bytes boundaries, and its size must be >= HUF_CTABLE_WORKSPACE_SIZE.
|
||||||
|
*/
|
||||||
|
#define HUF_CTABLE_WORKSPACE_SIZE_U32 (2*HUF_SYMBOLVALUE_MAX +1 +1)
|
||||||
|
#define HUF_CTABLE_WORKSPACE_SIZE (HUF_CTABLE_WORKSPACE_SIZE_U32 * sizeof(unsigned))
|
||||||
|
size_t HUF_buildCTable_wksp (HUF_CElt* tree,
|
||||||
|
const unsigned* count, U32 maxSymbolValue, U32 maxNbBits,
|
||||||
|
void* workSpace, size_t wkspSize);
|
||||||
|
|
||||||
|
/*! HUF_readStats() :
|
||||||
|
* Read compact Huffman tree, saved by HUF_writeCTable().
|
||||||
|
* `huffWeight` is destination buffer.
|
||||||
|
* @return : size read from `src` , or an error Code .
|
||||||
|
* Note : Needed by HUF_readCTable() and HUF_readDTableXn() . */
|
||||||
|
size_t HUF_readStats(BYTE* huffWeight, size_t hwSize,
|
||||||
|
U32* rankStats, U32* nbSymbolsPtr, U32* tableLogPtr,
|
||||||
|
const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
/** HUF_readCTable() :
|
||||||
|
* Loading a CTable saved with HUF_writeCTable() */
|
||||||
|
size_t HUF_readCTable (HUF_CElt* CTable, unsigned* maxSymbolValuePtr, const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
/** HUF_getNbBits() :
|
||||||
|
* Read nbBits from CTable symbolTable, for symbol `symbolValue` presumed <= HUF_SYMBOLVALUE_MAX
|
||||||
|
* Note 1 : is not inlined, as HUF_CElt definition is private
|
||||||
|
* Note 2 : const void* used, so that it can provide a statically allocated table as argument (which uses type U32) */
|
||||||
|
U32 HUF_getNbBits(const void* symbolTable, U32 symbolValue);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* HUF_decompress() does the following:
|
||||||
|
* 1. select the decompression algorithm (X1, X2) based on pre-computed heuristics
|
||||||
|
* 2. build Huffman table from save, using HUF_readDTableX?()
|
||||||
|
* 3. decode 1 or 4 segments in parallel using HUF_decompress?X?_usingDTable()
|
||||||
|
*/
|
||||||
|
|
||||||
|
/** HUF_selectDecoder() :
|
||||||
|
* Tells which decoder is likely to decode faster,
|
||||||
|
* based on a set of pre-computed metrics.
|
||||||
|
* @return : 0==HUF_decompress4X1, 1==HUF_decompress4X2 .
|
||||||
|
* Assumption : 0 < dstSize <= 128 KB */
|
||||||
|
U32 HUF_selectDecoder (size_t dstSize, size_t cSrcSize);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The minimum workspace size for the `workSpace` used in
|
||||||
|
* HUF_readDTableX1_wksp() and HUF_readDTableX2_wksp().
|
||||||
|
*
|
||||||
|
* The space used depends on HUF_TABLELOG_MAX, ranging from ~1500 bytes when
|
||||||
|
* HUF_TABLE_LOG_MAX=12 to ~1850 bytes when HUF_TABLE_LOG_MAX=15.
|
||||||
|
* Buffer overflow errors may potentially occur if code modifications result in
|
||||||
|
* a required workspace size greater than that specified in the following
|
||||||
|
* macro.
|
||||||
|
*/
|
||||||
|
#define HUF_DECOMPRESS_WORKSPACE_SIZE (2 << 10)
|
||||||
|
#define HUF_DECOMPRESS_WORKSPACE_SIZE_U32 (HUF_DECOMPRESS_WORKSPACE_SIZE / sizeof(U32))
|
||||||
|
|
||||||
|
#ifndef HUF_FORCE_DECOMPRESS_X2
|
||||||
|
size_t HUF_readDTableX1 (HUF_DTable* DTable, const void* src, size_t srcSize);
|
||||||
|
size_t HUF_readDTableX1_wksp (HUF_DTable* DTable, const void* src, size_t srcSize, void* workSpace, size_t wkspSize);
|
||||||
|
#endif
|
||||||
|
#ifndef HUF_FORCE_DECOMPRESS_X1
|
||||||
|
size_t HUF_readDTableX2 (HUF_DTable* DTable, const void* src, size_t srcSize);
|
||||||
|
size_t HUF_readDTableX2_wksp (HUF_DTable* DTable, const void* src, size_t srcSize, void* workSpace, size_t wkspSize);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
size_t HUF_decompress4X_usingDTable(void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize, const HUF_DTable* DTable);
|
||||||
|
#ifndef HUF_FORCE_DECOMPRESS_X2
|
||||||
|
size_t HUF_decompress4X1_usingDTable(void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize, const HUF_DTable* DTable);
|
||||||
|
#endif
|
||||||
|
#ifndef HUF_FORCE_DECOMPRESS_X1
|
||||||
|
size_t HUF_decompress4X2_usingDTable(void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize, const HUF_DTable* DTable);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* ====================== */
|
||||||
|
/* single stream variants */
|
||||||
|
/* ====================== */
|
||||||
|
|
||||||
|
size_t HUF_compress1X (void* dst, size_t dstSize, const void* src, size_t srcSize, unsigned maxSymbolValue, unsigned tableLog);
|
||||||
|
size_t HUF_compress1X_wksp (void* dst, size_t dstSize, const void* src, size_t srcSize, unsigned maxSymbolValue, unsigned tableLog, void* workSpace, size_t wkspSize); /**< `workSpace` must be a table of at least HUF_WORKSPACE_SIZE_U32 unsigned */
|
||||||
|
size_t HUF_compress1X_usingCTable(void* dst, size_t dstSize, const void* src, size_t srcSize, const HUF_CElt* CTable);
|
||||||
|
/** HUF_compress1X_repeat() :
|
||||||
|
* Same as HUF_compress1X_wksp(), but considers using hufTable if *repeat != HUF_repeat_none.
|
||||||
|
* If it uses hufTable it does not modify hufTable or repeat.
|
||||||
|
* If it doesn't, it sets *repeat = HUF_repeat_none, and it sets hufTable to the table used.
|
||||||
|
* If preferRepeat then the old table will always be used if valid. */
|
||||||
|
size_t HUF_compress1X_repeat(void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
unsigned maxSymbolValue, unsigned tableLog,
|
||||||
|
void* workSpace, size_t wkspSize, /**< `workSpace` must be aligned on 4-bytes boundaries, `wkspSize` must be >= HUF_WORKSPACE_SIZE */
|
||||||
|
HUF_CElt* hufTable, HUF_repeat* repeat, int preferRepeat, int bmi2);
|
||||||
|
|
||||||
|
size_t HUF_decompress1X1 (void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize); /* single-symbol decoder */
|
||||||
|
#ifndef HUF_FORCE_DECOMPRESS_X1
|
||||||
|
size_t HUF_decompress1X2 (void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize); /* double-symbol decoder */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
size_t HUF_decompress1X_DCtx (HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize);
|
||||||
|
size_t HUF_decompress1X_DCtx_wksp (HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize, void* workSpace, size_t wkspSize);
|
||||||
|
#ifndef HUF_FORCE_DECOMPRESS_X2
|
||||||
|
size_t HUF_decompress1X1_DCtx(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize); /**< single-symbol decoder */
|
||||||
|
size_t HUF_decompress1X1_DCtx_wksp(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize, void* workSpace, size_t wkspSize); /**< single-symbol decoder */
|
||||||
|
#endif
|
||||||
|
#ifndef HUF_FORCE_DECOMPRESS_X1
|
||||||
|
size_t HUF_decompress1X2_DCtx(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize); /**< double-symbols decoder */
|
||||||
|
size_t HUF_decompress1X2_DCtx_wksp(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize, void* workSpace, size_t wkspSize); /**< double-symbols decoder */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
size_t HUF_decompress1X_usingDTable(void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize, const HUF_DTable* DTable); /**< automatic selection of sing or double symbol decoder, based on DTable */
|
||||||
|
#ifndef HUF_FORCE_DECOMPRESS_X2
|
||||||
|
size_t HUF_decompress1X1_usingDTable(void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize, const HUF_DTable* DTable);
|
||||||
|
#endif
|
||||||
|
#ifndef HUF_FORCE_DECOMPRESS_X1
|
||||||
|
size_t HUF_decompress1X2_usingDTable(void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize, const HUF_DTable* DTable);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* BMI2 variants.
|
||||||
|
* If the CPU has BMI2 support, pass bmi2=1, otherwise pass bmi2=0.
|
||||||
|
*/
|
||||||
|
size_t HUF_decompress1X_usingDTable_bmi2(void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize, const HUF_DTable* DTable, int bmi2);
|
||||||
|
#ifndef HUF_FORCE_DECOMPRESS_X2
|
||||||
|
size_t HUF_decompress1X1_DCtx_wksp_bmi2(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize, void* workSpace, size_t wkspSize, int bmi2);
|
||||||
|
#endif
|
||||||
|
size_t HUF_decompress4X_usingDTable_bmi2(void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize, const HUF_DTable* DTable, int bmi2);
|
||||||
|
size_t HUF_decompress4X_hufOnly_wksp_bmi2(HUF_DTable* dctx, void* dst, size_t dstSize, const void* cSrc, size_t cSrcSize, void* workSpace, size_t wkspSize, int bmi2);
|
||||||
|
|
||||||
|
#endif /* HUF_STATIC_LINKING_ONLY */
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
798
vendor/github.com/DataDog/zstd/huf_compress.c
generated
vendored
Normal file
798
vendor/github.com/DataDog/zstd/huf_compress.c
generated
vendored
Normal file
@@ -0,0 +1,798 @@
|
|||||||
|
/* ******************************************************************
|
||||||
|
Huffman encoder, part of New Generation Entropy library
|
||||||
|
Copyright (C) 2013-2016, Yann Collet.
|
||||||
|
|
||||||
|
BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
You can contact the author at :
|
||||||
|
- FSE+HUF source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
||||||
|
- Public forum : https://groups.google.com/forum/#!forum/lz4c
|
||||||
|
****************************************************************** */
|
||||||
|
|
||||||
|
/* **************************************************************
|
||||||
|
* Compiler specifics
|
||||||
|
****************************************************************/
|
||||||
|
#ifdef _MSC_VER /* Visual Studio */
|
||||||
|
# pragma warning(disable : 4127) /* disable: C4127: conditional expression is constant */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************************************************
|
||||||
|
* Includes
|
||||||
|
****************************************************************/
|
||||||
|
#include <string.h> /* memcpy, memset */
|
||||||
|
#include <stdio.h> /* printf (debug) */
|
||||||
|
#include "compiler.h"
|
||||||
|
#include "bitstream.h"
|
||||||
|
#include "hist.h"
|
||||||
|
#define FSE_STATIC_LINKING_ONLY /* FSE_optimalTableLog_internal */
|
||||||
|
#include "fse.h" /* header compression */
|
||||||
|
#define HUF_STATIC_LINKING_ONLY
|
||||||
|
#include "huf.h"
|
||||||
|
#include "error_private.h"
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************************************************
|
||||||
|
* Error Management
|
||||||
|
****************************************************************/
|
||||||
|
#define HUF_isError ERR_isError
|
||||||
|
#define HUF_STATIC_ASSERT(c) DEBUG_STATIC_ASSERT(c) /* use only *after* variable declarations */
|
||||||
|
#define CHECK_V_F(e, f) size_t const e = f; if (ERR_isError(e)) return e
|
||||||
|
#define CHECK_F(f) { CHECK_V_F(_var_err__, f); }
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************************************************
|
||||||
|
* Utils
|
||||||
|
****************************************************************/
|
||||||
|
unsigned HUF_optimalTableLog(unsigned maxTableLog, size_t srcSize, unsigned maxSymbolValue)
|
||||||
|
{
|
||||||
|
return FSE_optimalTableLog_internal(maxTableLog, srcSize, maxSymbolValue, 1);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* *******************************************************
|
||||||
|
* HUF : Huffman block compression
|
||||||
|
*********************************************************/
|
||||||
|
/* HUF_compressWeights() :
|
||||||
|
* Same as FSE_compress(), but dedicated to huff0's weights compression.
|
||||||
|
* The use case needs much less stack memory.
|
||||||
|
* Note : all elements within weightTable are supposed to be <= HUF_TABLELOG_MAX.
|
||||||
|
*/
|
||||||
|
#define MAX_FSE_TABLELOG_FOR_HUFF_HEADER 6
|
||||||
|
static size_t HUF_compressWeights (void* dst, size_t dstSize, const void* weightTable, size_t wtSize)
|
||||||
|
{
|
||||||
|
BYTE* const ostart = (BYTE*) dst;
|
||||||
|
BYTE* op = ostart;
|
||||||
|
BYTE* const oend = ostart + dstSize;
|
||||||
|
|
||||||
|
unsigned maxSymbolValue = HUF_TABLELOG_MAX;
|
||||||
|
U32 tableLog = MAX_FSE_TABLELOG_FOR_HUFF_HEADER;
|
||||||
|
|
||||||
|
FSE_CTable CTable[FSE_CTABLE_SIZE_U32(MAX_FSE_TABLELOG_FOR_HUFF_HEADER, HUF_TABLELOG_MAX)];
|
||||||
|
BYTE scratchBuffer[1<<MAX_FSE_TABLELOG_FOR_HUFF_HEADER];
|
||||||
|
|
||||||
|
unsigned count[HUF_TABLELOG_MAX+1];
|
||||||
|
S16 norm[HUF_TABLELOG_MAX+1];
|
||||||
|
|
||||||
|
/* init conditions */
|
||||||
|
if (wtSize <= 1) return 0; /* Not compressible */
|
||||||
|
|
||||||
|
/* Scan input and build symbol stats */
|
||||||
|
{ unsigned const maxCount = HIST_count_simple(count, &maxSymbolValue, weightTable, wtSize); /* never fails */
|
||||||
|
if (maxCount == wtSize) return 1; /* only a single symbol in src : rle */
|
||||||
|
if (maxCount == 1) return 0; /* each symbol present maximum once => not compressible */
|
||||||
|
}
|
||||||
|
|
||||||
|
tableLog = FSE_optimalTableLog(tableLog, wtSize, maxSymbolValue);
|
||||||
|
CHECK_F( FSE_normalizeCount(norm, tableLog, count, wtSize, maxSymbolValue) );
|
||||||
|
|
||||||
|
/* Write table description header */
|
||||||
|
{ CHECK_V_F(hSize, FSE_writeNCount(op, oend-op, norm, maxSymbolValue, tableLog) );
|
||||||
|
op += hSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Compress */
|
||||||
|
CHECK_F( FSE_buildCTable_wksp(CTable, norm, maxSymbolValue, tableLog, scratchBuffer, sizeof(scratchBuffer)) );
|
||||||
|
{ CHECK_V_F(cSize, FSE_compress_usingCTable(op, oend - op, weightTable, wtSize, CTable) );
|
||||||
|
if (cSize == 0) return 0; /* not enough space for compressed data */
|
||||||
|
op += cSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
return op-ostart;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
struct HUF_CElt_s {
|
||||||
|
U16 val;
|
||||||
|
BYTE nbBits;
|
||||||
|
}; /* typedef'd to HUF_CElt within "huf.h" */
|
||||||
|
|
||||||
|
/*! HUF_writeCTable() :
|
||||||
|
`CTable` : Huffman tree to save, using huf representation.
|
||||||
|
@return : size of saved CTable */
|
||||||
|
size_t HUF_writeCTable (void* dst, size_t maxDstSize,
|
||||||
|
const HUF_CElt* CTable, unsigned maxSymbolValue, unsigned huffLog)
|
||||||
|
{
|
||||||
|
BYTE bitsToWeight[HUF_TABLELOG_MAX + 1]; /* precomputed conversion table */
|
||||||
|
BYTE huffWeight[HUF_SYMBOLVALUE_MAX];
|
||||||
|
BYTE* op = (BYTE*)dst;
|
||||||
|
U32 n;
|
||||||
|
|
||||||
|
/* check conditions */
|
||||||
|
if (maxSymbolValue > HUF_SYMBOLVALUE_MAX) return ERROR(maxSymbolValue_tooLarge);
|
||||||
|
|
||||||
|
/* convert to weight */
|
||||||
|
bitsToWeight[0] = 0;
|
||||||
|
for (n=1; n<huffLog+1; n++)
|
||||||
|
bitsToWeight[n] = (BYTE)(huffLog + 1 - n);
|
||||||
|
for (n=0; n<maxSymbolValue; n++)
|
||||||
|
huffWeight[n] = bitsToWeight[CTable[n].nbBits];
|
||||||
|
|
||||||
|
/* attempt weights compression by FSE */
|
||||||
|
{ CHECK_V_F(hSize, HUF_compressWeights(op+1, maxDstSize-1, huffWeight, maxSymbolValue) );
|
||||||
|
if ((hSize>1) & (hSize < maxSymbolValue/2)) { /* FSE compressed */
|
||||||
|
op[0] = (BYTE)hSize;
|
||||||
|
return hSize+1;
|
||||||
|
} }
|
||||||
|
|
||||||
|
/* write raw values as 4-bits (max : 15) */
|
||||||
|
if (maxSymbolValue > (256-128)) return ERROR(GENERIC); /* should not happen : likely means source cannot be compressed */
|
||||||
|
if (((maxSymbolValue+1)/2) + 1 > maxDstSize) return ERROR(dstSize_tooSmall); /* not enough space within dst buffer */
|
||||||
|
op[0] = (BYTE)(128 /*special case*/ + (maxSymbolValue-1));
|
||||||
|
huffWeight[maxSymbolValue] = 0; /* to be sure it doesn't cause msan issue in final combination */
|
||||||
|
for (n=0; n<maxSymbolValue; n+=2)
|
||||||
|
op[(n/2)+1] = (BYTE)((huffWeight[n] << 4) + huffWeight[n+1]);
|
||||||
|
return ((maxSymbolValue+1)/2) + 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t HUF_readCTable (HUF_CElt* CTable, unsigned* maxSymbolValuePtr, const void* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
BYTE huffWeight[HUF_SYMBOLVALUE_MAX + 1]; /* init not required, even though some static analyzer may complain */
|
||||||
|
U32 rankVal[HUF_TABLELOG_ABSOLUTEMAX + 1]; /* large enough for values from 0 to 16 */
|
||||||
|
U32 tableLog = 0;
|
||||||
|
U32 nbSymbols = 0;
|
||||||
|
|
||||||
|
/* get symbol weights */
|
||||||
|
CHECK_V_F(readSize, HUF_readStats(huffWeight, HUF_SYMBOLVALUE_MAX+1, rankVal, &nbSymbols, &tableLog, src, srcSize));
|
||||||
|
|
||||||
|
/* check result */
|
||||||
|
if (tableLog > HUF_TABLELOG_MAX) return ERROR(tableLog_tooLarge);
|
||||||
|
if (nbSymbols > *maxSymbolValuePtr+1) return ERROR(maxSymbolValue_tooSmall);
|
||||||
|
|
||||||
|
/* Prepare base value per rank */
|
||||||
|
{ U32 n, nextRankStart = 0;
|
||||||
|
for (n=1; n<=tableLog; n++) {
|
||||||
|
U32 current = nextRankStart;
|
||||||
|
nextRankStart += (rankVal[n] << (n-1));
|
||||||
|
rankVal[n] = current;
|
||||||
|
} }
|
||||||
|
|
||||||
|
/* fill nbBits */
|
||||||
|
{ U32 n; for (n=0; n<nbSymbols; n++) {
|
||||||
|
const U32 w = huffWeight[n];
|
||||||
|
CTable[n].nbBits = (BYTE)(tableLog + 1 - w);
|
||||||
|
} }
|
||||||
|
|
||||||
|
/* fill val */
|
||||||
|
{ U16 nbPerRank[HUF_TABLELOG_MAX+2] = {0}; /* support w=0=>n=tableLog+1 */
|
||||||
|
U16 valPerRank[HUF_TABLELOG_MAX+2] = {0};
|
||||||
|
{ U32 n; for (n=0; n<nbSymbols; n++) nbPerRank[CTable[n].nbBits]++; }
|
||||||
|
/* determine stating value per rank */
|
||||||
|
valPerRank[tableLog+1] = 0; /* for w==0 */
|
||||||
|
{ U16 min = 0;
|
||||||
|
U32 n; for (n=tableLog; n>0; n--) { /* start at n=tablelog <-> w=1 */
|
||||||
|
valPerRank[n] = min; /* get starting value within each rank */
|
||||||
|
min += nbPerRank[n];
|
||||||
|
min >>= 1;
|
||||||
|
} }
|
||||||
|
/* assign value within rank, symbol order */
|
||||||
|
{ U32 n; for (n=0; n<nbSymbols; n++) CTable[n].val = valPerRank[CTable[n].nbBits]++; }
|
||||||
|
}
|
||||||
|
|
||||||
|
*maxSymbolValuePtr = nbSymbols - 1;
|
||||||
|
return readSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
U32 HUF_getNbBits(const void* symbolTable, U32 symbolValue)
|
||||||
|
{
|
||||||
|
const HUF_CElt* table = (const HUF_CElt*)symbolTable;
|
||||||
|
assert(symbolValue <= HUF_SYMBOLVALUE_MAX);
|
||||||
|
return table[symbolValue].nbBits;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
typedef struct nodeElt_s {
|
||||||
|
U32 count;
|
||||||
|
U16 parent;
|
||||||
|
BYTE byte;
|
||||||
|
BYTE nbBits;
|
||||||
|
} nodeElt;
|
||||||
|
|
||||||
|
static U32 HUF_setMaxHeight(nodeElt* huffNode, U32 lastNonNull, U32 maxNbBits)
|
||||||
|
{
|
||||||
|
const U32 largestBits = huffNode[lastNonNull].nbBits;
|
||||||
|
if (largestBits <= maxNbBits) return largestBits; /* early exit : no elt > maxNbBits */
|
||||||
|
|
||||||
|
/* there are several too large elements (at least >= 2) */
|
||||||
|
{ int totalCost = 0;
|
||||||
|
const U32 baseCost = 1 << (largestBits - maxNbBits);
|
||||||
|
U32 n = lastNonNull;
|
||||||
|
|
||||||
|
while (huffNode[n].nbBits > maxNbBits) {
|
||||||
|
totalCost += baseCost - (1 << (largestBits - huffNode[n].nbBits));
|
||||||
|
huffNode[n].nbBits = (BYTE)maxNbBits;
|
||||||
|
n --;
|
||||||
|
} /* n stops at huffNode[n].nbBits <= maxNbBits */
|
||||||
|
while (huffNode[n].nbBits == maxNbBits) n--; /* n end at index of smallest symbol using < maxNbBits */
|
||||||
|
|
||||||
|
/* renorm totalCost */
|
||||||
|
totalCost >>= (largestBits - maxNbBits); /* note : totalCost is necessarily a multiple of baseCost */
|
||||||
|
|
||||||
|
/* repay normalized cost */
|
||||||
|
{ U32 const noSymbol = 0xF0F0F0F0;
|
||||||
|
U32 rankLast[HUF_TABLELOG_MAX+2];
|
||||||
|
int pos;
|
||||||
|
|
||||||
|
/* Get pos of last (smallest) symbol per rank */
|
||||||
|
memset(rankLast, 0xF0, sizeof(rankLast));
|
||||||
|
{ U32 currentNbBits = maxNbBits;
|
||||||
|
for (pos=n ; pos >= 0; pos--) {
|
||||||
|
if (huffNode[pos].nbBits >= currentNbBits) continue;
|
||||||
|
currentNbBits = huffNode[pos].nbBits; /* < maxNbBits */
|
||||||
|
rankLast[maxNbBits-currentNbBits] = pos;
|
||||||
|
} }
|
||||||
|
|
||||||
|
while (totalCost > 0) {
|
||||||
|
U32 nBitsToDecrease = BIT_highbit32(totalCost) + 1;
|
||||||
|
for ( ; nBitsToDecrease > 1; nBitsToDecrease--) {
|
||||||
|
U32 highPos = rankLast[nBitsToDecrease];
|
||||||
|
U32 lowPos = rankLast[nBitsToDecrease-1];
|
||||||
|
if (highPos == noSymbol) continue;
|
||||||
|
if (lowPos == noSymbol) break;
|
||||||
|
{ U32 const highTotal = huffNode[highPos].count;
|
||||||
|
U32 const lowTotal = 2 * huffNode[lowPos].count;
|
||||||
|
if (highTotal <= lowTotal) break;
|
||||||
|
} }
|
||||||
|
/* only triggered when no more rank 1 symbol left => find closest one (note : there is necessarily at least one !) */
|
||||||
|
/* HUF_MAX_TABLELOG test just to please gcc 5+; but it should not be necessary */
|
||||||
|
while ((nBitsToDecrease<=HUF_TABLELOG_MAX) && (rankLast[nBitsToDecrease] == noSymbol))
|
||||||
|
nBitsToDecrease ++;
|
||||||
|
totalCost -= 1 << (nBitsToDecrease-1);
|
||||||
|
if (rankLast[nBitsToDecrease-1] == noSymbol)
|
||||||
|
rankLast[nBitsToDecrease-1] = rankLast[nBitsToDecrease]; /* this rank is no longer empty */
|
||||||
|
huffNode[rankLast[nBitsToDecrease]].nbBits ++;
|
||||||
|
if (rankLast[nBitsToDecrease] == 0) /* special case, reached largest symbol */
|
||||||
|
rankLast[nBitsToDecrease] = noSymbol;
|
||||||
|
else {
|
||||||
|
rankLast[nBitsToDecrease]--;
|
||||||
|
if (huffNode[rankLast[nBitsToDecrease]].nbBits != maxNbBits-nBitsToDecrease)
|
||||||
|
rankLast[nBitsToDecrease] = noSymbol; /* this rank is now empty */
|
||||||
|
} } /* while (totalCost > 0) */
|
||||||
|
|
||||||
|
while (totalCost < 0) { /* Sometimes, cost correction overshoot */
|
||||||
|
if (rankLast[1] == noSymbol) { /* special case : no rank 1 symbol (using maxNbBits-1); let's create one from largest rank 0 (using maxNbBits) */
|
||||||
|
while (huffNode[n].nbBits == maxNbBits) n--;
|
||||||
|
huffNode[n+1].nbBits--;
|
||||||
|
rankLast[1] = n+1;
|
||||||
|
totalCost++;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
huffNode[ rankLast[1] + 1 ].nbBits--;
|
||||||
|
rankLast[1]++;
|
||||||
|
totalCost ++;
|
||||||
|
} } } /* there are several too large elements (at least >= 2) */
|
||||||
|
|
||||||
|
return maxNbBits;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
U32 base;
|
||||||
|
U32 current;
|
||||||
|
} rankPos;
|
||||||
|
|
||||||
|
static void HUF_sort(nodeElt* huffNode, const unsigned* count, U32 maxSymbolValue)
|
||||||
|
{
|
||||||
|
rankPos rank[32];
|
||||||
|
U32 n;
|
||||||
|
|
||||||
|
memset(rank, 0, sizeof(rank));
|
||||||
|
for (n=0; n<=maxSymbolValue; n++) {
|
||||||
|
U32 r = BIT_highbit32(count[n] + 1);
|
||||||
|
rank[r].base ++;
|
||||||
|
}
|
||||||
|
for (n=30; n>0; n--) rank[n-1].base += rank[n].base;
|
||||||
|
for (n=0; n<32; n++) rank[n].current = rank[n].base;
|
||||||
|
for (n=0; n<=maxSymbolValue; n++) {
|
||||||
|
U32 const c = count[n];
|
||||||
|
U32 const r = BIT_highbit32(c+1) + 1;
|
||||||
|
U32 pos = rank[r].current++;
|
||||||
|
while ((pos > rank[r].base) && (c > huffNode[pos-1].count)) {
|
||||||
|
huffNode[pos] = huffNode[pos-1];
|
||||||
|
pos--;
|
||||||
|
}
|
||||||
|
huffNode[pos].count = c;
|
||||||
|
huffNode[pos].byte = (BYTE)n;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/** HUF_buildCTable_wksp() :
|
||||||
|
* Same as HUF_buildCTable(), but using externally allocated scratch buffer.
|
||||||
|
* `workSpace` must be aligned on 4-bytes boundaries, and be at least as large as a table of HUF_CTABLE_WORKSPACE_SIZE_U32 unsigned.
|
||||||
|
*/
|
||||||
|
#define STARTNODE (HUF_SYMBOLVALUE_MAX+1)
|
||||||
|
typedef nodeElt huffNodeTable[HUF_CTABLE_WORKSPACE_SIZE_U32];
|
||||||
|
size_t HUF_buildCTable_wksp (HUF_CElt* tree, const unsigned* count, U32 maxSymbolValue, U32 maxNbBits, void* workSpace, size_t wkspSize)
|
||||||
|
{
|
||||||
|
nodeElt* const huffNode0 = (nodeElt*)workSpace;
|
||||||
|
nodeElt* const huffNode = huffNode0+1;
|
||||||
|
U32 n, nonNullRank;
|
||||||
|
int lowS, lowN;
|
||||||
|
U16 nodeNb = STARTNODE;
|
||||||
|
U32 nodeRoot;
|
||||||
|
|
||||||
|
/* safety checks */
|
||||||
|
if (((size_t)workSpace & 3) != 0) return ERROR(GENERIC); /* must be aligned on 4-bytes boundaries */
|
||||||
|
if (wkspSize < sizeof(huffNodeTable)) return ERROR(workSpace_tooSmall);
|
||||||
|
if (maxNbBits == 0) maxNbBits = HUF_TABLELOG_DEFAULT;
|
||||||
|
if (maxSymbolValue > HUF_SYMBOLVALUE_MAX) return ERROR(maxSymbolValue_tooLarge);
|
||||||
|
memset(huffNode0, 0, sizeof(huffNodeTable));
|
||||||
|
|
||||||
|
/* sort, decreasing order */
|
||||||
|
HUF_sort(huffNode, count, maxSymbolValue);
|
||||||
|
|
||||||
|
/* init for parents */
|
||||||
|
nonNullRank = maxSymbolValue;
|
||||||
|
while(huffNode[nonNullRank].count == 0) nonNullRank--;
|
||||||
|
lowS = nonNullRank; nodeRoot = nodeNb + lowS - 1; lowN = nodeNb;
|
||||||
|
huffNode[nodeNb].count = huffNode[lowS].count + huffNode[lowS-1].count;
|
||||||
|
huffNode[lowS].parent = huffNode[lowS-1].parent = nodeNb;
|
||||||
|
nodeNb++; lowS-=2;
|
||||||
|
for (n=nodeNb; n<=nodeRoot; n++) huffNode[n].count = (U32)(1U<<30);
|
||||||
|
huffNode0[0].count = (U32)(1U<<31); /* fake entry, strong barrier */
|
||||||
|
|
||||||
|
/* create parents */
|
||||||
|
while (nodeNb <= nodeRoot) {
|
||||||
|
U32 n1 = (huffNode[lowS].count < huffNode[lowN].count) ? lowS-- : lowN++;
|
||||||
|
U32 n2 = (huffNode[lowS].count < huffNode[lowN].count) ? lowS-- : lowN++;
|
||||||
|
huffNode[nodeNb].count = huffNode[n1].count + huffNode[n2].count;
|
||||||
|
huffNode[n1].parent = huffNode[n2].parent = nodeNb;
|
||||||
|
nodeNb++;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* distribute weights (unlimited tree height) */
|
||||||
|
huffNode[nodeRoot].nbBits = 0;
|
||||||
|
for (n=nodeRoot-1; n>=STARTNODE; n--)
|
||||||
|
huffNode[n].nbBits = huffNode[ huffNode[n].parent ].nbBits + 1;
|
||||||
|
for (n=0; n<=nonNullRank; n++)
|
||||||
|
huffNode[n].nbBits = huffNode[ huffNode[n].parent ].nbBits + 1;
|
||||||
|
|
||||||
|
/* enforce maxTableLog */
|
||||||
|
maxNbBits = HUF_setMaxHeight(huffNode, nonNullRank, maxNbBits);
|
||||||
|
|
||||||
|
/* fill result into tree (val, nbBits) */
|
||||||
|
{ U16 nbPerRank[HUF_TABLELOG_MAX+1] = {0};
|
||||||
|
U16 valPerRank[HUF_TABLELOG_MAX+1] = {0};
|
||||||
|
if (maxNbBits > HUF_TABLELOG_MAX) return ERROR(GENERIC); /* check fit into table */
|
||||||
|
for (n=0; n<=nonNullRank; n++)
|
||||||
|
nbPerRank[huffNode[n].nbBits]++;
|
||||||
|
/* determine stating value per rank */
|
||||||
|
{ U16 min = 0;
|
||||||
|
for (n=maxNbBits; n>0; n--) {
|
||||||
|
valPerRank[n] = min; /* get starting value within each rank */
|
||||||
|
min += nbPerRank[n];
|
||||||
|
min >>= 1;
|
||||||
|
} }
|
||||||
|
for (n=0; n<=maxSymbolValue; n++)
|
||||||
|
tree[huffNode[n].byte].nbBits = huffNode[n].nbBits; /* push nbBits per symbol, symbol order */
|
||||||
|
for (n=0; n<=maxSymbolValue; n++)
|
||||||
|
tree[n].val = valPerRank[tree[n].nbBits]++; /* assign value within rank, symbol order */
|
||||||
|
}
|
||||||
|
|
||||||
|
return maxNbBits;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** HUF_buildCTable() :
|
||||||
|
* @return : maxNbBits
|
||||||
|
* Note : count is used before tree is written, so they can safely overlap
|
||||||
|
*/
|
||||||
|
size_t HUF_buildCTable (HUF_CElt* tree, const unsigned* count, unsigned maxSymbolValue, unsigned maxNbBits)
|
||||||
|
{
|
||||||
|
huffNodeTable nodeTable;
|
||||||
|
return HUF_buildCTable_wksp(tree, count, maxSymbolValue, maxNbBits, nodeTable, sizeof(nodeTable));
|
||||||
|
}
|
||||||
|
|
||||||
|
static size_t HUF_estimateCompressedSize(HUF_CElt* CTable, const unsigned* count, unsigned maxSymbolValue)
|
||||||
|
{
|
||||||
|
size_t nbBits = 0;
|
||||||
|
int s;
|
||||||
|
for (s = 0; s <= (int)maxSymbolValue; ++s) {
|
||||||
|
nbBits += CTable[s].nbBits * count[s];
|
||||||
|
}
|
||||||
|
return nbBits >> 3;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int HUF_validateCTable(const HUF_CElt* CTable, const unsigned* count, unsigned maxSymbolValue) {
|
||||||
|
int bad = 0;
|
||||||
|
int s;
|
||||||
|
for (s = 0; s <= (int)maxSymbolValue; ++s) {
|
||||||
|
bad |= (count[s] != 0) & (CTable[s].nbBits == 0);
|
||||||
|
}
|
||||||
|
return !bad;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t HUF_compressBound(size_t size) { return HUF_COMPRESSBOUND(size); }
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE void
|
||||||
|
HUF_encodeSymbol(BIT_CStream_t* bitCPtr, U32 symbol, const HUF_CElt* CTable)
|
||||||
|
{
|
||||||
|
BIT_addBitsFast(bitCPtr, CTable[symbol].val, CTable[symbol].nbBits);
|
||||||
|
}
|
||||||
|
|
||||||
|
#define HUF_FLUSHBITS(s) BIT_flushBits(s)
|
||||||
|
|
||||||
|
#define HUF_FLUSHBITS_1(stream) \
|
||||||
|
if (sizeof((stream)->bitContainer)*8 < HUF_TABLELOG_MAX*2+7) HUF_FLUSHBITS(stream)
|
||||||
|
|
||||||
|
#define HUF_FLUSHBITS_2(stream) \
|
||||||
|
if (sizeof((stream)->bitContainer)*8 < HUF_TABLELOG_MAX*4+7) HUF_FLUSHBITS(stream)
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE size_t
|
||||||
|
HUF_compress1X_usingCTable_internal_body(void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const HUF_CElt* CTable)
|
||||||
|
{
|
||||||
|
const BYTE* ip = (const BYTE*) src;
|
||||||
|
BYTE* const ostart = (BYTE*)dst;
|
||||||
|
BYTE* const oend = ostart + dstSize;
|
||||||
|
BYTE* op = ostart;
|
||||||
|
size_t n;
|
||||||
|
BIT_CStream_t bitC;
|
||||||
|
|
||||||
|
/* init */
|
||||||
|
if (dstSize < 8) return 0; /* not enough space to compress */
|
||||||
|
{ size_t const initErr = BIT_initCStream(&bitC, op, oend-op);
|
||||||
|
if (HUF_isError(initErr)) return 0; }
|
||||||
|
|
||||||
|
n = srcSize & ~3; /* join to mod 4 */
|
||||||
|
switch (srcSize & 3)
|
||||||
|
{
|
||||||
|
case 3 : HUF_encodeSymbol(&bitC, ip[n+ 2], CTable);
|
||||||
|
HUF_FLUSHBITS_2(&bitC);
|
||||||
|
/* fall-through */
|
||||||
|
case 2 : HUF_encodeSymbol(&bitC, ip[n+ 1], CTable);
|
||||||
|
HUF_FLUSHBITS_1(&bitC);
|
||||||
|
/* fall-through */
|
||||||
|
case 1 : HUF_encodeSymbol(&bitC, ip[n+ 0], CTable);
|
||||||
|
HUF_FLUSHBITS(&bitC);
|
||||||
|
/* fall-through */
|
||||||
|
case 0 : /* fall-through */
|
||||||
|
default: break;
|
||||||
|
}
|
||||||
|
|
||||||
|
for (; n>0; n-=4) { /* note : n&3==0 at this stage */
|
||||||
|
HUF_encodeSymbol(&bitC, ip[n- 1], CTable);
|
||||||
|
HUF_FLUSHBITS_1(&bitC);
|
||||||
|
HUF_encodeSymbol(&bitC, ip[n- 2], CTable);
|
||||||
|
HUF_FLUSHBITS_2(&bitC);
|
||||||
|
HUF_encodeSymbol(&bitC, ip[n- 3], CTable);
|
||||||
|
HUF_FLUSHBITS_1(&bitC);
|
||||||
|
HUF_encodeSymbol(&bitC, ip[n- 4], CTable);
|
||||||
|
HUF_FLUSHBITS(&bitC);
|
||||||
|
}
|
||||||
|
|
||||||
|
return BIT_closeCStream(&bitC);
|
||||||
|
}
|
||||||
|
|
||||||
|
#if DYNAMIC_BMI2
|
||||||
|
|
||||||
|
static TARGET_ATTRIBUTE("bmi2") size_t
|
||||||
|
HUF_compress1X_usingCTable_internal_bmi2(void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const HUF_CElt* CTable)
|
||||||
|
{
|
||||||
|
return HUF_compress1X_usingCTable_internal_body(dst, dstSize, src, srcSize, CTable);
|
||||||
|
}
|
||||||
|
|
||||||
|
static size_t
|
||||||
|
HUF_compress1X_usingCTable_internal_default(void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const HUF_CElt* CTable)
|
||||||
|
{
|
||||||
|
return HUF_compress1X_usingCTable_internal_body(dst, dstSize, src, srcSize, CTable);
|
||||||
|
}
|
||||||
|
|
||||||
|
static size_t
|
||||||
|
HUF_compress1X_usingCTable_internal(void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const HUF_CElt* CTable, const int bmi2)
|
||||||
|
{
|
||||||
|
if (bmi2) {
|
||||||
|
return HUF_compress1X_usingCTable_internal_bmi2(dst, dstSize, src, srcSize, CTable);
|
||||||
|
}
|
||||||
|
return HUF_compress1X_usingCTable_internal_default(dst, dstSize, src, srcSize, CTable);
|
||||||
|
}
|
||||||
|
|
||||||
|
#else
|
||||||
|
|
||||||
|
static size_t
|
||||||
|
HUF_compress1X_usingCTable_internal(void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const HUF_CElt* CTable, const int bmi2)
|
||||||
|
{
|
||||||
|
(void)bmi2;
|
||||||
|
return HUF_compress1X_usingCTable_internal_body(dst, dstSize, src, srcSize, CTable);
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif
|
||||||
|
|
||||||
|
size_t HUF_compress1X_usingCTable(void* dst, size_t dstSize, const void* src, size_t srcSize, const HUF_CElt* CTable)
|
||||||
|
{
|
||||||
|
return HUF_compress1X_usingCTable_internal(dst, dstSize, src, srcSize, CTable, /* bmi2 */ 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static size_t
|
||||||
|
HUF_compress4X_usingCTable_internal(void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const HUF_CElt* CTable, int bmi2)
|
||||||
|
{
|
||||||
|
size_t const segmentSize = (srcSize+3)/4; /* first 3 segments */
|
||||||
|
const BYTE* ip = (const BYTE*) src;
|
||||||
|
const BYTE* const iend = ip + srcSize;
|
||||||
|
BYTE* const ostart = (BYTE*) dst;
|
||||||
|
BYTE* const oend = ostart + dstSize;
|
||||||
|
BYTE* op = ostart;
|
||||||
|
|
||||||
|
if (dstSize < 6 + 1 + 1 + 1 + 8) return 0; /* minimum space to compress successfully */
|
||||||
|
if (srcSize < 12) return 0; /* no saving possible : too small input */
|
||||||
|
op += 6; /* jumpTable */
|
||||||
|
|
||||||
|
{ CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, oend-op, ip, segmentSize, CTable, bmi2) );
|
||||||
|
if (cSize==0) return 0;
|
||||||
|
assert(cSize <= 65535);
|
||||||
|
MEM_writeLE16(ostart, (U16)cSize);
|
||||||
|
op += cSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
ip += segmentSize;
|
||||||
|
{ CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, oend-op, ip, segmentSize, CTable, bmi2) );
|
||||||
|
if (cSize==0) return 0;
|
||||||
|
assert(cSize <= 65535);
|
||||||
|
MEM_writeLE16(ostart+2, (U16)cSize);
|
||||||
|
op += cSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
ip += segmentSize;
|
||||||
|
{ CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, oend-op, ip, segmentSize, CTable, bmi2) );
|
||||||
|
if (cSize==0) return 0;
|
||||||
|
assert(cSize <= 65535);
|
||||||
|
MEM_writeLE16(ostart+4, (U16)cSize);
|
||||||
|
op += cSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
ip += segmentSize;
|
||||||
|
{ CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, oend-op, ip, iend-ip, CTable, bmi2) );
|
||||||
|
if (cSize==0) return 0;
|
||||||
|
op += cSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
return op-ostart;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t HUF_compress4X_usingCTable(void* dst, size_t dstSize, const void* src, size_t srcSize, const HUF_CElt* CTable)
|
||||||
|
{
|
||||||
|
return HUF_compress4X_usingCTable_internal(dst, dstSize, src, srcSize, CTable, /* bmi2 */ 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
typedef enum { HUF_singleStream, HUF_fourStreams } HUF_nbStreams_e;
|
||||||
|
|
||||||
|
static size_t HUF_compressCTable_internal(
|
||||||
|
BYTE* const ostart, BYTE* op, BYTE* const oend,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
HUF_nbStreams_e nbStreams, const HUF_CElt* CTable, const int bmi2)
|
||||||
|
{
|
||||||
|
size_t const cSize = (nbStreams==HUF_singleStream) ?
|
||||||
|
HUF_compress1X_usingCTable_internal(op, oend - op, src, srcSize, CTable, bmi2) :
|
||||||
|
HUF_compress4X_usingCTable_internal(op, oend - op, src, srcSize, CTable, bmi2);
|
||||||
|
if (HUF_isError(cSize)) { return cSize; }
|
||||||
|
if (cSize==0) { return 0; } /* uncompressible */
|
||||||
|
op += cSize;
|
||||||
|
/* check compressibility */
|
||||||
|
if ((size_t)(op-ostart) >= srcSize-1) { return 0; }
|
||||||
|
return op-ostart;
|
||||||
|
}
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
unsigned count[HUF_SYMBOLVALUE_MAX + 1];
|
||||||
|
HUF_CElt CTable[HUF_SYMBOLVALUE_MAX + 1];
|
||||||
|
huffNodeTable nodeTable;
|
||||||
|
} HUF_compress_tables_t;
|
||||||
|
|
||||||
|
/* HUF_compress_internal() :
|
||||||
|
* `workSpace` must a table of at least HUF_WORKSPACE_SIZE_U32 unsigned */
|
||||||
|
static size_t
|
||||||
|
HUF_compress_internal (void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
unsigned maxSymbolValue, unsigned huffLog,
|
||||||
|
HUF_nbStreams_e nbStreams,
|
||||||
|
void* workSpace, size_t wkspSize,
|
||||||
|
HUF_CElt* oldHufTable, HUF_repeat* repeat, int preferRepeat,
|
||||||
|
const int bmi2)
|
||||||
|
{
|
||||||
|
HUF_compress_tables_t* const table = (HUF_compress_tables_t*)workSpace;
|
||||||
|
BYTE* const ostart = (BYTE*)dst;
|
||||||
|
BYTE* const oend = ostart + dstSize;
|
||||||
|
BYTE* op = ostart;
|
||||||
|
|
||||||
|
/* checks & inits */
|
||||||
|
if (((size_t)workSpace & 3) != 0) return ERROR(GENERIC); /* must be aligned on 4-bytes boundaries */
|
||||||
|
if (wkspSize < HUF_WORKSPACE_SIZE) return ERROR(workSpace_tooSmall);
|
||||||
|
if (!srcSize) return 0; /* Uncompressed */
|
||||||
|
if (!dstSize) return 0; /* cannot fit anything within dst budget */
|
||||||
|
if (srcSize > HUF_BLOCKSIZE_MAX) return ERROR(srcSize_wrong); /* current block size limit */
|
||||||
|
if (huffLog > HUF_TABLELOG_MAX) return ERROR(tableLog_tooLarge);
|
||||||
|
if (maxSymbolValue > HUF_SYMBOLVALUE_MAX) return ERROR(maxSymbolValue_tooLarge);
|
||||||
|
if (!maxSymbolValue) maxSymbolValue = HUF_SYMBOLVALUE_MAX;
|
||||||
|
if (!huffLog) huffLog = HUF_TABLELOG_DEFAULT;
|
||||||
|
|
||||||
|
/* Heuristic : If old table is valid, use it for small inputs */
|
||||||
|
if (preferRepeat && repeat && *repeat == HUF_repeat_valid) {
|
||||||
|
return HUF_compressCTable_internal(ostart, op, oend,
|
||||||
|
src, srcSize,
|
||||||
|
nbStreams, oldHufTable, bmi2);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Scan input and build symbol stats */
|
||||||
|
{ CHECK_V_F(largest, HIST_count_wksp (table->count, &maxSymbolValue, (const BYTE*)src, srcSize, workSpace, wkspSize) );
|
||||||
|
if (largest == srcSize) { *ostart = ((const BYTE*)src)[0]; return 1; } /* single symbol, rle */
|
||||||
|
if (largest <= (srcSize >> 7)+4) return 0; /* heuristic : probably not compressible enough */
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Check validity of previous table */
|
||||||
|
if ( repeat
|
||||||
|
&& *repeat == HUF_repeat_check
|
||||||
|
&& !HUF_validateCTable(oldHufTable, table->count, maxSymbolValue)) {
|
||||||
|
*repeat = HUF_repeat_none;
|
||||||
|
}
|
||||||
|
/* Heuristic : use existing table for small inputs */
|
||||||
|
if (preferRepeat && repeat && *repeat != HUF_repeat_none) {
|
||||||
|
return HUF_compressCTable_internal(ostart, op, oend,
|
||||||
|
src, srcSize,
|
||||||
|
nbStreams, oldHufTable, bmi2);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Build Huffman Tree */
|
||||||
|
huffLog = HUF_optimalTableLog(huffLog, srcSize, maxSymbolValue);
|
||||||
|
{ size_t const maxBits = HUF_buildCTable_wksp(table->CTable, table->count,
|
||||||
|
maxSymbolValue, huffLog,
|
||||||
|
table->nodeTable, sizeof(table->nodeTable));
|
||||||
|
CHECK_F(maxBits);
|
||||||
|
huffLog = (U32)maxBits;
|
||||||
|
/* Zero unused symbols in CTable, so we can check it for validity */
|
||||||
|
memset(table->CTable + (maxSymbolValue + 1), 0,
|
||||||
|
sizeof(table->CTable) - ((maxSymbolValue + 1) * sizeof(HUF_CElt)));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Write table description header */
|
||||||
|
{ CHECK_V_F(hSize, HUF_writeCTable (op, dstSize, table->CTable, maxSymbolValue, huffLog) );
|
||||||
|
/* Check if using previous huffman table is beneficial */
|
||||||
|
if (repeat && *repeat != HUF_repeat_none) {
|
||||||
|
size_t const oldSize = HUF_estimateCompressedSize(oldHufTable, table->count, maxSymbolValue);
|
||||||
|
size_t const newSize = HUF_estimateCompressedSize(table->CTable, table->count, maxSymbolValue);
|
||||||
|
if (oldSize <= hSize + newSize || hSize + 12 >= srcSize) {
|
||||||
|
return HUF_compressCTable_internal(ostart, op, oend,
|
||||||
|
src, srcSize,
|
||||||
|
nbStreams, oldHufTable, bmi2);
|
||||||
|
} }
|
||||||
|
|
||||||
|
/* Use the new huffman table */
|
||||||
|
if (hSize + 12ul >= srcSize) { return 0; }
|
||||||
|
op += hSize;
|
||||||
|
if (repeat) { *repeat = HUF_repeat_none; }
|
||||||
|
if (oldHufTable)
|
||||||
|
memcpy(oldHufTable, table->CTable, sizeof(table->CTable)); /* Save new table */
|
||||||
|
}
|
||||||
|
return HUF_compressCTable_internal(ostart, op, oend,
|
||||||
|
src, srcSize,
|
||||||
|
nbStreams, table->CTable, bmi2);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t HUF_compress1X_wksp (void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
unsigned maxSymbolValue, unsigned huffLog,
|
||||||
|
void* workSpace, size_t wkspSize)
|
||||||
|
{
|
||||||
|
return HUF_compress_internal(dst, dstSize, src, srcSize,
|
||||||
|
maxSymbolValue, huffLog, HUF_singleStream,
|
||||||
|
workSpace, wkspSize,
|
||||||
|
NULL, NULL, 0, 0 /*bmi2*/);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t HUF_compress1X_repeat (void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
unsigned maxSymbolValue, unsigned huffLog,
|
||||||
|
void* workSpace, size_t wkspSize,
|
||||||
|
HUF_CElt* hufTable, HUF_repeat* repeat, int preferRepeat, int bmi2)
|
||||||
|
{
|
||||||
|
return HUF_compress_internal(dst, dstSize, src, srcSize,
|
||||||
|
maxSymbolValue, huffLog, HUF_singleStream,
|
||||||
|
workSpace, wkspSize, hufTable,
|
||||||
|
repeat, preferRepeat, bmi2);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t HUF_compress1X (void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
unsigned maxSymbolValue, unsigned huffLog)
|
||||||
|
{
|
||||||
|
unsigned workSpace[HUF_WORKSPACE_SIZE_U32];
|
||||||
|
return HUF_compress1X_wksp(dst, dstSize, src, srcSize, maxSymbolValue, huffLog, workSpace, sizeof(workSpace));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* HUF_compress4X_repeat():
|
||||||
|
* compress input using 4 streams.
|
||||||
|
* provide workspace to generate compression tables */
|
||||||
|
size_t HUF_compress4X_wksp (void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
unsigned maxSymbolValue, unsigned huffLog,
|
||||||
|
void* workSpace, size_t wkspSize)
|
||||||
|
{
|
||||||
|
return HUF_compress_internal(dst, dstSize, src, srcSize,
|
||||||
|
maxSymbolValue, huffLog, HUF_fourStreams,
|
||||||
|
workSpace, wkspSize,
|
||||||
|
NULL, NULL, 0, 0 /*bmi2*/);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* HUF_compress4X_repeat():
|
||||||
|
* compress input using 4 streams.
|
||||||
|
* re-use an existing huffman compression table */
|
||||||
|
size_t HUF_compress4X_repeat (void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
unsigned maxSymbolValue, unsigned huffLog,
|
||||||
|
void* workSpace, size_t wkspSize,
|
||||||
|
HUF_CElt* hufTable, HUF_repeat* repeat, int preferRepeat, int bmi2)
|
||||||
|
{
|
||||||
|
return HUF_compress_internal(dst, dstSize, src, srcSize,
|
||||||
|
maxSymbolValue, huffLog, HUF_fourStreams,
|
||||||
|
workSpace, wkspSize,
|
||||||
|
hufTable, repeat, preferRepeat, bmi2);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t HUF_compress2 (void* dst, size_t dstSize,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
unsigned maxSymbolValue, unsigned huffLog)
|
||||||
|
{
|
||||||
|
unsigned workSpace[HUF_WORKSPACE_SIZE_U32];
|
||||||
|
return HUF_compress4X_wksp(dst, dstSize, src, srcSize, maxSymbolValue, huffLog, workSpace, sizeof(workSpace));
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t HUF_compress (void* dst, size_t maxDstSize, const void* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
return HUF_compress2(dst, maxDstSize, src, srcSize, 255, HUF_TABLELOG_DEFAULT);
|
||||||
|
}
|
||||||
1234
vendor/github.com/DataDog/zstd/huf_decompress.c
generated
vendored
Normal file
1234
vendor/github.com/DataDog/zstd/huf_decompress.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
453
vendor/github.com/DataDog/zstd/mem.h
generated
vendored
Normal file
453
vendor/github.com/DataDog/zstd/mem.h
generated
vendored
Normal file
@@ -0,0 +1,453 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef MEM_H_MODULE
|
||||||
|
#define MEM_H_MODULE
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*-****************************************
|
||||||
|
* Dependencies
|
||||||
|
******************************************/
|
||||||
|
#include <stddef.h> /* size_t, ptrdiff_t */
|
||||||
|
#include <string.h> /* memcpy */
|
||||||
|
|
||||||
|
|
||||||
|
/*-****************************************
|
||||||
|
* Compiler specifics
|
||||||
|
******************************************/
|
||||||
|
#if defined(_MSC_VER) /* Visual Studio */
|
||||||
|
# include <stdlib.h> /* _byteswap_ulong */
|
||||||
|
# include <intrin.h> /* _byteswap_* */
|
||||||
|
#endif
|
||||||
|
#if defined(__GNUC__)
|
||||||
|
# define MEM_STATIC static __inline __attribute__((unused))
|
||||||
|
#elif defined (__cplusplus) || (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L) /* C99 */)
|
||||||
|
# define MEM_STATIC static inline
|
||||||
|
#elif defined(_MSC_VER)
|
||||||
|
# define MEM_STATIC static __inline
|
||||||
|
#else
|
||||||
|
# define MEM_STATIC static /* this version may generate warnings for unused static functions; disable the relevant warning */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#ifndef __has_builtin
|
||||||
|
# define __has_builtin(x) 0 /* compat. with non-clang compilers */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* code only tested on 32 and 64 bits systems */
|
||||||
|
#define MEM_STATIC_ASSERT(c) { enum { MEM_static_assert = 1/(int)(!!(c)) }; }
|
||||||
|
MEM_STATIC void MEM_check(void) { MEM_STATIC_ASSERT((sizeof(size_t)==4) || (sizeof(size_t)==8)); }
|
||||||
|
|
||||||
|
/* detects whether we are being compiled under msan */
|
||||||
|
#if defined (__has_feature)
|
||||||
|
# if __has_feature(memory_sanitizer)
|
||||||
|
# define MEMORY_SANITIZER 1
|
||||||
|
# endif
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#if defined (MEMORY_SANITIZER)
|
||||||
|
/* Not all platforms that support msan provide sanitizers/msan_interface.h.
|
||||||
|
* We therefore declare the functions we need ourselves, rather than trying to
|
||||||
|
* include the header file... */
|
||||||
|
|
||||||
|
#include <stdint.h> /* intptr_t */
|
||||||
|
|
||||||
|
/* Make memory region fully initialized (without changing its contents). */
|
||||||
|
void __msan_unpoison(const volatile void *a, size_t size);
|
||||||
|
|
||||||
|
/* Make memory region fully uninitialized (without changing its contents).
|
||||||
|
This is a legacy interface that does not update origin information. Use
|
||||||
|
__msan_allocated_memory() instead. */
|
||||||
|
void __msan_poison(const volatile void *a, size_t size);
|
||||||
|
|
||||||
|
/* Returns the offset of the first (at least partially) poisoned byte in the
|
||||||
|
memory range, or -1 if the whole range is good. */
|
||||||
|
intptr_t __msan_test_shadow(const volatile void *x, size_t size);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* detects whether we are being compiled under asan */
|
||||||
|
#if defined (__has_feature)
|
||||||
|
# if __has_feature(address_sanitizer)
|
||||||
|
# define ADDRESS_SANITIZER 1
|
||||||
|
# endif
|
||||||
|
#elif defined(__SANITIZE_ADDRESS__)
|
||||||
|
# define ADDRESS_SANITIZER 1
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#if defined (ADDRESS_SANITIZER)
|
||||||
|
/* Not all platforms that support asan provide sanitizers/asan_interface.h.
|
||||||
|
* We therefore declare the functions we need ourselves, rather than trying to
|
||||||
|
* include the header file... */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Marks a memory region (<c>[addr, addr+size)</c>) as unaddressable.
|
||||||
|
*
|
||||||
|
* This memory must be previously allocated by your program. Instrumented
|
||||||
|
* code is forbidden from accessing addresses in this region until it is
|
||||||
|
* unpoisoned. This function is not guaranteed to poison the entire region -
|
||||||
|
* it could poison only a subregion of <c>[addr, addr+size)</c> due to ASan
|
||||||
|
* alignment restrictions.
|
||||||
|
*
|
||||||
|
* \note This function is not thread-safe because no two threads can poison or
|
||||||
|
* unpoison memory in the same memory region simultaneously.
|
||||||
|
*
|
||||||
|
* \param addr Start of memory region.
|
||||||
|
* \param size Size of memory region. */
|
||||||
|
void __asan_poison_memory_region(void const volatile *addr, size_t size);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Marks a memory region (<c>[addr, addr+size)</c>) as addressable.
|
||||||
|
*
|
||||||
|
* This memory must be previously allocated by your program. Accessing
|
||||||
|
* addresses in this region is allowed until this region is poisoned again.
|
||||||
|
* This function could unpoison a super-region of <c>[addr, addr+size)</c> due
|
||||||
|
* to ASan alignment restrictions.
|
||||||
|
*
|
||||||
|
* \note This function is not thread-safe because no two threads can
|
||||||
|
* poison or unpoison memory in the same memory region simultaneously.
|
||||||
|
*
|
||||||
|
* \param addr Start of memory region.
|
||||||
|
* \param size Size of memory region. */
|
||||||
|
void __asan_unpoison_memory_region(void const volatile *addr, size_t size);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/*-**************************************************************
|
||||||
|
* Basic Types
|
||||||
|
*****************************************************************/
|
||||||
|
#if !defined (__VMS) && (defined (__cplusplus) || (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L) /* C99 */) )
|
||||||
|
# include <stdint.h>
|
||||||
|
typedef uint8_t BYTE;
|
||||||
|
typedef uint16_t U16;
|
||||||
|
typedef int16_t S16;
|
||||||
|
typedef uint32_t U32;
|
||||||
|
typedef int32_t S32;
|
||||||
|
typedef uint64_t U64;
|
||||||
|
typedef int64_t S64;
|
||||||
|
#else
|
||||||
|
# include <limits.h>
|
||||||
|
#if CHAR_BIT != 8
|
||||||
|
# error "this implementation requires char to be exactly 8-bit type"
|
||||||
|
#endif
|
||||||
|
typedef unsigned char BYTE;
|
||||||
|
#if USHRT_MAX != 65535
|
||||||
|
# error "this implementation requires short to be exactly 16-bit type"
|
||||||
|
#endif
|
||||||
|
typedef unsigned short U16;
|
||||||
|
typedef signed short S16;
|
||||||
|
#if UINT_MAX != 4294967295
|
||||||
|
# error "this implementation requires int to be exactly 32-bit type"
|
||||||
|
#endif
|
||||||
|
typedef unsigned int U32;
|
||||||
|
typedef signed int S32;
|
||||||
|
/* note : there are no limits defined for long long type in C90.
|
||||||
|
* limits exist in C99, however, in such case, <stdint.h> is preferred */
|
||||||
|
typedef unsigned long long U64;
|
||||||
|
typedef signed long long S64;
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/*-**************************************************************
|
||||||
|
* Memory I/O
|
||||||
|
*****************************************************************/
|
||||||
|
/* MEM_FORCE_MEMORY_ACCESS :
|
||||||
|
* By default, access to unaligned memory is controlled by `memcpy()`, which is safe and portable.
|
||||||
|
* Unfortunately, on some target/compiler combinations, the generated assembly is sub-optimal.
|
||||||
|
* The below switch allow to select different access method for improved performance.
|
||||||
|
* Method 0 (default) : use `memcpy()`. Safe and portable.
|
||||||
|
* Method 1 : `__packed` statement. It depends on compiler extension (i.e., not portable).
|
||||||
|
* This method is safe if your compiler supports it, and *generally* as fast or faster than `memcpy`.
|
||||||
|
* Method 2 : direct access. This method is portable but violate C standard.
|
||||||
|
* It can generate buggy code on targets depending on alignment.
|
||||||
|
* In some circumstances, it's the only known way to get the most performance (i.e. GCC + ARMv6)
|
||||||
|
* See http://fastcompression.blogspot.fr/2015/08/accessing-unaligned-memory.html for details.
|
||||||
|
* Prefer these methods in priority order (0 > 1 > 2)
|
||||||
|
*/
|
||||||
|
#ifndef MEM_FORCE_MEMORY_ACCESS /* can be defined externally, on command line for example */
|
||||||
|
# if defined(__GNUC__) && ( defined(__ARM_ARCH_6__) || defined(__ARM_ARCH_6J__) || defined(__ARM_ARCH_6K__) || defined(__ARM_ARCH_6Z__) || defined(__ARM_ARCH_6ZK__) || defined(__ARM_ARCH_6T2__) )
|
||||||
|
# define MEM_FORCE_MEMORY_ACCESS 2
|
||||||
|
# elif defined(__INTEL_COMPILER) || defined(__GNUC__) || defined(__ICCARM__)
|
||||||
|
# define MEM_FORCE_MEMORY_ACCESS 1
|
||||||
|
# endif
|
||||||
|
#endif
|
||||||
|
|
||||||
|
MEM_STATIC unsigned MEM_32bits(void) { return sizeof(size_t)==4; }
|
||||||
|
MEM_STATIC unsigned MEM_64bits(void) { return sizeof(size_t)==8; }
|
||||||
|
|
||||||
|
MEM_STATIC unsigned MEM_isLittleEndian(void)
|
||||||
|
{
|
||||||
|
const union { U32 u; BYTE c[4]; } one = { 1 }; /* don't use static : performance detrimental */
|
||||||
|
return one.c[0];
|
||||||
|
}
|
||||||
|
|
||||||
|
#if defined(MEM_FORCE_MEMORY_ACCESS) && (MEM_FORCE_MEMORY_ACCESS==2)
|
||||||
|
|
||||||
|
/* violates C standard, by lying on structure alignment.
|
||||||
|
Only use if no other choice to achieve best performance on target platform */
|
||||||
|
MEM_STATIC U16 MEM_read16(const void* memPtr) { return *(const U16*) memPtr; }
|
||||||
|
MEM_STATIC U32 MEM_read32(const void* memPtr) { return *(const U32*) memPtr; }
|
||||||
|
MEM_STATIC U64 MEM_read64(const void* memPtr) { return *(const U64*) memPtr; }
|
||||||
|
MEM_STATIC size_t MEM_readST(const void* memPtr) { return *(const size_t*) memPtr; }
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_write16(void* memPtr, U16 value) { *(U16*)memPtr = value; }
|
||||||
|
MEM_STATIC void MEM_write32(void* memPtr, U32 value) { *(U32*)memPtr = value; }
|
||||||
|
MEM_STATIC void MEM_write64(void* memPtr, U64 value) { *(U64*)memPtr = value; }
|
||||||
|
|
||||||
|
#elif defined(MEM_FORCE_MEMORY_ACCESS) && (MEM_FORCE_MEMORY_ACCESS==1)
|
||||||
|
|
||||||
|
/* __pack instructions are safer, but compiler specific, hence potentially problematic for some compilers */
|
||||||
|
/* currently only defined for gcc and icc */
|
||||||
|
#if defined(_MSC_VER) || (defined(__INTEL_COMPILER) && defined(WIN32))
|
||||||
|
__pragma( pack(push, 1) )
|
||||||
|
typedef struct { U16 v; } unalign16;
|
||||||
|
typedef struct { U32 v; } unalign32;
|
||||||
|
typedef struct { U64 v; } unalign64;
|
||||||
|
typedef struct { size_t v; } unalignArch;
|
||||||
|
__pragma( pack(pop) )
|
||||||
|
#else
|
||||||
|
typedef struct { U16 v; } __attribute__((packed)) unalign16;
|
||||||
|
typedef struct { U32 v; } __attribute__((packed)) unalign32;
|
||||||
|
typedef struct { U64 v; } __attribute__((packed)) unalign64;
|
||||||
|
typedef struct { size_t v; } __attribute__((packed)) unalignArch;
|
||||||
|
#endif
|
||||||
|
|
||||||
|
MEM_STATIC U16 MEM_read16(const void* ptr) { return ((const unalign16*)ptr)->v; }
|
||||||
|
MEM_STATIC U32 MEM_read32(const void* ptr) { return ((const unalign32*)ptr)->v; }
|
||||||
|
MEM_STATIC U64 MEM_read64(const void* ptr) { return ((const unalign64*)ptr)->v; }
|
||||||
|
MEM_STATIC size_t MEM_readST(const void* ptr) { return ((const unalignArch*)ptr)->v; }
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_write16(void* memPtr, U16 value) { ((unalign16*)memPtr)->v = value; }
|
||||||
|
MEM_STATIC void MEM_write32(void* memPtr, U32 value) { ((unalign32*)memPtr)->v = value; }
|
||||||
|
MEM_STATIC void MEM_write64(void* memPtr, U64 value) { ((unalign64*)memPtr)->v = value; }
|
||||||
|
|
||||||
|
#else
|
||||||
|
|
||||||
|
/* default method, safe and standard.
|
||||||
|
can sometimes prove slower */
|
||||||
|
|
||||||
|
MEM_STATIC U16 MEM_read16(const void* memPtr)
|
||||||
|
{
|
||||||
|
U16 val; memcpy(&val, memPtr, sizeof(val)); return val;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC U32 MEM_read32(const void* memPtr)
|
||||||
|
{
|
||||||
|
U32 val; memcpy(&val, memPtr, sizeof(val)); return val;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC U64 MEM_read64(const void* memPtr)
|
||||||
|
{
|
||||||
|
U64 val; memcpy(&val, memPtr, sizeof(val)); return val;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC size_t MEM_readST(const void* memPtr)
|
||||||
|
{
|
||||||
|
size_t val; memcpy(&val, memPtr, sizeof(val)); return val;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_write16(void* memPtr, U16 value)
|
||||||
|
{
|
||||||
|
memcpy(memPtr, &value, sizeof(value));
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_write32(void* memPtr, U32 value)
|
||||||
|
{
|
||||||
|
memcpy(memPtr, &value, sizeof(value));
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_write64(void* memPtr, U64 value)
|
||||||
|
{
|
||||||
|
memcpy(memPtr, &value, sizeof(value));
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif /* MEM_FORCE_MEMORY_ACCESS */
|
||||||
|
|
||||||
|
MEM_STATIC U32 MEM_swap32(U32 in)
|
||||||
|
{
|
||||||
|
#if defined(_MSC_VER) /* Visual Studio */
|
||||||
|
return _byteswap_ulong(in);
|
||||||
|
#elif (defined (__GNUC__) && (__GNUC__ * 100 + __GNUC_MINOR__ >= 403)) \
|
||||||
|
|| (defined(__clang__) && __has_builtin(__builtin_bswap32))
|
||||||
|
return __builtin_bswap32(in);
|
||||||
|
#else
|
||||||
|
return ((in << 24) & 0xff000000 ) |
|
||||||
|
((in << 8) & 0x00ff0000 ) |
|
||||||
|
((in >> 8) & 0x0000ff00 ) |
|
||||||
|
((in >> 24) & 0x000000ff );
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC U64 MEM_swap64(U64 in)
|
||||||
|
{
|
||||||
|
#if defined(_MSC_VER) /* Visual Studio */
|
||||||
|
return _byteswap_uint64(in);
|
||||||
|
#elif (defined (__GNUC__) && (__GNUC__ * 100 + __GNUC_MINOR__ >= 403)) \
|
||||||
|
|| (defined(__clang__) && __has_builtin(__builtin_bswap64))
|
||||||
|
return __builtin_bswap64(in);
|
||||||
|
#else
|
||||||
|
return ((in << 56) & 0xff00000000000000ULL) |
|
||||||
|
((in << 40) & 0x00ff000000000000ULL) |
|
||||||
|
((in << 24) & 0x0000ff0000000000ULL) |
|
||||||
|
((in << 8) & 0x000000ff00000000ULL) |
|
||||||
|
((in >> 8) & 0x00000000ff000000ULL) |
|
||||||
|
((in >> 24) & 0x0000000000ff0000ULL) |
|
||||||
|
((in >> 40) & 0x000000000000ff00ULL) |
|
||||||
|
((in >> 56) & 0x00000000000000ffULL);
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC size_t MEM_swapST(size_t in)
|
||||||
|
{
|
||||||
|
if (MEM_32bits())
|
||||||
|
return (size_t)MEM_swap32((U32)in);
|
||||||
|
else
|
||||||
|
return (size_t)MEM_swap64((U64)in);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*=== Little endian r/w ===*/
|
||||||
|
|
||||||
|
MEM_STATIC U16 MEM_readLE16(const void* memPtr)
|
||||||
|
{
|
||||||
|
if (MEM_isLittleEndian())
|
||||||
|
return MEM_read16(memPtr);
|
||||||
|
else {
|
||||||
|
const BYTE* p = (const BYTE*)memPtr;
|
||||||
|
return (U16)(p[0] + (p[1]<<8));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_writeLE16(void* memPtr, U16 val)
|
||||||
|
{
|
||||||
|
if (MEM_isLittleEndian()) {
|
||||||
|
MEM_write16(memPtr, val);
|
||||||
|
} else {
|
||||||
|
BYTE* p = (BYTE*)memPtr;
|
||||||
|
p[0] = (BYTE)val;
|
||||||
|
p[1] = (BYTE)(val>>8);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC U32 MEM_readLE24(const void* memPtr)
|
||||||
|
{
|
||||||
|
return MEM_readLE16(memPtr) + (((const BYTE*)memPtr)[2] << 16);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_writeLE24(void* memPtr, U32 val)
|
||||||
|
{
|
||||||
|
MEM_writeLE16(memPtr, (U16)val);
|
||||||
|
((BYTE*)memPtr)[2] = (BYTE)(val>>16);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC U32 MEM_readLE32(const void* memPtr)
|
||||||
|
{
|
||||||
|
if (MEM_isLittleEndian())
|
||||||
|
return MEM_read32(memPtr);
|
||||||
|
else
|
||||||
|
return MEM_swap32(MEM_read32(memPtr));
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_writeLE32(void* memPtr, U32 val32)
|
||||||
|
{
|
||||||
|
if (MEM_isLittleEndian())
|
||||||
|
MEM_write32(memPtr, val32);
|
||||||
|
else
|
||||||
|
MEM_write32(memPtr, MEM_swap32(val32));
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC U64 MEM_readLE64(const void* memPtr)
|
||||||
|
{
|
||||||
|
if (MEM_isLittleEndian())
|
||||||
|
return MEM_read64(memPtr);
|
||||||
|
else
|
||||||
|
return MEM_swap64(MEM_read64(memPtr));
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_writeLE64(void* memPtr, U64 val64)
|
||||||
|
{
|
||||||
|
if (MEM_isLittleEndian())
|
||||||
|
MEM_write64(memPtr, val64);
|
||||||
|
else
|
||||||
|
MEM_write64(memPtr, MEM_swap64(val64));
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC size_t MEM_readLEST(const void* memPtr)
|
||||||
|
{
|
||||||
|
if (MEM_32bits())
|
||||||
|
return (size_t)MEM_readLE32(memPtr);
|
||||||
|
else
|
||||||
|
return (size_t)MEM_readLE64(memPtr);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_writeLEST(void* memPtr, size_t val)
|
||||||
|
{
|
||||||
|
if (MEM_32bits())
|
||||||
|
MEM_writeLE32(memPtr, (U32)val);
|
||||||
|
else
|
||||||
|
MEM_writeLE64(memPtr, (U64)val);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*=== Big endian r/w ===*/
|
||||||
|
|
||||||
|
MEM_STATIC U32 MEM_readBE32(const void* memPtr)
|
||||||
|
{
|
||||||
|
if (MEM_isLittleEndian())
|
||||||
|
return MEM_swap32(MEM_read32(memPtr));
|
||||||
|
else
|
||||||
|
return MEM_read32(memPtr);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_writeBE32(void* memPtr, U32 val32)
|
||||||
|
{
|
||||||
|
if (MEM_isLittleEndian())
|
||||||
|
MEM_write32(memPtr, MEM_swap32(val32));
|
||||||
|
else
|
||||||
|
MEM_write32(memPtr, val32);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC U64 MEM_readBE64(const void* memPtr)
|
||||||
|
{
|
||||||
|
if (MEM_isLittleEndian())
|
||||||
|
return MEM_swap64(MEM_read64(memPtr));
|
||||||
|
else
|
||||||
|
return MEM_read64(memPtr);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_writeBE64(void* memPtr, U64 val64)
|
||||||
|
{
|
||||||
|
if (MEM_isLittleEndian())
|
||||||
|
MEM_write64(memPtr, MEM_swap64(val64));
|
||||||
|
else
|
||||||
|
MEM_write64(memPtr, val64);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC size_t MEM_readBEST(const void* memPtr)
|
||||||
|
{
|
||||||
|
if (MEM_32bits())
|
||||||
|
return (size_t)MEM_readBE32(memPtr);
|
||||||
|
else
|
||||||
|
return (size_t)MEM_readBE64(memPtr);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void MEM_writeBEST(void* memPtr, size_t val)
|
||||||
|
{
|
||||||
|
if (MEM_32bits())
|
||||||
|
MEM_writeBE32(memPtr, (U32)val);
|
||||||
|
else
|
||||||
|
MEM_writeBE64(memPtr, (U64)val);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* MEM_H_MODULE */
|
||||||
344
vendor/github.com/DataDog/zstd/pool.c
generated
vendored
Normal file
344
vendor/github.com/DataDog/zstd/pool.c
generated
vendored
Normal file
@@ -0,0 +1,344 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/* ====== Dependencies ======= */
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
#include "debug.h" /* assert */
|
||||||
|
#include "zstd_internal.h" /* ZSTD_malloc, ZSTD_free */
|
||||||
|
#include "pool.h"
|
||||||
|
|
||||||
|
/* ====== Compiler specifics ====== */
|
||||||
|
#if defined(_MSC_VER)
|
||||||
|
# pragma warning(disable : 4204) /* disable: C4204: non-constant aggregate initializer */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
#ifdef ZSTD_MULTITHREAD
|
||||||
|
|
||||||
|
#include "threading.h" /* pthread adaptation */
|
||||||
|
|
||||||
|
/* A job is a function and an opaque argument */
|
||||||
|
typedef struct POOL_job_s {
|
||||||
|
POOL_function function;
|
||||||
|
void *opaque;
|
||||||
|
} POOL_job;
|
||||||
|
|
||||||
|
struct POOL_ctx_s {
|
||||||
|
ZSTD_customMem customMem;
|
||||||
|
/* Keep track of the threads */
|
||||||
|
ZSTD_pthread_t* threads;
|
||||||
|
size_t threadCapacity;
|
||||||
|
size_t threadLimit;
|
||||||
|
|
||||||
|
/* The queue is a circular buffer */
|
||||||
|
POOL_job *queue;
|
||||||
|
size_t queueHead;
|
||||||
|
size_t queueTail;
|
||||||
|
size_t queueSize;
|
||||||
|
|
||||||
|
/* The number of threads working on jobs */
|
||||||
|
size_t numThreadsBusy;
|
||||||
|
/* Indicates if the queue is empty */
|
||||||
|
int queueEmpty;
|
||||||
|
|
||||||
|
/* The mutex protects the queue */
|
||||||
|
ZSTD_pthread_mutex_t queueMutex;
|
||||||
|
/* Condition variable for pushers to wait on when the queue is full */
|
||||||
|
ZSTD_pthread_cond_t queuePushCond;
|
||||||
|
/* Condition variables for poppers to wait on when the queue is empty */
|
||||||
|
ZSTD_pthread_cond_t queuePopCond;
|
||||||
|
/* Indicates if the queue is shutting down */
|
||||||
|
int shutdown;
|
||||||
|
};
|
||||||
|
|
||||||
|
/* POOL_thread() :
|
||||||
|
* Work thread for the thread pool.
|
||||||
|
* Waits for jobs and executes them.
|
||||||
|
* @returns : NULL on failure else non-null.
|
||||||
|
*/
|
||||||
|
static void* POOL_thread(void* opaque) {
|
||||||
|
POOL_ctx* const ctx = (POOL_ctx*)opaque;
|
||||||
|
if (!ctx) { return NULL; }
|
||||||
|
for (;;) {
|
||||||
|
/* Lock the mutex and wait for a non-empty queue or until shutdown */
|
||||||
|
ZSTD_pthread_mutex_lock(&ctx->queueMutex);
|
||||||
|
|
||||||
|
while ( ctx->queueEmpty
|
||||||
|
|| (ctx->numThreadsBusy >= ctx->threadLimit) ) {
|
||||||
|
if (ctx->shutdown) {
|
||||||
|
/* even if !queueEmpty, (possible if numThreadsBusy >= threadLimit),
|
||||||
|
* a few threads will be shutdown while !queueEmpty,
|
||||||
|
* but enough threads will remain active to finish the queue */
|
||||||
|
ZSTD_pthread_mutex_unlock(&ctx->queueMutex);
|
||||||
|
return opaque;
|
||||||
|
}
|
||||||
|
ZSTD_pthread_cond_wait(&ctx->queuePopCond, &ctx->queueMutex);
|
||||||
|
}
|
||||||
|
/* Pop a job off the queue */
|
||||||
|
{ POOL_job const job = ctx->queue[ctx->queueHead];
|
||||||
|
ctx->queueHead = (ctx->queueHead + 1) % ctx->queueSize;
|
||||||
|
ctx->numThreadsBusy++;
|
||||||
|
ctx->queueEmpty = ctx->queueHead == ctx->queueTail;
|
||||||
|
/* Unlock the mutex, signal a pusher, and run the job */
|
||||||
|
ZSTD_pthread_cond_signal(&ctx->queuePushCond);
|
||||||
|
ZSTD_pthread_mutex_unlock(&ctx->queueMutex);
|
||||||
|
|
||||||
|
job.function(job.opaque);
|
||||||
|
|
||||||
|
/* If the intended queue size was 0, signal after finishing job */
|
||||||
|
ZSTD_pthread_mutex_lock(&ctx->queueMutex);
|
||||||
|
ctx->numThreadsBusy--;
|
||||||
|
if (ctx->queueSize == 1) {
|
||||||
|
ZSTD_pthread_cond_signal(&ctx->queuePushCond);
|
||||||
|
}
|
||||||
|
ZSTD_pthread_mutex_unlock(&ctx->queueMutex);
|
||||||
|
}
|
||||||
|
} /* for (;;) */
|
||||||
|
assert(0); /* Unreachable */
|
||||||
|
}
|
||||||
|
|
||||||
|
POOL_ctx* POOL_create(size_t numThreads, size_t queueSize) {
|
||||||
|
return POOL_create_advanced(numThreads, queueSize, ZSTD_defaultCMem);
|
||||||
|
}
|
||||||
|
|
||||||
|
POOL_ctx* POOL_create_advanced(size_t numThreads, size_t queueSize,
|
||||||
|
ZSTD_customMem customMem) {
|
||||||
|
POOL_ctx* ctx;
|
||||||
|
/* Check parameters */
|
||||||
|
if (!numThreads) { return NULL; }
|
||||||
|
/* Allocate the context and zero initialize */
|
||||||
|
ctx = (POOL_ctx*)ZSTD_calloc(sizeof(POOL_ctx), customMem);
|
||||||
|
if (!ctx) { return NULL; }
|
||||||
|
/* Initialize the job queue.
|
||||||
|
* It needs one extra space since one space is wasted to differentiate
|
||||||
|
* empty and full queues.
|
||||||
|
*/
|
||||||
|
ctx->queueSize = queueSize + 1;
|
||||||
|
ctx->queue = (POOL_job*)ZSTD_malloc(ctx->queueSize * sizeof(POOL_job), customMem);
|
||||||
|
ctx->queueHead = 0;
|
||||||
|
ctx->queueTail = 0;
|
||||||
|
ctx->numThreadsBusy = 0;
|
||||||
|
ctx->queueEmpty = 1;
|
||||||
|
{
|
||||||
|
int error = 0;
|
||||||
|
error |= ZSTD_pthread_mutex_init(&ctx->queueMutex, NULL);
|
||||||
|
error |= ZSTD_pthread_cond_init(&ctx->queuePushCond, NULL);
|
||||||
|
error |= ZSTD_pthread_cond_init(&ctx->queuePopCond, NULL);
|
||||||
|
if (error) { POOL_free(ctx); return NULL; }
|
||||||
|
}
|
||||||
|
ctx->shutdown = 0;
|
||||||
|
/* Allocate space for the thread handles */
|
||||||
|
ctx->threads = (ZSTD_pthread_t*)ZSTD_malloc(numThreads * sizeof(ZSTD_pthread_t), customMem);
|
||||||
|
ctx->threadCapacity = 0;
|
||||||
|
ctx->customMem = customMem;
|
||||||
|
/* Check for errors */
|
||||||
|
if (!ctx->threads || !ctx->queue) { POOL_free(ctx); return NULL; }
|
||||||
|
/* Initialize the threads */
|
||||||
|
{ size_t i;
|
||||||
|
for (i = 0; i < numThreads; ++i) {
|
||||||
|
if (ZSTD_pthread_create(&ctx->threads[i], NULL, &POOL_thread, ctx)) {
|
||||||
|
ctx->threadCapacity = i;
|
||||||
|
POOL_free(ctx);
|
||||||
|
return NULL;
|
||||||
|
} }
|
||||||
|
ctx->threadCapacity = numThreads;
|
||||||
|
ctx->threadLimit = numThreads;
|
||||||
|
}
|
||||||
|
return ctx;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! POOL_join() :
|
||||||
|
Shutdown the queue, wake any sleeping threads, and join all of the threads.
|
||||||
|
*/
|
||||||
|
static void POOL_join(POOL_ctx* ctx) {
|
||||||
|
/* Shut down the queue */
|
||||||
|
ZSTD_pthread_mutex_lock(&ctx->queueMutex);
|
||||||
|
ctx->shutdown = 1;
|
||||||
|
ZSTD_pthread_mutex_unlock(&ctx->queueMutex);
|
||||||
|
/* Wake up sleeping threads */
|
||||||
|
ZSTD_pthread_cond_broadcast(&ctx->queuePushCond);
|
||||||
|
ZSTD_pthread_cond_broadcast(&ctx->queuePopCond);
|
||||||
|
/* Join all of the threads */
|
||||||
|
{ size_t i;
|
||||||
|
for (i = 0; i < ctx->threadCapacity; ++i) {
|
||||||
|
ZSTD_pthread_join(ctx->threads[i], NULL); /* note : could fail */
|
||||||
|
} }
|
||||||
|
}
|
||||||
|
|
||||||
|
void POOL_free(POOL_ctx *ctx) {
|
||||||
|
if (!ctx) { return; }
|
||||||
|
POOL_join(ctx);
|
||||||
|
ZSTD_pthread_mutex_destroy(&ctx->queueMutex);
|
||||||
|
ZSTD_pthread_cond_destroy(&ctx->queuePushCond);
|
||||||
|
ZSTD_pthread_cond_destroy(&ctx->queuePopCond);
|
||||||
|
ZSTD_free(ctx->queue, ctx->customMem);
|
||||||
|
ZSTD_free(ctx->threads, ctx->customMem);
|
||||||
|
ZSTD_free(ctx, ctx->customMem);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
size_t POOL_sizeof(POOL_ctx *ctx) {
|
||||||
|
if (ctx==NULL) return 0; /* supports sizeof NULL */
|
||||||
|
return sizeof(*ctx)
|
||||||
|
+ ctx->queueSize * sizeof(POOL_job)
|
||||||
|
+ ctx->threadCapacity * sizeof(ZSTD_pthread_t);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* @return : 0 on success, 1 on error */
|
||||||
|
static int POOL_resize_internal(POOL_ctx* ctx, size_t numThreads)
|
||||||
|
{
|
||||||
|
if (numThreads <= ctx->threadCapacity) {
|
||||||
|
if (!numThreads) return 1;
|
||||||
|
ctx->threadLimit = numThreads;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
/* numThreads > threadCapacity */
|
||||||
|
{ ZSTD_pthread_t* const threadPool = (ZSTD_pthread_t*)ZSTD_malloc(numThreads * sizeof(ZSTD_pthread_t), ctx->customMem);
|
||||||
|
if (!threadPool) return 1;
|
||||||
|
/* replace existing thread pool */
|
||||||
|
memcpy(threadPool, ctx->threads, ctx->threadCapacity * sizeof(*threadPool));
|
||||||
|
ZSTD_free(ctx->threads, ctx->customMem);
|
||||||
|
ctx->threads = threadPool;
|
||||||
|
/* Initialize additional threads */
|
||||||
|
{ size_t threadId;
|
||||||
|
for (threadId = ctx->threadCapacity; threadId < numThreads; ++threadId) {
|
||||||
|
if (ZSTD_pthread_create(&threadPool[threadId], NULL, &POOL_thread, ctx)) {
|
||||||
|
ctx->threadCapacity = threadId;
|
||||||
|
return 1;
|
||||||
|
} }
|
||||||
|
} }
|
||||||
|
/* successfully expanded */
|
||||||
|
ctx->threadCapacity = numThreads;
|
||||||
|
ctx->threadLimit = numThreads;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* @return : 0 on success, 1 on error */
|
||||||
|
int POOL_resize(POOL_ctx* ctx, size_t numThreads)
|
||||||
|
{
|
||||||
|
int result;
|
||||||
|
if (ctx==NULL) return 1;
|
||||||
|
ZSTD_pthread_mutex_lock(&ctx->queueMutex);
|
||||||
|
result = POOL_resize_internal(ctx, numThreads);
|
||||||
|
ZSTD_pthread_cond_broadcast(&ctx->queuePopCond);
|
||||||
|
ZSTD_pthread_mutex_unlock(&ctx->queueMutex);
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns 1 if the queue is full and 0 otherwise.
|
||||||
|
*
|
||||||
|
* When queueSize is 1 (pool was created with an intended queueSize of 0),
|
||||||
|
* then a queue is empty if there is a thread free _and_ no job is waiting.
|
||||||
|
*/
|
||||||
|
static int isQueueFull(POOL_ctx const* ctx) {
|
||||||
|
if (ctx->queueSize > 1) {
|
||||||
|
return ctx->queueHead == ((ctx->queueTail + 1) % ctx->queueSize);
|
||||||
|
} else {
|
||||||
|
return (ctx->numThreadsBusy == ctx->threadLimit) ||
|
||||||
|
!ctx->queueEmpty;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static void POOL_add_internal(POOL_ctx* ctx, POOL_function function, void *opaque)
|
||||||
|
{
|
||||||
|
POOL_job const job = {function, opaque};
|
||||||
|
assert(ctx != NULL);
|
||||||
|
if (ctx->shutdown) return;
|
||||||
|
|
||||||
|
ctx->queueEmpty = 0;
|
||||||
|
ctx->queue[ctx->queueTail] = job;
|
||||||
|
ctx->queueTail = (ctx->queueTail + 1) % ctx->queueSize;
|
||||||
|
ZSTD_pthread_cond_signal(&ctx->queuePopCond);
|
||||||
|
}
|
||||||
|
|
||||||
|
void POOL_add(POOL_ctx* ctx, POOL_function function, void* opaque)
|
||||||
|
{
|
||||||
|
assert(ctx != NULL);
|
||||||
|
ZSTD_pthread_mutex_lock(&ctx->queueMutex);
|
||||||
|
/* Wait until there is space in the queue for the new job */
|
||||||
|
while (isQueueFull(ctx) && (!ctx->shutdown)) {
|
||||||
|
ZSTD_pthread_cond_wait(&ctx->queuePushCond, &ctx->queueMutex);
|
||||||
|
}
|
||||||
|
POOL_add_internal(ctx, function, opaque);
|
||||||
|
ZSTD_pthread_mutex_unlock(&ctx->queueMutex);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
int POOL_tryAdd(POOL_ctx* ctx, POOL_function function, void* opaque)
|
||||||
|
{
|
||||||
|
assert(ctx != NULL);
|
||||||
|
ZSTD_pthread_mutex_lock(&ctx->queueMutex);
|
||||||
|
if (isQueueFull(ctx)) {
|
||||||
|
ZSTD_pthread_mutex_unlock(&ctx->queueMutex);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
POOL_add_internal(ctx, function, opaque);
|
||||||
|
ZSTD_pthread_mutex_unlock(&ctx->queueMutex);
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#else /* ZSTD_MULTITHREAD not defined */
|
||||||
|
|
||||||
|
/* ========================== */
|
||||||
|
/* No multi-threading support */
|
||||||
|
/* ========================== */
|
||||||
|
|
||||||
|
|
||||||
|
/* We don't need any data, but if it is empty, malloc() might return NULL. */
|
||||||
|
struct POOL_ctx_s {
|
||||||
|
int dummy;
|
||||||
|
};
|
||||||
|
static POOL_ctx g_ctx;
|
||||||
|
|
||||||
|
POOL_ctx* POOL_create(size_t numThreads, size_t queueSize) {
|
||||||
|
return POOL_create_advanced(numThreads, queueSize, ZSTD_defaultCMem);
|
||||||
|
}
|
||||||
|
|
||||||
|
POOL_ctx* POOL_create_advanced(size_t numThreads, size_t queueSize, ZSTD_customMem customMem) {
|
||||||
|
(void)numThreads;
|
||||||
|
(void)queueSize;
|
||||||
|
(void)customMem;
|
||||||
|
return &g_ctx;
|
||||||
|
}
|
||||||
|
|
||||||
|
void POOL_free(POOL_ctx* ctx) {
|
||||||
|
assert(!ctx || ctx == &g_ctx);
|
||||||
|
(void)ctx;
|
||||||
|
}
|
||||||
|
|
||||||
|
int POOL_resize(POOL_ctx* ctx, size_t numThreads) {
|
||||||
|
(void)ctx; (void)numThreads;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
void POOL_add(POOL_ctx* ctx, POOL_function function, void* opaque) {
|
||||||
|
(void)ctx;
|
||||||
|
function(opaque);
|
||||||
|
}
|
||||||
|
|
||||||
|
int POOL_tryAdd(POOL_ctx* ctx, POOL_function function, void* opaque) {
|
||||||
|
(void)ctx;
|
||||||
|
function(opaque);
|
||||||
|
return 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t POOL_sizeof(POOL_ctx* ctx) {
|
||||||
|
if (ctx==NULL) return 0; /* supports sizeof NULL */
|
||||||
|
assert(ctx == &g_ctx);
|
||||||
|
return sizeof(*ctx);
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif /* ZSTD_MULTITHREAD */
|
||||||
84
vendor/github.com/DataDog/zstd/pool.h
generated
vendored
Normal file
84
vendor/github.com/DataDog/zstd/pool.h
generated
vendored
Normal file
@@ -0,0 +1,84 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef POOL_H
|
||||||
|
#define POOL_H
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
#define ZSTD_STATIC_LINKING_ONLY /* ZSTD_customMem */
|
||||||
|
#include "zstd.h"
|
||||||
|
|
||||||
|
typedef struct POOL_ctx_s POOL_ctx;
|
||||||
|
|
||||||
|
/*! POOL_create() :
|
||||||
|
* Create a thread pool with at most `numThreads` threads.
|
||||||
|
* `numThreads` must be at least 1.
|
||||||
|
* The maximum number of queued jobs before blocking is `queueSize`.
|
||||||
|
* @return : POOL_ctx pointer on success, else NULL.
|
||||||
|
*/
|
||||||
|
POOL_ctx* POOL_create(size_t numThreads, size_t queueSize);
|
||||||
|
|
||||||
|
POOL_ctx* POOL_create_advanced(size_t numThreads, size_t queueSize,
|
||||||
|
ZSTD_customMem customMem);
|
||||||
|
|
||||||
|
/*! POOL_free() :
|
||||||
|
* Free a thread pool returned by POOL_create().
|
||||||
|
*/
|
||||||
|
void POOL_free(POOL_ctx* ctx);
|
||||||
|
|
||||||
|
/*! POOL_resize() :
|
||||||
|
* Expands or shrinks pool's number of threads.
|
||||||
|
* This is more efficient than releasing + creating a new context,
|
||||||
|
* since it tries to preserve and re-use existing threads.
|
||||||
|
* `numThreads` must be at least 1.
|
||||||
|
* @return : 0 when resize was successful,
|
||||||
|
* !0 (typically 1) if there is an error.
|
||||||
|
* note : only numThreads can be resized, queueSize remains unchanged.
|
||||||
|
*/
|
||||||
|
int POOL_resize(POOL_ctx* ctx, size_t numThreads);
|
||||||
|
|
||||||
|
/*! POOL_sizeof() :
|
||||||
|
* @return threadpool memory usage
|
||||||
|
* note : compatible with NULL (returns 0 in this case)
|
||||||
|
*/
|
||||||
|
size_t POOL_sizeof(POOL_ctx* ctx);
|
||||||
|
|
||||||
|
/*! POOL_function :
|
||||||
|
* The function type that can be added to a thread pool.
|
||||||
|
*/
|
||||||
|
typedef void (*POOL_function)(void*);
|
||||||
|
|
||||||
|
/*! POOL_add() :
|
||||||
|
* Add the job `function(opaque)` to the thread pool. `ctx` must be valid.
|
||||||
|
* Possibly blocks until there is room in the queue.
|
||||||
|
* Note : The function may be executed asynchronously,
|
||||||
|
* therefore, `opaque` must live until function has been completed.
|
||||||
|
*/
|
||||||
|
void POOL_add(POOL_ctx* ctx, POOL_function function, void* opaque);
|
||||||
|
|
||||||
|
|
||||||
|
/*! POOL_tryAdd() :
|
||||||
|
* Add the job `function(opaque)` to thread pool _if_ a worker is available.
|
||||||
|
* Returns immediately even if not (does not block).
|
||||||
|
* @return : 1 if successful, 0 if not.
|
||||||
|
*/
|
||||||
|
int POOL_tryAdd(POOL_ctx* ctx, POOL_function function, void* opaque);
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif
|
||||||
120
vendor/github.com/DataDog/zstd/threading.c
generated
vendored
Normal file
120
vendor/github.com/DataDog/zstd/threading.c
generated
vendored
Normal file
@@ -0,0 +1,120 @@
|
|||||||
|
/**
|
||||||
|
* Copyright (c) 2016 Tino Reichardt
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
*
|
||||||
|
* You can contact the author at:
|
||||||
|
* - zstdmt source repository: https://github.com/mcmilk/zstdmt
|
||||||
|
*/
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This file will hold wrapper for systems, which do not support pthreads
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include "threading.h"
|
||||||
|
|
||||||
|
/* create fake symbol to avoid empty translation unit warning */
|
||||||
|
int g_ZSTD_threading_useless_symbol;
|
||||||
|
|
||||||
|
#if defined(ZSTD_MULTITHREAD) && defined(_WIN32)
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Windows minimalist Pthread Wrapper, based on :
|
||||||
|
* http://www.cse.wustl.edu/~schmidt/win32-cv-1.html
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/* === Dependencies === */
|
||||||
|
#include <process.h>
|
||||||
|
#include <errno.h>
|
||||||
|
|
||||||
|
|
||||||
|
/* === Implementation === */
|
||||||
|
|
||||||
|
static unsigned __stdcall worker(void *arg)
|
||||||
|
{
|
||||||
|
ZSTD_pthread_t* const thread = (ZSTD_pthread_t*) arg;
|
||||||
|
thread->arg = thread->start_routine(thread->arg);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
int ZSTD_pthread_create(ZSTD_pthread_t* thread, const void* unused,
|
||||||
|
void* (*start_routine) (void*), void* arg)
|
||||||
|
{
|
||||||
|
(void)unused;
|
||||||
|
thread->arg = arg;
|
||||||
|
thread->start_routine = start_routine;
|
||||||
|
thread->handle = (HANDLE) _beginthreadex(NULL, 0, worker, thread, 0, NULL);
|
||||||
|
|
||||||
|
if (!thread->handle)
|
||||||
|
return errno;
|
||||||
|
else
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
int ZSTD_pthread_join(ZSTD_pthread_t thread, void **value_ptr)
|
||||||
|
{
|
||||||
|
DWORD result;
|
||||||
|
|
||||||
|
if (!thread.handle) return 0;
|
||||||
|
|
||||||
|
result = WaitForSingleObject(thread.handle, INFINITE);
|
||||||
|
switch (result) {
|
||||||
|
case WAIT_OBJECT_0:
|
||||||
|
if (value_ptr) *value_ptr = thread.arg;
|
||||||
|
return 0;
|
||||||
|
case WAIT_ABANDONED:
|
||||||
|
return EINVAL;
|
||||||
|
default:
|
||||||
|
return GetLastError();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif /* ZSTD_MULTITHREAD */
|
||||||
|
|
||||||
|
#if defined(ZSTD_MULTITHREAD) && DEBUGLEVEL >= 1 && !defined(_WIN32)
|
||||||
|
|
||||||
|
#include <stdlib.h>
|
||||||
|
|
||||||
|
int ZSTD_pthread_mutex_init(ZSTD_pthread_mutex_t* mutex, pthread_mutexattr_t const* attr)
|
||||||
|
{
|
||||||
|
*mutex = (pthread_mutex_t*)malloc(sizeof(pthread_mutex_t));
|
||||||
|
if (!*mutex)
|
||||||
|
return 1;
|
||||||
|
return pthread_mutex_init(*mutex, attr);
|
||||||
|
}
|
||||||
|
|
||||||
|
int ZSTD_pthread_mutex_destroy(ZSTD_pthread_mutex_t* mutex)
|
||||||
|
{
|
||||||
|
if (!*mutex)
|
||||||
|
return 0;
|
||||||
|
{
|
||||||
|
int const ret = pthread_mutex_destroy(*mutex);
|
||||||
|
free(*mutex);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
int ZSTD_pthread_cond_init(ZSTD_pthread_cond_t* cond, pthread_condattr_t const* attr)
|
||||||
|
{
|
||||||
|
*cond = (pthread_cond_t*)malloc(sizeof(pthread_cond_t));
|
||||||
|
if (!*cond)
|
||||||
|
return 1;
|
||||||
|
return pthread_cond_init(*cond, attr);
|
||||||
|
}
|
||||||
|
|
||||||
|
int ZSTD_pthread_cond_destroy(ZSTD_pthread_cond_t* cond)
|
||||||
|
{
|
||||||
|
if (!*cond)
|
||||||
|
return 0;
|
||||||
|
{
|
||||||
|
int const ret = pthread_cond_destroy(*cond);
|
||||||
|
free(*cond);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif
|
||||||
154
vendor/github.com/DataDog/zstd/threading.h
generated
vendored
Normal file
154
vendor/github.com/DataDog/zstd/threading.h
generated
vendored
Normal file
@@ -0,0 +1,154 @@
|
|||||||
|
/**
|
||||||
|
* Copyright (c) 2016 Tino Reichardt
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
*
|
||||||
|
* You can contact the author at:
|
||||||
|
* - zstdmt source repository: https://github.com/mcmilk/zstdmt
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef THREADING_H_938743
|
||||||
|
#define THREADING_H_938743
|
||||||
|
|
||||||
|
#include "debug.h"
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#if defined(ZSTD_MULTITHREAD) && defined(_WIN32)
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Windows minimalist Pthread Wrapper, based on :
|
||||||
|
* http://www.cse.wustl.edu/~schmidt/win32-cv-1.html
|
||||||
|
*/
|
||||||
|
#ifdef WINVER
|
||||||
|
# undef WINVER
|
||||||
|
#endif
|
||||||
|
#define WINVER 0x0600
|
||||||
|
|
||||||
|
#ifdef _WIN32_WINNT
|
||||||
|
# undef _WIN32_WINNT
|
||||||
|
#endif
|
||||||
|
#define _WIN32_WINNT 0x0600
|
||||||
|
|
||||||
|
#ifndef WIN32_LEAN_AND_MEAN
|
||||||
|
# define WIN32_LEAN_AND_MEAN
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#undef ERROR /* reported already defined on VS 2015 (Rich Geldreich) */
|
||||||
|
#include <windows.h>
|
||||||
|
#undef ERROR
|
||||||
|
#define ERROR(name) ZSTD_ERROR(name)
|
||||||
|
|
||||||
|
|
||||||
|
/* mutex */
|
||||||
|
#define ZSTD_pthread_mutex_t CRITICAL_SECTION
|
||||||
|
#define ZSTD_pthread_mutex_init(a, b) ((void)(b), InitializeCriticalSection((a)), 0)
|
||||||
|
#define ZSTD_pthread_mutex_destroy(a) DeleteCriticalSection((a))
|
||||||
|
#define ZSTD_pthread_mutex_lock(a) EnterCriticalSection((a))
|
||||||
|
#define ZSTD_pthread_mutex_unlock(a) LeaveCriticalSection((a))
|
||||||
|
|
||||||
|
/* condition variable */
|
||||||
|
#define ZSTD_pthread_cond_t CONDITION_VARIABLE
|
||||||
|
#define ZSTD_pthread_cond_init(a, b) ((void)(b), InitializeConditionVariable((a)), 0)
|
||||||
|
#define ZSTD_pthread_cond_destroy(a) ((void)(a))
|
||||||
|
#define ZSTD_pthread_cond_wait(a, b) SleepConditionVariableCS((a), (b), INFINITE)
|
||||||
|
#define ZSTD_pthread_cond_signal(a) WakeConditionVariable((a))
|
||||||
|
#define ZSTD_pthread_cond_broadcast(a) WakeAllConditionVariable((a))
|
||||||
|
|
||||||
|
/* ZSTD_pthread_create() and ZSTD_pthread_join() */
|
||||||
|
typedef struct {
|
||||||
|
HANDLE handle;
|
||||||
|
void* (*start_routine)(void*);
|
||||||
|
void* arg;
|
||||||
|
} ZSTD_pthread_t;
|
||||||
|
|
||||||
|
int ZSTD_pthread_create(ZSTD_pthread_t* thread, const void* unused,
|
||||||
|
void* (*start_routine) (void*), void* arg);
|
||||||
|
|
||||||
|
int ZSTD_pthread_join(ZSTD_pthread_t thread, void** value_ptr);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* add here more wrappers as required
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
#elif defined(ZSTD_MULTITHREAD) /* posix assumed ; need a better detection method */
|
||||||
|
/* === POSIX Systems === */
|
||||||
|
# include <pthread.h>
|
||||||
|
|
||||||
|
#if DEBUGLEVEL < 1
|
||||||
|
|
||||||
|
#define ZSTD_pthread_mutex_t pthread_mutex_t
|
||||||
|
#define ZSTD_pthread_mutex_init(a, b) pthread_mutex_init((a), (b))
|
||||||
|
#define ZSTD_pthread_mutex_destroy(a) pthread_mutex_destroy((a))
|
||||||
|
#define ZSTD_pthread_mutex_lock(a) pthread_mutex_lock((a))
|
||||||
|
#define ZSTD_pthread_mutex_unlock(a) pthread_mutex_unlock((a))
|
||||||
|
|
||||||
|
#define ZSTD_pthread_cond_t pthread_cond_t
|
||||||
|
#define ZSTD_pthread_cond_init(a, b) pthread_cond_init((a), (b))
|
||||||
|
#define ZSTD_pthread_cond_destroy(a) pthread_cond_destroy((a))
|
||||||
|
#define ZSTD_pthread_cond_wait(a, b) pthread_cond_wait((a), (b))
|
||||||
|
#define ZSTD_pthread_cond_signal(a) pthread_cond_signal((a))
|
||||||
|
#define ZSTD_pthread_cond_broadcast(a) pthread_cond_broadcast((a))
|
||||||
|
|
||||||
|
#define ZSTD_pthread_t pthread_t
|
||||||
|
#define ZSTD_pthread_create(a, b, c, d) pthread_create((a), (b), (c), (d))
|
||||||
|
#define ZSTD_pthread_join(a, b) pthread_join((a),(b))
|
||||||
|
|
||||||
|
#else /* DEBUGLEVEL >= 1 */
|
||||||
|
|
||||||
|
/* Debug implementation of threading.
|
||||||
|
* In this implementation we use pointers for mutexes and condition variables.
|
||||||
|
* This way, if we forget to init/destroy them the program will crash or ASAN
|
||||||
|
* will report leaks.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#define ZSTD_pthread_mutex_t pthread_mutex_t*
|
||||||
|
int ZSTD_pthread_mutex_init(ZSTD_pthread_mutex_t* mutex, pthread_mutexattr_t const* attr);
|
||||||
|
int ZSTD_pthread_mutex_destroy(ZSTD_pthread_mutex_t* mutex);
|
||||||
|
#define ZSTD_pthread_mutex_lock(a) pthread_mutex_lock(*(a))
|
||||||
|
#define ZSTD_pthread_mutex_unlock(a) pthread_mutex_unlock(*(a))
|
||||||
|
|
||||||
|
#define ZSTD_pthread_cond_t pthread_cond_t*
|
||||||
|
int ZSTD_pthread_cond_init(ZSTD_pthread_cond_t* cond, pthread_condattr_t const* attr);
|
||||||
|
int ZSTD_pthread_cond_destroy(ZSTD_pthread_cond_t* cond);
|
||||||
|
#define ZSTD_pthread_cond_wait(a, b) pthread_cond_wait(*(a), *(b))
|
||||||
|
#define ZSTD_pthread_cond_signal(a) pthread_cond_signal(*(a))
|
||||||
|
#define ZSTD_pthread_cond_broadcast(a) pthread_cond_broadcast(*(a))
|
||||||
|
|
||||||
|
#define ZSTD_pthread_t pthread_t
|
||||||
|
#define ZSTD_pthread_create(a, b, c, d) pthread_create((a), (b), (c), (d))
|
||||||
|
#define ZSTD_pthread_join(a, b) pthread_join((a),(b))
|
||||||
|
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#else /* ZSTD_MULTITHREAD not defined */
|
||||||
|
/* No multithreading support */
|
||||||
|
|
||||||
|
typedef int ZSTD_pthread_mutex_t;
|
||||||
|
#define ZSTD_pthread_mutex_init(a, b) ((void)(a), (void)(b), 0)
|
||||||
|
#define ZSTD_pthread_mutex_destroy(a) ((void)(a))
|
||||||
|
#define ZSTD_pthread_mutex_lock(a) ((void)(a))
|
||||||
|
#define ZSTD_pthread_mutex_unlock(a) ((void)(a))
|
||||||
|
|
||||||
|
typedef int ZSTD_pthread_cond_t;
|
||||||
|
#define ZSTD_pthread_cond_init(a, b) ((void)(a), (void)(b), 0)
|
||||||
|
#define ZSTD_pthread_cond_destroy(a) ((void)(a))
|
||||||
|
#define ZSTD_pthread_cond_wait(a, b) ((void)(a), (void)(b))
|
||||||
|
#define ZSTD_pthread_cond_signal(a) ((void)(a))
|
||||||
|
#define ZSTD_pthread_cond_broadcast(a) ((void)(a))
|
||||||
|
|
||||||
|
/* do not use ZSTD_pthread_t */
|
||||||
|
|
||||||
|
#endif /* ZSTD_MULTITHREAD */
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* THREADING_H_938743 */
|
||||||
17
vendor/github.com/DataDog/zstd/travis_test_32.sh
generated
vendored
Normal file
17
vendor/github.com/DataDog/zstd/travis_test_32.sh
generated
vendored
Normal file
@@ -0,0 +1,17 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# Get utilities
|
||||||
|
yum -y -q -e 0 install wget tar unzip gcc
|
||||||
|
|
||||||
|
# Get Go
|
||||||
|
wget -q https://dl.google.com/go/go1.13.linux-386.tar.gz
|
||||||
|
tar -C /usr/local -xzf go1.13.linux-386.tar.gz
|
||||||
|
export PATH=$PATH:/usr/local/go/bin
|
||||||
|
|
||||||
|
# Get payload
|
||||||
|
wget -q https://github.com/DataDog/zstd/files/2246767/mr.zip
|
||||||
|
unzip mr.zip
|
||||||
|
|
||||||
|
# Build and run tests
|
||||||
|
go build
|
||||||
|
PAYLOAD=$(pwd)/mr go test -v
|
||||||
|
PAYLOAD=$(pwd)/mr go test -bench .
|
||||||
56
vendor/github.com/DataDog/zstd/update.txt
generated
vendored
Normal file
56
vendor/github.com/DataDog/zstd/update.txt
generated
vendored
Normal file
@@ -0,0 +1,56 @@
|
|||||||
|
./lib/common/bitstream.h
|
||||||
|
./lib/common/compiler.h
|
||||||
|
./lib/compress/zstd_compress_internal.h
|
||||||
|
./lib/compress/zstd_fast.h
|
||||||
|
./lib/compress/zstd_double_fast.h
|
||||||
|
./lib/compress/zstd_lazy.h
|
||||||
|
./lib/compress/zstd_ldm.h
|
||||||
|
./lib/dictBuilder/cover.c
|
||||||
|
./lib/dictBuilder/divsufsort.c
|
||||||
|
./lib/dictBuilder/divsufsort.h
|
||||||
|
./lib/common/entropy_common.c
|
||||||
|
./lib/common/error_private.c
|
||||||
|
./lib/common/error_private.h
|
||||||
|
./lib/compress/fse_compress.c
|
||||||
|
./lib/common/fse_decompress.c
|
||||||
|
./lib/common/fse.h
|
||||||
|
./lib/compress/huf_compress.c
|
||||||
|
./lib/decompress/huf_decompress.c
|
||||||
|
./lib/common/huf.h
|
||||||
|
./lib/common/mem.h
|
||||||
|
./lib/common/pool.c
|
||||||
|
./lib/common/pool.h
|
||||||
|
./lib/common/threading.c
|
||||||
|
./lib/common/threading.h
|
||||||
|
./lib/common/xxhash.c
|
||||||
|
./lib/common/xxhash.h
|
||||||
|
./lib/deprecated/zbuff_common.c
|
||||||
|
./lib/deprecated/zbuff_compress.c
|
||||||
|
./lib/deprecated/zbuff_decompress.c
|
||||||
|
./lib/deprecated/zbuff.h
|
||||||
|
./lib/dictBuilder/zdict.c
|
||||||
|
./lib/dictBuilder/zdict.h
|
||||||
|
./lib/common/zstd_common.c
|
||||||
|
./lib/compress/zstd_compress.c
|
||||||
|
./lib/decompress/zstd_decompress.c
|
||||||
|
./lib/common/zstd_errors.h
|
||||||
|
./lib/zstd.h
|
||||||
|
./lib/common/zstd_internal.h
|
||||||
|
./lib/legacy/zstd_legacy.h
|
||||||
|
./lib/compress/zstd_opt.c
|
||||||
|
./lib/compress/zstd_opt.h
|
||||||
|
./lib/legacy/zstd_v01.c
|
||||||
|
./lib/legacy/zstd_v01.h
|
||||||
|
./lib/legacy/zstd_v02.c
|
||||||
|
./lib/legacy/zstd_v02.h
|
||||||
|
./lib/legacy/zstd_v03.c
|
||||||
|
./lib/legacy/zstd_v03.h
|
||||||
|
./lib/legacy/zstd_v04.c
|
||||||
|
./lib/legacy/zstd_v04.h
|
||||||
|
./lib/legacy/zstd_v05.c
|
||||||
|
./lib/legacy/zstd_v05.h
|
||||||
|
./lib/legacy/zstd_v06.c
|
||||||
|
./lib/legacy/zstd_v06.h
|
||||||
|
./lib/legacy/zstd_v07.c
|
||||||
|
./lib/legacy/zstd_v07.h
|
||||||
|
|
||||||
882
vendor/github.com/DataDog/zstd/xxhash.c
generated
vendored
Normal file
882
vendor/github.com/DataDog/zstd/xxhash.c
generated
vendored
Normal file
@@ -0,0 +1,882 @@
|
|||||||
|
/*
|
||||||
|
* xxHash - Fast Hash algorithm
|
||||||
|
* Copyright (C) 2012-2016, Yann Collet
|
||||||
|
*
|
||||||
|
* BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
*
|
||||||
|
* Redistribution and use in source and binary forms, with or without
|
||||||
|
* modification, are permitted provided that the following conditions are
|
||||||
|
* met:
|
||||||
|
*
|
||||||
|
* * Redistributions of source code must retain the above copyright
|
||||||
|
* notice, this list of conditions and the following disclaimer.
|
||||||
|
* * Redistributions in binary form must reproduce the above
|
||||||
|
* copyright notice, this list of conditions and the following disclaimer
|
||||||
|
* in the documentation and/or other materials provided with the
|
||||||
|
* distribution.
|
||||||
|
*
|
||||||
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
*
|
||||||
|
* You can contact the author at :
|
||||||
|
* - xxHash homepage: http://www.xxhash.com
|
||||||
|
* - xxHash source repository : https://github.com/Cyan4973/xxHash
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Tuning parameters
|
||||||
|
***************************************/
|
||||||
|
/*!XXH_FORCE_MEMORY_ACCESS :
|
||||||
|
* By default, access to unaligned memory is controlled by `memcpy()`, which is safe and portable.
|
||||||
|
* Unfortunately, on some target/compiler combinations, the generated assembly is sub-optimal.
|
||||||
|
* The below switch allow to select different access method for improved performance.
|
||||||
|
* Method 0 (default) : use `memcpy()`. Safe and portable.
|
||||||
|
* Method 1 : `__packed` statement. It depends on compiler extension (ie, not portable).
|
||||||
|
* This method is safe if your compiler supports it, and *generally* as fast or faster than `memcpy`.
|
||||||
|
* Method 2 : direct access. This method doesn't depend on compiler but violate C standard.
|
||||||
|
* It can generate buggy code on targets which do not support unaligned memory accesses.
|
||||||
|
* But in some circumstances, it's the only known way to get the most performance (ie GCC + ARMv6)
|
||||||
|
* See http://stackoverflow.com/a/32095106/646947 for details.
|
||||||
|
* Prefer these methods in priority order (0 > 1 > 2)
|
||||||
|
*/
|
||||||
|
#ifndef XXH_FORCE_MEMORY_ACCESS /* can be defined externally, on command line for example */
|
||||||
|
# if defined(__GNUC__) && ( defined(__ARM_ARCH_6__) || defined(__ARM_ARCH_6J__) || defined(__ARM_ARCH_6K__) || defined(__ARM_ARCH_6Z__) || defined(__ARM_ARCH_6ZK__) || defined(__ARM_ARCH_6T2__) )
|
||||||
|
# define XXH_FORCE_MEMORY_ACCESS 2
|
||||||
|
# elif (defined(__INTEL_COMPILER) && !defined(WIN32)) || \
|
||||||
|
(defined(__GNUC__) && ( defined(__ARM_ARCH_7__) || defined(__ARM_ARCH_7A__) || defined(__ARM_ARCH_7R__) || defined(__ARM_ARCH_7M__) || defined(__ARM_ARCH_7S__) )) || \
|
||||||
|
defined(__ICCARM__)
|
||||||
|
# define XXH_FORCE_MEMORY_ACCESS 1
|
||||||
|
# endif
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*!XXH_ACCEPT_NULL_INPUT_POINTER :
|
||||||
|
* If the input pointer is a null pointer, xxHash default behavior is to trigger a memory access error, since it is a bad pointer.
|
||||||
|
* When this option is enabled, xxHash output for null input pointers will be the same as a null-length input.
|
||||||
|
* By default, this option is disabled. To enable it, uncomment below define :
|
||||||
|
*/
|
||||||
|
/* #define XXH_ACCEPT_NULL_INPUT_POINTER 1 */
|
||||||
|
|
||||||
|
/*!XXH_FORCE_NATIVE_FORMAT :
|
||||||
|
* By default, xxHash library provides endian-independent Hash values, based on little-endian convention.
|
||||||
|
* Results are therefore identical for little-endian and big-endian CPU.
|
||||||
|
* This comes at a performance cost for big-endian CPU, since some swapping is required to emulate little-endian format.
|
||||||
|
* Should endian-independence be of no importance for your application, you may set the #define below to 1,
|
||||||
|
* to improve speed for Big-endian CPU.
|
||||||
|
* This option has no impact on Little_Endian CPU.
|
||||||
|
*/
|
||||||
|
#ifndef XXH_FORCE_NATIVE_FORMAT /* can be defined externally */
|
||||||
|
# define XXH_FORCE_NATIVE_FORMAT 0
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*!XXH_FORCE_ALIGN_CHECK :
|
||||||
|
* This is a minor performance trick, only useful with lots of very small keys.
|
||||||
|
* It means : check for aligned/unaligned input.
|
||||||
|
* The check costs one initial branch per hash; set to 0 when the input data
|
||||||
|
* is guaranteed to be aligned.
|
||||||
|
*/
|
||||||
|
#ifndef XXH_FORCE_ALIGN_CHECK /* can be defined externally */
|
||||||
|
# if defined(__i386) || defined(_M_IX86) || defined(__x86_64__) || defined(_M_X64)
|
||||||
|
# define XXH_FORCE_ALIGN_CHECK 0
|
||||||
|
# else
|
||||||
|
# define XXH_FORCE_ALIGN_CHECK 1
|
||||||
|
# endif
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Includes & Memory related functions
|
||||||
|
***************************************/
|
||||||
|
/* Modify the local functions below should you wish to use some other memory routines */
|
||||||
|
/* for malloc(), free() */
|
||||||
|
#include <stdlib.h>
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
static void* XXH_malloc(size_t s) { return malloc(s); }
|
||||||
|
static void XXH_free (void* p) { free(p); }
|
||||||
|
/* for memcpy() */
|
||||||
|
#include <string.h>
|
||||||
|
static void* XXH_memcpy(void* dest, const void* src, size_t size) { return memcpy(dest,src,size); }
|
||||||
|
|
||||||
|
#ifndef XXH_STATIC_LINKING_ONLY
|
||||||
|
# define XXH_STATIC_LINKING_ONLY
|
||||||
|
#endif
|
||||||
|
#include "xxhash.h"
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Compiler Specific Options
|
||||||
|
***************************************/
|
||||||
|
#if defined (__GNUC__) || defined(__cplusplus) || defined(__STDC_VERSION__) && __STDC_VERSION__ >= 199901L /* C99 */
|
||||||
|
# define INLINE_KEYWORD inline
|
||||||
|
#else
|
||||||
|
# define INLINE_KEYWORD
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#if defined(__GNUC__) || defined(__ICCARM__)
|
||||||
|
# define FORCE_INLINE_ATTR __attribute__((always_inline))
|
||||||
|
#elif defined(_MSC_VER)
|
||||||
|
# define FORCE_INLINE_ATTR __forceinline
|
||||||
|
#else
|
||||||
|
# define FORCE_INLINE_ATTR
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#define FORCE_INLINE_TEMPLATE static INLINE_KEYWORD FORCE_INLINE_ATTR
|
||||||
|
|
||||||
|
|
||||||
|
#ifdef _MSC_VER
|
||||||
|
# pragma warning(disable : 4127) /* disable: C4127: conditional expression is constant */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Basic Types
|
||||||
|
***************************************/
|
||||||
|
#ifndef MEM_MODULE
|
||||||
|
# define MEM_MODULE
|
||||||
|
# if !defined (__VMS) && (defined (__cplusplus) || (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L) /* C99 */) )
|
||||||
|
# include <stdint.h>
|
||||||
|
typedef uint8_t BYTE;
|
||||||
|
typedef uint16_t U16;
|
||||||
|
typedef uint32_t U32;
|
||||||
|
typedef int32_t S32;
|
||||||
|
typedef uint64_t U64;
|
||||||
|
# else
|
||||||
|
typedef unsigned char BYTE;
|
||||||
|
typedef unsigned short U16;
|
||||||
|
typedef unsigned int U32;
|
||||||
|
typedef signed int S32;
|
||||||
|
typedef unsigned long long U64; /* if your compiler doesn't support unsigned long long, replace by another 64-bit type here. Note that xxhash.h will also need to be updated. */
|
||||||
|
# endif
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
#if (defined(XXH_FORCE_MEMORY_ACCESS) && (XXH_FORCE_MEMORY_ACCESS==2))
|
||||||
|
|
||||||
|
/* Force direct memory access. Only works on CPU which support unaligned memory access in hardware */
|
||||||
|
static U32 XXH_read32(const void* memPtr) { return *(const U32*) memPtr; }
|
||||||
|
static U64 XXH_read64(const void* memPtr) { return *(const U64*) memPtr; }
|
||||||
|
|
||||||
|
#elif (defined(XXH_FORCE_MEMORY_ACCESS) && (XXH_FORCE_MEMORY_ACCESS==1))
|
||||||
|
|
||||||
|
/* __pack instructions are safer, but compiler specific, hence potentially problematic for some compilers */
|
||||||
|
/* currently only defined for gcc and icc */
|
||||||
|
typedef union { U32 u32; U64 u64; } __attribute__((packed)) unalign;
|
||||||
|
|
||||||
|
static U32 XXH_read32(const void* ptr) { return ((const unalign*)ptr)->u32; }
|
||||||
|
static U64 XXH_read64(const void* ptr) { return ((const unalign*)ptr)->u64; }
|
||||||
|
|
||||||
|
#else
|
||||||
|
|
||||||
|
/* portable and safe solution. Generally efficient.
|
||||||
|
* see : http://stackoverflow.com/a/32095106/646947
|
||||||
|
*/
|
||||||
|
|
||||||
|
static U32 XXH_read32(const void* memPtr)
|
||||||
|
{
|
||||||
|
U32 val;
|
||||||
|
memcpy(&val, memPtr, sizeof(val));
|
||||||
|
return val;
|
||||||
|
}
|
||||||
|
|
||||||
|
static U64 XXH_read64(const void* memPtr)
|
||||||
|
{
|
||||||
|
U64 val;
|
||||||
|
memcpy(&val, memPtr, sizeof(val));
|
||||||
|
return val;
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif /* XXH_FORCE_DIRECT_MEMORY_ACCESS */
|
||||||
|
|
||||||
|
|
||||||
|
/* ****************************************
|
||||||
|
* Compiler-specific Functions and Macros
|
||||||
|
******************************************/
|
||||||
|
#define GCC_VERSION (__GNUC__ * 100 + __GNUC_MINOR__)
|
||||||
|
|
||||||
|
/* Note : although _rotl exists for minGW (GCC under windows), performance seems poor */
|
||||||
|
#if defined(_MSC_VER)
|
||||||
|
# define XXH_rotl32(x,r) _rotl(x,r)
|
||||||
|
# define XXH_rotl64(x,r) _rotl64(x,r)
|
||||||
|
#else
|
||||||
|
#if defined(__ICCARM__)
|
||||||
|
# include <intrinsics.h>
|
||||||
|
# define XXH_rotl32(x,r) __ROR(x,(32 - r))
|
||||||
|
#else
|
||||||
|
# define XXH_rotl32(x,r) ((x << r) | (x >> (32 - r)))
|
||||||
|
#endif
|
||||||
|
# define XXH_rotl64(x,r) ((x << r) | (x >> (64 - r)))
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#if defined(_MSC_VER) /* Visual Studio */
|
||||||
|
# define XXH_swap32 _byteswap_ulong
|
||||||
|
# define XXH_swap64 _byteswap_uint64
|
||||||
|
#elif GCC_VERSION >= 403
|
||||||
|
# define XXH_swap32 __builtin_bswap32
|
||||||
|
# define XXH_swap64 __builtin_bswap64
|
||||||
|
#else
|
||||||
|
static U32 XXH_swap32 (U32 x)
|
||||||
|
{
|
||||||
|
return ((x << 24) & 0xff000000 ) |
|
||||||
|
((x << 8) & 0x00ff0000 ) |
|
||||||
|
((x >> 8) & 0x0000ff00 ) |
|
||||||
|
((x >> 24) & 0x000000ff );
|
||||||
|
}
|
||||||
|
static U64 XXH_swap64 (U64 x)
|
||||||
|
{
|
||||||
|
return ((x << 56) & 0xff00000000000000ULL) |
|
||||||
|
((x << 40) & 0x00ff000000000000ULL) |
|
||||||
|
((x << 24) & 0x0000ff0000000000ULL) |
|
||||||
|
((x << 8) & 0x000000ff00000000ULL) |
|
||||||
|
((x >> 8) & 0x00000000ff000000ULL) |
|
||||||
|
((x >> 24) & 0x0000000000ff0000ULL) |
|
||||||
|
((x >> 40) & 0x000000000000ff00ULL) |
|
||||||
|
((x >> 56) & 0x00000000000000ffULL);
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Architecture Macros
|
||||||
|
***************************************/
|
||||||
|
typedef enum { XXH_bigEndian=0, XXH_littleEndian=1 } XXH_endianess;
|
||||||
|
|
||||||
|
/* XXH_CPU_LITTLE_ENDIAN can be defined externally, for example on the compiler command line */
|
||||||
|
#ifndef XXH_CPU_LITTLE_ENDIAN
|
||||||
|
static const int g_one = 1;
|
||||||
|
# define XXH_CPU_LITTLE_ENDIAN (*(const char*)(&g_one))
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* ***************************
|
||||||
|
* Memory reads
|
||||||
|
*****************************/
|
||||||
|
typedef enum { XXH_aligned, XXH_unaligned } XXH_alignment;
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE U32 XXH_readLE32_align(const void* ptr, XXH_endianess endian, XXH_alignment align)
|
||||||
|
{
|
||||||
|
if (align==XXH_unaligned)
|
||||||
|
return endian==XXH_littleEndian ? XXH_read32(ptr) : XXH_swap32(XXH_read32(ptr));
|
||||||
|
else
|
||||||
|
return endian==XXH_littleEndian ? *(const U32*)ptr : XXH_swap32(*(const U32*)ptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE U32 XXH_readLE32(const void* ptr, XXH_endianess endian)
|
||||||
|
{
|
||||||
|
return XXH_readLE32_align(ptr, endian, XXH_unaligned);
|
||||||
|
}
|
||||||
|
|
||||||
|
static U32 XXH_readBE32(const void* ptr)
|
||||||
|
{
|
||||||
|
return XXH_CPU_LITTLE_ENDIAN ? XXH_swap32(XXH_read32(ptr)) : XXH_read32(ptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE U64 XXH_readLE64_align(const void* ptr, XXH_endianess endian, XXH_alignment align)
|
||||||
|
{
|
||||||
|
if (align==XXH_unaligned)
|
||||||
|
return endian==XXH_littleEndian ? XXH_read64(ptr) : XXH_swap64(XXH_read64(ptr));
|
||||||
|
else
|
||||||
|
return endian==XXH_littleEndian ? *(const U64*)ptr : XXH_swap64(*(const U64*)ptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE U64 XXH_readLE64(const void* ptr, XXH_endianess endian)
|
||||||
|
{
|
||||||
|
return XXH_readLE64_align(ptr, endian, XXH_unaligned);
|
||||||
|
}
|
||||||
|
|
||||||
|
static U64 XXH_readBE64(const void* ptr)
|
||||||
|
{
|
||||||
|
return XXH_CPU_LITTLE_ENDIAN ? XXH_swap64(XXH_read64(ptr)) : XXH_read64(ptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Macros
|
||||||
|
***************************************/
|
||||||
|
#define XXH_STATIC_ASSERT(c) { enum { XXH_static_assert = 1/(int)(!!(c)) }; } /* use only *after* variable declarations */
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Constants
|
||||||
|
***************************************/
|
||||||
|
static const U32 PRIME32_1 = 2654435761U;
|
||||||
|
static const U32 PRIME32_2 = 2246822519U;
|
||||||
|
static const U32 PRIME32_3 = 3266489917U;
|
||||||
|
static const U32 PRIME32_4 = 668265263U;
|
||||||
|
static const U32 PRIME32_5 = 374761393U;
|
||||||
|
|
||||||
|
static const U64 PRIME64_1 = 11400714785074694791ULL;
|
||||||
|
static const U64 PRIME64_2 = 14029467366897019727ULL;
|
||||||
|
static const U64 PRIME64_3 = 1609587929392839161ULL;
|
||||||
|
static const U64 PRIME64_4 = 9650029242287828579ULL;
|
||||||
|
static const U64 PRIME64_5 = 2870177450012600261ULL;
|
||||||
|
|
||||||
|
XXH_PUBLIC_API unsigned XXH_versionNumber (void) { return XXH_VERSION_NUMBER; }
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************
|
||||||
|
* Utils
|
||||||
|
****************************/
|
||||||
|
XXH_PUBLIC_API void XXH32_copyState(XXH32_state_t* restrict dstState, const XXH32_state_t* restrict srcState)
|
||||||
|
{
|
||||||
|
memcpy(dstState, srcState, sizeof(*dstState));
|
||||||
|
}
|
||||||
|
|
||||||
|
XXH_PUBLIC_API void XXH64_copyState(XXH64_state_t* restrict dstState, const XXH64_state_t* restrict srcState)
|
||||||
|
{
|
||||||
|
memcpy(dstState, srcState, sizeof(*dstState));
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* ***************************
|
||||||
|
* Simple Hash Functions
|
||||||
|
*****************************/
|
||||||
|
|
||||||
|
static U32 XXH32_round(U32 seed, U32 input)
|
||||||
|
{
|
||||||
|
seed += input * PRIME32_2;
|
||||||
|
seed = XXH_rotl32(seed, 13);
|
||||||
|
seed *= PRIME32_1;
|
||||||
|
return seed;
|
||||||
|
}
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE U32 XXH32_endian_align(const void* input, size_t len, U32 seed, XXH_endianess endian, XXH_alignment align)
|
||||||
|
{
|
||||||
|
const BYTE* p = (const BYTE*)input;
|
||||||
|
const BYTE* bEnd = p + len;
|
||||||
|
U32 h32;
|
||||||
|
#define XXH_get32bits(p) XXH_readLE32_align(p, endian, align)
|
||||||
|
|
||||||
|
#ifdef XXH_ACCEPT_NULL_INPUT_POINTER
|
||||||
|
if (p==NULL) {
|
||||||
|
len=0;
|
||||||
|
bEnd=p=(const BYTE*)(size_t)16;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
if (len>=16) {
|
||||||
|
const BYTE* const limit = bEnd - 16;
|
||||||
|
U32 v1 = seed + PRIME32_1 + PRIME32_2;
|
||||||
|
U32 v2 = seed + PRIME32_2;
|
||||||
|
U32 v3 = seed + 0;
|
||||||
|
U32 v4 = seed - PRIME32_1;
|
||||||
|
|
||||||
|
do {
|
||||||
|
v1 = XXH32_round(v1, XXH_get32bits(p)); p+=4;
|
||||||
|
v2 = XXH32_round(v2, XXH_get32bits(p)); p+=4;
|
||||||
|
v3 = XXH32_round(v3, XXH_get32bits(p)); p+=4;
|
||||||
|
v4 = XXH32_round(v4, XXH_get32bits(p)); p+=4;
|
||||||
|
} while (p<=limit);
|
||||||
|
|
||||||
|
h32 = XXH_rotl32(v1, 1) + XXH_rotl32(v2, 7) + XXH_rotl32(v3, 12) + XXH_rotl32(v4, 18);
|
||||||
|
} else {
|
||||||
|
h32 = seed + PRIME32_5;
|
||||||
|
}
|
||||||
|
|
||||||
|
h32 += (U32) len;
|
||||||
|
|
||||||
|
while (p+4<=bEnd) {
|
||||||
|
h32 += XXH_get32bits(p) * PRIME32_3;
|
||||||
|
h32 = XXH_rotl32(h32, 17) * PRIME32_4 ;
|
||||||
|
p+=4;
|
||||||
|
}
|
||||||
|
|
||||||
|
while (p<bEnd) {
|
||||||
|
h32 += (*p) * PRIME32_5;
|
||||||
|
h32 = XXH_rotl32(h32, 11) * PRIME32_1 ;
|
||||||
|
p++;
|
||||||
|
}
|
||||||
|
|
||||||
|
h32 ^= h32 >> 15;
|
||||||
|
h32 *= PRIME32_2;
|
||||||
|
h32 ^= h32 >> 13;
|
||||||
|
h32 *= PRIME32_3;
|
||||||
|
h32 ^= h32 >> 16;
|
||||||
|
|
||||||
|
return h32;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
XXH_PUBLIC_API unsigned int XXH32 (const void* input, size_t len, unsigned int seed)
|
||||||
|
{
|
||||||
|
#if 0
|
||||||
|
/* Simple version, good for code maintenance, but unfortunately slow for small inputs */
|
||||||
|
XXH32_CREATESTATE_STATIC(state);
|
||||||
|
XXH32_reset(state, seed);
|
||||||
|
XXH32_update(state, input, len);
|
||||||
|
return XXH32_digest(state);
|
||||||
|
#else
|
||||||
|
XXH_endianess endian_detected = (XXH_endianess)XXH_CPU_LITTLE_ENDIAN;
|
||||||
|
|
||||||
|
if (XXH_FORCE_ALIGN_CHECK) {
|
||||||
|
if ((((size_t)input) & 3) == 0) { /* Input is 4-bytes aligned, leverage the speed benefit */
|
||||||
|
if ((endian_detected==XXH_littleEndian) || XXH_FORCE_NATIVE_FORMAT)
|
||||||
|
return XXH32_endian_align(input, len, seed, XXH_littleEndian, XXH_aligned);
|
||||||
|
else
|
||||||
|
return XXH32_endian_align(input, len, seed, XXH_bigEndian, XXH_aligned);
|
||||||
|
} }
|
||||||
|
|
||||||
|
if ((endian_detected==XXH_littleEndian) || XXH_FORCE_NATIVE_FORMAT)
|
||||||
|
return XXH32_endian_align(input, len, seed, XXH_littleEndian, XXH_unaligned);
|
||||||
|
else
|
||||||
|
return XXH32_endian_align(input, len, seed, XXH_bigEndian, XXH_unaligned);
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static U64 XXH64_round(U64 acc, U64 input)
|
||||||
|
{
|
||||||
|
acc += input * PRIME64_2;
|
||||||
|
acc = XXH_rotl64(acc, 31);
|
||||||
|
acc *= PRIME64_1;
|
||||||
|
return acc;
|
||||||
|
}
|
||||||
|
|
||||||
|
static U64 XXH64_mergeRound(U64 acc, U64 val)
|
||||||
|
{
|
||||||
|
val = XXH64_round(0, val);
|
||||||
|
acc ^= val;
|
||||||
|
acc = acc * PRIME64_1 + PRIME64_4;
|
||||||
|
return acc;
|
||||||
|
}
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE U64 XXH64_endian_align(const void* input, size_t len, U64 seed, XXH_endianess endian, XXH_alignment align)
|
||||||
|
{
|
||||||
|
const BYTE* p = (const BYTE*)input;
|
||||||
|
const BYTE* const bEnd = p + len;
|
||||||
|
U64 h64;
|
||||||
|
#define XXH_get64bits(p) XXH_readLE64_align(p, endian, align)
|
||||||
|
|
||||||
|
#ifdef XXH_ACCEPT_NULL_INPUT_POINTER
|
||||||
|
if (p==NULL) {
|
||||||
|
len=0;
|
||||||
|
bEnd=p=(const BYTE*)(size_t)32;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
if (len>=32) {
|
||||||
|
const BYTE* const limit = bEnd - 32;
|
||||||
|
U64 v1 = seed + PRIME64_1 + PRIME64_2;
|
||||||
|
U64 v2 = seed + PRIME64_2;
|
||||||
|
U64 v3 = seed + 0;
|
||||||
|
U64 v4 = seed - PRIME64_1;
|
||||||
|
|
||||||
|
do {
|
||||||
|
v1 = XXH64_round(v1, XXH_get64bits(p)); p+=8;
|
||||||
|
v2 = XXH64_round(v2, XXH_get64bits(p)); p+=8;
|
||||||
|
v3 = XXH64_round(v3, XXH_get64bits(p)); p+=8;
|
||||||
|
v4 = XXH64_round(v4, XXH_get64bits(p)); p+=8;
|
||||||
|
} while (p<=limit);
|
||||||
|
|
||||||
|
h64 = XXH_rotl64(v1, 1) + XXH_rotl64(v2, 7) + XXH_rotl64(v3, 12) + XXH_rotl64(v4, 18);
|
||||||
|
h64 = XXH64_mergeRound(h64, v1);
|
||||||
|
h64 = XXH64_mergeRound(h64, v2);
|
||||||
|
h64 = XXH64_mergeRound(h64, v3);
|
||||||
|
h64 = XXH64_mergeRound(h64, v4);
|
||||||
|
|
||||||
|
} else {
|
||||||
|
h64 = seed + PRIME64_5;
|
||||||
|
}
|
||||||
|
|
||||||
|
h64 += (U64) len;
|
||||||
|
|
||||||
|
while (p+8<=bEnd) {
|
||||||
|
U64 const k1 = XXH64_round(0, XXH_get64bits(p));
|
||||||
|
h64 ^= k1;
|
||||||
|
h64 = XXH_rotl64(h64,27) * PRIME64_1 + PRIME64_4;
|
||||||
|
p+=8;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (p+4<=bEnd) {
|
||||||
|
h64 ^= (U64)(XXH_get32bits(p)) * PRIME64_1;
|
||||||
|
h64 = XXH_rotl64(h64, 23) * PRIME64_2 + PRIME64_3;
|
||||||
|
p+=4;
|
||||||
|
}
|
||||||
|
|
||||||
|
while (p<bEnd) {
|
||||||
|
h64 ^= (*p) * PRIME64_5;
|
||||||
|
h64 = XXH_rotl64(h64, 11) * PRIME64_1;
|
||||||
|
p++;
|
||||||
|
}
|
||||||
|
|
||||||
|
h64 ^= h64 >> 33;
|
||||||
|
h64 *= PRIME64_2;
|
||||||
|
h64 ^= h64 >> 29;
|
||||||
|
h64 *= PRIME64_3;
|
||||||
|
h64 ^= h64 >> 32;
|
||||||
|
|
||||||
|
return h64;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
XXH_PUBLIC_API unsigned long long XXH64 (const void* input, size_t len, unsigned long long seed)
|
||||||
|
{
|
||||||
|
#if 0
|
||||||
|
/* Simple version, good for code maintenance, but unfortunately slow for small inputs */
|
||||||
|
XXH64_CREATESTATE_STATIC(state);
|
||||||
|
XXH64_reset(state, seed);
|
||||||
|
XXH64_update(state, input, len);
|
||||||
|
return XXH64_digest(state);
|
||||||
|
#else
|
||||||
|
XXH_endianess endian_detected = (XXH_endianess)XXH_CPU_LITTLE_ENDIAN;
|
||||||
|
|
||||||
|
if (XXH_FORCE_ALIGN_CHECK) {
|
||||||
|
if ((((size_t)input) & 7)==0) { /* Input is aligned, let's leverage the speed advantage */
|
||||||
|
if ((endian_detected==XXH_littleEndian) || XXH_FORCE_NATIVE_FORMAT)
|
||||||
|
return XXH64_endian_align(input, len, seed, XXH_littleEndian, XXH_aligned);
|
||||||
|
else
|
||||||
|
return XXH64_endian_align(input, len, seed, XXH_bigEndian, XXH_aligned);
|
||||||
|
} }
|
||||||
|
|
||||||
|
if ((endian_detected==XXH_littleEndian) || XXH_FORCE_NATIVE_FORMAT)
|
||||||
|
return XXH64_endian_align(input, len, seed, XXH_littleEndian, XXH_unaligned);
|
||||||
|
else
|
||||||
|
return XXH64_endian_align(input, len, seed, XXH_bigEndian, XXH_unaligned);
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************************************
|
||||||
|
* Advanced Hash Functions
|
||||||
|
****************************************************/
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH32_state_t* XXH32_createState(void)
|
||||||
|
{
|
||||||
|
return (XXH32_state_t*)XXH_malloc(sizeof(XXH32_state_t));
|
||||||
|
}
|
||||||
|
XXH_PUBLIC_API XXH_errorcode XXH32_freeState(XXH32_state_t* statePtr)
|
||||||
|
{
|
||||||
|
XXH_free(statePtr);
|
||||||
|
return XXH_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH64_state_t* XXH64_createState(void)
|
||||||
|
{
|
||||||
|
return (XXH64_state_t*)XXH_malloc(sizeof(XXH64_state_t));
|
||||||
|
}
|
||||||
|
XXH_PUBLIC_API XXH_errorcode XXH64_freeState(XXH64_state_t* statePtr)
|
||||||
|
{
|
||||||
|
XXH_free(statePtr);
|
||||||
|
return XXH_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/*** Hash feed ***/
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH_errorcode XXH32_reset(XXH32_state_t* statePtr, unsigned int seed)
|
||||||
|
{
|
||||||
|
XXH32_state_t state; /* using a local state to memcpy() in order to avoid strict-aliasing warnings */
|
||||||
|
memset(&state, 0, sizeof(state)-4); /* do not write into reserved, for future removal */
|
||||||
|
state.v1 = seed + PRIME32_1 + PRIME32_2;
|
||||||
|
state.v2 = seed + PRIME32_2;
|
||||||
|
state.v3 = seed + 0;
|
||||||
|
state.v4 = seed - PRIME32_1;
|
||||||
|
memcpy(statePtr, &state, sizeof(state));
|
||||||
|
return XXH_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH_errorcode XXH64_reset(XXH64_state_t* statePtr, unsigned long long seed)
|
||||||
|
{
|
||||||
|
XXH64_state_t state; /* using a local state to memcpy() in order to avoid strict-aliasing warnings */
|
||||||
|
memset(&state, 0, sizeof(state)-8); /* do not write into reserved, for future removal */
|
||||||
|
state.v1 = seed + PRIME64_1 + PRIME64_2;
|
||||||
|
state.v2 = seed + PRIME64_2;
|
||||||
|
state.v3 = seed + 0;
|
||||||
|
state.v4 = seed - PRIME64_1;
|
||||||
|
memcpy(statePtr, &state, sizeof(state));
|
||||||
|
return XXH_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE XXH_errorcode XXH32_update_endian (XXH32_state_t* state, const void* input, size_t len, XXH_endianess endian)
|
||||||
|
{
|
||||||
|
const BYTE* p = (const BYTE*)input;
|
||||||
|
const BYTE* const bEnd = p + len;
|
||||||
|
|
||||||
|
#ifdef XXH_ACCEPT_NULL_INPUT_POINTER
|
||||||
|
if (input==NULL) return XXH_ERROR;
|
||||||
|
#endif
|
||||||
|
|
||||||
|
state->total_len_32 += (unsigned)len;
|
||||||
|
state->large_len |= (len>=16) | (state->total_len_32>=16);
|
||||||
|
|
||||||
|
if (state->memsize + len < 16) { /* fill in tmp buffer */
|
||||||
|
XXH_memcpy((BYTE*)(state->mem32) + state->memsize, input, len);
|
||||||
|
state->memsize += (unsigned)len;
|
||||||
|
return XXH_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (state->memsize) { /* some data left from previous update */
|
||||||
|
XXH_memcpy((BYTE*)(state->mem32) + state->memsize, input, 16-state->memsize);
|
||||||
|
{ const U32* p32 = state->mem32;
|
||||||
|
state->v1 = XXH32_round(state->v1, XXH_readLE32(p32, endian)); p32++;
|
||||||
|
state->v2 = XXH32_round(state->v2, XXH_readLE32(p32, endian)); p32++;
|
||||||
|
state->v3 = XXH32_round(state->v3, XXH_readLE32(p32, endian)); p32++;
|
||||||
|
state->v4 = XXH32_round(state->v4, XXH_readLE32(p32, endian)); p32++;
|
||||||
|
}
|
||||||
|
p += 16-state->memsize;
|
||||||
|
state->memsize = 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (p <= bEnd-16) {
|
||||||
|
const BYTE* const limit = bEnd - 16;
|
||||||
|
U32 v1 = state->v1;
|
||||||
|
U32 v2 = state->v2;
|
||||||
|
U32 v3 = state->v3;
|
||||||
|
U32 v4 = state->v4;
|
||||||
|
|
||||||
|
do {
|
||||||
|
v1 = XXH32_round(v1, XXH_readLE32(p, endian)); p+=4;
|
||||||
|
v2 = XXH32_round(v2, XXH_readLE32(p, endian)); p+=4;
|
||||||
|
v3 = XXH32_round(v3, XXH_readLE32(p, endian)); p+=4;
|
||||||
|
v4 = XXH32_round(v4, XXH_readLE32(p, endian)); p+=4;
|
||||||
|
} while (p<=limit);
|
||||||
|
|
||||||
|
state->v1 = v1;
|
||||||
|
state->v2 = v2;
|
||||||
|
state->v3 = v3;
|
||||||
|
state->v4 = v4;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (p < bEnd) {
|
||||||
|
XXH_memcpy(state->mem32, p, (size_t)(bEnd-p));
|
||||||
|
state->memsize = (unsigned)(bEnd-p);
|
||||||
|
}
|
||||||
|
|
||||||
|
return XXH_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH_errorcode XXH32_update (XXH32_state_t* state_in, const void* input, size_t len)
|
||||||
|
{
|
||||||
|
XXH_endianess endian_detected = (XXH_endianess)XXH_CPU_LITTLE_ENDIAN;
|
||||||
|
|
||||||
|
if ((endian_detected==XXH_littleEndian) || XXH_FORCE_NATIVE_FORMAT)
|
||||||
|
return XXH32_update_endian(state_in, input, len, XXH_littleEndian);
|
||||||
|
else
|
||||||
|
return XXH32_update_endian(state_in, input, len, XXH_bigEndian);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE U32 XXH32_digest_endian (const XXH32_state_t* state, XXH_endianess endian)
|
||||||
|
{
|
||||||
|
const BYTE * p = (const BYTE*)state->mem32;
|
||||||
|
const BYTE* const bEnd = (const BYTE*)(state->mem32) + state->memsize;
|
||||||
|
U32 h32;
|
||||||
|
|
||||||
|
if (state->large_len) {
|
||||||
|
h32 = XXH_rotl32(state->v1, 1) + XXH_rotl32(state->v2, 7) + XXH_rotl32(state->v3, 12) + XXH_rotl32(state->v4, 18);
|
||||||
|
} else {
|
||||||
|
h32 = state->v3 /* == seed */ + PRIME32_5;
|
||||||
|
}
|
||||||
|
|
||||||
|
h32 += state->total_len_32;
|
||||||
|
|
||||||
|
while (p+4<=bEnd) {
|
||||||
|
h32 += XXH_readLE32(p, endian) * PRIME32_3;
|
||||||
|
h32 = XXH_rotl32(h32, 17) * PRIME32_4;
|
||||||
|
p+=4;
|
||||||
|
}
|
||||||
|
|
||||||
|
while (p<bEnd) {
|
||||||
|
h32 += (*p) * PRIME32_5;
|
||||||
|
h32 = XXH_rotl32(h32, 11) * PRIME32_1;
|
||||||
|
p++;
|
||||||
|
}
|
||||||
|
|
||||||
|
h32 ^= h32 >> 15;
|
||||||
|
h32 *= PRIME32_2;
|
||||||
|
h32 ^= h32 >> 13;
|
||||||
|
h32 *= PRIME32_3;
|
||||||
|
h32 ^= h32 >> 16;
|
||||||
|
|
||||||
|
return h32;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
XXH_PUBLIC_API unsigned int XXH32_digest (const XXH32_state_t* state_in)
|
||||||
|
{
|
||||||
|
XXH_endianess endian_detected = (XXH_endianess)XXH_CPU_LITTLE_ENDIAN;
|
||||||
|
|
||||||
|
if ((endian_detected==XXH_littleEndian) || XXH_FORCE_NATIVE_FORMAT)
|
||||||
|
return XXH32_digest_endian(state_in, XXH_littleEndian);
|
||||||
|
else
|
||||||
|
return XXH32_digest_endian(state_in, XXH_bigEndian);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/* **** XXH64 **** */
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE XXH_errorcode XXH64_update_endian (XXH64_state_t* state, const void* input, size_t len, XXH_endianess endian)
|
||||||
|
{
|
||||||
|
const BYTE* p = (const BYTE*)input;
|
||||||
|
const BYTE* const bEnd = p + len;
|
||||||
|
|
||||||
|
#ifdef XXH_ACCEPT_NULL_INPUT_POINTER
|
||||||
|
if (input==NULL) return XXH_ERROR;
|
||||||
|
#endif
|
||||||
|
|
||||||
|
state->total_len += len;
|
||||||
|
|
||||||
|
if (state->memsize + len < 32) { /* fill in tmp buffer */
|
||||||
|
XXH_memcpy(((BYTE*)state->mem64) + state->memsize, input, len);
|
||||||
|
state->memsize += (U32)len;
|
||||||
|
return XXH_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (state->memsize) { /* tmp buffer is full */
|
||||||
|
XXH_memcpy(((BYTE*)state->mem64) + state->memsize, input, 32-state->memsize);
|
||||||
|
state->v1 = XXH64_round(state->v1, XXH_readLE64(state->mem64+0, endian));
|
||||||
|
state->v2 = XXH64_round(state->v2, XXH_readLE64(state->mem64+1, endian));
|
||||||
|
state->v3 = XXH64_round(state->v3, XXH_readLE64(state->mem64+2, endian));
|
||||||
|
state->v4 = XXH64_round(state->v4, XXH_readLE64(state->mem64+3, endian));
|
||||||
|
p += 32-state->memsize;
|
||||||
|
state->memsize = 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (p+32 <= bEnd) {
|
||||||
|
const BYTE* const limit = bEnd - 32;
|
||||||
|
U64 v1 = state->v1;
|
||||||
|
U64 v2 = state->v2;
|
||||||
|
U64 v3 = state->v3;
|
||||||
|
U64 v4 = state->v4;
|
||||||
|
|
||||||
|
do {
|
||||||
|
v1 = XXH64_round(v1, XXH_readLE64(p, endian)); p+=8;
|
||||||
|
v2 = XXH64_round(v2, XXH_readLE64(p, endian)); p+=8;
|
||||||
|
v3 = XXH64_round(v3, XXH_readLE64(p, endian)); p+=8;
|
||||||
|
v4 = XXH64_round(v4, XXH_readLE64(p, endian)); p+=8;
|
||||||
|
} while (p<=limit);
|
||||||
|
|
||||||
|
state->v1 = v1;
|
||||||
|
state->v2 = v2;
|
||||||
|
state->v3 = v3;
|
||||||
|
state->v4 = v4;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (p < bEnd) {
|
||||||
|
XXH_memcpy(state->mem64, p, (size_t)(bEnd-p));
|
||||||
|
state->memsize = (unsigned)(bEnd-p);
|
||||||
|
}
|
||||||
|
|
||||||
|
return XXH_OK;
|
||||||
|
}
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH_errorcode XXH64_update (XXH64_state_t* state_in, const void* input, size_t len)
|
||||||
|
{
|
||||||
|
XXH_endianess endian_detected = (XXH_endianess)XXH_CPU_LITTLE_ENDIAN;
|
||||||
|
|
||||||
|
if ((endian_detected==XXH_littleEndian) || XXH_FORCE_NATIVE_FORMAT)
|
||||||
|
return XXH64_update_endian(state_in, input, len, XXH_littleEndian);
|
||||||
|
else
|
||||||
|
return XXH64_update_endian(state_in, input, len, XXH_bigEndian);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE U64 XXH64_digest_endian (const XXH64_state_t* state, XXH_endianess endian)
|
||||||
|
{
|
||||||
|
const BYTE * p = (const BYTE*)state->mem64;
|
||||||
|
const BYTE* const bEnd = (const BYTE*)state->mem64 + state->memsize;
|
||||||
|
U64 h64;
|
||||||
|
|
||||||
|
if (state->total_len >= 32) {
|
||||||
|
U64 const v1 = state->v1;
|
||||||
|
U64 const v2 = state->v2;
|
||||||
|
U64 const v3 = state->v3;
|
||||||
|
U64 const v4 = state->v4;
|
||||||
|
|
||||||
|
h64 = XXH_rotl64(v1, 1) + XXH_rotl64(v2, 7) + XXH_rotl64(v3, 12) + XXH_rotl64(v4, 18);
|
||||||
|
h64 = XXH64_mergeRound(h64, v1);
|
||||||
|
h64 = XXH64_mergeRound(h64, v2);
|
||||||
|
h64 = XXH64_mergeRound(h64, v3);
|
||||||
|
h64 = XXH64_mergeRound(h64, v4);
|
||||||
|
} else {
|
||||||
|
h64 = state->v3 + PRIME64_5;
|
||||||
|
}
|
||||||
|
|
||||||
|
h64 += (U64) state->total_len;
|
||||||
|
|
||||||
|
while (p+8<=bEnd) {
|
||||||
|
U64 const k1 = XXH64_round(0, XXH_readLE64(p, endian));
|
||||||
|
h64 ^= k1;
|
||||||
|
h64 = XXH_rotl64(h64,27) * PRIME64_1 + PRIME64_4;
|
||||||
|
p+=8;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (p+4<=bEnd) {
|
||||||
|
h64 ^= (U64)(XXH_readLE32(p, endian)) * PRIME64_1;
|
||||||
|
h64 = XXH_rotl64(h64, 23) * PRIME64_2 + PRIME64_3;
|
||||||
|
p+=4;
|
||||||
|
}
|
||||||
|
|
||||||
|
while (p<bEnd) {
|
||||||
|
h64 ^= (*p) * PRIME64_5;
|
||||||
|
h64 = XXH_rotl64(h64, 11) * PRIME64_1;
|
||||||
|
p++;
|
||||||
|
}
|
||||||
|
|
||||||
|
h64 ^= h64 >> 33;
|
||||||
|
h64 *= PRIME64_2;
|
||||||
|
h64 ^= h64 >> 29;
|
||||||
|
h64 *= PRIME64_3;
|
||||||
|
h64 ^= h64 >> 32;
|
||||||
|
|
||||||
|
return h64;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
XXH_PUBLIC_API unsigned long long XXH64_digest (const XXH64_state_t* state_in)
|
||||||
|
{
|
||||||
|
XXH_endianess endian_detected = (XXH_endianess)XXH_CPU_LITTLE_ENDIAN;
|
||||||
|
|
||||||
|
if ((endian_detected==XXH_littleEndian) || XXH_FORCE_NATIVE_FORMAT)
|
||||||
|
return XXH64_digest_endian(state_in, XXH_littleEndian);
|
||||||
|
else
|
||||||
|
return XXH64_digest_endian(state_in, XXH_bigEndian);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************
|
||||||
|
* Canonical representation
|
||||||
|
****************************/
|
||||||
|
|
||||||
|
/*! Default XXH result types are basic unsigned 32 and 64 bits.
|
||||||
|
* The canonical representation follows human-readable write convention, aka big-endian (large digits first).
|
||||||
|
* These functions allow transformation of hash result into and from its canonical format.
|
||||||
|
* This way, hash values can be written into a file or buffer, and remain comparable across different systems and programs.
|
||||||
|
*/
|
||||||
|
|
||||||
|
XXH_PUBLIC_API void XXH32_canonicalFromHash(XXH32_canonical_t* dst, XXH32_hash_t hash)
|
||||||
|
{
|
||||||
|
XXH_STATIC_ASSERT(sizeof(XXH32_canonical_t) == sizeof(XXH32_hash_t));
|
||||||
|
if (XXH_CPU_LITTLE_ENDIAN) hash = XXH_swap32(hash);
|
||||||
|
memcpy(dst, &hash, sizeof(*dst));
|
||||||
|
}
|
||||||
|
|
||||||
|
XXH_PUBLIC_API void XXH64_canonicalFromHash(XXH64_canonical_t* dst, XXH64_hash_t hash)
|
||||||
|
{
|
||||||
|
XXH_STATIC_ASSERT(sizeof(XXH64_canonical_t) == sizeof(XXH64_hash_t));
|
||||||
|
if (XXH_CPU_LITTLE_ENDIAN) hash = XXH_swap64(hash);
|
||||||
|
memcpy(dst, &hash, sizeof(*dst));
|
||||||
|
}
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH32_hash_t XXH32_hashFromCanonical(const XXH32_canonical_t* src)
|
||||||
|
{
|
||||||
|
return XXH_readBE32(src);
|
||||||
|
}
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH64_hash_t XXH64_hashFromCanonical(const XXH64_canonical_t* src)
|
||||||
|
{
|
||||||
|
return XXH_readBE64(src);
|
||||||
|
}
|
||||||
305
vendor/github.com/DataDog/zstd/xxhash.h
generated
vendored
Normal file
305
vendor/github.com/DataDog/zstd/xxhash.h
generated
vendored
Normal file
@@ -0,0 +1,305 @@
|
|||||||
|
/*
|
||||||
|
xxHash - Extremely Fast Hash algorithm
|
||||||
|
Header File
|
||||||
|
Copyright (C) 2012-2016, Yann Collet.
|
||||||
|
|
||||||
|
BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php)
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
|
||||||
|
You can contact the author at :
|
||||||
|
- xxHash source repository : https://github.com/Cyan4973/xxHash
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* Notice extracted from xxHash homepage :
|
||||||
|
|
||||||
|
xxHash is an extremely fast Hash algorithm, running at RAM speed limits.
|
||||||
|
It also successfully passes all tests from the SMHasher suite.
|
||||||
|
|
||||||
|
Comparison (single thread, Windows Seven 32 bits, using SMHasher on a Core 2 Duo @3GHz)
|
||||||
|
|
||||||
|
Name Speed Q.Score Author
|
||||||
|
xxHash 5.4 GB/s 10
|
||||||
|
CrapWow 3.2 GB/s 2 Andrew
|
||||||
|
MumurHash 3a 2.7 GB/s 10 Austin Appleby
|
||||||
|
SpookyHash 2.0 GB/s 10 Bob Jenkins
|
||||||
|
SBox 1.4 GB/s 9 Bret Mulvey
|
||||||
|
Lookup3 1.2 GB/s 9 Bob Jenkins
|
||||||
|
SuperFastHash 1.2 GB/s 1 Paul Hsieh
|
||||||
|
CityHash64 1.05 GB/s 10 Pike & Alakuijala
|
||||||
|
FNV 0.55 GB/s 5 Fowler, Noll, Vo
|
||||||
|
CRC32 0.43 GB/s 9
|
||||||
|
MD5-32 0.33 GB/s 10 Ronald L. Rivest
|
||||||
|
SHA1-32 0.28 GB/s 10
|
||||||
|
|
||||||
|
Q.Score is a measure of quality of the hash function.
|
||||||
|
It depends on successfully passing SMHasher test set.
|
||||||
|
10 is a perfect score.
|
||||||
|
|
||||||
|
A 64-bits version, named XXH64, is available since r35.
|
||||||
|
It offers much better speed, but for 64-bits applications only.
|
||||||
|
Name Speed on 64 bits Speed on 32 bits
|
||||||
|
XXH64 13.8 GB/s 1.9 GB/s
|
||||||
|
XXH32 6.8 GB/s 6.0 GB/s
|
||||||
|
*/
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#ifndef XXHASH_H_5627135585666179
|
||||||
|
#define XXHASH_H_5627135585666179 1
|
||||||
|
|
||||||
|
|
||||||
|
/* ****************************
|
||||||
|
* Definitions
|
||||||
|
******************************/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
typedef enum { XXH_OK=0, XXH_ERROR } XXH_errorcode;
|
||||||
|
|
||||||
|
|
||||||
|
/* ****************************
|
||||||
|
* API modifier
|
||||||
|
******************************/
|
||||||
|
/** XXH_PRIVATE_API
|
||||||
|
* This is useful if you want to include xxhash functions in `static` mode
|
||||||
|
* in order to inline them, and remove their symbol from the public list.
|
||||||
|
* Methodology :
|
||||||
|
* #define XXH_PRIVATE_API
|
||||||
|
* #include "xxhash.h"
|
||||||
|
* `xxhash.c` is automatically included.
|
||||||
|
* It's not useful to compile and link it as a separate module anymore.
|
||||||
|
*/
|
||||||
|
#ifdef XXH_PRIVATE_API
|
||||||
|
# ifndef XXH_STATIC_LINKING_ONLY
|
||||||
|
# define XXH_STATIC_LINKING_ONLY
|
||||||
|
# endif
|
||||||
|
# if defined(__GNUC__)
|
||||||
|
# define XXH_PUBLIC_API static __inline __attribute__((unused))
|
||||||
|
# elif defined (__cplusplus) || (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L) /* C99 */)
|
||||||
|
# define XXH_PUBLIC_API static inline
|
||||||
|
# elif defined(_MSC_VER)
|
||||||
|
# define XXH_PUBLIC_API static __inline
|
||||||
|
# else
|
||||||
|
# define XXH_PUBLIC_API static /* this version may generate warnings for unused static functions; disable the relevant warning */
|
||||||
|
# endif
|
||||||
|
#else
|
||||||
|
# define XXH_PUBLIC_API /* do nothing */
|
||||||
|
#endif /* XXH_PRIVATE_API */
|
||||||
|
|
||||||
|
/*!XXH_NAMESPACE, aka Namespace Emulation :
|
||||||
|
|
||||||
|
If you want to include _and expose_ xxHash functions from within your own library,
|
||||||
|
but also want to avoid symbol collisions with another library which also includes xxHash,
|
||||||
|
|
||||||
|
you can use XXH_NAMESPACE, to automatically prefix any public symbol from xxhash library
|
||||||
|
with the value of XXH_NAMESPACE (so avoid to keep it NULL and avoid numeric values).
|
||||||
|
|
||||||
|
Note that no change is required within the calling program as long as it includes `xxhash.h` :
|
||||||
|
regular symbol name will be automatically translated by this header.
|
||||||
|
*/
|
||||||
|
#ifdef XXH_NAMESPACE
|
||||||
|
# define XXH_CAT(A,B) A##B
|
||||||
|
# define XXH_NAME2(A,B) XXH_CAT(A,B)
|
||||||
|
# define XXH32 XXH_NAME2(XXH_NAMESPACE, XXH32)
|
||||||
|
# define XXH64 XXH_NAME2(XXH_NAMESPACE, XXH64)
|
||||||
|
# define XXH_versionNumber XXH_NAME2(XXH_NAMESPACE, XXH_versionNumber)
|
||||||
|
# define XXH32_createState XXH_NAME2(XXH_NAMESPACE, XXH32_createState)
|
||||||
|
# define XXH64_createState XXH_NAME2(XXH_NAMESPACE, XXH64_createState)
|
||||||
|
# define XXH32_freeState XXH_NAME2(XXH_NAMESPACE, XXH32_freeState)
|
||||||
|
# define XXH64_freeState XXH_NAME2(XXH_NAMESPACE, XXH64_freeState)
|
||||||
|
# define XXH32_reset XXH_NAME2(XXH_NAMESPACE, XXH32_reset)
|
||||||
|
# define XXH64_reset XXH_NAME2(XXH_NAMESPACE, XXH64_reset)
|
||||||
|
# define XXH32_update XXH_NAME2(XXH_NAMESPACE, XXH32_update)
|
||||||
|
# define XXH64_update XXH_NAME2(XXH_NAMESPACE, XXH64_update)
|
||||||
|
# define XXH32_digest XXH_NAME2(XXH_NAMESPACE, XXH32_digest)
|
||||||
|
# define XXH64_digest XXH_NAME2(XXH_NAMESPACE, XXH64_digest)
|
||||||
|
# define XXH32_copyState XXH_NAME2(XXH_NAMESPACE, XXH32_copyState)
|
||||||
|
# define XXH64_copyState XXH_NAME2(XXH_NAMESPACE, XXH64_copyState)
|
||||||
|
# define XXH32_canonicalFromHash XXH_NAME2(XXH_NAMESPACE, XXH32_canonicalFromHash)
|
||||||
|
# define XXH64_canonicalFromHash XXH_NAME2(XXH_NAMESPACE, XXH64_canonicalFromHash)
|
||||||
|
# define XXH32_hashFromCanonical XXH_NAME2(XXH_NAMESPACE, XXH32_hashFromCanonical)
|
||||||
|
# define XXH64_hashFromCanonical XXH_NAME2(XXH_NAMESPACE, XXH64_hashFromCanonical)
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Version
|
||||||
|
***************************************/
|
||||||
|
#define XXH_VERSION_MAJOR 0
|
||||||
|
#define XXH_VERSION_MINOR 6
|
||||||
|
#define XXH_VERSION_RELEASE 2
|
||||||
|
#define XXH_VERSION_NUMBER (XXH_VERSION_MAJOR *100*100 + XXH_VERSION_MINOR *100 + XXH_VERSION_RELEASE)
|
||||||
|
XXH_PUBLIC_API unsigned XXH_versionNumber (void);
|
||||||
|
|
||||||
|
|
||||||
|
/* ****************************
|
||||||
|
* Simple Hash Functions
|
||||||
|
******************************/
|
||||||
|
typedef unsigned int XXH32_hash_t;
|
||||||
|
typedef unsigned long long XXH64_hash_t;
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH32_hash_t XXH32 (const void* input, size_t length, unsigned int seed);
|
||||||
|
XXH_PUBLIC_API XXH64_hash_t XXH64 (const void* input, size_t length, unsigned long long seed);
|
||||||
|
|
||||||
|
/*!
|
||||||
|
XXH32() :
|
||||||
|
Calculate the 32-bits hash of sequence "length" bytes stored at memory address "input".
|
||||||
|
The memory between input & input+length must be valid (allocated and read-accessible).
|
||||||
|
"seed" can be used to alter the result predictably.
|
||||||
|
Speed on Core 2 Duo @ 3 GHz (single thread, SMHasher benchmark) : 5.4 GB/s
|
||||||
|
XXH64() :
|
||||||
|
Calculate the 64-bits hash of sequence of length "len" stored at memory address "input".
|
||||||
|
"seed" can be used to alter the result predictably.
|
||||||
|
This function runs 2x faster on 64-bits systems, but slower on 32-bits systems (see benchmark).
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/* ****************************
|
||||||
|
* Streaming Hash Functions
|
||||||
|
******************************/
|
||||||
|
typedef struct XXH32_state_s XXH32_state_t; /* incomplete type */
|
||||||
|
typedef struct XXH64_state_s XXH64_state_t; /* incomplete type */
|
||||||
|
|
||||||
|
/*! State allocation, compatible with dynamic libraries */
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH32_state_t* XXH32_createState(void);
|
||||||
|
XXH_PUBLIC_API XXH_errorcode XXH32_freeState(XXH32_state_t* statePtr);
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH64_state_t* XXH64_createState(void);
|
||||||
|
XXH_PUBLIC_API XXH_errorcode XXH64_freeState(XXH64_state_t* statePtr);
|
||||||
|
|
||||||
|
|
||||||
|
/* hash streaming */
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH_errorcode XXH32_reset (XXH32_state_t* statePtr, unsigned int seed);
|
||||||
|
XXH_PUBLIC_API XXH_errorcode XXH32_update (XXH32_state_t* statePtr, const void* input, size_t length);
|
||||||
|
XXH_PUBLIC_API XXH32_hash_t XXH32_digest (const XXH32_state_t* statePtr);
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH_errorcode XXH64_reset (XXH64_state_t* statePtr, unsigned long long seed);
|
||||||
|
XXH_PUBLIC_API XXH_errorcode XXH64_update (XXH64_state_t* statePtr, const void* input, size_t length);
|
||||||
|
XXH_PUBLIC_API XXH64_hash_t XXH64_digest (const XXH64_state_t* statePtr);
|
||||||
|
|
||||||
|
/*
|
||||||
|
These functions generate the xxHash of an input provided in multiple segments.
|
||||||
|
Note that, for small input, they are slower than single-call functions, due to state management.
|
||||||
|
For small input, prefer `XXH32()` and `XXH64()` .
|
||||||
|
|
||||||
|
XXH state must first be allocated, using XXH*_createState() .
|
||||||
|
|
||||||
|
Start a new hash by initializing state with a seed, using XXH*_reset().
|
||||||
|
|
||||||
|
Then, feed the hash state by calling XXH*_update() as many times as necessary.
|
||||||
|
Obviously, input must be allocated and read accessible.
|
||||||
|
The function returns an error code, with 0 meaning OK, and any other value meaning there is an error.
|
||||||
|
|
||||||
|
Finally, a hash value can be produced anytime, by using XXH*_digest().
|
||||||
|
This function returns the nn-bits hash as an int or long long.
|
||||||
|
|
||||||
|
It's still possible to continue inserting input into the hash state after a digest,
|
||||||
|
and generate some new hashes later on, by calling again XXH*_digest().
|
||||||
|
|
||||||
|
When done, free XXH state space if it was allocated dynamically.
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************
|
||||||
|
* Utils
|
||||||
|
****************************/
|
||||||
|
#if !(defined(__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L)) /* ! C99 */
|
||||||
|
# define restrict /* disable restrict */
|
||||||
|
#endif
|
||||||
|
|
||||||
|
XXH_PUBLIC_API void XXH32_copyState(XXH32_state_t* restrict dst_state, const XXH32_state_t* restrict src_state);
|
||||||
|
XXH_PUBLIC_API void XXH64_copyState(XXH64_state_t* restrict dst_state, const XXH64_state_t* restrict src_state);
|
||||||
|
|
||||||
|
|
||||||
|
/* **************************
|
||||||
|
* Canonical representation
|
||||||
|
****************************/
|
||||||
|
/* Default result type for XXH functions are primitive unsigned 32 and 64 bits.
|
||||||
|
* The canonical representation uses human-readable write convention, aka big-endian (large digits first).
|
||||||
|
* These functions allow transformation of hash result into and from its canonical format.
|
||||||
|
* This way, hash values can be written into a file / memory, and remain comparable on different systems and programs.
|
||||||
|
*/
|
||||||
|
typedef struct { unsigned char digest[4]; } XXH32_canonical_t;
|
||||||
|
typedef struct { unsigned char digest[8]; } XXH64_canonical_t;
|
||||||
|
|
||||||
|
XXH_PUBLIC_API void XXH32_canonicalFromHash(XXH32_canonical_t* dst, XXH32_hash_t hash);
|
||||||
|
XXH_PUBLIC_API void XXH64_canonicalFromHash(XXH64_canonical_t* dst, XXH64_hash_t hash);
|
||||||
|
|
||||||
|
XXH_PUBLIC_API XXH32_hash_t XXH32_hashFromCanonical(const XXH32_canonical_t* src);
|
||||||
|
XXH_PUBLIC_API XXH64_hash_t XXH64_hashFromCanonical(const XXH64_canonical_t* src);
|
||||||
|
|
||||||
|
#endif /* XXHASH_H_5627135585666179 */
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/* ================================================================================================
|
||||||
|
This section contains definitions which are not guaranteed to remain stable.
|
||||||
|
They may change in future versions, becoming incompatible with a different version of the library.
|
||||||
|
They shall only be used with static linking.
|
||||||
|
Never use these definitions in association with dynamic linking !
|
||||||
|
=================================================================================================== */
|
||||||
|
#if defined(XXH_STATIC_LINKING_ONLY) && !defined(XXH_STATIC_H_3543687687345)
|
||||||
|
#define XXH_STATIC_H_3543687687345
|
||||||
|
|
||||||
|
/* These definitions are only meant to allow allocation of XXH state
|
||||||
|
statically, on stack, or in a struct for example.
|
||||||
|
Do not use members directly. */
|
||||||
|
|
||||||
|
struct XXH32_state_s {
|
||||||
|
unsigned total_len_32;
|
||||||
|
unsigned large_len;
|
||||||
|
unsigned v1;
|
||||||
|
unsigned v2;
|
||||||
|
unsigned v3;
|
||||||
|
unsigned v4;
|
||||||
|
unsigned mem32[4]; /* buffer defined as U32 for alignment */
|
||||||
|
unsigned memsize;
|
||||||
|
unsigned reserved; /* never read nor write, will be removed in a future version */
|
||||||
|
}; /* typedef'd to XXH32_state_t */
|
||||||
|
|
||||||
|
struct XXH64_state_s {
|
||||||
|
unsigned long long total_len;
|
||||||
|
unsigned long long v1;
|
||||||
|
unsigned long long v2;
|
||||||
|
unsigned long long v3;
|
||||||
|
unsigned long long v4;
|
||||||
|
unsigned long long mem64[4]; /* buffer defined as U64 for alignment */
|
||||||
|
unsigned memsize;
|
||||||
|
unsigned reserved[2]; /* never read nor write, will be removed in a future version */
|
||||||
|
}; /* typedef'd to XXH64_state_t */
|
||||||
|
|
||||||
|
|
||||||
|
# ifdef XXH_PRIVATE_API
|
||||||
|
# include "xxhash.c" /* include xxhash functions as `static`, for inlining */
|
||||||
|
# endif
|
||||||
|
|
||||||
|
#endif /* XXH_STATIC_LINKING_ONLY && XXH_STATIC_H_3543687687345 */
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
214
vendor/github.com/DataDog/zstd/zbuff.h
generated
vendored
Normal file
214
vendor/github.com/DataDog/zstd/zbuff.h
generated
vendored
Normal file
@@ -0,0 +1,214 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* ***************************************************************
|
||||||
|
* NOTES/WARNINGS
|
||||||
|
******************************************************************/
|
||||||
|
/* The streaming API defined here is deprecated.
|
||||||
|
* Consider migrating towards ZSTD_compressStream() API in `zstd.h`
|
||||||
|
* See 'lib/README.md'.
|
||||||
|
*****************************************************************/
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#ifndef ZSTD_BUFFERED_H_23987
|
||||||
|
#define ZSTD_BUFFERED_H_23987
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Dependencies
|
||||||
|
***************************************/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
#include "zstd.h" /* ZSTD_CStream, ZSTD_DStream, ZSTDLIB_API */
|
||||||
|
|
||||||
|
|
||||||
|
/* ***************************************************************
|
||||||
|
* Compiler specifics
|
||||||
|
*****************************************************************/
|
||||||
|
/* Deprecation warnings */
|
||||||
|
/* Should these warnings be a problem,
|
||||||
|
* it is generally possible to disable them,
|
||||||
|
* typically with -Wno-deprecated-declarations for gcc
|
||||||
|
* or _CRT_SECURE_NO_WARNINGS in Visual.
|
||||||
|
* Otherwise, it's also possible to define ZBUFF_DISABLE_DEPRECATE_WARNINGS
|
||||||
|
*/
|
||||||
|
#ifdef ZBUFF_DISABLE_DEPRECATE_WARNINGS
|
||||||
|
# define ZBUFF_DEPRECATED(message) ZSTDLIB_API /* disable deprecation warnings */
|
||||||
|
#else
|
||||||
|
# if defined (__cplusplus) && (__cplusplus >= 201402) /* C++14 or greater */
|
||||||
|
# define ZBUFF_DEPRECATED(message) [[deprecated(message)]] ZSTDLIB_API
|
||||||
|
# elif (defined(GNUC) && (GNUC > 4 || (GNUC == 4 && GNUC_MINOR >= 5))) || defined(__clang__)
|
||||||
|
# define ZBUFF_DEPRECATED(message) ZSTDLIB_API __attribute__((deprecated(message)))
|
||||||
|
# elif defined(__GNUC__) && (__GNUC__ >= 3)
|
||||||
|
# define ZBUFF_DEPRECATED(message) ZSTDLIB_API __attribute__((deprecated))
|
||||||
|
# elif defined(_MSC_VER)
|
||||||
|
# define ZBUFF_DEPRECATED(message) ZSTDLIB_API __declspec(deprecated(message))
|
||||||
|
# else
|
||||||
|
# pragma message("WARNING: You need to implement ZBUFF_DEPRECATED for this compiler")
|
||||||
|
# define ZBUFF_DEPRECATED(message) ZSTDLIB_API
|
||||||
|
# endif
|
||||||
|
#endif /* ZBUFF_DISABLE_DEPRECATE_WARNINGS */
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Streaming functions
|
||||||
|
***************************************/
|
||||||
|
/* This is the easier "buffered" streaming API,
|
||||||
|
* using an internal buffer to lift all restrictions on user-provided buffers
|
||||||
|
* which can be any size, any place, for both input and output.
|
||||||
|
* ZBUFF and ZSTD are 100% interoperable,
|
||||||
|
* frames created by one can be decoded by the other one */
|
||||||
|
|
||||||
|
typedef ZSTD_CStream ZBUFF_CCtx;
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_createCStream") ZBUFF_CCtx* ZBUFF_createCCtx(void);
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_freeCStream") size_t ZBUFF_freeCCtx(ZBUFF_CCtx* cctx);
|
||||||
|
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_initCStream") size_t ZBUFF_compressInit(ZBUFF_CCtx* cctx, int compressionLevel);
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_initCStream_usingDict") size_t ZBUFF_compressInitDictionary(ZBUFF_CCtx* cctx, const void* dict, size_t dictSize, int compressionLevel);
|
||||||
|
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_compressStream") size_t ZBUFF_compressContinue(ZBUFF_CCtx* cctx, void* dst, size_t* dstCapacityPtr, const void* src, size_t* srcSizePtr);
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_flushStream") size_t ZBUFF_compressFlush(ZBUFF_CCtx* cctx, void* dst, size_t* dstCapacityPtr);
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_endStream") size_t ZBUFF_compressEnd(ZBUFF_CCtx* cctx, void* dst, size_t* dstCapacityPtr);
|
||||||
|
|
||||||
|
/*-*************************************************
|
||||||
|
* Streaming compression - howto
|
||||||
|
*
|
||||||
|
* A ZBUFF_CCtx object is required to track streaming operation.
|
||||||
|
* Use ZBUFF_createCCtx() and ZBUFF_freeCCtx() to create/release resources.
|
||||||
|
* ZBUFF_CCtx objects can be reused multiple times.
|
||||||
|
*
|
||||||
|
* Start by initializing ZBUF_CCtx.
|
||||||
|
* Use ZBUFF_compressInit() to start a new compression operation.
|
||||||
|
* Use ZBUFF_compressInitDictionary() for a compression which requires a dictionary.
|
||||||
|
*
|
||||||
|
* Use ZBUFF_compressContinue() repetitively to consume input stream.
|
||||||
|
* *srcSizePtr and *dstCapacityPtr can be any size.
|
||||||
|
* The function will report how many bytes were read or written within *srcSizePtr and *dstCapacityPtr.
|
||||||
|
* Note that it may not consume the entire input, in which case it's up to the caller to present again remaining data.
|
||||||
|
* The content of `dst` will be overwritten (up to *dstCapacityPtr) at each call, so save its content if it matters or change @dst .
|
||||||
|
* @return : a hint to preferred nb of bytes to use as input for next function call (it's just a hint, to improve latency)
|
||||||
|
* or an error code, which can be tested using ZBUFF_isError().
|
||||||
|
*
|
||||||
|
* At any moment, it's possible to flush whatever data remains within buffer, using ZBUFF_compressFlush().
|
||||||
|
* The nb of bytes written into `dst` will be reported into *dstCapacityPtr.
|
||||||
|
* Note that the function cannot output more than *dstCapacityPtr,
|
||||||
|
* therefore, some content might still be left into internal buffer if *dstCapacityPtr is too small.
|
||||||
|
* @return : nb of bytes still present into internal buffer (0 if it's empty)
|
||||||
|
* or an error code, which can be tested using ZBUFF_isError().
|
||||||
|
*
|
||||||
|
* ZBUFF_compressEnd() instructs to finish a frame.
|
||||||
|
* It will perform a flush and write frame epilogue.
|
||||||
|
* The epilogue is required for decoders to consider a frame completed.
|
||||||
|
* Similar to ZBUFF_compressFlush(), it may not be able to output the entire internal buffer content if *dstCapacityPtr is too small.
|
||||||
|
* In which case, call again ZBUFF_compressFlush() to complete the flush.
|
||||||
|
* @return : nb of bytes still present into internal buffer (0 if it's empty)
|
||||||
|
* or an error code, which can be tested using ZBUFF_isError().
|
||||||
|
*
|
||||||
|
* Hint : _recommended buffer_ sizes (not compulsory) : ZBUFF_recommendedCInSize() / ZBUFF_recommendedCOutSize()
|
||||||
|
* input : ZBUFF_recommendedCInSize==128 KB block size is the internal unit, use this value to reduce intermediate stages (better latency)
|
||||||
|
* output : ZBUFF_recommendedCOutSize==ZSTD_compressBound(128 KB) + 3 + 3 : ensures it's always possible to write/flush/end a full block. Skip some buffering.
|
||||||
|
* By using both, it ensures that input will be entirely consumed, and output will always contain the result, reducing intermediate buffering.
|
||||||
|
* **************************************************/
|
||||||
|
|
||||||
|
|
||||||
|
typedef ZSTD_DStream ZBUFF_DCtx;
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_createDStream") ZBUFF_DCtx* ZBUFF_createDCtx(void);
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_freeDStream") size_t ZBUFF_freeDCtx(ZBUFF_DCtx* dctx);
|
||||||
|
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_initDStream") size_t ZBUFF_decompressInit(ZBUFF_DCtx* dctx);
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_initDStream_usingDict") size_t ZBUFF_decompressInitDictionary(ZBUFF_DCtx* dctx, const void* dict, size_t dictSize);
|
||||||
|
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_decompressStream") size_t ZBUFF_decompressContinue(ZBUFF_DCtx* dctx,
|
||||||
|
void* dst, size_t* dstCapacityPtr,
|
||||||
|
const void* src, size_t* srcSizePtr);
|
||||||
|
|
||||||
|
/*-***************************************************************************
|
||||||
|
* Streaming decompression howto
|
||||||
|
*
|
||||||
|
* A ZBUFF_DCtx object is required to track streaming operations.
|
||||||
|
* Use ZBUFF_createDCtx() and ZBUFF_freeDCtx() to create/release resources.
|
||||||
|
* Use ZBUFF_decompressInit() to start a new decompression operation,
|
||||||
|
* or ZBUFF_decompressInitDictionary() if decompression requires a dictionary.
|
||||||
|
* Note that ZBUFF_DCtx objects can be re-init multiple times.
|
||||||
|
*
|
||||||
|
* Use ZBUFF_decompressContinue() repetitively to consume your input.
|
||||||
|
* *srcSizePtr and *dstCapacityPtr can be any size.
|
||||||
|
* The function will report how many bytes were read or written by modifying *srcSizePtr and *dstCapacityPtr.
|
||||||
|
* Note that it may not consume the entire input, in which case it's up to the caller to present remaining input again.
|
||||||
|
* The content of `dst` will be overwritten (up to *dstCapacityPtr) at each function call, so save its content if it matters, or change `dst`.
|
||||||
|
* @return : 0 when a frame is completely decoded and fully flushed,
|
||||||
|
* 1 when there is still some data left within internal buffer to flush,
|
||||||
|
* >1 when more data is expected, with value being a suggested next input size (it's just a hint, which helps latency),
|
||||||
|
* or an error code, which can be tested using ZBUFF_isError().
|
||||||
|
*
|
||||||
|
* Hint : recommended buffer sizes (not compulsory) : ZBUFF_recommendedDInSize() and ZBUFF_recommendedDOutSize()
|
||||||
|
* output : ZBUFF_recommendedDOutSize== 128 KB block size is the internal unit, it ensures it's always possible to write a full block when decoded.
|
||||||
|
* input : ZBUFF_recommendedDInSize == 128KB + 3;
|
||||||
|
* just follow indications from ZBUFF_decompressContinue() to minimize latency. It should always be <= 128 KB + 3 .
|
||||||
|
* *******************************************************************************/
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Tool functions
|
||||||
|
***************************************/
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_isError") unsigned ZBUFF_isError(size_t errorCode);
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_getErrorName") const char* ZBUFF_getErrorName(size_t errorCode);
|
||||||
|
|
||||||
|
/** Functions below provide recommended buffer sizes for Compression or Decompression operations.
|
||||||
|
* These sizes are just hints, they tend to offer better latency */
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_CStreamInSize") size_t ZBUFF_recommendedCInSize(void);
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_CStreamOutSize") size_t ZBUFF_recommendedCOutSize(void);
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_DStreamInSize") size_t ZBUFF_recommendedDInSize(void);
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_DStreamOutSize") size_t ZBUFF_recommendedDOutSize(void);
|
||||||
|
|
||||||
|
#endif /* ZSTD_BUFFERED_H_23987 */
|
||||||
|
|
||||||
|
|
||||||
|
#ifdef ZBUFF_STATIC_LINKING_ONLY
|
||||||
|
#ifndef ZBUFF_STATIC_H_30298098432
|
||||||
|
#define ZBUFF_STATIC_H_30298098432
|
||||||
|
|
||||||
|
/* ====================================================================================
|
||||||
|
* The definitions in this section are considered experimental.
|
||||||
|
* They should never be used in association with a dynamic library, as they may change in the future.
|
||||||
|
* They are provided for advanced usages.
|
||||||
|
* Use them only in association with static linking.
|
||||||
|
* ==================================================================================== */
|
||||||
|
|
||||||
|
/*--- Dependency ---*/
|
||||||
|
#define ZSTD_STATIC_LINKING_ONLY /* ZSTD_parameters, ZSTD_customMem */
|
||||||
|
#include "zstd.h"
|
||||||
|
|
||||||
|
|
||||||
|
/*--- Custom memory allocator ---*/
|
||||||
|
/*! ZBUFF_createCCtx_advanced() :
|
||||||
|
* Create a ZBUFF compression context using external alloc and free functions */
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_createCStream_advanced") ZBUFF_CCtx* ZBUFF_createCCtx_advanced(ZSTD_customMem customMem);
|
||||||
|
|
||||||
|
/*! ZBUFF_createDCtx_advanced() :
|
||||||
|
* Create a ZBUFF decompression context using external alloc and free functions */
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_createDStream_advanced") ZBUFF_DCtx* ZBUFF_createDCtx_advanced(ZSTD_customMem customMem);
|
||||||
|
|
||||||
|
|
||||||
|
/*--- Advanced Streaming Initialization ---*/
|
||||||
|
ZBUFF_DEPRECATED("use ZSTD_initDStream_usingDict") size_t ZBUFF_compressInit_advanced(ZBUFF_CCtx* zbc,
|
||||||
|
const void* dict, size_t dictSize,
|
||||||
|
ZSTD_parameters params, unsigned long long pledgedSrcSize);
|
||||||
|
|
||||||
|
|
||||||
|
#endif /* ZBUFF_STATIC_H_30298098432 */
|
||||||
|
#endif /* ZBUFF_STATIC_LINKING_ONLY */
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
26
vendor/github.com/DataDog/zstd/zbuff_common.c
generated
vendored
Normal file
26
vendor/github.com/DataDog/zstd/zbuff_common.c
generated
vendored
Normal file
@@ -0,0 +1,26 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Dependencies
|
||||||
|
***************************************/
|
||||||
|
#include "error_private.h"
|
||||||
|
#include "zbuff.h"
|
||||||
|
|
||||||
|
/*-****************************************
|
||||||
|
* ZBUFF Error Management (deprecated)
|
||||||
|
******************************************/
|
||||||
|
|
||||||
|
/*! ZBUFF_isError() :
|
||||||
|
* tells if a return value is an error code */
|
||||||
|
unsigned ZBUFF_isError(size_t errorCode) { return ERR_isError(errorCode); }
|
||||||
|
/*! ZBUFF_getErrorName() :
|
||||||
|
* provides error code string from function result (useful for debugging) */
|
||||||
|
const char* ZBUFF_getErrorName(size_t errorCode) { return ERR_getErrorName(errorCode); }
|
||||||
147
vendor/github.com/DataDog/zstd/zbuff_compress.c
generated
vendored
Normal file
147
vendor/github.com/DataDog/zstd/zbuff_compress.c
generated
vendored
Normal file
@@ -0,0 +1,147 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Dependencies
|
||||||
|
***************************************/
|
||||||
|
#define ZBUFF_STATIC_LINKING_ONLY
|
||||||
|
#include "zbuff.h"
|
||||||
|
|
||||||
|
|
||||||
|
/*-***********************************************************
|
||||||
|
* Streaming compression
|
||||||
|
*
|
||||||
|
* A ZBUFF_CCtx object is required to track streaming operation.
|
||||||
|
* Use ZBUFF_createCCtx() and ZBUFF_freeCCtx() to create/release resources.
|
||||||
|
* Use ZBUFF_compressInit() to start a new compression operation.
|
||||||
|
* ZBUFF_CCtx objects can be reused multiple times.
|
||||||
|
*
|
||||||
|
* Use ZBUFF_compressContinue() repetitively to consume your input.
|
||||||
|
* *srcSizePtr and *dstCapacityPtr can be any size.
|
||||||
|
* The function will report how many bytes were read or written by modifying *srcSizePtr and *dstCapacityPtr.
|
||||||
|
* Note that it may not consume the entire input, in which case it's up to the caller to call again the function with remaining input.
|
||||||
|
* The content of dst will be overwritten (up to *dstCapacityPtr) at each function call, so save its content if it matters or change dst .
|
||||||
|
* @return : a hint to preferred nb of bytes to use as input for next function call (it's only a hint, to improve latency)
|
||||||
|
* or an error code, which can be tested using ZBUFF_isError().
|
||||||
|
*
|
||||||
|
* ZBUFF_compressFlush() can be used to instruct ZBUFF to compress and output whatever remains within its buffer.
|
||||||
|
* Note that it will not output more than *dstCapacityPtr.
|
||||||
|
* Therefore, some content might still be left into its internal buffer if dst buffer is too small.
|
||||||
|
* @return : nb of bytes still present into internal buffer (0 if it's empty)
|
||||||
|
* or an error code, which can be tested using ZBUFF_isError().
|
||||||
|
*
|
||||||
|
* ZBUFF_compressEnd() instructs to finish a frame.
|
||||||
|
* It will perform a flush and write frame epilogue.
|
||||||
|
* Similar to ZBUFF_compressFlush(), it may not be able to output the entire internal buffer content if *dstCapacityPtr is too small.
|
||||||
|
* @return : nb of bytes still present into internal buffer (0 if it's empty)
|
||||||
|
* or an error code, which can be tested using ZBUFF_isError().
|
||||||
|
*
|
||||||
|
* Hint : recommended buffer sizes (not compulsory)
|
||||||
|
* input : ZSTD_BLOCKSIZE_MAX (128 KB), internal unit size, it improves latency to use this value.
|
||||||
|
* output : ZSTD_compressBound(ZSTD_BLOCKSIZE_MAX) + ZSTD_blockHeaderSize + ZBUFF_endFrameSize : ensures it's always possible to write/flush/end a full block at best speed.
|
||||||
|
* ***********************************************************/
|
||||||
|
|
||||||
|
ZBUFF_CCtx* ZBUFF_createCCtx(void)
|
||||||
|
{
|
||||||
|
return ZSTD_createCStream();
|
||||||
|
}
|
||||||
|
|
||||||
|
ZBUFF_CCtx* ZBUFF_createCCtx_advanced(ZSTD_customMem customMem)
|
||||||
|
{
|
||||||
|
return ZSTD_createCStream_advanced(customMem);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZBUFF_freeCCtx(ZBUFF_CCtx* zbc)
|
||||||
|
{
|
||||||
|
return ZSTD_freeCStream(zbc);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* ====== Initialization ====== */
|
||||||
|
|
||||||
|
size_t ZBUFF_compressInit_advanced(ZBUFF_CCtx* zbc,
|
||||||
|
const void* dict, size_t dictSize,
|
||||||
|
ZSTD_parameters params, unsigned long long pledgedSrcSize)
|
||||||
|
{
|
||||||
|
if (pledgedSrcSize==0) pledgedSrcSize = ZSTD_CONTENTSIZE_UNKNOWN; /* preserve "0 == unknown" behavior */
|
||||||
|
return ZSTD_initCStream_advanced(zbc, dict, dictSize, params, pledgedSrcSize);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t ZBUFF_compressInitDictionary(ZBUFF_CCtx* zbc, const void* dict, size_t dictSize, int compressionLevel)
|
||||||
|
{
|
||||||
|
return ZSTD_initCStream_usingDict(zbc, dict, dictSize, compressionLevel);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZBUFF_compressInit(ZBUFF_CCtx* zbc, int compressionLevel)
|
||||||
|
{
|
||||||
|
return ZSTD_initCStream(zbc, compressionLevel);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ====== Compression ====== */
|
||||||
|
|
||||||
|
|
||||||
|
size_t ZBUFF_compressContinue(ZBUFF_CCtx* zbc,
|
||||||
|
void* dst, size_t* dstCapacityPtr,
|
||||||
|
const void* src, size_t* srcSizePtr)
|
||||||
|
{
|
||||||
|
size_t result;
|
||||||
|
ZSTD_outBuffer outBuff;
|
||||||
|
ZSTD_inBuffer inBuff;
|
||||||
|
outBuff.dst = dst;
|
||||||
|
outBuff.pos = 0;
|
||||||
|
outBuff.size = *dstCapacityPtr;
|
||||||
|
inBuff.src = src;
|
||||||
|
inBuff.pos = 0;
|
||||||
|
inBuff.size = *srcSizePtr;
|
||||||
|
result = ZSTD_compressStream(zbc, &outBuff, &inBuff);
|
||||||
|
*dstCapacityPtr = outBuff.pos;
|
||||||
|
*srcSizePtr = inBuff.pos;
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/* ====== Finalize ====== */
|
||||||
|
|
||||||
|
size_t ZBUFF_compressFlush(ZBUFF_CCtx* zbc, void* dst, size_t* dstCapacityPtr)
|
||||||
|
{
|
||||||
|
size_t result;
|
||||||
|
ZSTD_outBuffer outBuff;
|
||||||
|
outBuff.dst = dst;
|
||||||
|
outBuff.pos = 0;
|
||||||
|
outBuff.size = *dstCapacityPtr;
|
||||||
|
result = ZSTD_flushStream(zbc, &outBuff);
|
||||||
|
*dstCapacityPtr = outBuff.pos;
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t ZBUFF_compressEnd(ZBUFF_CCtx* zbc, void* dst, size_t* dstCapacityPtr)
|
||||||
|
{
|
||||||
|
size_t result;
|
||||||
|
ZSTD_outBuffer outBuff;
|
||||||
|
outBuff.dst = dst;
|
||||||
|
outBuff.pos = 0;
|
||||||
|
outBuff.size = *dstCapacityPtr;
|
||||||
|
result = ZSTD_endStream(zbc, &outBuff);
|
||||||
|
*dstCapacityPtr = outBuff.pos;
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Tool functions
|
||||||
|
***************************************/
|
||||||
|
size_t ZBUFF_recommendedCInSize(void) { return ZSTD_CStreamInSize(); }
|
||||||
|
size_t ZBUFF_recommendedCOutSize(void) { return ZSTD_CStreamOutSize(); }
|
||||||
75
vendor/github.com/DataDog/zstd/zbuff_decompress.c
generated
vendored
Normal file
75
vendor/github.com/DataDog/zstd/zbuff_decompress.c
generated
vendored
Normal file
@@ -0,0 +1,75 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Dependencies
|
||||||
|
***************************************/
|
||||||
|
#define ZBUFF_STATIC_LINKING_ONLY
|
||||||
|
#include "zbuff.h"
|
||||||
|
|
||||||
|
|
||||||
|
ZBUFF_DCtx* ZBUFF_createDCtx(void)
|
||||||
|
{
|
||||||
|
return ZSTD_createDStream();
|
||||||
|
}
|
||||||
|
|
||||||
|
ZBUFF_DCtx* ZBUFF_createDCtx_advanced(ZSTD_customMem customMem)
|
||||||
|
{
|
||||||
|
return ZSTD_createDStream_advanced(customMem);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZBUFF_freeDCtx(ZBUFF_DCtx* zbd)
|
||||||
|
{
|
||||||
|
return ZSTD_freeDStream(zbd);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* *** Initialization *** */
|
||||||
|
|
||||||
|
size_t ZBUFF_decompressInitDictionary(ZBUFF_DCtx* zbd, const void* dict, size_t dictSize)
|
||||||
|
{
|
||||||
|
return ZSTD_initDStream_usingDict(zbd, dict, dictSize);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZBUFF_decompressInit(ZBUFF_DCtx* zbd)
|
||||||
|
{
|
||||||
|
return ZSTD_initDStream(zbd);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* *** Decompression *** */
|
||||||
|
|
||||||
|
size_t ZBUFF_decompressContinue(ZBUFF_DCtx* zbd,
|
||||||
|
void* dst, size_t* dstCapacityPtr,
|
||||||
|
const void* src, size_t* srcSizePtr)
|
||||||
|
{
|
||||||
|
ZSTD_outBuffer outBuff;
|
||||||
|
ZSTD_inBuffer inBuff;
|
||||||
|
size_t result;
|
||||||
|
outBuff.dst = dst;
|
||||||
|
outBuff.pos = 0;
|
||||||
|
outBuff.size = *dstCapacityPtr;
|
||||||
|
inBuff.src = src;
|
||||||
|
inBuff.pos = 0;
|
||||||
|
inBuff.size = *srcSizePtr;
|
||||||
|
result = ZSTD_decompressStream(zbd, &outBuff, &inBuff);
|
||||||
|
*dstCapacityPtr = outBuff.pos;
|
||||||
|
*srcSizePtr = inBuff.pos;
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Tool functions
|
||||||
|
***************************************/
|
||||||
|
size_t ZBUFF_recommendedDInSize(void) { return ZSTD_DStreamInSize(); }
|
||||||
|
size_t ZBUFF_recommendedDOutSize(void) { return ZSTD_DStreamOutSize(); }
|
||||||
1111
vendor/github.com/DataDog/zstd/zdict.c
generated
vendored
Normal file
1111
vendor/github.com/DataDog/zstd/zdict.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
282
vendor/github.com/DataDog/zstd/zdict.h
generated
vendored
Normal file
282
vendor/github.com/DataDog/zstd/zdict.h
generated
vendored
Normal file
@@ -0,0 +1,282 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef DICTBUILDER_H_001
|
||||||
|
#define DICTBUILDER_H_001
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/*====== Dependencies ======*/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
|
||||||
|
|
||||||
|
/* ===== ZDICTLIB_API : control library symbols visibility ===== */
|
||||||
|
#ifndef ZDICTLIB_VISIBILITY
|
||||||
|
# if defined(__GNUC__) && (__GNUC__ >= 4)
|
||||||
|
# define ZDICTLIB_VISIBILITY __attribute__ ((visibility ("default")))
|
||||||
|
# else
|
||||||
|
# define ZDICTLIB_VISIBILITY
|
||||||
|
# endif
|
||||||
|
#endif
|
||||||
|
#if defined(ZSTD_DLL_EXPORT) && (ZSTD_DLL_EXPORT==1)
|
||||||
|
# define ZDICTLIB_API __declspec(dllexport) ZDICTLIB_VISIBILITY
|
||||||
|
#elif defined(ZSTD_DLL_IMPORT) && (ZSTD_DLL_IMPORT==1)
|
||||||
|
# define ZDICTLIB_API __declspec(dllimport) ZDICTLIB_VISIBILITY /* It isn't required but allows to generate better code, saving a function pointer load from the IAT and an indirect jump.*/
|
||||||
|
#else
|
||||||
|
# define ZDICTLIB_API ZDICTLIB_VISIBILITY
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/*! ZDICT_trainFromBuffer():
|
||||||
|
* Train a dictionary from an array of samples.
|
||||||
|
* Redirect towards ZDICT_optimizeTrainFromBuffer_fastCover() single-threaded, with d=8, steps=4,
|
||||||
|
* f=20, and accel=1.
|
||||||
|
* Samples must be stored concatenated in a single flat buffer `samplesBuffer`,
|
||||||
|
* supplied with an array of sizes `samplesSizes`, providing the size of each sample, in order.
|
||||||
|
* The resulting dictionary will be saved into `dictBuffer`.
|
||||||
|
* @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`)
|
||||||
|
* or an error code, which can be tested with ZDICT_isError().
|
||||||
|
* Note: Dictionary training will fail if there are not enough samples to construct a
|
||||||
|
* dictionary, or if most of the samples are too small (< 8 bytes being the lower limit).
|
||||||
|
* If dictionary training fails, you should use zstd without a dictionary, as the dictionary
|
||||||
|
* would've been ineffective anyways. If you believe your samples would benefit from a dictionary
|
||||||
|
* please open an issue with details, and we can look into it.
|
||||||
|
* Note: ZDICT_trainFromBuffer()'s memory usage is about 6 MB.
|
||||||
|
* Tips: In general, a reasonable dictionary has a size of ~ 100 KB.
|
||||||
|
* It's possible to select smaller or larger size, just by specifying `dictBufferCapacity`.
|
||||||
|
* In general, it's recommended to provide a few thousands samples, though this can vary a lot.
|
||||||
|
* It's recommended that total size of all samples be about ~x100 times the target size of dictionary.
|
||||||
|
*/
|
||||||
|
ZDICTLIB_API size_t ZDICT_trainFromBuffer(void* dictBuffer, size_t dictBufferCapacity,
|
||||||
|
const void* samplesBuffer,
|
||||||
|
const size_t* samplesSizes, unsigned nbSamples);
|
||||||
|
|
||||||
|
|
||||||
|
/*====== Helper functions ======*/
|
||||||
|
ZDICTLIB_API unsigned ZDICT_getDictID(const void* dictBuffer, size_t dictSize); /**< extracts dictID; @return zero if error (not a valid dictionary) */
|
||||||
|
ZDICTLIB_API unsigned ZDICT_isError(size_t errorCode);
|
||||||
|
ZDICTLIB_API const char* ZDICT_getErrorName(size_t errorCode);
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
#ifdef ZDICT_STATIC_LINKING_ONLY
|
||||||
|
|
||||||
|
/* ====================================================================================
|
||||||
|
* The definitions in this section are considered experimental.
|
||||||
|
* They should never be used with a dynamic library, as they may change in the future.
|
||||||
|
* They are provided for advanced usages.
|
||||||
|
* Use them only in association with static linking.
|
||||||
|
* ==================================================================================== */
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
int compressionLevel; /* optimize for a specific zstd compression level; 0 means default */
|
||||||
|
unsigned notificationLevel; /* Write log to stderr; 0 = none (default); 1 = errors; 2 = progression; 3 = details; 4 = debug; */
|
||||||
|
unsigned dictID; /* force dictID value; 0 means auto mode (32-bits random value) */
|
||||||
|
} ZDICT_params_t;
|
||||||
|
|
||||||
|
/*! ZDICT_cover_params_t:
|
||||||
|
* k and d are the only required parameters.
|
||||||
|
* For others, value 0 means default.
|
||||||
|
*/
|
||||||
|
typedef struct {
|
||||||
|
unsigned k; /* Segment size : constraint: 0 < k : Reasonable range [16, 2048+] */
|
||||||
|
unsigned d; /* dmer size : constraint: 0 < d <= k : Reasonable range [6, 16] */
|
||||||
|
unsigned steps; /* Number of steps : Only used for optimization : 0 means default (40) : Higher means more parameters checked */
|
||||||
|
unsigned nbThreads; /* Number of threads : constraint: 0 < nbThreads : 1 means single-threaded : Only used for optimization : Ignored if ZSTD_MULTITHREAD is not defined */
|
||||||
|
double splitPoint; /* Percentage of samples used for training: Only used for optimization : the first nbSamples * splitPoint samples will be used to training, the last nbSamples * (1 - splitPoint) samples will be used for testing, 0 means default (1.0), 1.0 when all samples are used for both training and testing */
|
||||||
|
unsigned shrinkDict; /* Train dictionaries to shrink in size starting from the minimum size and selects the smallest dictionary that is shrinkDictMaxRegression% worse than the largest dictionary. 0 means no shrinking and 1 means shrinking */
|
||||||
|
unsigned shrinkDictMaxRegression; /* Sets shrinkDictMaxRegression so that a smaller dictionary can be at worse shrinkDictMaxRegression% worse than the max dict size dictionary. */
|
||||||
|
ZDICT_params_t zParams;
|
||||||
|
} ZDICT_cover_params_t;
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
unsigned k; /* Segment size : constraint: 0 < k : Reasonable range [16, 2048+] */
|
||||||
|
unsigned d; /* dmer size : constraint: 0 < d <= k : Reasonable range [6, 16] */
|
||||||
|
unsigned f; /* log of size of frequency array : constraint: 0 < f <= 31 : 1 means default(20)*/
|
||||||
|
unsigned steps; /* Number of steps : Only used for optimization : 0 means default (40) : Higher means more parameters checked */
|
||||||
|
unsigned nbThreads; /* Number of threads : constraint: 0 < nbThreads : 1 means single-threaded : Only used for optimization : Ignored if ZSTD_MULTITHREAD is not defined */
|
||||||
|
double splitPoint; /* Percentage of samples used for training: Only used for optimization : the first nbSamples * splitPoint samples will be used to training, the last nbSamples * (1 - splitPoint) samples will be used for testing, 0 means default (0.75), 1.0 when all samples are used for both training and testing */
|
||||||
|
unsigned accel; /* Acceleration level: constraint: 0 < accel <= 10, higher means faster and less accurate, 0 means default(1) */
|
||||||
|
unsigned shrinkDict; /* Train dictionaries to shrink in size starting from the minimum size and selects the smallest dictionary that is shrinkDictMaxRegression% worse than the largest dictionary. 0 means no shrinking and 1 means shrinking */
|
||||||
|
unsigned shrinkDictMaxRegression; /* Sets shrinkDictMaxRegression so that a smaller dictionary can be at worse shrinkDictMaxRegression% worse than the max dict size dictionary. */
|
||||||
|
|
||||||
|
ZDICT_params_t zParams;
|
||||||
|
} ZDICT_fastCover_params_t;
|
||||||
|
|
||||||
|
/*! ZDICT_trainFromBuffer_cover():
|
||||||
|
* Train a dictionary from an array of samples using the COVER algorithm.
|
||||||
|
* Samples must be stored concatenated in a single flat buffer `samplesBuffer`,
|
||||||
|
* supplied with an array of sizes `samplesSizes`, providing the size of each sample, in order.
|
||||||
|
* The resulting dictionary will be saved into `dictBuffer`.
|
||||||
|
* @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`)
|
||||||
|
* or an error code, which can be tested with ZDICT_isError().
|
||||||
|
* See ZDICT_trainFromBuffer() for details on failure modes.
|
||||||
|
* Note: ZDICT_trainFromBuffer_cover() requires about 9 bytes of memory for each input byte.
|
||||||
|
* Tips: In general, a reasonable dictionary has a size of ~ 100 KB.
|
||||||
|
* It's possible to select smaller or larger size, just by specifying `dictBufferCapacity`.
|
||||||
|
* In general, it's recommended to provide a few thousands samples, though this can vary a lot.
|
||||||
|
* It's recommended that total size of all samples be about ~x100 times the target size of dictionary.
|
||||||
|
*/
|
||||||
|
ZDICTLIB_API size_t ZDICT_trainFromBuffer_cover(
|
||||||
|
void *dictBuffer, size_t dictBufferCapacity,
|
||||||
|
const void *samplesBuffer, const size_t *samplesSizes, unsigned nbSamples,
|
||||||
|
ZDICT_cover_params_t parameters);
|
||||||
|
|
||||||
|
/*! ZDICT_optimizeTrainFromBuffer_cover():
|
||||||
|
* The same requirements as above hold for all the parameters except `parameters`.
|
||||||
|
* This function tries many parameter combinations and picks the best parameters.
|
||||||
|
* `*parameters` is filled with the best parameters found,
|
||||||
|
* dictionary constructed with those parameters is stored in `dictBuffer`.
|
||||||
|
*
|
||||||
|
* All of the parameters d, k, steps are optional.
|
||||||
|
* If d is non-zero then we don't check multiple values of d, otherwise we check d = {6, 8}.
|
||||||
|
* if steps is zero it defaults to its default value.
|
||||||
|
* If k is non-zero then we don't check multiple values of k, otherwise we check steps values in [50, 2000].
|
||||||
|
*
|
||||||
|
* @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`)
|
||||||
|
* or an error code, which can be tested with ZDICT_isError().
|
||||||
|
* On success `*parameters` contains the parameters selected.
|
||||||
|
* See ZDICT_trainFromBuffer() for details on failure modes.
|
||||||
|
* Note: ZDICT_optimizeTrainFromBuffer_cover() requires about 8 bytes of memory for each input byte and additionally another 5 bytes of memory for each byte of memory for each thread.
|
||||||
|
*/
|
||||||
|
ZDICTLIB_API size_t ZDICT_optimizeTrainFromBuffer_cover(
|
||||||
|
void* dictBuffer, size_t dictBufferCapacity,
|
||||||
|
const void* samplesBuffer, const size_t* samplesSizes, unsigned nbSamples,
|
||||||
|
ZDICT_cover_params_t* parameters);
|
||||||
|
|
||||||
|
/*! ZDICT_trainFromBuffer_fastCover():
|
||||||
|
* Train a dictionary from an array of samples using a modified version of COVER algorithm.
|
||||||
|
* Samples must be stored concatenated in a single flat buffer `samplesBuffer`,
|
||||||
|
* supplied with an array of sizes `samplesSizes`, providing the size of each sample, in order.
|
||||||
|
* d and k are required.
|
||||||
|
* All other parameters are optional, will use default values if not provided
|
||||||
|
* The resulting dictionary will be saved into `dictBuffer`.
|
||||||
|
* @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`)
|
||||||
|
* or an error code, which can be tested with ZDICT_isError().
|
||||||
|
* See ZDICT_trainFromBuffer() for details on failure modes.
|
||||||
|
* Note: ZDICT_trainFromBuffer_fastCover() requires 6 * 2^f bytes of memory.
|
||||||
|
* Tips: In general, a reasonable dictionary has a size of ~ 100 KB.
|
||||||
|
* It's possible to select smaller or larger size, just by specifying `dictBufferCapacity`.
|
||||||
|
* In general, it's recommended to provide a few thousands samples, though this can vary a lot.
|
||||||
|
* It's recommended that total size of all samples be about ~x100 times the target size of dictionary.
|
||||||
|
*/
|
||||||
|
ZDICTLIB_API size_t ZDICT_trainFromBuffer_fastCover(void *dictBuffer,
|
||||||
|
size_t dictBufferCapacity, const void *samplesBuffer,
|
||||||
|
const size_t *samplesSizes, unsigned nbSamples,
|
||||||
|
ZDICT_fastCover_params_t parameters);
|
||||||
|
|
||||||
|
/*! ZDICT_optimizeTrainFromBuffer_fastCover():
|
||||||
|
* The same requirements as above hold for all the parameters except `parameters`.
|
||||||
|
* This function tries many parameter combinations (specifically, k and d combinations)
|
||||||
|
* and picks the best parameters. `*parameters` is filled with the best parameters found,
|
||||||
|
* dictionary constructed with those parameters is stored in `dictBuffer`.
|
||||||
|
* All of the parameters d, k, steps, f, and accel are optional.
|
||||||
|
* If d is non-zero then we don't check multiple values of d, otherwise we check d = {6, 8}.
|
||||||
|
* if steps is zero it defaults to its default value.
|
||||||
|
* If k is non-zero then we don't check multiple values of k, otherwise we check steps values in [50, 2000].
|
||||||
|
* If f is zero, default value of 20 is used.
|
||||||
|
* If accel is zero, default value of 1 is used.
|
||||||
|
*
|
||||||
|
* @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`)
|
||||||
|
* or an error code, which can be tested with ZDICT_isError().
|
||||||
|
* On success `*parameters` contains the parameters selected.
|
||||||
|
* See ZDICT_trainFromBuffer() for details on failure modes.
|
||||||
|
* Note: ZDICT_optimizeTrainFromBuffer_fastCover() requires about 6 * 2^f bytes of memory for each thread.
|
||||||
|
*/
|
||||||
|
ZDICTLIB_API size_t ZDICT_optimizeTrainFromBuffer_fastCover(void* dictBuffer,
|
||||||
|
size_t dictBufferCapacity, const void* samplesBuffer,
|
||||||
|
const size_t* samplesSizes, unsigned nbSamples,
|
||||||
|
ZDICT_fastCover_params_t* parameters);
|
||||||
|
|
||||||
|
/*! ZDICT_finalizeDictionary():
|
||||||
|
* Given a custom content as a basis for dictionary, and a set of samples,
|
||||||
|
* finalize dictionary by adding headers and statistics.
|
||||||
|
*
|
||||||
|
* Samples must be stored concatenated in a flat buffer `samplesBuffer`,
|
||||||
|
* supplied with an array of sizes `samplesSizes`, providing the size of each sample in order.
|
||||||
|
*
|
||||||
|
* dictContentSize must be >= ZDICT_CONTENTSIZE_MIN bytes.
|
||||||
|
* maxDictSize must be >= dictContentSize, and must be >= ZDICT_DICTSIZE_MIN bytes.
|
||||||
|
*
|
||||||
|
* @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`),
|
||||||
|
* or an error code, which can be tested by ZDICT_isError().
|
||||||
|
* Note: ZDICT_finalizeDictionary() will push notifications into stderr if instructed to, using notificationLevel>0.
|
||||||
|
* Note 2: dictBuffer and dictContent can overlap
|
||||||
|
*/
|
||||||
|
#define ZDICT_CONTENTSIZE_MIN 128
|
||||||
|
#define ZDICT_DICTSIZE_MIN 256
|
||||||
|
ZDICTLIB_API size_t ZDICT_finalizeDictionary(void* dictBuffer, size_t dictBufferCapacity,
|
||||||
|
const void* dictContent, size_t dictContentSize,
|
||||||
|
const void* samplesBuffer, const size_t* samplesSizes, unsigned nbSamples,
|
||||||
|
ZDICT_params_t parameters);
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
unsigned selectivityLevel; /* 0 means default; larger => select more => larger dictionary */
|
||||||
|
ZDICT_params_t zParams;
|
||||||
|
} ZDICT_legacy_params_t;
|
||||||
|
|
||||||
|
/*! ZDICT_trainFromBuffer_legacy():
|
||||||
|
* Train a dictionary from an array of samples.
|
||||||
|
* Samples must be stored concatenated in a single flat buffer `samplesBuffer`,
|
||||||
|
* supplied with an array of sizes `samplesSizes`, providing the size of each sample, in order.
|
||||||
|
* The resulting dictionary will be saved into `dictBuffer`.
|
||||||
|
* `parameters` is optional and can be provided with values set to 0 to mean "default".
|
||||||
|
* @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`)
|
||||||
|
* or an error code, which can be tested with ZDICT_isError().
|
||||||
|
* See ZDICT_trainFromBuffer() for details on failure modes.
|
||||||
|
* Tips: In general, a reasonable dictionary has a size of ~ 100 KB.
|
||||||
|
* It's possible to select smaller or larger size, just by specifying `dictBufferCapacity`.
|
||||||
|
* In general, it's recommended to provide a few thousands samples, though this can vary a lot.
|
||||||
|
* It's recommended that total size of all samples be about ~x100 times the target size of dictionary.
|
||||||
|
* Note: ZDICT_trainFromBuffer_legacy() will send notifications into stderr if instructed to, using notificationLevel>0.
|
||||||
|
*/
|
||||||
|
ZDICTLIB_API size_t ZDICT_trainFromBuffer_legacy(
|
||||||
|
void *dictBuffer, size_t dictBufferCapacity,
|
||||||
|
const void *samplesBuffer, const size_t *samplesSizes, unsigned nbSamples,
|
||||||
|
ZDICT_legacy_params_t parameters);
|
||||||
|
|
||||||
|
/* Deprecation warnings */
|
||||||
|
/* It is generally possible to disable deprecation warnings from compiler,
|
||||||
|
for example with -Wno-deprecated-declarations for gcc
|
||||||
|
or _CRT_SECURE_NO_WARNINGS in Visual.
|
||||||
|
Otherwise, it's also possible to manually define ZDICT_DISABLE_DEPRECATE_WARNINGS */
|
||||||
|
#ifdef ZDICT_DISABLE_DEPRECATE_WARNINGS
|
||||||
|
# define ZDICT_DEPRECATED(message) ZDICTLIB_API /* disable deprecation warnings */
|
||||||
|
#else
|
||||||
|
# define ZDICT_GCC_VERSION (__GNUC__ * 100 + __GNUC_MINOR__)
|
||||||
|
# if defined (__cplusplus) && (__cplusplus >= 201402) /* C++14 or greater */
|
||||||
|
# define ZDICT_DEPRECATED(message) [[deprecated(message)]] ZDICTLIB_API
|
||||||
|
# elif (ZDICT_GCC_VERSION >= 405) || defined(__clang__)
|
||||||
|
# define ZDICT_DEPRECATED(message) ZDICTLIB_API __attribute__((deprecated(message)))
|
||||||
|
# elif (ZDICT_GCC_VERSION >= 301)
|
||||||
|
# define ZDICT_DEPRECATED(message) ZDICTLIB_API __attribute__((deprecated))
|
||||||
|
# elif defined(_MSC_VER)
|
||||||
|
# define ZDICT_DEPRECATED(message) ZDICTLIB_API __declspec(deprecated(message))
|
||||||
|
# else
|
||||||
|
# pragma message("WARNING: You need to implement ZDICT_DEPRECATED for this compiler")
|
||||||
|
# define ZDICT_DEPRECATED(message) ZDICTLIB_API
|
||||||
|
# endif
|
||||||
|
#endif /* ZDICT_DISABLE_DEPRECATE_WARNINGS */
|
||||||
|
|
||||||
|
ZDICT_DEPRECATED("use ZDICT_finalizeDictionary() instead")
|
||||||
|
size_t ZDICT_addEntropyTablesFromBuffer(void* dictBuffer, size_t dictContentSize, size_t dictBufferCapacity,
|
||||||
|
const void* samplesBuffer, const size_t* samplesSizes, unsigned nbSamples);
|
||||||
|
|
||||||
|
|
||||||
|
#endif /* ZDICT_STATIC_LINKING_ONLY */
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* DICTBUILDER_H_001 */
|
||||||
147
vendor/github.com/DataDog/zstd/zstd.go
generated
vendored
Normal file
147
vendor/github.com/DataDog/zstd/zstd.go
generated
vendored
Normal file
@@ -0,0 +1,147 @@
|
|||||||
|
package zstd
|
||||||
|
|
||||||
|
/*
|
||||||
|
#define ZSTD_STATIC_LINKING_ONLY
|
||||||
|
#include "zstd.h"
|
||||||
|
#include "stdint.h" // for uintptr_t
|
||||||
|
|
||||||
|
// The following *_wrapper function are used for removing superflouos
|
||||||
|
// memory allocations when calling the wrapped functions from Go code.
|
||||||
|
// See https://github.com/golang/go/issues/24450 for details.
|
||||||
|
|
||||||
|
static size_t ZSTD_compress_wrapper(uintptr_t dst, size_t maxDstSize, const uintptr_t src, size_t srcSize, int compressionLevel) {
|
||||||
|
return ZSTD_compress((void*)dst, maxDstSize, (const void*)src, srcSize, compressionLevel);
|
||||||
|
}
|
||||||
|
|
||||||
|
static size_t ZSTD_decompress_wrapper(uintptr_t dst, size_t maxDstSize, uintptr_t src, size_t srcSize) {
|
||||||
|
return ZSTD_decompress((void*)dst, maxDstSize, (const void *)src, srcSize);
|
||||||
|
}
|
||||||
|
|
||||||
|
*/
|
||||||
|
import "C"
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"errors"
|
||||||
|
"io/ioutil"
|
||||||
|
"runtime"
|
||||||
|
"unsafe"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Defines best and standard values for zstd cli
|
||||||
|
const (
|
||||||
|
BestSpeed = 1
|
||||||
|
BestCompression = 20
|
||||||
|
DefaultCompression = 5
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
// ErrEmptySlice is returned when there is nothing to compress
|
||||||
|
ErrEmptySlice = errors.New("Bytes slice is empty")
|
||||||
|
)
|
||||||
|
|
||||||
|
// CompressBound returns the worst case size needed for a destination buffer,
|
||||||
|
// which can be used to preallocate a destination buffer or select a previously
|
||||||
|
// allocated buffer from a pool.
|
||||||
|
// See zstd.h to mirror implementation of ZSTD_COMPRESSBOUND
|
||||||
|
func CompressBound(srcSize int) int {
|
||||||
|
lowLimit := 128 << 10 // 128 kB
|
||||||
|
var margin int
|
||||||
|
if srcSize < lowLimit {
|
||||||
|
margin = (lowLimit - srcSize) >> 11
|
||||||
|
}
|
||||||
|
return srcSize + (srcSize >> 8) + margin
|
||||||
|
}
|
||||||
|
|
||||||
|
// cCompressBound is a cgo call to check the go implementation above against the c code.
|
||||||
|
func cCompressBound(srcSize int) int {
|
||||||
|
return int(C.ZSTD_compressBound(C.size_t(srcSize)))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Compress src into dst. If you have a buffer to use, you can pass it to
|
||||||
|
// prevent allocation. If it is too small, or if nil is passed, a new buffer
|
||||||
|
// will be allocated and returned.
|
||||||
|
func Compress(dst, src []byte) ([]byte, error) {
|
||||||
|
return CompressLevel(dst, src, DefaultCompression)
|
||||||
|
}
|
||||||
|
|
||||||
|
// CompressLevel is the same as Compress but you can pass a compression level
|
||||||
|
func CompressLevel(dst, src []byte, level int) ([]byte, error) {
|
||||||
|
bound := CompressBound(len(src))
|
||||||
|
if cap(dst) >= bound {
|
||||||
|
dst = dst[0:bound] // Reuse dst buffer
|
||||||
|
} else {
|
||||||
|
dst = make([]byte, bound)
|
||||||
|
}
|
||||||
|
|
||||||
|
srcPtr := C.uintptr_t(uintptr(0)) // Do not point anywhere, if src is empty
|
||||||
|
if len(src) > 0 {
|
||||||
|
srcPtr = C.uintptr_t(uintptr(unsafe.Pointer(&src[0])))
|
||||||
|
}
|
||||||
|
|
||||||
|
cWritten := C.ZSTD_compress_wrapper(
|
||||||
|
C.uintptr_t(uintptr(unsafe.Pointer(&dst[0]))),
|
||||||
|
C.size_t(len(dst)),
|
||||||
|
srcPtr,
|
||||||
|
C.size_t(len(src)),
|
||||||
|
C.int(level))
|
||||||
|
|
||||||
|
runtime.KeepAlive(src)
|
||||||
|
written := int(cWritten)
|
||||||
|
// Check if the return is an Error code
|
||||||
|
if err := getError(written); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return dst[:written], nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Decompress src into dst. If you have a buffer to use, you can pass it to
|
||||||
|
// prevent allocation. If it is too small, or if nil is passed, a new buffer
|
||||||
|
// will be allocated and returned.
|
||||||
|
func Decompress(dst, src []byte) ([]byte, error) {
|
||||||
|
if len(src) == 0 {
|
||||||
|
return []byte{}, ErrEmptySlice
|
||||||
|
}
|
||||||
|
decompress := func(dst, src []byte) ([]byte, error) {
|
||||||
|
|
||||||
|
cWritten := C.ZSTD_decompress_wrapper(
|
||||||
|
C.uintptr_t(uintptr(unsafe.Pointer(&dst[0]))),
|
||||||
|
C.size_t(len(dst)),
|
||||||
|
C.uintptr_t(uintptr(unsafe.Pointer(&src[0]))),
|
||||||
|
C.size_t(len(src)))
|
||||||
|
|
||||||
|
runtime.KeepAlive(src)
|
||||||
|
written := int(cWritten)
|
||||||
|
// Check error
|
||||||
|
if err := getError(written); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return dst[:written], nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(dst) == 0 {
|
||||||
|
// Attempt to use zStd to determine decompressed size (may result in error or 0)
|
||||||
|
size := int(C.size_t(C.ZSTD_getDecompressedSize(unsafe.Pointer(&src[0]), C.size_t(len(src)))))
|
||||||
|
|
||||||
|
if err := getError(size); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if size > 0 {
|
||||||
|
dst = make([]byte, size)
|
||||||
|
} else {
|
||||||
|
dst = make([]byte, len(src)*3) // starting guess
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for i := 0; i < 3; i++ { // 3 tries to allocate a bigger buffer
|
||||||
|
result, err := decompress(dst, src)
|
||||||
|
if !IsDstSizeTooSmallError(err) {
|
||||||
|
return result, err
|
||||||
|
}
|
||||||
|
dst = make([]byte, len(dst)*2) // Grow buffer by 2
|
||||||
|
}
|
||||||
|
|
||||||
|
// We failed getting a dst buffer of correct size, use stream API
|
||||||
|
r := NewReader(bytes.NewReader(src))
|
||||||
|
defer r.Close()
|
||||||
|
return ioutil.ReadAll(r)
|
||||||
|
}
|
||||||
2049
vendor/github.com/DataDog/zstd/zstd.h
generated
vendored
Normal file
2049
vendor/github.com/DataDog/zstd/zstd.h
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
83
vendor/github.com/DataDog/zstd/zstd_common.c
generated
vendored
Normal file
83
vendor/github.com/DataDog/zstd/zstd_common.c
generated
vendored
Normal file
@@ -0,0 +1,83 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Dependencies
|
||||||
|
***************************************/
|
||||||
|
#include <stdlib.h> /* malloc, calloc, free */
|
||||||
|
#include <string.h> /* memset */
|
||||||
|
#include "error_private.h"
|
||||||
|
#include "zstd_internal.h"
|
||||||
|
|
||||||
|
|
||||||
|
/*-****************************************
|
||||||
|
* Version
|
||||||
|
******************************************/
|
||||||
|
unsigned ZSTD_versionNumber(void) { return ZSTD_VERSION_NUMBER; }
|
||||||
|
|
||||||
|
const char* ZSTD_versionString(void) { return ZSTD_VERSION_STRING; }
|
||||||
|
|
||||||
|
|
||||||
|
/*-****************************************
|
||||||
|
* ZSTD Error Management
|
||||||
|
******************************************/
|
||||||
|
#undef ZSTD_isError /* defined within zstd_internal.h */
|
||||||
|
/*! ZSTD_isError() :
|
||||||
|
* tells if a return value is an error code
|
||||||
|
* symbol is required for external callers */
|
||||||
|
unsigned ZSTD_isError(size_t code) { return ERR_isError(code); }
|
||||||
|
|
||||||
|
/*! ZSTD_getErrorName() :
|
||||||
|
* provides error code string from function result (useful for debugging) */
|
||||||
|
const char* ZSTD_getErrorName(size_t code) { return ERR_getErrorName(code); }
|
||||||
|
|
||||||
|
/*! ZSTD_getError() :
|
||||||
|
* convert a `size_t` function result into a proper ZSTD_errorCode enum */
|
||||||
|
ZSTD_ErrorCode ZSTD_getErrorCode(size_t code) { return ERR_getErrorCode(code); }
|
||||||
|
|
||||||
|
/*! ZSTD_getErrorString() :
|
||||||
|
* provides error code string from enum */
|
||||||
|
const char* ZSTD_getErrorString(ZSTD_ErrorCode code) { return ERR_getErrorString(code); }
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/*=**************************************************************
|
||||||
|
* Custom allocator
|
||||||
|
****************************************************************/
|
||||||
|
void* ZSTD_malloc(size_t size, ZSTD_customMem customMem)
|
||||||
|
{
|
||||||
|
if (customMem.customAlloc)
|
||||||
|
return customMem.customAlloc(customMem.opaque, size);
|
||||||
|
return malloc(size);
|
||||||
|
}
|
||||||
|
|
||||||
|
void* ZSTD_calloc(size_t size, ZSTD_customMem customMem)
|
||||||
|
{
|
||||||
|
if (customMem.customAlloc) {
|
||||||
|
/* calloc implemented as malloc+memset;
|
||||||
|
* not as efficient as calloc, but next best guess for custom malloc */
|
||||||
|
void* const ptr = customMem.customAlloc(customMem.opaque, size);
|
||||||
|
memset(ptr, 0, size);
|
||||||
|
return ptr;
|
||||||
|
}
|
||||||
|
return calloc(1, size);
|
||||||
|
}
|
||||||
|
|
||||||
|
void ZSTD_free(void* ptr, ZSTD_customMem customMem)
|
||||||
|
{
|
||||||
|
if (ptr!=NULL) {
|
||||||
|
if (customMem.customFree)
|
||||||
|
customMem.customFree(customMem.opaque, ptr);
|
||||||
|
else
|
||||||
|
free(ptr);
|
||||||
|
}
|
||||||
|
}
|
||||||
4103
vendor/github.com/DataDog/zstd/zstd_compress.c
generated
vendored
Normal file
4103
vendor/github.com/DataDog/zstd/zstd_compress.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
1003
vendor/github.com/DataDog/zstd/zstd_compress_internal.h
generated
vendored
Normal file
1003
vendor/github.com/DataDog/zstd/zstd_compress_internal.h
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
154
vendor/github.com/DataDog/zstd/zstd_compress_literals.c
generated
vendored
Normal file
154
vendor/github.com/DataDog/zstd/zstd_compress_literals.c
generated
vendored
Normal file
@@ -0,0 +1,154 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Dependencies
|
||||||
|
***************************************/
|
||||||
|
#include "zstd_compress_literals.h"
|
||||||
|
|
||||||
|
size_t ZSTD_noCompressLiterals (void* dst, size_t dstCapacity, const void* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
BYTE* const ostart = (BYTE* const)dst;
|
||||||
|
U32 const flSize = 1 + (srcSize>31) + (srcSize>4095);
|
||||||
|
|
||||||
|
RETURN_ERROR_IF(srcSize + flSize > dstCapacity, dstSize_tooSmall);
|
||||||
|
|
||||||
|
switch(flSize)
|
||||||
|
{
|
||||||
|
case 1: /* 2 - 1 - 5 */
|
||||||
|
ostart[0] = (BYTE)((U32)set_basic + (srcSize<<3));
|
||||||
|
break;
|
||||||
|
case 2: /* 2 - 2 - 12 */
|
||||||
|
MEM_writeLE16(ostart, (U16)((U32)set_basic + (1<<2) + (srcSize<<4)));
|
||||||
|
break;
|
||||||
|
case 3: /* 2 - 2 - 20 */
|
||||||
|
MEM_writeLE32(ostart, (U32)((U32)set_basic + (3<<2) + (srcSize<<4)));
|
||||||
|
break;
|
||||||
|
default: /* not necessary : flSize is {1,2,3} */
|
||||||
|
assert(0);
|
||||||
|
}
|
||||||
|
|
||||||
|
memcpy(ostart + flSize, src, srcSize);
|
||||||
|
return srcSize + flSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZSTD_compressRleLiteralsBlock (void* dst, size_t dstCapacity, const void* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
BYTE* const ostart = (BYTE* const)dst;
|
||||||
|
U32 const flSize = 1 + (srcSize>31) + (srcSize>4095);
|
||||||
|
|
||||||
|
(void)dstCapacity; /* dstCapacity already guaranteed to be >=4, hence large enough */
|
||||||
|
|
||||||
|
switch(flSize)
|
||||||
|
{
|
||||||
|
case 1: /* 2 - 1 - 5 */
|
||||||
|
ostart[0] = (BYTE)((U32)set_rle + (srcSize<<3));
|
||||||
|
break;
|
||||||
|
case 2: /* 2 - 2 - 12 */
|
||||||
|
MEM_writeLE16(ostart, (U16)((U32)set_rle + (1<<2) + (srcSize<<4)));
|
||||||
|
break;
|
||||||
|
case 3: /* 2 - 2 - 20 */
|
||||||
|
MEM_writeLE32(ostart, (U32)((U32)set_rle + (3<<2) + (srcSize<<4)));
|
||||||
|
break;
|
||||||
|
default: /* not necessary : flSize is {1,2,3} */
|
||||||
|
assert(0);
|
||||||
|
}
|
||||||
|
|
||||||
|
ostart[flSize] = *(const BYTE*)src;
|
||||||
|
return flSize+1;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZSTD_compressLiterals (ZSTD_hufCTables_t const* prevHuf,
|
||||||
|
ZSTD_hufCTables_t* nextHuf,
|
||||||
|
ZSTD_strategy strategy, int disableLiteralCompression,
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
void* entropyWorkspace, size_t entropyWorkspaceSize,
|
||||||
|
const int bmi2)
|
||||||
|
{
|
||||||
|
size_t const minGain = ZSTD_minGain(srcSize, strategy);
|
||||||
|
size_t const lhSize = 3 + (srcSize >= 1 KB) + (srcSize >= 16 KB);
|
||||||
|
BYTE* const ostart = (BYTE*)dst;
|
||||||
|
U32 singleStream = srcSize < 256;
|
||||||
|
symbolEncodingType_e hType = set_compressed;
|
||||||
|
size_t cLitSize;
|
||||||
|
|
||||||
|
DEBUGLOG(5,"ZSTD_compressLiterals (disableLiteralCompression=%i)",
|
||||||
|
disableLiteralCompression);
|
||||||
|
|
||||||
|
/* Prepare nextEntropy assuming reusing the existing table */
|
||||||
|
memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
|
||||||
|
|
||||||
|
if (disableLiteralCompression)
|
||||||
|
return ZSTD_noCompressLiterals(dst, dstCapacity, src, srcSize);
|
||||||
|
|
||||||
|
/* small ? don't even attempt compression (speed opt) */
|
||||||
|
# define COMPRESS_LITERALS_SIZE_MIN 63
|
||||||
|
{ size_t const minLitSize = (prevHuf->repeatMode == HUF_repeat_valid) ? 6 : COMPRESS_LITERALS_SIZE_MIN;
|
||||||
|
if (srcSize <= minLitSize) return ZSTD_noCompressLiterals(dst, dstCapacity, src, srcSize);
|
||||||
|
}
|
||||||
|
|
||||||
|
RETURN_ERROR_IF(dstCapacity < lhSize+1, dstSize_tooSmall, "not enough space for compression");
|
||||||
|
{ HUF_repeat repeat = prevHuf->repeatMode;
|
||||||
|
int const preferRepeat = strategy < ZSTD_lazy ? srcSize <= 1024 : 0;
|
||||||
|
if (repeat == HUF_repeat_valid && lhSize == 3) singleStream = 1;
|
||||||
|
cLitSize = singleStream ?
|
||||||
|
HUF_compress1X_repeat(
|
||||||
|
ostart+lhSize, dstCapacity-lhSize, src, srcSize,
|
||||||
|
255, 11, entropyWorkspace, entropyWorkspaceSize,
|
||||||
|
(HUF_CElt*)nextHuf->CTable, &repeat, preferRepeat, bmi2) :
|
||||||
|
HUF_compress4X_repeat(
|
||||||
|
ostart+lhSize, dstCapacity-lhSize, src, srcSize,
|
||||||
|
255, 11, entropyWorkspace, entropyWorkspaceSize,
|
||||||
|
(HUF_CElt*)nextHuf->CTable, &repeat, preferRepeat, bmi2);
|
||||||
|
if (repeat != HUF_repeat_none) {
|
||||||
|
/* reused the existing table */
|
||||||
|
hType = set_repeat;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if ((cLitSize==0) | (cLitSize >= srcSize - minGain) | ERR_isError(cLitSize)) {
|
||||||
|
memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
|
||||||
|
return ZSTD_noCompressLiterals(dst, dstCapacity, src, srcSize);
|
||||||
|
}
|
||||||
|
if (cLitSize==1) {
|
||||||
|
memcpy(nextHuf, prevHuf, sizeof(*prevHuf));
|
||||||
|
return ZSTD_compressRleLiteralsBlock(dst, dstCapacity, src, srcSize);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (hType == set_compressed) {
|
||||||
|
/* using a newly constructed table */
|
||||||
|
nextHuf->repeatMode = HUF_repeat_check;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Build header */
|
||||||
|
switch(lhSize)
|
||||||
|
{
|
||||||
|
case 3: /* 2 - 2 - 10 - 10 */
|
||||||
|
{ U32 const lhc = hType + ((!singleStream) << 2) + ((U32)srcSize<<4) + ((U32)cLitSize<<14);
|
||||||
|
MEM_writeLE24(ostart, lhc);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
case 4: /* 2 - 2 - 14 - 14 */
|
||||||
|
{ U32 const lhc = hType + (2 << 2) + ((U32)srcSize<<4) + ((U32)cLitSize<<18);
|
||||||
|
MEM_writeLE32(ostart, lhc);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
case 5: /* 2 - 2 - 18 - 18 */
|
||||||
|
{ U32 const lhc = hType + (3 << 2) + ((U32)srcSize<<4) + ((U32)cLitSize<<22);
|
||||||
|
MEM_writeLE32(ostart, lhc);
|
||||||
|
ostart[4] = (BYTE)(cLitSize >> 10);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
default: /* not possible : lhSize is {3,4,5} */
|
||||||
|
assert(0);
|
||||||
|
}
|
||||||
|
return lhSize+cLitSize;
|
||||||
|
}
|
||||||
29
vendor/github.com/DataDog/zstd/zstd_compress_literals.h
generated
vendored
Normal file
29
vendor/github.com/DataDog/zstd/zstd_compress_literals.h
generated
vendored
Normal file
@@ -0,0 +1,29 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_COMPRESS_LITERALS_H
|
||||||
|
#define ZSTD_COMPRESS_LITERALS_H
|
||||||
|
|
||||||
|
#include "zstd_compress_internal.h" /* ZSTD_hufCTables_t, ZSTD_minGain() */
|
||||||
|
|
||||||
|
|
||||||
|
size_t ZSTD_noCompressLiterals (void* dst, size_t dstCapacity, const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
size_t ZSTD_compressRleLiteralsBlock (void* dst, size_t dstCapacity, const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
size_t ZSTD_compressLiterals (ZSTD_hufCTables_t const* prevHuf,
|
||||||
|
ZSTD_hufCTables_t* nextHuf,
|
||||||
|
ZSTD_strategy strategy, int disableLiteralCompression,
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
void* entropyWorkspace, size_t entropyWorkspaceSize,
|
||||||
|
const int bmi2);
|
||||||
|
|
||||||
|
#endif /* ZSTD_COMPRESS_LITERALS_H */
|
||||||
415
vendor/github.com/DataDog/zstd/zstd_compress_sequences.c
generated
vendored
Normal file
415
vendor/github.com/DataDog/zstd/zstd_compress_sequences.c
generated
vendored
Normal file
@@ -0,0 +1,415 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Dependencies
|
||||||
|
***************************************/
|
||||||
|
#include "zstd_compress_sequences.h"
|
||||||
|
|
||||||
|
/**
|
||||||
|
* -log2(x / 256) lookup table for x in [0, 256).
|
||||||
|
* If x == 0: Return 0
|
||||||
|
* Else: Return floor(-log2(x / 256) * 256)
|
||||||
|
*/
|
||||||
|
static unsigned const kInverseProbabilityLog256[256] = {
|
||||||
|
0, 2048, 1792, 1642, 1536, 1453, 1386, 1329, 1280, 1236, 1197, 1162,
|
||||||
|
1130, 1100, 1073, 1047, 1024, 1001, 980, 960, 941, 923, 906, 889,
|
||||||
|
874, 859, 844, 830, 817, 804, 791, 779, 768, 756, 745, 734,
|
||||||
|
724, 714, 704, 694, 685, 676, 667, 658, 650, 642, 633, 626,
|
||||||
|
618, 610, 603, 595, 588, 581, 574, 567, 561, 554, 548, 542,
|
||||||
|
535, 529, 523, 517, 512, 506, 500, 495, 489, 484, 478, 473,
|
||||||
|
468, 463, 458, 453, 448, 443, 438, 434, 429, 424, 420, 415,
|
||||||
|
411, 407, 402, 398, 394, 390, 386, 382, 377, 373, 370, 366,
|
||||||
|
362, 358, 354, 350, 347, 343, 339, 336, 332, 329, 325, 322,
|
||||||
|
318, 315, 311, 308, 305, 302, 298, 295, 292, 289, 286, 282,
|
||||||
|
279, 276, 273, 270, 267, 264, 261, 258, 256, 253, 250, 247,
|
||||||
|
244, 241, 239, 236, 233, 230, 228, 225, 222, 220, 217, 215,
|
||||||
|
212, 209, 207, 204, 202, 199, 197, 194, 192, 190, 187, 185,
|
||||||
|
182, 180, 178, 175, 173, 171, 168, 166, 164, 162, 159, 157,
|
||||||
|
155, 153, 151, 149, 146, 144, 142, 140, 138, 136, 134, 132,
|
||||||
|
130, 128, 126, 123, 121, 119, 117, 115, 114, 112, 110, 108,
|
||||||
|
106, 104, 102, 100, 98, 96, 94, 93, 91, 89, 87, 85,
|
||||||
|
83, 82, 80, 78, 76, 74, 73, 71, 69, 67, 66, 64,
|
||||||
|
62, 61, 59, 57, 55, 54, 52, 50, 49, 47, 46, 44,
|
||||||
|
42, 41, 39, 37, 36, 34, 33, 31, 30, 28, 26, 25,
|
||||||
|
23, 22, 20, 19, 17, 16, 14, 13, 11, 10, 8, 7,
|
||||||
|
5, 4, 2, 1,
|
||||||
|
};
|
||||||
|
|
||||||
|
static unsigned ZSTD_getFSEMaxSymbolValue(FSE_CTable const* ctable) {
|
||||||
|
void const* ptr = ctable;
|
||||||
|
U16 const* u16ptr = (U16 const*)ptr;
|
||||||
|
U32 const maxSymbolValue = MEM_read16(u16ptr + 1);
|
||||||
|
return maxSymbolValue;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns the cost in bytes of encoding the normalized count header.
|
||||||
|
* Returns an error if any of the helper functions return an error.
|
||||||
|
*/
|
||||||
|
static size_t ZSTD_NCountCost(unsigned const* count, unsigned const max,
|
||||||
|
size_t const nbSeq, unsigned const FSELog)
|
||||||
|
{
|
||||||
|
BYTE wksp[FSE_NCOUNTBOUND];
|
||||||
|
S16 norm[MaxSeq + 1];
|
||||||
|
const U32 tableLog = FSE_optimalTableLog(FSELog, nbSeq, max);
|
||||||
|
FORWARD_IF_ERROR(FSE_normalizeCount(norm, tableLog, count, nbSeq, max));
|
||||||
|
return FSE_writeNCount(wksp, sizeof(wksp), norm, max, tableLog);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns the cost in bits of encoding the distribution described by count
|
||||||
|
* using the entropy bound.
|
||||||
|
*/
|
||||||
|
static size_t ZSTD_entropyCost(unsigned const* count, unsigned const max, size_t const total)
|
||||||
|
{
|
||||||
|
unsigned cost = 0;
|
||||||
|
unsigned s;
|
||||||
|
for (s = 0; s <= max; ++s) {
|
||||||
|
unsigned norm = (unsigned)((256 * count[s]) / total);
|
||||||
|
if (count[s] != 0 && norm == 0)
|
||||||
|
norm = 1;
|
||||||
|
assert(count[s] < total);
|
||||||
|
cost += count[s] * kInverseProbabilityLog256[norm];
|
||||||
|
}
|
||||||
|
return cost >> 8;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns the cost in bits of encoding the distribution in count using ctable.
|
||||||
|
* Returns an error if ctable cannot represent all the symbols in count.
|
||||||
|
*/
|
||||||
|
static size_t ZSTD_fseBitCost(
|
||||||
|
FSE_CTable const* ctable,
|
||||||
|
unsigned const* count,
|
||||||
|
unsigned const max)
|
||||||
|
{
|
||||||
|
unsigned const kAccuracyLog = 8;
|
||||||
|
size_t cost = 0;
|
||||||
|
unsigned s;
|
||||||
|
FSE_CState_t cstate;
|
||||||
|
FSE_initCState(&cstate, ctable);
|
||||||
|
RETURN_ERROR_IF(ZSTD_getFSEMaxSymbolValue(ctable) < max, GENERIC,
|
||||||
|
"Repeat FSE_CTable has maxSymbolValue %u < %u",
|
||||||
|
ZSTD_getFSEMaxSymbolValue(ctable), max);
|
||||||
|
for (s = 0; s <= max; ++s) {
|
||||||
|
unsigned const tableLog = cstate.stateLog;
|
||||||
|
unsigned const badCost = (tableLog + 1) << kAccuracyLog;
|
||||||
|
unsigned const bitCost = FSE_bitCost(cstate.symbolTT, tableLog, s, kAccuracyLog);
|
||||||
|
if (count[s] == 0)
|
||||||
|
continue;
|
||||||
|
RETURN_ERROR_IF(bitCost >= badCost, GENERIC,
|
||||||
|
"Repeat FSE_CTable has Prob[%u] == 0", s);
|
||||||
|
cost += count[s] * bitCost;
|
||||||
|
}
|
||||||
|
return cost >> kAccuracyLog;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns the cost in bits of encoding the distribution in count using the
|
||||||
|
* table described by norm. The max symbol support by norm is assumed >= max.
|
||||||
|
* norm must be valid for every symbol with non-zero probability in count.
|
||||||
|
*/
|
||||||
|
static size_t ZSTD_crossEntropyCost(short const* norm, unsigned accuracyLog,
|
||||||
|
unsigned const* count, unsigned const max)
|
||||||
|
{
|
||||||
|
unsigned const shift = 8 - accuracyLog;
|
||||||
|
size_t cost = 0;
|
||||||
|
unsigned s;
|
||||||
|
assert(accuracyLog <= 8);
|
||||||
|
for (s = 0; s <= max; ++s) {
|
||||||
|
unsigned const normAcc = norm[s] != -1 ? norm[s] : 1;
|
||||||
|
unsigned const norm256 = normAcc << shift;
|
||||||
|
assert(norm256 > 0);
|
||||||
|
assert(norm256 < 256);
|
||||||
|
cost += count[s] * kInverseProbabilityLog256[norm256];
|
||||||
|
}
|
||||||
|
return cost >> 8;
|
||||||
|
}
|
||||||
|
|
||||||
|
symbolEncodingType_e
|
||||||
|
ZSTD_selectEncodingType(
|
||||||
|
FSE_repeat* repeatMode, unsigned const* count, unsigned const max,
|
||||||
|
size_t const mostFrequent, size_t nbSeq, unsigned const FSELog,
|
||||||
|
FSE_CTable const* prevCTable,
|
||||||
|
short const* defaultNorm, U32 defaultNormLog,
|
||||||
|
ZSTD_defaultPolicy_e const isDefaultAllowed,
|
||||||
|
ZSTD_strategy const strategy)
|
||||||
|
{
|
||||||
|
ZSTD_STATIC_ASSERT(ZSTD_defaultDisallowed == 0 && ZSTD_defaultAllowed != 0);
|
||||||
|
if (mostFrequent == nbSeq) {
|
||||||
|
*repeatMode = FSE_repeat_none;
|
||||||
|
if (isDefaultAllowed && nbSeq <= 2) {
|
||||||
|
/* Prefer set_basic over set_rle when there are 2 or less symbols,
|
||||||
|
* since RLE uses 1 byte, but set_basic uses 5-6 bits per symbol.
|
||||||
|
* If basic encoding isn't possible, always choose RLE.
|
||||||
|
*/
|
||||||
|
DEBUGLOG(5, "Selected set_basic");
|
||||||
|
return set_basic;
|
||||||
|
}
|
||||||
|
DEBUGLOG(5, "Selected set_rle");
|
||||||
|
return set_rle;
|
||||||
|
}
|
||||||
|
if (strategy < ZSTD_lazy) {
|
||||||
|
if (isDefaultAllowed) {
|
||||||
|
size_t const staticFse_nbSeq_max = 1000;
|
||||||
|
size_t const mult = 10 - strategy;
|
||||||
|
size_t const baseLog = 3;
|
||||||
|
size_t const dynamicFse_nbSeq_min = (((size_t)1 << defaultNormLog) * mult) >> baseLog; /* 28-36 for offset, 56-72 for lengths */
|
||||||
|
assert(defaultNormLog >= 5 && defaultNormLog <= 6); /* xx_DEFAULTNORMLOG */
|
||||||
|
assert(mult <= 9 && mult >= 7);
|
||||||
|
if ( (*repeatMode == FSE_repeat_valid)
|
||||||
|
&& (nbSeq < staticFse_nbSeq_max) ) {
|
||||||
|
DEBUGLOG(5, "Selected set_repeat");
|
||||||
|
return set_repeat;
|
||||||
|
}
|
||||||
|
if ( (nbSeq < dynamicFse_nbSeq_min)
|
||||||
|
|| (mostFrequent < (nbSeq >> (defaultNormLog-1))) ) {
|
||||||
|
DEBUGLOG(5, "Selected set_basic");
|
||||||
|
/* The format allows default tables to be repeated, but it isn't useful.
|
||||||
|
* When using simple heuristics to select encoding type, we don't want
|
||||||
|
* to confuse these tables with dictionaries. When running more careful
|
||||||
|
* analysis, we don't need to waste time checking both repeating tables
|
||||||
|
* and default tables.
|
||||||
|
*/
|
||||||
|
*repeatMode = FSE_repeat_none;
|
||||||
|
return set_basic;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
size_t const basicCost = isDefaultAllowed ? ZSTD_crossEntropyCost(defaultNorm, defaultNormLog, count, max) : ERROR(GENERIC);
|
||||||
|
size_t const repeatCost = *repeatMode != FSE_repeat_none ? ZSTD_fseBitCost(prevCTable, count, max) : ERROR(GENERIC);
|
||||||
|
size_t const NCountCost = ZSTD_NCountCost(count, max, nbSeq, FSELog);
|
||||||
|
size_t const compressedCost = (NCountCost << 3) + ZSTD_entropyCost(count, max, nbSeq);
|
||||||
|
|
||||||
|
if (isDefaultAllowed) {
|
||||||
|
assert(!ZSTD_isError(basicCost));
|
||||||
|
assert(!(*repeatMode == FSE_repeat_valid && ZSTD_isError(repeatCost)));
|
||||||
|
}
|
||||||
|
assert(!ZSTD_isError(NCountCost));
|
||||||
|
assert(compressedCost < ERROR(maxCode));
|
||||||
|
DEBUGLOG(5, "Estimated bit costs: basic=%u\trepeat=%u\tcompressed=%u",
|
||||||
|
(unsigned)basicCost, (unsigned)repeatCost, (unsigned)compressedCost);
|
||||||
|
if (basicCost <= repeatCost && basicCost <= compressedCost) {
|
||||||
|
DEBUGLOG(5, "Selected set_basic");
|
||||||
|
assert(isDefaultAllowed);
|
||||||
|
*repeatMode = FSE_repeat_none;
|
||||||
|
return set_basic;
|
||||||
|
}
|
||||||
|
if (repeatCost <= compressedCost) {
|
||||||
|
DEBUGLOG(5, "Selected set_repeat");
|
||||||
|
assert(!ZSTD_isError(repeatCost));
|
||||||
|
return set_repeat;
|
||||||
|
}
|
||||||
|
assert(compressedCost < basicCost && compressedCost < repeatCost);
|
||||||
|
}
|
||||||
|
DEBUGLOG(5, "Selected set_compressed");
|
||||||
|
*repeatMode = FSE_repeat_check;
|
||||||
|
return set_compressed;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t
|
||||||
|
ZSTD_buildCTable(void* dst, size_t dstCapacity,
|
||||||
|
FSE_CTable* nextCTable, U32 FSELog, symbolEncodingType_e type,
|
||||||
|
unsigned* count, U32 max,
|
||||||
|
const BYTE* codeTable, size_t nbSeq,
|
||||||
|
const S16* defaultNorm, U32 defaultNormLog, U32 defaultMax,
|
||||||
|
const FSE_CTable* prevCTable, size_t prevCTableSize,
|
||||||
|
void* entropyWorkspace, size_t entropyWorkspaceSize)
|
||||||
|
{
|
||||||
|
BYTE* op = (BYTE*)dst;
|
||||||
|
const BYTE* const oend = op + dstCapacity;
|
||||||
|
DEBUGLOG(6, "ZSTD_buildCTable (dstCapacity=%u)", (unsigned)dstCapacity);
|
||||||
|
|
||||||
|
switch (type) {
|
||||||
|
case set_rle:
|
||||||
|
FORWARD_IF_ERROR(FSE_buildCTable_rle(nextCTable, (BYTE)max));
|
||||||
|
RETURN_ERROR_IF(dstCapacity==0, dstSize_tooSmall);
|
||||||
|
*op = codeTable[0];
|
||||||
|
return 1;
|
||||||
|
case set_repeat:
|
||||||
|
memcpy(nextCTable, prevCTable, prevCTableSize);
|
||||||
|
return 0;
|
||||||
|
case set_basic:
|
||||||
|
FORWARD_IF_ERROR(FSE_buildCTable_wksp(nextCTable, defaultNorm, defaultMax, defaultNormLog, entropyWorkspace, entropyWorkspaceSize)); /* note : could be pre-calculated */
|
||||||
|
return 0;
|
||||||
|
case set_compressed: {
|
||||||
|
S16 norm[MaxSeq + 1];
|
||||||
|
size_t nbSeq_1 = nbSeq;
|
||||||
|
const U32 tableLog = FSE_optimalTableLog(FSELog, nbSeq, max);
|
||||||
|
if (count[codeTable[nbSeq-1]] > 1) {
|
||||||
|
count[codeTable[nbSeq-1]]--;
|
||||||
|
nbSeq_1--;
|
||||||
|
}
|
||||||
|
assert(nbSeq_1 > 1);
|
||||||
|
FORWARD_IF_ERROR(FSE_normalizeCount(norm, tableLog, count, nbSeq_1, max));
|
||||||
|
{ size_t const NCountSize = FSE_writeNCount(op, oend - op, norm, max, tableLog); /* overflow protected */
|
||||||
|
FORWARD_IF_ERROR(NCountSize);
|
||||||
|
FORWARD_IF_ERROR(FSE_buildCTable_wksp(nextCTable, norm, max, tableLog, entropyWorkspace, entropyWorkspaceSize));
|
||||||
|
return NCountSize;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
default: assert(0); RETURN_ERROR(GENERIC);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE size_t
|
||||||
|
ZSTD_encodeSequences_body(
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
FSE_CTable const* CTable_MatchLength, BYTE const* mlCodeTable,
|
||||||
|
FSE_CTable const* CTable_OffsetBits, BYTE const* ofCodeTable,
|
||||||
|
FSE_CTable const* CTable_LitLength, BYTE const* llCodeTable,
|
||||||
|
seqDef const* sequences, size_t nbSeq, int longOffsets)
|
||||||
|
{
|
||||||
|
BIT_CStream_t blockStream;
|
||||||
|
FSE_CState_t stateMatchLength;
|
||||||
|
FSE_CState_t stateOffsetBits;
|
||||||
|
FSE_CState_t stateLitLength;
|
||||||
|
|
||||||
|
RETURN_ERROR_IF(
|
||||||
|
ERR_isError(BIT_initCStream(&blockStream, dst, dstCapacity)),
|
||||||
|
dstSize_tooSmall, "not enough space remaining");
|
||||||
|
DEBUGLOG(6, "available space for bitstream : %i (dstCapacity=%u)",
|
||||||
|
(int)(blockStream.endPtr - blockStream.startPtr),
|
||||||
|
(unsigned)dstCapacity);
|
||||||
|
|
||||||
|
/* first symbols */
|
||||||
|
FSE_initCState2(&stateMatchLength, CTable_MatchLength, mlCodeTable[nbSeq-1]);
|
||||||
|
FSE_initCState2(&stateOffsetBits, CTable_OffsetBits, ofCodeTable[nbSeq-1]);
|
||||||
|
FSE_initCState2(&stateLitLength, CTable_LitLength, llCodeTable[nbSeq-1]);
|
||||||
|
BIT_addBits(&blockStream, sequences[nbSeq-1].litLength, LL_bits[llCodeTable[nbSeq-1]]);
|
||||||
|
if (MEM_32bits()) BIT_flushBits(&blockStream);
|
||||||
|
BIT_addBits(&blockStream, sequences[nbSeq-1].matchLength, ML_bits[mlCodeTable[nbSeq-1]]);
|
||||||
|
if (MEM_32bits()) BIT_flushBits(&blockStream);
|
||||||
|
if (longOffsets) {
|
||||||
|
U32 const ofBits = ofCodeTable[nbSeq-1];
|
||||||
|
int const extraBits = ofBits - MIN(ofBits, STREAM_ACCUMULATOR_MIN-1);
|
||||||
|
if (extraBits) {
|
||||||
|
BIT_addBits(&blockStream, sequences[nbSeq-1].offset, extraBits);
|
||||||
|
BIT_flushBits(&blockStream);
|
||||||
|
}
|
||||||
|
BIT_addBits(&blockStream, sequences[nbSeq-1].offset >> extraBits,
|
||||||
|
ofBits - extraBits);
|
||||||
|
} else {
|
||||||
|
BIT_addBits(&blockStream, sequences[nbSeq-1].offset, ofCodeTable[nbSeq-1]);
|
||||||
|
}
|
||||||
|
BIT_flushBits(&blockStream);
|
||||||
|
|
||||||
|
{ size_t n;
|
||||||
|
for (n=nbSeq-2 ; n<nbSeq ; n--) { /* intentional underflow */
|
||||||
|
BYTE const llCode = llCodeTable[n];
|
||||||
|
BYTE const ofCode = ofCodeTable[n];
|
||||||
|
BYTE const mlCode = mlCodeTable[n];
|
||||||
|
U32 const llBits = LL_bits[llCode];
|
||||||
|
U32 const ofBits = ofCode;
|
||||||
|
U32 const mlBits = ML_bits[mlCode];
|
||||||
|
DEBUGLOG(6, "encoding: litlen:%2u - matchlen:%2u - offCode:%7u",
|
||||||
|
(unsigned)sequences[n].litLength,
|
||||||
|
(unsigned)sequences[n].matchLength + MINMATCH,
|
||||||
|
(unsigned)sequences[n].offset);
|
||||||
|
/* 32b*/ /* 64b*/
|
||||||
|
/* (7)*/ /* (7)*/
|
||||||
|
FSE_encodeSymbol(&blockStream, &stateOffsetBits, ofCode); /* 15 */ /* 15 */
|
||||||
|
FSE_encodeSymbol(&blockStream, &stateMatchLength, mlCode); /* 24 */ /* 24 */
|
||||||
|
if (MEM_32bits()) BIT_flushBits(&blockStream); /* (7)*/
|
||||||
|
FSE_encodeSymbol(&blockStream, &stateLitLength, llCode); /* 16 */ /* 33 */
|
||||||
|
if (MEM_32bits() || (ofBits+mlBits+llBits >= 64-7-(LLFSELog+MLFSELog+OffFSELog)))
|
||||||
|
BIT_flushBits(&blockStream); /* (7)*/
|
||||||
|
BIT_addBits(&blockStream, sequences[n].litLength, llBits);
|
||||||
|
if (MEM_32bits() && ((llBits+mlBits)>24)) BIT_flushBits(&blockStream);
|
||||||
|
BIT_addBits(&blockStream, sequences[n].matchLength, mlBits);
|
||||||
|
if (MEM_32bits() || (ofBits+mlBits+llBits > 56)) BIT_flushBits(&blockStream);
|
||||||
|
if (longOffsets) {
|
||||||
|
int const extraBits = ofBits - MIN(ofBits, STREAM_ACCUMULATOR_MIN-1);
|
||||||
|
if (extraBits) {
|
||||||
|
BIT_addBits(&blockStream, sequences[n].offset, extraBits);
|
||||||
|
BIT_flushBits(&blockStream); /* (7)*/
|
||||||
|
}
|
||||||
|
BIT_addBits(&blockStream, sequences[n].offset >> extraBits,
|
||||||
|
ofBits - extraBits); /* 31 */
|
||||||
|
} else {
|
||||||
|
BIT_addBits(&blockStream, sequences[n].offset, ofBits); /* 31 */
|
||||||
|
}
|
||||||
|
BIT_flushBits(&blockStream); /* (7)*/
|
||||||
|
DEBUGLOG(7, "remaining space : %i", (int)(blockStream.endPtr - blockStream.ptr));
|
||||||
|
} }
|
||||||
|
|
||||||
|
DEBUGLOG(6, "ZSTD_encodeSequences: flushing ML state with %u bits", stateMatchLength.stateLog);
|
||||||
|
FSE_flushCState(&blockStream, &stateMatchLength);
|
||||||
|
DEBUGLOG(6, "ZSTD_encodeSequences: flushing Off state with %u bits", stateOffsetBits.stateLog);
|
||||||
|
FSE_flushCState(&blockStream, &stateOffsetBits);
|
||||||
|
DEBUGLOG(6, "ZSTD_encodeSequences: flushing LL state with %u bits", stateLitLength.stateLog);
|
||||||
|
FSE_flushCState(&blockStream, &stateLitLength);
|
||||||
|
|
||||||
|
{ size_t const streamSize = BIT_closeCStream(&blockStream);
|
||||||
|
RETURN_ERROR_IF(streamSize==0, dstSize_tooSmall, "not enough space");
|
||||||
|
return streamSize;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static size_t
|
||||||
|
ZSTD_encodeSequences_default(
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
FSE_CTable const* CTable_MatchLength, BYTE const* mlCodeTable,
|
||||||
|
FSE_CTable const* CTable_OffsetBits, BYTE const* ofCodeTable,
|
||||||
|
FSE_CTable const* CTable_LitLength, BYTE const* llCodeTable,
|
||||||
|
seqDef const* sequences, size_t nbSeq, int longOffsets)
|
||||||
|
{
|
||||||
|
return ZSTD_encodeSequences_body(dst, dstCapacity,
|
||||||
|
CTable_MatchLength, mlCodeTable,
|
||||||
|
CTable_OffsetBits, ofCodeTable,
|
||||||
|
CTable_LitLength, llCodeTable,
|
||||||
|
sequences, nbSeq, longOffsets);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#if DYNAMIC_BMI2
|
||||||
|
|
||||||
|
static TARGET_ATTRIBUTE("bmi2") size_t
|
||||||
|
ZSTD_encodeSequences_bmi2(
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
FSE_CTable const* CTable_MatchLength, BYTE const* mlCodeTable,
|
||||||
|
FSE_CTable const* CTable_OffsetBits, BYTE const* ofCodeTable,
|
||||||
|
FSE_CTable const* CTable_LitLength, BYTE const* llCodeTable,
|
||||||
|
seqDef const* sequences, size_t nbSeq, int longOffsets)
|
||||||
|
{
|
||||||
|
return ZSTD_encodeSequences_body(dst, dstCapacity,
|
||||||
|
CTable_MatchLength, mlCodeTable,
|
||||||
|
CTable_OffsetBits, ofCodeTable,
|
||||||
|
CTable_LitLength, llCodeTable,
|
||||||
|
sequences, nbSeq, longOffsets);
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif
|
||||||
|
|
||||||
|
size_t ZSTD_encodeSequences(
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
FSE_CTable const* CTable_MatchLength, BYTE const* mlCodeTable,
|
||||||
|
FSE_CTable const* CTable_OffsetBits, BYTE const* ofCodeTable,
|
||||||
|
FSE_CTable const* CTable_LitLength, BYTE const* llCodeTable,
|
||||||
|
seqDef const* sequences, size_t nbSeq, int longOffsets, int bmi2)
|
||||||
|
{
|
||||||
|
DEBUGLOG(5, "ZSTD_encodeSequences: dstCapacity = %u", (unsigned)dstCapacity);
|
||||||
|
#if DYNAMIC_BMI2
|
||||||
|
if (bmi2) {
|
||||||
|
return ZSTD_encodeSequences_bmi2(dst, dstCapacity,
|
||||||
|
CTable_MatchLength, mlCodeTable,
|
||||||
|
CTable_OffsetBits, ofCodeTable,
|
||||||
|
CTable_LitLength, llCodeTable,
|
||||||
|
sequences, nbSeq, longOffsets);
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
(void)bmi2;
|
||||||
|
return ZSTD_encodeSequences_default(dst, dstCapacity,
|
||||||
|
CTable_MatchLength, mlCodeTable,
|
||||||
|
CTable_OffsetBits, ofCodeTable,
|
||||||
|
CTable_LitLength, llCodeTable,
|
||||||
|
sequences, nbSeq, longOffsets);
|
||||||
|
}
|
||||||
47
vendor/github.com/DataDog/zstd/zstd_compress_sequences.h
generated
vendored
Normal file
47
vendor/github.com/DataDog/zstd/zstd_compress_sequences.h
generated
vendored
Normal file
@@ -0,0 +1,47 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_COMPRESS_SEQUENCES_H
|
||||||
|
#define ZSTD_COMPRESS_SEQUENCES_H
|
||||||
|
|
||||||
|
#include "fse.h" /* FSE_repeat, FSE_CTable */
|
||||||
|
#include "zstd_internal.h" /* symbolEncodingType_e, ZSTD_strategy */
|
||||||
|
|
||||||
|
typedef enum {
|
||||||
|
ZSTD_defaultDisallowed = 0,
|
||||||
|
ZSTD_defaultAllowed = 1
|
||||||
|
} ZSTD_defaultPolicy_e;
|
||||||
|
|
||||||
|
symbolEncodingType_e
|
||||||
|
ZSTD_selectEncodingType(
|
||||||
|
FSE_repeat* repeatMode, unsigned const* count, unsigned const max,
|
||||||
|
size_t const mostFrequent, size_t nbSeq, unsigned const FSELog,
|
||||||
|
FSE_CTable const* prevCTable,
|
||||||
|
short const* defaultNorm, U32 defaultNormLog,
|
||||||
|
ZSTD_defaultPolicy_e const isDefaultAllowed,
|
||||||
|
ZSTD_strategy const strategy);
|
||||||
|
|
||||||
|
size_t
|
||||||
|
ZSTD_buildCTable(void* dst, size_t dstCapacity,
|
||||||
|
FSE_CTable* nextCTable, U32 FSELog, symbolEncodingType_e type,
|
||||||
|
unsigned* count, U32 max,
|
||||||
|
const BYTE* codeTable, size_t nbSeq,
|
||||||
|
const S16* defaultNorm, U32 defaultNormLog, U32 defaultMax,
|
||||||
|
const FSE_CTable* prevCTable, size_t prevCTableSize,
|
||||||
|
void* entropyWorkspace, size_t entropyWorkspaceSize);
|
||||||
|
|
||||||
|
size_t ZSTD_encodeSequences(
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
FSE_CTable const* CTable_MatchLength, BYTE const* mlCodeTable,
|
||||||
|
FSE_CTable const* CTable_OffsetBits, BYTE const* ofCodeTable,
|
||||||
|
FSE_CTable const* CTable_LitLength, BYTE const* llCodeTable,
|
||||||
|
seqDef const* sequences, size_t nbSeq, int longOffsets, int bmi2);
|
||||||
|
|
||||||
|
#endif /* ZSTD_COMPRESS_SEQUENCES_H */
|
||||||
535
vendor/github.com/DataDog/zstd/zstd_cwksp.h
generated
vendored
Normal file
535
vendor/github.com/DataDog/zstd/zstd_cwksp.h
generated
vendored
Normal file
@@ -0,0 +1,535 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_CWKSP_H
|
||||||
|
#define ZSTD_CWKSP_H
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Dependencies
|
||||||
|
***************************************/
|
||||||
|
#include "zstd_internal.h"
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Constants
|
||||||
|
***************************************/
|
||||||
|
|
||||||
|
/* define "workspace is too large" as this number of times larger than needed */
|
||||||
|
#define ZSTD_WORKSPACETOOLARGE_FACTOR 3
|
||||||
|
|
||||||
|
/* when workspace is continuously too large
|
||||||
|
* during at least this number of times,
|
||||||
|
* context's memory usage is considered wasteful,
|
||||||
|
* because it's sized to handle a worst case scenario which rarely happens.
|
||||||
|
* In which case, resize it down to free some memory */
|
||||||
|
#define ZSTD_WORKSPACETOOLARGE_MAXDURATION 128
|
||||||
|
|
||||||
|
/* Since the workspace is effectively its own little malloc implementation /
|
||||||
|
* arena, when we run under ASAN, we should similarly insert redzones between
|
||||||
|
* each internal element of the workspace, so ASAN will catch overruns that
|
||||||
|
* reach outside an object but that stay inside the workspace.
|
||||||
|
*
|
||||||
|
* This defines the size of that redzone.
|
||||||
|
*/
|
||||||
|
#ifndef ZSTD_CWKSP_ASAN_REDZONE_SIZE
|
||||||
|
#define ZSTD_CWKSP_ASAN_REDZONE_SIZE 128
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Structures
|
||||||
|
***************************************/
|
||||||
|
typedef enum {
|
||||||
|
ZSTD_cwksp_alloc_objects,
|
||||||
|
ZSTD_cwksp_alloc_buffers,
|
||||||
|
ZSTD_cwksp_alloc_aligned
|
||||||
|
} ZSTD_cwksp_alloc_phase_e;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Zstd fits all its internal datastructures into a single continuous buffer,
|
||||||
|
* so that it only needs to perform a single OS allocation (or so that a buffer
|
||||||
|
* can be provided to it and it can perform no allocations at all). This buffer
|
||||||
|
* is called the workspace.
|
||||||
|
*
|
||||||
|
* Several optimizations complicate that process of allocating memory ranges
|
||||||
|
* from this workspace for each internal datastructure:
|
||||||
|
*
|
||||||
|
* - These different internal datastructures have different setup requirements:
|
||||||
|
*
|
||||||
|
* - The static objects need to be cleared once and can then be trivially
|
||||||
|
* reused for each compression.
|
||||||
|
*
|
||||||
|
* - Various buffers don't need to be initialized at all--they are always
|
||||||
|
* written into before they're read.
|
||||||
|
*
|
||||||
|
* - The matchstate tables have a unique requirement that they don't need
|
||||||
|
* their memory to be totally cleared, but they do need the memory to have
|
||||||
|
* some bound, i.e., a guarantee that all values in the memory they've been
|
||||||
|
* allocated is less than some maximum value (which is the starting value
|
||||||
|
* for the indices that they will then use for compression). When this
|
||||||
|
* guarantee is provided to them, they can use the memory without any setup
|
||||||
|
* work. When it can't, they have to clear the area.
|
||||||
|
*
|
||||||
|
* - These buffers also have different alignment requirements.
|
||||||
|
*
|
||||||
|
* - We would like to reuse the objects in the workspace for multiple
|
||||||
|
* compressions without having to perform any expensive reallocation or
|
||||||
|
* reinitialization work.
|
||||||
|
*
|
||||||
|
* - We would like to be able to efficiently reuse the workspace across
|
||||||
|
* multiple compressions **even when the compression parameters change** and
|
||||||
|
* we need to resize some of the objects (where possible).
|
||||||
|
*
|
||||||
|
* To attempt to manage this buffer, given these constraints, the ZSTD_cwksp
|
||||||
|
* abstraction was created. It works as follows:
|
||||||
|
*
|
||||||
|
* Workspace Layout:
|
||||||
|
*
|
||||||
|
* [ ... workspace ... ]
|
||||||
|
* [objects][tables ... ->] free space [<- ... aligned][<- ... buffers]
|
||||||
|
*
|
||||||
|
* The various objects that live in the workspace are divided into the
|
||||||
|
* following categories, and are allocated separately:
|
||||||
|
*
|
||||||
|
* - Static objects: this is optionally the enclosing ZSTD_CCtx or ZSTD_CDict,
|
||||||
|
* so that literally everything fits in a single buffer. Note: if present,
|
||||||
|
* this must be the first object in the workspace, since ZSTD_free{CCtx,
|
||||||
|
* CDict}() rely on a pointer comparison to see whether one or two frees are
|
||||||
|
* required.
|
||||||
|
*
|
||||||
|
* - Fixed size objects: these are fixed-size, fixed-count objects that are
|
||||||
|
* nonetheless "dynamically" allocated in the workspace so that we can
|
||||||
|
* control how they're initialized separately from the broader ZSTD_CCtx.
|
||||||
|
* Examples:
|
||||||
|
* - Entropy Workspace
|
||||||
|
* - 2 x ZSTD_compressedBlockState_t
|
||||||
|
* - CDict dictionary contents
|
||||||
|
*
|
||||||
|
* - Tables: these are any of several different datastructures (hash tables,
|
||||||
|
* chain tables, binary trees) that all respect a common format: they are
|
||||||
|
* uint32_t arrays, all of whose values are between 0 and (nextSrc - base).
|
||||||
|
* Their sizes depend on the cparams.
|
||||||
|
*
|
||||||
|
* - Aligned: these buffers are used for various purposes that require 4 byte
|
||||||
|
* alignment, but don't require any initialization before they're used.
|
||||||
|
*
|
||||||
|
* - Buffers: these buffers are used for various purposes that don't require
|
||||||
|
* any alignment or initialization before they're used. This means they can
|
||||||
|
* be moved around at no cost for a new compression.
|
||||||
|
*
|
||||||
|
* Allocating Memory:
|
||||||
|
*
|
||||||
|
* The various types of objects must be allocated in order, so they can be
|
||||||
|
* correctly packed into the workspace buffer. That order is:
|
||||||
|
*
|
||||||
|
* 1. Objects
|
||||||
|
* 2. Buffers
|
||||||
|
* 3. Aligned
|
||||||
|
* 4. Tables
|
||||||
|
*
|
||||||
|
* Attempts to reserve objects of different types out of order will fail.
|
||||||
|
*/
|
||||||
|
typedef struct {
|
||||||
|
void* workspace;
|
||||||
|
void* workspaceEnd;
|
||||||
|
|
||||||
|
void* objectEnd;
|
||||||
|
void* tableEnd;
|
||||||
|
void* tableValidEnd;
|
||||||
|
void* allocStart;
|
||||||
|
|
||||||
|
int allocFailed;
|
||||||
|
int workspaceOversizedDuration;
|
||||||
|
ZSTD_cwksp_alloc_phase_e phase;
|
||||||
|
} ZSTD_cwksp;
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Functions
|
||||||
|
***************************************/
|
||||||
|
|
||||||
|
MEM_STATIC size_t ZSTD_cwksp_available_space(ZSTD_cwksp* ws);
|
||||||
|
|
||||||
|
MEM_STATIC void ZSTD_cwksp_assert_internal_consistency(ZSTD_cwksp* ws) {
|
||||||
|
(void)ws;
|
||||||
|
assert(ws->workspace <= ws->objectEnd);
|
||||||
|
assert(ws->objectEnd <= ws->tableEnd);
|
||||||
|
assert(ws->objectEnd <= ws->tableValidEnd);
|
||||||
|
assert(ws->tableEnd <= ws->allocStart);
|
||||||
|
assert(ws->tableValidEnd <= ws->allocStart);
|
||||||
|
assert(ws->allocStart <= ws->workspaceEnd);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Align must be a power of 2.
|
||||||
|
*/
|
||||||
|
MEM_STATIC size_t ZSTD_cwksp_align(size_t size, size_t const align) {
|
||||||
|
size_t const mask = align - 1;
|
||||||
|
assert((align & mask) == 0);
|
||||||
|
return (size + mask) & ~mask;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Use this to determine how much space in the workspace we will consume to
|
||||||
|
* allocate this object. (Normally it should be exactly the size of the object,
|
||||||
|
* but under special conditions, like ASAN, where we pad each object, it might
|
||||||
|
* be larger.)
|
||||||
|
*
|
||||||
|
* Since tables aren't currently redzoned, you don't need to call through this
|
||||||
|
* to figure out how much space you need for the matchState tables. Everything
|
||||||
|
* else is though.
|
||||||
|
*/
|
||||||
|
MEM_STATIC size_t ZSTD_cwksp_alloc_size(size_t size) {
|
||||||
|
#if defined (ADDRESS_SANITIZER) && !defined (ZSTD_ASAN_DONT_POISON_WORKSPACE)
|
||||||
|
return size + 2 * ZSTD_CWKSP_ASAN_REDZONE_SIZE;
|
||||||
|
#else
|
||||||
|
return size;
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void ZSTD_cwksp_internal_advance_phase(
|
||||||
|
ZSTD_cwksp* ws, ZSTD_cwksp_alloc_phase_e phase) {
|
||||||
|
assert(phase >= ws->phase);
|
||||||
|
if (phase > ws->phase) {
|
||||||
|
if (ws->phase < ZSTD_cwksp_alloc_buffers &&
|
||||||
|
phase >= ZSTD_cwksp_alloc_buffers) {
|
||||||
|
ws->tableValidEnd = ws->objectEnd;
|
||||||
|
}
|
||||||
|
if (ws->phase < ZSTD_cwksp_alloc_aligned &&
|
||||||
|
phase >= ZSTD_cwksp_alloc_aligned) {
|
||||||
|
/* If unaligned allocations down from a too-large top have left us
|
||||||
|
* unaligned, we need to realign our alloc ptr. Technically, this
|
||||||
|
* can consume space that is unaccounted for in the neededSpace
|
||||||
|
* calculation. However, I believe this can only happen when the
|
||||||
|
* workspace is too large, and specifically when it is too large
|
||||||
|
* by a larger margin than the space that will be consumed. */
|
||||||
|
/* TODO: cleaner, compiler warning friendly way to do this??? */
|
||||||
|
ws->allocStart = (BYTE*)ws->allocStart - ((size_t)ws->allocStart & (sizeof(U32)-1));
|
||||||
|
if (ws->allocStart < ws->tableValidEnd) {
|
||||||
|
ws->tableValidEnd = ws->allocStart;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
ws->phase = phase;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns whether this object/buffer/etc was allocated in this workspace.
|
||||||
|
*/
|
||||||
|
MEM_STATIC int ZSTD_cwksp_owns_buffer(const ZSTD_cwksp* ws, const void* ptr) {
|
||||||
|
return (ptr != NULL) && (ws->workspace <= ptr) && (ptr <= ws->workspaceEnd);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Internal function. Do not use directly.
|
||||||
|
*/
|
||||||
|
MEM_STATIC void* ZSTD_cwksp_reserve_internal(
|
||||||
|
ZSTD_cwksp* ws, size_t bytes, ZSTD_cwksp_alloc_phase_e phase) {
|
||||||
|
void* alloc;
|
||||||
|
void* bottom = ws->tableEnd;
|
||||||
|
ZSTD_cwksp_internal_advance_phase(ws, phase);
|
||||||
|
alloc = (BYTE *)ws->allocStart - bytes;
|
||||||
|
|
||||||
|
#if defined (ADDRESS_SANITIZER) && !defined (ZSTD_ASAN_DONT_POISON_WORKSPACE)
|
||||||
|
/* over-reserve space */
|
||||||
|
alloc = (BYTE *)alloc - 2 * ZSTD_CWKSP_ASAN_REDZONE_SIZE;
|
||||||
|
#endif
|
||||||
|
|
||||||
|
DEBUGLOG(5, "cwksp: reserving %p %zd bytes, %zd bytes remaining",
|
||||||
|
alloc, bytes, ZSTD_cwksp_available_space(ws) - bytes);
|
||||||
|
ZSTD_cwksp_assert_internal_consistency(ws);
|
||||||
|
assert(alloc >= bottom);
|
||||||
|
if (alloc < bottom) {
|
||||||
|
DEBUGLOG(4, "cwksp: alloc failed!");
|
||||||
|
ws->allocFailed = 1;
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
if (alloc < ws->tableValidEnd) {
|
||||||
|
ws->tableValidEnd = alloc;
|
||||||
|
}
|
||||||
|
ws->allocStart = alloc;
|
||||||
|
|
||||||
|
#if defined (ADDRESS_SANITIZER) && !defined (ZSTD_ASAN_DONT_POISON_WORKSPACE)
|
||||||
|
/* Move alloc so there's ZSTD_CWKSP_ASAN_REDZONE_SIZE unused space on
|
||||||
|
* either size. */
|
||||||
|
alloc = (BYTE *)alloc + ZSTD_CWKSP_ASAN_REDZONE_SIZE;
|
||||||
|
__asan_unpoison_memory_region(alloc, bytes);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
return alloc;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Reserves and returns unaligned memory.
|
||||||
|
*/
|
||||||
|
MEM_STATIC BYTE* ZSTD_cwksp_reserve_buffer(ZSTD_cwksp* ws, size_t bytes) {
|
||||||
|
return (BYTE*)ZSTD_cwksp_reserve_internal(ws, bytes, ZSTD_cwksp_alloc_buffers);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Reserves and returns memory sized on and aligned on sizeof(unsigned).
|
||||||
|
*/
|
||||||
|
MEM_STATIC void* ZSTD_cwksp_reserve_aligned(ZSTD_cwksp* ws, size_t bytes) {
|
||||||
|
assert((bytes & (sizeof(U32)-1)) == 0);
|
||||||
|
return ZSTD_cwksp_reserve_internal(ws, ZSTD_cwksp_align(bytes, sizeof(U32)), ZSTD_cwksp_alloc_aligned);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Aligned on sizeof(unsigned). These buffers have the special property that
|
||||||
|
* their values remain constrained, allowing us to re-use them without
|
||||||
|
* memset()-ing them.
|
||||||
|
*/
|
||||||
|
MEM_STATIC void* ZSTD_cwksp_reserve_table(ZSTD_cwksp* ws, size_t bytes) {
|
||||||
|
const ZSTD_cwksp_alloc_phase_e phase = ZSTD_cwksp_alloc_aligned;
|
||||||
|
void* alloc = ws->tableEnd;
|
||||||
|
void* end = (BYTE *)alloc + bytes;
|
||||||
|
void* top = ws->allocStart;
|
||||||
|
|
||||||
|
DEBUGLOG(5, "cwksp: reserving %p table %zd bytes, %zd bytes remaining",
|
||||||
|
alloc, bytes, ZSTD_cwksp_available_space(ws) - bytes);
|
||||||
|
assert((bytes & (sizeof(U32)-1)) == 0);
|
||||||
|
ZSTD_cwksp_internal_advance_phase(ws, phase);
|
||||||
|
ZSTD_cwksp_assert_internal_consistency(ws);
|
||||||
|
assert(end <= top);
|
||||||
|
if (end > top) {
|
||||||
|
DEBUGLOG(4, "cwksp: table alloc failed!");
|
||||||
|
ws->allocFailed = 1;
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
ws->tableEnd = end;
|
||||||
|
|
||||||
|
#if defined (ADDRESS_SANITIZER) && !defined (ZSTD_ASAN_DONT_POISON_WORKSPACE)
|
||||||
|
__asan_unpoison_memory_region(alloc, bytes);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
return alloc;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Aligned on sizeof(void*).
|
||||||
|
*/
|
||||||
|
MEM_STATIC void* ZSTD_cwksp_reserve_object(ZSTD_cwksp* ws, size_t bytes) {
|
||||||
|
size_t roundedBytes = ZSTD_cwksp_align(bytes, sizeof(void*));
|
||||||
|
void* alloc = ws->objectEnd;
|
||||||
|
void* end = (BYTE*)alloc + roundedBytes;
|
||||||
|
|
||||||
|
#if defined (ADDRESS_SANITIZER) && !defined (ZSTD_ASAN_DONT_POISON_WORKSPACE)
|
||||||
|
/* over-reserve space */
|
||||||
|
end = (BYTE *)end + 2 * ZSTD_CWKSP_ASAN_REDZONE_SIZE;
|
||||||
|
#endif
|
||||||
|
|
||||||
|
DEBUGLOG(5,
|
||||||
|
"cwksp: reserving %p object %zd bytes (rounded to %zd), %zd bytes remaining",
|
||||||
|
alloc, bytes, roundedBytes, ZSTD_cwksp_available_space(ws) - roundedBytes);
|
||||||
|
assert(((size_t)alloc & (sizeof(void*)-1)) == 0);
|
||||||
|
assert((bytes & (sizeof(void*)-1)) == 0);
|
||||||
|
ZSTD_cwksp_assert_internal_consistency(ws);
|
||||||
|
/* we must be in the first phase, no advance is possible */
|
||||||
|
if (ws->phase != ZSTD_cwksp_alloc_objects || end > ws->workspaceEnd) {
|
||||||
|
DEBUGLOG(4, "cwksp: object alloc failed!");
|
||||||
|
ws->allocFailed = 1;
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
ws->objectEnd = end;
|
||||||
|
ws->tableEnd = end;
|
||||||
|
ws->tableValidEnd = end;
|
||||||
|
|
||||||
|
#if defined (ADDRESS_SANITIZER) && !defined (ZSTD_ASAN_DONT_POISON_WORKSPACE)
|
||||||
|
/* Move alloc so there's ZSTD_CWKSP_ASAN_REDZONE_SIZE unused space on
|
||||||
|
* either size. */
|
||||||
|
alloc = (BYTE *)alloc + ZSTD_CWKSP_ASAN_REDZONE_SIZE;
|
||||||
|
__asan_unpoison_memory_region(alloc, bytes);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
return alloc;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void ZSTD_cwksp_mark_tables_dirty(ZSTD_cwksp* ws) {
|
||||||
|
DEBUGLOG(4, "cwksp: ZSTD_cwksp_mark_tables_dirty");
|
||||||
|
|
||||||
|
#if defined (MEMORY_SANITIZER) && !defined (ZSTD_MSAN_DONT_POISON_WORKSPACE)
|
||||||
|
/* To validate that the table re-use logic is sound, and that we don't
|
||||||
|
* access table space that we haven't cleaned, we re-"poison" the table
|
||||||
|
* space every time we mark it dirty. */
|
||||||
|
{
|
||||||
|
size_t size = (BYTE*)ws->tableValidEnd - (BYTE*)ws->objectEnd;
|
||||||
|
assert(__msan_test_shadow(ws->objectEnd, size) == -1);
|
||||||
|
__msan_poison(ws->objectEnd, size);
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
assert(ws->tableValidEnd >= ws->objectEnd);
|
||||||
|
assert(ws->tableValidEnd <= ws->allocStart);
|
||||||
|
ws->tableValidEnd = ws->objectEnd;
|
||||||
|
ZSTD_cwksp_assert_internal_consistency(ws);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void ZSTD_cwksp_mark_tables_clean(ZSTD_cwksp* ws) {
|
||||||
|
DEBUGLOG(4, "cwksp: ZSTD_cwksp_mark_tables_clean");
|
||||||
|
assert(ws->tableValidEnd >= ws->objectEnd);
|
||||||
|
assert(ws->tableValidEnd <= ws->allocStart);
|
||||||
|
if (ws->tableValidEnd < ws->tableEnd) {
|
||||||
|
ws->tableValidEnd = ws->tableEnd;
|
||||||
|
}
|
||||||
|
ZSTD_cwksp_assert_internal_consistency(ws);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Zero the part of the allocated tables not already marked clean.
|
||||||
|
*/
|
||||||
|
MEM_STATIC void ZSTD_cwksp_clean_tables(ZSTD_cwksp* ws) {
|
||||||
|
DEBUGLOG(4, "cwksp: ZSTD_cwksp_clean_tables");
|
||||||
|
assert(ws->tableValidEnd >= ws->objectEnd);
|
||||||
|
assert(ws->tableValidEnd <= ws->allocStart);
|
||||||
|
if (ws->tableValidEnd < ws->tableEnd) {
|
||||||
|
memset(ws->tableValidEnd, 0, (BYTE*)ws->tableEnd - (BYTE*)ws->tableValidEnd);
|
||||||
|
}
|
||||||
|
ZSTD_cwksp_mark_tables_clean(ws);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Invalidates table allocations.
|
||||||
|
* All other allocations remain valid.
|
||||||
|
*/
|
||||||
|
MEM_STATIC void ZSTD_cwksp_clear_tables(ZSTD_cwksp* ws) {
|
||||||
|
DEBUGLOG(4, "cwksp: clearing tables!");
|
||||||
|
|
||||||
|
#if defined (ADDRESS_SANITIZER) && !defined (ZSTD_ASAN_DONT_POISON_WORKSPACE)
|
||||||
|
{
|
||||||
|
size_t size = (BYTE*)ws->tableValidEnd - (BYTE*)ws->objectEnd;
|
||||||
|
__asan_poison_memory_region(ws->objectEnd, size);
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
ws->tableEnd = ws->objectEnd;
|
||||||
|
ZSTD_cwksp_assert_internal_consistency(ws);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Invalidates all buffer, aligned, and table allocations.
|
||||||
|
* Object allocations remain valid.
|
||||||
|
*/
|
||||||
|
MEM_STATIC void ZSTD_cwksp_clear(ZSTD_cwksp* ws) {
|
||||||
|
DEBUGLOG(4, "cwksp: clearing!");
|
||||||
|
|
||||||
|
#if defined (MEMORY_SANITIZER) && !defined (ZSTD_MSAN_DONT_POISON_WORKSPACE)
|
||||||
|
/* To validate that the context re-use logic is sound, and that we don't
|
||||||
|
* access stuff that this compression hasn't initialized, we re-"poison"
|
||||||
|
* the workspace (or at least the non-static, non-table parts of it)
|
||||||
|
* every time we start a new compression. */
|
||||||
|
{
|
||||||
|
size_t size = (BYTE*)ws->workspaceEnd - (BYTE*)ws->tableValidEnd;
|
||||||
|
__msan_poison(ws->tableValidEnd, size);
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#if defined (ADDRESS_SANITIZER) && !defined (ZSTD_ASAN_DONT_POISON_WORKSPACE)
|
||||||
|
{
|
||||||
|
size_t size = (BYTE*)ws->workspaceEnd - (BYTE*)ws->objectEnd;
|
||||||
|
__asan_poison_memory_region(ws->objectEnd, size);
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
ws->tableEnd = ws->objectEnd;
|
||||||
|
ws->allocStart = ws->workspaceEnd;
|
||||||
|
ws->allocFailed = 0;
|
||||||
|
if (ws->phase > ZSTD_cwksp_alloc_buffers) {
|
||||||
|
ws->phase = ZSTD_cwksp_alloc_buffers;
|
||||||
|
}
|
||||||
|
ZSTD_cwksp_assert_internal_consistency(ws);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The provided workspace takes ownership of the buffer [start, start+size).
|
||||||
|
* Any existing values in the workspace are ignored (the previously managed
|
||||||
|
* buffer, if present, must be separately freed).
|
||||||
|
*/
|
||||||
|
MEM_STATIC void ZSTD_cwksp_init(ZSTD_cwksp* ws, void* start, size_t size) {
|
||||||
|
DEBUGLOG(4, "cwksp: init'ing workspace with %zd bytes", size);
|
||||||
|
assert(((size_t)start & (sizeof(void*)-1)) == 0); /* ensure correct alignment */
|
||||||
|
ws->workspace = start;
|
||||||
|
ws->workspaceEnd = (BYTE*)start + size;
|
||||||
|
ws->objectEnd = ws->workspace;
|
||||||
|
ws->tableValidEnd = ws->objectEnd;
|
||||||
|
ws->phase = ZSTD_cwksp_alloc_objects;
|
||||||
|
ZSTD_cwksp_clear(ws);
|
||||||
|
ws->workspaceOversizedDuration = 0;
|
||||||
|
ZSTD_cwksp_assert_internal_consistency(ws);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC size_t ZSTD_cwksp_create(ZSTD_cwksp* ws, size_t size, ZSTD_customMem customMem) {
|
||||||
|
void* workspace = ZSTD_malloc(size, customMem);
|
||||||
|
DEBUGLOG(4, "cwksp: creating new workspace with %zd bytes", size);
|
||||||
|
RETURN_ERROR_IF(workspace == NULL, memory_allocation);
|
||||||
|
ZSTD_cwksp_init(ws, workspace, size);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void ZSTD_cwksp_free(ZSTD_cwksp* ws, ZSTD_customMem customMem) {
|
||||||
|
void *ptr = ws->workspace;
|
||||||
|
DEBUGLOG(4, "cwksp: freeing workspace");
|
||||||
|
memset(ws, 0, sizeof(ZSTD_cwksp));
|
||||||
|
ZSTD_free(ptr, customMem);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Moves the management of a workspace from one cwksp to another. The src cwksp
|
||||||
|
* is left in an invalid state (src must be re-init()'ed before its used again).
|
||||||
|
*/
|
||||||
|
MEM_STATIC void ZSTD_cwksp_move(ZSTD_cwksp* dst, ZSTD_cwksp* src) {
|
||||||
|
*dst = *src;
|
||||||
|
memset(src, 0, sizeof(ZSTD_cwksp));
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC size_t ZSTD_cwksp_sizeof(const ZSTD_cwksp* ws) {
|
||||||
|
return (size_t)((BYTE*)ws->workspaceEnd - (BYTE*)ws->workspace);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC int ZSTD_cwksp_reserve_failed(const ZSTD_cwksp* ws) {
|
||||||
|
return ws->allocFailed;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Functions Checking Free Space
|
||||||
|
***************************************/
|
||||||
|
|
||||||
|
MEM_STATIC size_t ZSTD_cwksp_available_space(ZSTD_cwksp* ws) {
|
||||||
|
return (size_t)((BYTE*)ws->allocStart - (BYTE*)ws->tableEnd);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC int ZSTD_cwksp_check_available(ZSTD_cwksp* ws, size_t additionalNeededSpace) {
|
||||||
|
return ZSTD_cwksp_available_space(ws) >= additionalNeededSpace;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC int ZSTD_cwksp_check_too_large(ZSTD_cwksp* ws, size_t additionalNeededSpace) {
|
||||||
|
return ZSTD_cwksp_check_available(
|
||||||
|
ws, additionalNeededSpace * ZSTD_WORKSPACETOOLARGE_FACTOR);
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC int ZSTD_cwksp_check_wasteful(ZSTD_cwksp* ws, size_t additionalNeededSpace) {
|
||||||
|
return ZSTD_cwksp_check_too_large(ws, additionalNeededSpace)
|
||||||
|
&& ws->workspaceOversizedDuration > ZSTD_WORKSPACETOOLARGE_MAXDURATION;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC void ZSTD_cwksp_bump_oversized_duration(
|
||||||
|
ZSTD_cwksp* ws, size_t additionalNeededSpace) {
|
||||||
|
if (ZSTD_cwksp_check_too_large(ws, additionalNeededSpace)) {
|
||||||
|
ws->workspaceOversizedDuration++;
|
||||||
|
} else {
|
||||||
|
ws->workspaceOversizedDuration = 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_CWKSP_H */
|
||||||
240
vendor/github.com/DataDog/zstd/zstd_ddict.c
generated
vendored
Normal file
240
vendor/github.com/DataDog/zstd/zstd_ddict.c
generated
vendored
Normal file
@@ -0,0 +1,240 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* zstd_ddict.c :
|
||||||
|
* concentrates all logic that needs to know the internals of ZSTD_DDict object */
|
||||||
|
|
||||||
|
/*-*******************************************************
|
||||||
|
* Dependencies
|
||||||
|
*********************************************************/
|
||||||
|
#include <string.h> /* memcpy, memmove, memset */
|
||||||
|
#include "cpu.h" /* bmi2 */
|
||||||
|
#include "mem.h" /* low level memory routines */
|
||||||
|
#define FSE_STATIC_LINKING_ONLY
|
||||||
|
#include "fse.h"
|
||||||
|
#define HUF_STATIC_LINKING_ONLY
|
||||||
|
#include "huf.h"
|
||||||
|
#include "zstd_decompress_internal.h"
|
||||||
|
#include "zstd_ddict.h"
|
||||||
|
|
||||||
|
#if defined(ZSTD_LEGACY_SUPPORT) && (ZSTD_LEGACY_SUPPORT>=1)
|
||||||
|
# include "zstd_legacy.h"
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/*-*******************************************************
|
||||||
|
* Types
|
||||||
|
*********************************************************/
|
||||||
|
struct ZSTD_DDict_s {
|
||||||
|
void* dictBuffer;
|
||||||
|
const void* dictContent;
|
||||||
|
size_t dictSize;
|
||||||
|
ZSTD_entropyDTables_t entropy;
|
||||||
|
U32 dictID;
|
||||||
|
U32 entropyPresent;
|
||||||
|
ZSTD_customMem cMem;
|
||||||
|
}; /* typedef'd to ZSTD_DDict within "zstd.h" */
|
||||||
|
|
||||||
|
const void* ZSTD_DDict_dictContent(const ZSTD_DDict* ddict)
|
||||||
|
{
|
||||||
|
assert(ddict != NULL);
|
||||||
|
return ddict->dictContent;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZSTD_DDict_dictSize(const ZSTD_DDict* ddict)
|
||||||
|
{
|
||||||
|
assert(ddict != NULL);
|
||||||
|
return ddict->dictSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
void ZSTD_copyDDictParameters(ZSTD_DCtx* dctx, const ZSTD_DDict* ddict)
|
||||||
|
{
|
||||||
|
DEBUGLOG(4, "ZSTD_copyDDictParameters");
|
||||||
|
assert(dctx != NULL);
|
||||||
|
assert(ddict != NULL);
|
||||||
|
dctx->dictID = ddict->dictID;
|
||||||
|
dctx->prefixStart = ddict->dictContent;
|
||||||
|
dctx->virtualStart = ddict->dictContent;
|
||||||
|
dctx->dictEnd = (const BYTE*)ddict->dictContent + ddict->dictSize;
|
||||||
|
dctx->previousDstEnd = dctx->dictEnd;
|
||||||
|
if (ddict->entropyPresent) {
|
||||||
|
dctx->litEntropy = 1;
|
||||||
|
dctx->fseEntropy = 1;
|
||||||
|
dctx->LLTptr = ddict->entropy.LLTable;
|
||||||
|
dctx->MLTptr = ddict->entropy.MLTable;
|
||||||
|
dctx->OFTptr = ddict->entropy.OFTable;
|
||||||
|
dctx->HUFptr = ddict->entropy.hufTable;
|
||||||
|
dctx->entropy.rep[0] = ddict->entropy.rep[0];
|
||||||
|
dctx->entropy.rep[1] = ddict->entropy.rep[1];
|
||||||
|
dctx->entropy.rep[2] = ddict->entropy.rep[2];
|
||||||
|
} else {
|
||||||
|
dctx->litEntropy = 0;
|
||||||
|
dctx->fseEntropy = 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static size_t
|
||||||
|
ZSTD_loadEntropy_intoDDict(ZSTD_DDict* ddict,
|
||||||
|
ZSTD_dictContentType_e dictContentType)
|
||||||
|
{
|
||||||
|
ddict->dictID = 0;
|
||||||
|
ddict->entropyPresent = 0;
|
||||||
|
if (dictContentType == ZSTD_dct_rawContent) return 0;
|
||||||
|
|
||||||
|
if (ddict->dictSize < 8) {
|
||||||
|
if (dictContentType == ZSTD_dct_fullDict)
|
||||||
|
return ERROR(dictionary_corrupted); /* only accept specified dictionaries */
|
||||||
|
return 0; /* pure content mode */
|
||||||
|
}
|
||||||
|
{ U32 const magic = MEM_readLE32(ddict->dictContent);
|
||||||
|
if (magic != ZSTD_MAGIC_DICTIONARY) {
|
||||||
|
if (dictContentType == ZSTD_dct_fullDict)
|
||||||
|
return ERROR(dictionary_corrupted); /* only accept specified dictionaries */
|
||||||
|
return 0; /* pure content mode */
|
||||||
|
}
|
||||||
|
}
|
||||||
|
ddict->dictID = MEM_readLE32((const char*)ddict->dictContent + ZSTD_FRAMEIDSIZE);
|
||||||
|
|
||||||
|
/* load entropy tables */
|
||||||
|
RETURN_ERROR_IF(ZSTD_isError(ZSTD_loadDEntropy(
|
||||||
|
&ddict->entropy, ddict->dictContent, ddict->dictSize)),
|
||||||
|
dictionary_corrupted);
|
||||||
|
ddict->entropyPresent = 1;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static size_t ZSTD_initDDict_internal(ZSTD_DDict* ddict,
|
||||||
|
const void* dict, size_t dictSize,
|
||||||
|
ZSTD_dictLoadMethod_e dictLoadMethod,
|
||||||
|
ZSTD_dictContentType_e dictContentType)
|
||||||
|
{
|
||||||
|
if ((dictLoadMethod == ZSTD_dlm_byRef) || (!dict) || (!dictSize)) {
|
||||||
|
ddict->dictBuffer = NULL;
|
||||||
|
ddict->dictContent = dict;
|
||||||
|
if (!dict) dictSize = 0;
|
||||||
|
} else {
|
||||||
|
void* const internalBuffer = ZSTD_malloc(dictSize, ddict->cMem);
|
||||||
|
ddict->dictBuffer = internalBuffer;
|
||||||
|
ddict->dictContent = internalBuffer;
|
||||||
|
if (!internalBuffer) return ERROR(memory_allocation);
|
||||||
|
memcpy(internalBuffer, dict, dictSize);
|
||||||
|
}
|
||||||
|
ddict->dictSize = dictSize;
|
||||||
|
ddict->entropy.hufTable[0] = (HUF_DTable)((HufLog)*0x1000001); /* cover both little and big endian */
|
||||||
|
|
||||||
|
/* parse dictionary content */
|
||||||
|
FORWARD_IF_ERROR( ZSTD_loadEntropy_intoDDict(ddict, dictContentType) );
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
ZSTD_DDict* ZSTD_createDDict_advanced(const void* dict, size_t dictSize,
|
||||||
|
ZSTD_dictLoadMethod_e dictLoadMethod,
|
||||||
|
ZSTD_dictContentType_e dictContentType,
|
||||||
|
ZSTD_customMem customMem)
|
||||||
|
{
|
||||||
|
if (!customMem.customAlloc ^ !customMem.customFree) return NULL;
|
||||||
|
|
||||||
|
{ ZSTD_DDict* const ddict = (ZSTD_DDict*) ZSTD_malloc(sizeof(ZSTD_DDict), customMem);
|
||||||
|
if (ddict == NULL) return NULL;
|
||||||
|
ddict->cMem = customMem;
|
||||||
|
{ size_t const initResult = ZSTD_initDDict_internal(ddict,
|
||||||
|
dict, dictSize,
|
||||||
|
dictLoadMethod, dictContentType);
|
||||||
|
if (ZSTD_isError(initResult)) {
|
||||||
|
ZSTD_freeDDict(ddict);
|
||||||
|
return NULL;
|
||||||
|
} }
|
||||||
|
return ddict;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! ZSTD_createDDict() :
|
||||||
|
* Create a digested dictionary, to start decompression without startup delay.
|
||||||
|
* `dict` content is copied inside DDict.
|
||||||
|
* Consequently, `dict` can be released after `ZSTD_DDict` creation */
|
||||||
|
ZSTD_DDict* ZSTD_createDDict(const void* dict, size_t dictSize)
|
||||||
|
{
|
||||||
|
ZSTD_customMem const allocator = { NULL, NULL, NULL };
|
||||||
|
return ZSTD_createDDict_advanced(dict, dictSize, ZSTD_dlm_byCopy, ZSTD_dct_auto, allocator);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! ZSTD_createDDict_byReference() :
|
||||||
|
* Create a digested dictionary, to start decompression without startup delay.
|
||||||
|
* Dictionary content is simply referenced, it will be accessed during decompression.
|
||||||
|
* Warning : dictBuffer must outlive DDict (DDict must be freed before dictBuffer) */
|
||||||
|
ZSTD_DDict* ZSTD_createDDict_byReference(const void* dictBuffer, size_t dictSize)
|
||||||
|
{
|
||||||
|
ZSTD_customMem const allocator = { NULL, NULL, NULL };
|
||||||
|
return ZSTD_createDDict_advanced(dictBuffer, dictSize, ZSTD_dlm_byRef, ZSTD_dct_auto, allocator);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
const ZSTD_DDict* ZSTD_initStaticDDict(
|
||||||
|
void* sBuffer, size_t sBufferSize,
|
||||||
|
const void* dict, size_t dictSize,
|
||||||
|
ZSTD_dictLoadMethod_e dictLoadMethod,
|
||||||
|
ZSTD_dictContentType_e dictContentType)
|
||||||
|
{
|
||||||
|
size_t const neededSpace = sizeof(ZSTD_DDict)
|
||||||
|
+ (dictLoadMethod == ZSTD_dlm_byRef ? 0 : dictSize);
|
||||||
|
ZSTD_DDict* const ddict = (ZSTD_DDict*)sBuffer;
|
||||||
|
assert(sBuffer != NULL);
|
||||||
|
assert(dict != NULL);
|
||||||
|
if ((size_t)sBuffer & 7) return NULL; /* 8-aligned */
|
||||||
|
if (sBufferSize < neededSpace) return NULL;
|
||||||
|
if (dictLoadMethod == ZSTD_dlm_byCopy) {
|
||||||
|
memcpy(ddict+1, dict, dictSize); /* local copy */
|
||||||
|
dict = ddict+1;
|
||||||
|
}
|
||||||
|
if (ZSTD_isError( ZSTD_initDDict_internal(ddict,
|
||||||
|
dict, dictSize,
|
||||||
|
ZSTD_dlm_byRef, dictContentType) ))
|
||||||
|
return NULL;
|
||||||
|
return ddict;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t ZSTD_freeDDict(ZSTD_DDict* ddict)
|
||||||
|
{
|
||||||
|
if (ddict==NULL) return 0; /* support free on NULL */
|
||||||
|
{ ZSTD_customMem const cMem = ddict->cMem;
|
||||||
|
ZSTD_free(ddict->dictBuffer, cMem);
|
||||||
|
ZSTD_free(ddict, cMem);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! ZSTD_estimateDDictSize() :
|
||||||
|
* Estimate amount of memory that will be needed to create a dictionary for decompression.
|
||||||
|
* Note : dictionary created by reference using ZSTD_dlm_byRef are smaller */
|
||||||
|
size_t ZSTD_estimateDDictSize(size_t dictSize, ZSTD_dictLoadMethod_e dictLoadMethod)
|
||||||
|
{
|
||||||
|
return sizeof(ZSTD_DDict) + (dictLoadMethod == ZSTD_dlm_byRef ? 0 : dictSize);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZSTD_sizeof_DDict(const ZSTD_DDict* ddict)
|
||||||
|
{
|
||||||
|
if (ddict==NULL) return 0; /* support sizeof on NULL */
|
||||||
|
return sizeof(*ddict) + (ddict->dictBuffer ? ddict->dictSize : 0) ;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! ZSTD_getDictID_fromDDict() :
|
||||||
|
* Provides the dictID of the dictionary loaded into `ddict`.
|
||||||
|
* If @return == 0, the dictionary is not conformant to Zstandard specification, or empty.
|
||||||
|
* Non-conformant dictionaries can still be loaded, but as content-only dictionaries. */
|
||||||
|
unsigned ZSTD_getDictID_fromDDict(const ZSTD_DDict* ddict)
|
||||||
|
{
|
||||||
|
if (ddict==NULL) return 0;
|
||||||
|
return ZSTD_getDictID_fromDict(ddict->dictContent, ddict->dictSize);
|
||||||
|
}
|
||||||
44
vendor/github.com/DataDog/zstd/zstd_ddict.h
generated
vendored
Normal file
44
vendor/github.com/DataDog/zstd/zstd_ddict.h
generated
vendored
Normal file
@@ -0,0 +1,44 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
#ifndef ZSTD_DDICT_H
|
||||||
|
#define ZSTD_DDICT_H
|
||||||
|
|
||||||
|
/*-*******************************************************
|
||||||
|
* Dependencies
|
||||||
|
*********************************************************/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
#include "zstd.h" /* ZSTD_DDict, and several public functions */
|
||||||
|
|
||||||
|
|
||||||
|
/*-*******************************************************
|
||||||
|
* Interface
|
||||||
|
*********************************************************/
|
||||||
|
|
||||||
|
/* note: several prototypes are already published in `zstd.h` :
|
||||||
|
* ZSTD_createDDict()
|
||||||
|
* ZSTD_createDDict_byReference()
|
||||||
|
* ZSTD_createDDict_advanced()
|
||||||
|
* ZSTD_freeDDict()
|
||||||
|
* ZSTD_initStaticDDict()
|
||||||
|
* ZSTD_sizeof_DDict()
|
||||||
|
* ZSTD_estimateDDictSize()
|
||||||
|
* ZSTD_getDictID_fromDict()
|
||||||
|
*/
|
||||||
|
|
||||||
|
const void* ZSTD_DDict_dictContent(const ZSTD_DDict* ddict);
|
||||||
|
size_t ZSTD_DDict_dictSize(const ZSTD_DDict* ddict);
|
||||||
|
|
||||||
|
void ZSTD_copyDDictParameters(ZSTD_DCtx* dctx, const ZSTD_DDict* ddict);
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
#endif /* ZSTD_DDICT_H */
|
||||||
1769
vendor/github.com/DataDog/zstd/zstd_decompress.c
generated
vendored
Normal file
1769
vendor/github.com/DataDog/zstd/zstd_decompress.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
1323
vendor/github.com/DataDog/zstd/zstd_decompress_block.c
generated
vendored
Normal file
1323
vendor/github.com/DataDog/zstd/zstd_decompress_block.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
59
vendor/github.com/DataDog/zstd/zstd_decompress_block.h
generated
vendored
Normal file
59
vendor/github.com/DataDog/zstd/zstd_decompress_block.h
generated
vendored
Normal file
@@ -0,0 +1,59 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
#ifndef ZSTD_DEC_BLOCK_H
|
||||||
|
#define ZSTD_DEC_BLOCK_H
|
||||||
|
|
||||||
|
/*-*******************************************************
|
||||||
|
* Dependencies
|
||||||
|
*********************************************************/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
#include "zstd.h" /* DCtx, and some public functions */
|
||||||
|
#include "zstd_internal.h" /* blockProperties_t, and some public functions */
|
||||||
|
#include "zstd_decompress_internal.h" /* ZSTD_seqSymbol */
|
||||||
|
|
||||||
|
|
||||||
|
/* === Prototypes === */
|
||||||
|
|
||||||
|
/* note: prototypes already published within `zstd.h` :
|
||||||
|
* ZSTD_decompressBlock()
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* note: prototypes already published within `zstd_internal.h` :
|
||||||
|
* ZSTD_getcBlockSize()
|
||||||
|
* ZSTD_decodeSeqHeaders()
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/* ZSTD_decompressBlock_internal() :
|
||||||
|
* decompress block, starting at `src`,
|
||||||
|
* into destination buffer `dst`.
|
||||||
|
* @return : decompressed block size,
|
||||||
|
* or an error code (which can be tested using ZSTD_isError())
|
||||||
|
*/
|
||||||
|
size_t ZSTD_decompressBlock_internal(ZSTD_DCtx* dctx,
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize, const int frame);
|
||||||
|
|
||||||
|
/* ZSTD_buildFSETable() :
|
||||||
|
* generate FSE decoding table for one symbol (ll, ml or off)
|
||||||
|
* this function must be called with valid parameters only
|
||||||
|
* (dt is large enough, normalizedCounter distribution total is a power of 2, max is within range, etc.)
|
||||||
|
* in which case it cannot fail.
|
||||||
|
* Internal use only.
|
||||||
|
*/
|
||||||
|
void ZSTD_buildFSETable(ZSTD_seqSymbol* dt,
|
||||||
|
const short* normalizedCounter, unsigned maxSymbolValue,
|
||||||
|
const U32* baseValue, const U32* nbAdditionalBits,
|
||||||
|
unsigned tableLog);
|
||||||
|
|
||||||
|
|
||||||
|
#endif /* ZSTD_DEC_BLOCK_H */
|
||||||
175
vendor/github.com/DataDog/zstd/zstd_decompress_internal.h
generated
vendored
Normal file
175
vendor/github.com/DataDog/zstd/zstd_decompress_internal.h
generated
vendored
Normal file
@@ -0,0 +1,175 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/* zstd_decompress_internal:
|
||||||
|
* objects and definitions shared within lib/decompress modules */
|
||||||
|
|
||||||
|
#ifndef ZSTD_DECOMPRESS_INTERNAL_H
|
||||||
|
#define ZSTD_DECOMPRESS_INTERNAL_H
|
||||||
|
|
||||||
|
|
||||||
|
/*-*******************************************************
|
||||||
|
* Dependencies
|
||||||
|
*********************************************************/
|
||||||
|
#include "mem.h" /* BYTE, U16, U32 */
|
||||||
|
#include "zstd_internal.h" /* ZSTD_seqSymbol */
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/*-*******************************************************
|
||||||
|
* Constants
|
||||||
|
*********************************************************/
|
||||||
|
static const U32 LL_base[MaxLL+1] = {
|
||||||
|
0, 1, 2, 3, 4, 5, 6, 7,
|
||||||
|
8, 9, 10, 11, 12, 13, 14, 15,
|
||||||
|
16, 18, 20, 22, 24, 28, 32, 40,
|
||||||
|
48, 64, 0x80, 0x100, 0x200, 0x400, 0x800, 0x1000,
|
||||||
|
0x2000, 0x4000, 0x8000, 0x10000 };
|
||||||
|
|
||||||
|
static const U32 OF_base[MaxOff+1] = {
|
||||||
|
0, 1, 1, 5, 0xD, 0x1D, 0x3D, 0x7D,
|
||||||
|
0xFD, 0x1FD, 0x3FD, 0x7FD, 0xFFD, 0x1FFD, 0x3FFD, 0x7FFD,
|
||||||
|
0xFFFD, 0x1FFFD, 0x3FFFD, 0x7FFFD, 0xFFFFD, 0x1FFFFD, 0x3FFFFD, 0x7FFFFD,
|
||||||
|
0xFFFFFD, 0x1FFFFFD, 0x3FFFFFD, 0x7FFFFFD, 0xFFFFFFD, 0x1FFFFFFD, 0x3FFFFFFD, 0x7FFFFFFD };
|
||||||
|
|
||||||
|
static const U32 OF_bits[MaxOff+1] = {
|
||||||
|
0, 1, 2, 3, 4, 5, 6, 7,
|
||||||
|
8, 9, 10, 11, 12, 13, 14, 15,
|
||||||
|
16, 17, 18, 19, 20, 21, 22, 23,
|
||||||
|
24, 25, 26, 27, 28, 29, 30, 31 };
|
||||||
|
|
||||||
|
static const U32 ML_base[MaxML+1] = {
|
||||||
|
3, 4, 5, 6, 7, 8, 9, 10,
|
||||||
|
11, 12, 13, 14, 15, 16, 17, 18,
|
||||||
|
19, 20, 21, 22, 23, 24, 25, 26,
|
||||||
|
27, 28, 29, 30, 31, 32, 33, 34,
|
||||||
|
35, 37, 39, 41, 43, 47, 51, 59,
|
||||||
|
67, 83, 99, 0x83, 0x103, 0x203, 0x403, 0x803,
|
||||||
|
0x1003, 0x2003, 0x4003, 0x8003, 0x10003 };
|
||||||
|
|
||||||
|
|
||||||
|
/*-*******************************************************
|
||||||
|
* Decompression types
|
||||||
|
*********************************************************/
|
||||||
|
typedef struct {
|
||||||
|
U32 fastMode;
|
||||||
|
U32 tableLog;
|
||||||
|
} ZSTD_seqSymbol_header;
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
U16 nextState;
|
||||||
|
BYTE nbAdditionalBits;
|
||||||
|
BYTE nbBits;
|
||||||
|
U32 baseValue;
|
||||||
|
} ZSTD_seqSymbol;
|
||||||
|
|
||||||
|
#define SEQSYMBOL_TABLE_SIZE(log) (1 + (1 << (log)))
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
ZSTD_seqSymbol LLTable[SEQSYMBOL_TABLE_SIZE(LLFSELog)]; /* Note : Space reserved for FSE Tables */
|
||||||
|
ZSTD_seqSymbol OFTable[SEQSYMBOL_TABLE_SIZE(OffFSELog)]; /* is also used as temporary workspace while building hufTable during DDict creation */
|
||||||
|
ZSTD_seqSymbol MLTable[SEQSYMBOL_TABLE_SIZE(MLFSELog)]; /* and therefore must be at least HUF_DECOMPRESS_WORKSPACE_SIZE large */
|
||||||
|
HUF_DTable hufTable[HUF_DTABLE_SIZE(HufLog)]; /* can accommodate HUF_decompress4X */
|
||||||
|
U32 rep[ZSTD_REP_NUM];
|
||||||
|
} ZSTD_entropyDTables_t;
|
||||||
|
|
||||||
|
typedef enum { ZSTDds_getFrameHeaderSize, ZSTDds_decodeFrameHeader,
|
||||||
|
ZSTDds_decodeBlockHeader, ZSTDds_decompressBlock,
|
||||||
|
ZSTDds_decompressLastBlock, ZSTDds_checkChecksum,
|
||||||
|
ZSTDds_decodeSkippableHeader, ZSTDds_skipFrame } ZSTD_dStage;
|
||||||
|
|
||||||
|
typedef enum { zdss_init=0, zdss_loadHeader,
|
||||||
|
zdss_read, zdss_load, zdss_flush } ZSTD_dStreamStage;
|
||||||
|
|
||||||
|
typedef enum {
|
||||||
|
ZSTD_use_indefinitely = -1, /* Use the dictionary indefinitely */
|
||||||
|
ZSTD_dont_use = 0, /* Do not use the dictionary (if one exists free it) */
|
||||||
|
ZSTD_use_once = 1 /* Use the dictionary once and set to ZSTD_dont_use */
|
||||||
|
} ZSTD_dictUses_e;
|
||||||
|
|
||||||
|
struct ZSTD_DCtx_s
|
||||||
|
{
|
||||||
|
const ZSTD_seqSymbol* LLTptr;
|
||||||
|
const ZSTD_seqSymbol* MLTptr;
|
||||||
|
const ZSTD_seqSymbol* OFTptr;
|
||||||
|
const HUF_DTable* HUFptr;
|
||||||
|
ZSTD_entropyDTables_t entropy;
|
||||||
|
U32 workspace[HUF_DECOMPRESS_WORKSPACE_SIZE_U32]; /* space needed when building huffman tables */
|
||||||
|
const void* previousDstEnd; /* detect continuity */
|
||||||
|
const void* prefixStart; /* start of current segment */
|
||||||
|
const void* virtualStart; /* virtual start of previous segment if it was just before current one */
|
||||||
|
const void* dictEnd; /* end of previous segment */
|
||||||
|
size_t expected;
|
||||||
|
ZSTD_frameHeader fParams;
|
||||||
|
U64 decodedSize;
|
||||||
|
blockType_e bType; /* used in ZSTD_decompressContinue(), store blockType between block header decoding and block decompression stages */
|
||||||
|
ZSTD_dStage stage;
|
||||||
|
U32 litEntropy;
|
||||||
|
U32 fseEntropy;
|
||||||
|
XXH64_state_t xxhState;
|
||||||
|
size_t headerSize;
|
||||||
|
ZSTD_format_e format;
|
||||||
|
const BYTE* litPtr;
|
||||||
|
ZSTD_customMem customMem;
|
||||||
|
size_t litSize;
|
||||||
|
size_t rleSize;
|
||||||
|
size_t staticSize;
|
||||||
|
int bmi2; /* == 1 if the CPU supports BMI2 and 0 otherwise. CPU support is determined dynamically once per context lifetime. */
|
||||||
|
|
||||||
|
/* dictionary */
|
||||||
|
ZSTD_DDict* ddictLocal;
|
||||||
|
const ZSTD_DDict* ddict; /* set by ZSTD_initDStream_usingDDict(), or ZSTD_DCtx_refDDict() */
|
||||||
|
U32 dictID;
|
||||||
|
int ddictIsCold; /* if == 1 : dictionary is "new" for working context, and presumed "cold" (not in cpu cache) */
|
||||||
|
ZSTD_dictUses_e dictUses;
|
||||||
|
|
||||||
|
/* streaming */
|
||||||
|
ZSTD_dStreamStage streamStage;
|
||||||
|
char* inBuff;
|
||||||
|
size_t inBuffSize;
|
||||||
|
size_t inPos;
|
||||||
|
size_t maxWindowSize;
|
||||||
|
char* outBuff;
|
||||||
|
size_t outBuffSize;
|
||||||
|
size_t outStart;
|
||||||
|
size_t outEnd;
|
||||||
|
size_t lhSize;
|
||||||
|
void* legacyContext;
|
||||||
|
U32 previousLegacyVersion;
|
||||||
|
U32 legacyVersion;
|
||||||
|
U32 hostageByte;
|
||||||
|
int noForwardProgress;
|
||||||
|
|
||||||
|
/* workspace */
|
||||||
|
BYTE litBuffer[ZSTD_BLOCKSIZE_MAX + WILDCOPY_OVERLENGTH];
|
||||||
|
BYTE headerBuffer[ZSTD_FRAMEHEADERSIZE_MAX];
|
||||||
|
}; /* typedef'd to ZSTD_DCtx within "zstd.h" */
|
||||||
|
|
||||||
|
|
||||||
|
/*-*******************************************************
|
||||||
|
* Shared internal functions
|
||||||
|
*********************************************************/
|
||||||
|
|
||||||
|
/*! ZSTD_loadDEntropy() :
|
||||||
|
* dict : must point at beginning of a valid zstd dictionary.
|
||||||
|
* @return : size of entropy tables read */
|
||||||
|
size_t ZSTD_loadDEntropy(ZSTD_entropyDTables_t* entropy,
|
||||||
|
const void* const dict, size_t const dictSize);
|
||||||
|
|
||||||
|
/*! ZSTD_checkContinuity() :
|
||||||
|
* check if next `dst` follows previous position, where decompression ended.
|
||||||
|
* If yes, do nothing (continue on current segment).
|
||||||
|
* If not, classify previous segment as "external dictionary", and start a new segment.
|
||||||
|
* This function cannot fail. */
|
||||||
|
void ZSTD_checkContinuity(ZSTD_DCtx* dctx, const void* dst);
|
||||||
|
|
||||||
|
|
||||||
|
#endif /* ZSTD_DECOMPRESS_INTERNAL_H */
|
||||||
518
vendor/github.com/DataDog/zstd/zstd_double_fast.c
generated
vendored
Normal file
518
vendor/github.com/DataDog/zstd/zstd_double_fast.c
generated
vendored
Normal file
@@ -0,0 +1,518 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include "zstd_compress_internal.h"
|
||||||
|
#include "zstd_double_fast.h"
|
||||||
|
|
||||||
|
|
||||||
|
void ZSTD_fillDoubleHashTable(ZSTD_matchState_t* ms,
|
||||||
|
void const* end, ZSTD_dictTableLoadMethod_e dtlm)
|
||||||
|
{
|
||||||
|
const ZSTD_compressionParameters* const cParams = &ms->cParams;
|
||||||
|
U32* const hashLarge = ms->hashTable;
|
||||||
|
U32 const hBitsL = cParams->hashLog;
|
||||||
|
U32 const mls = cParams->minMatch;
|
||||||
|
U32* const hashSmall = ms->chainTable;
|
||||||
|
U32 const hBitsS = cParams->chainLog;
|
||||||
|
const BYTE* const base = ms->window.base;
|
||||||
|
const BYTE* ip = base + ms->nextToUpdate;
|
||||||
|
const BYTE* const iend = ((const BYTE*)end) - HASH_READ_SIZE;
|
||||||
|
const U32 fastHashFillStep = 3;
|
||||||
|
|
||||||
|
/* Always insert every fastHashFillStep position into the hash tables.
|
||||||
|
* Insert the other positions into the large hash table if their entry
|
||||||
|
* is empty.
|
||||||
|
*/
|
||||||
|
for (; ip + fastHashFillStep - 1 <= iend; ip += fastHashFillStep) {
|
||||||
|
U32 const current = (U32)(ip - base);
|
||||||
|
U32 i;
|
||||||
|
for (i = 0; i < fastHashFillStep; ++i) {
|
||||||
|
size_t const smHash = ZSTD_hashPtr(ip + i, hBitsS, mls);
|
||||||
|
size_t const lgHash = ZSTD_hashPtr(ip + i, hBitsL, 8);
|
||||||
|
if (i == 0)
|
||||||
|
hashSmall[smHash] = current + i;
|
||||||
|
if (i == 0 || hashLarge[lgHash] == 0)
|
||||||
|
hashLarge[lgHash] = current + i;
|
||||||
|
/* Only load extra positions for ZSTD_dtlm_full */
|
||||||
|
if (dtlm == ZSTD_dtlm_fast)
|
||||||
|
break;
|
||||||
|
} }
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE
|
||||||
|
size_t ZSTD_compressBlock_doubleFast_generic(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize,
|
||||||
|
U32 const mls /* template */, ZSTD_dictMode_e const dictMode)
|
||||||
|
{
|
||||||
|
ZSTD_compressionParameters const* cParams = &ms->cParams;
|
||||||
|
U32* const hashLong = ms->hashTable;
|
||||||
|
const U32 hBitsL = cParams->hashLog;
|
||||||
|
U32* const hashSmall = ms->chainTable;
|
||||||
|
const U32 hBitsS = cParams->chainLog;
|
||||||
|
const BYTE* const base = ms->window.base;
|
||||||
|
const BYTE* const istart = (const BYTE*)src;
|
||||||
|
const BYTE* ip = istart;
|
||||||
|
const BYTE* anchor = istart;
|
||||||
|
const U32 endIndex = (U32)((size_t)(istart - base) + srcSize);
|
||||||
|
const U32 lowestValid = ms->window.dictLimit;
|
||||||
|
const U32 maxDistance = 1U << cParams->windowLog;
|
||||||
|
/* presumes that, if there is a dictionary, it must be using Attach mode */
|
||||||
|
const U32 prefixLowestIndex = (endIndex - lowestValid > maxDistance) ? endIndex - maxDistance : lowestValid;
|
||||||
|
const BYTE* const prefixLowest = base + prefixLowestIndex;
|
||||||
|
const BYTE* const iend = istart + srcSize;
|
||||||
|
const BYTE* const ilimit = iend - HASH_READ_SIZE;
|
||||||
|
U32 offset_1=rep[0], offset_2=rep[1];
|
||||||
|
U32 offsetSaved = 0;
|
||||||
|
|
||||||
|
const ZSTD_matchState_t* const dms = ms->dictMatchState;
|
||||||
|
const ZSTD_compressionParameters* const dictCParams =
|
||||||
|
dictMode == ZSTD_dictMatchState ?
|
||||||
|
&dms->cParams : NULL;
|
||||||
|
const U32* const dictHashLong = dictMode == ZSTD_dictMatchState ?
|
||||||
|
dms->hashTable : NULL;
|
||||||
|
const U32* const dictHashSmall = dictMode == ZSTD_dictMatchState ?
|
||||||
|
dms->chainTable : NULL;
|
||||||
|
const U32 dictStartIndex = dictMode == ZSTD_dictMatchState ?
|
||||||
|
dms->window.dictLimit : 0;
|
||||||
|
const BYTE* const dictBase = dictMode == ZSTD_dictMatchState ?
|
||||||
|
dms->window.base : NULL;
|
||||||
|
const BYTE* const dictStart = dictMode == ZSTD_dictMatchState ?
|
||||||
|
dictBase + dictStartIndex : NULL;
|
||||||
|
const BYTE* const dictEnd = dictMode == ZSTD_dictMatchState ?
|
||||||
|
dms->window.nextSrc : NULL;
|
||||||
|
const U32 dictIndexDelta = dictMode == ZSTD_dictMatchState ?
|
||||||
|
prefixLowestIndex - (U32)(dictEnd - dictBase) :
|
||||||
|
0;
|
||||||
|
const U32 dictHBitsL = dictMode == ZSTD_dictMatchState ?
|
||||||
|
dictCParams->hashLog : hBitsL;
|
||||||
|
const U32 dictHBitsS = dictMode == ZSTD_dictMatchState ?
|
||||||
|
dictCParams->chainLog : hBitsS;
|
||||||
|
const U32 dictAndPrefixLength = (U32)(ip - prefixLowest + dictEnd - dictStart);
|
||||||
|
|
||||||
|
DEBUGLOG(5, "ZSTD_compressBlock_doubleFast_generic");
|
||||||
|
|
||||||
|
assert(dictMode == ZSTD_noDict || dictMode == ZSTD_dictMatchState);
|
||||||
|
|
||||||
|
/* if a dictionary is attached, it must be within window range */
|
||||||
|
if (dictMode == ZSTD_dictMatchState) {
|
||||||
|
assert(lowestValid + maxDistance >= endIndex);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* init */
|
||||||
|
ip += (dictAndPrefixLength == 0);
|
||||||
|
if (dictMode == ZSTD_noDict) {
|
||||||
|
U32 const maxRep = (U32)(ip - prefixLowest);
|
||||||
|
if (offset_2 > maxRep) offsetSaved = offset_2, offset_2 = 0;
|
||||||
|
if (offset_1 > maxRep) offsetSaved = offset_1, offset_1 = 0;
|
||||||
|
}
|
||||||
|
if (dictMode == ZSTD_dictMatchState) {
|
||||||
|
/* dictMatchState repCode checks don't currently handle repCode == 0
|
||||||
|
* disabling. */
|
||||||
|
assert(offset_1 <= dictAndPrefixLength);
|
||||||
|
assert(offset_2 <= dictAndPrefixLength);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Main Search Loop */
|
||||||
|
while (ip < ilimit) { /* < instead of <=, because repcode check at (ip+1) */
|
||||||
|
size_t mLength;
|
||||||
|
U32 offset;
|
||||||
|
size_t const h2 = ZSTD_hashPtr(ip, hBitsL, 8);
|
||||||
|
size_t const h = ZSTD_hashPtr(ip, hBitsS, mls);
|
||||||
|
size_t const dictHL = ZSTD_hashPtr(ip, dictHBitsL, 8);
|
||||||
|
size_t const dictHS = ZSTD_hashPtr(ip, dictHBitsS, mls);
|
||||||
|
U32 const current = (U32)(ip-base);
|
||||||
|
U32 const matchIndexL = hashLong[h2];
|
||||||
|
U32 matchIndexS = hashSmall[h];
|
||||||
|
const BYTE* matchLong = base + matchIndexL;
|
||||||
|
const BYTE* match = base + matchIndexS;
|
||||||
|
const U32 repIndex = current + 1 - offset_1;
|
||||||
|
const BYTE* repMatch = (dictMode == ZSTD_dictMatchState
|
||||||
|
&& repIndex < prefixLowestIndex) ?
|
||||||
|
dictBase + (repIndex - dictIndexDelta) :
|
||||||
|
base + repIndex;
|
||||||
|
hashLong[h2] = hashSmall[h] = current; /* update hash tables */
|
||||||
|
|
||||||
|
/* check dictMatchState repcode */
|
||||||
|
if (dictMode == ZSTD_dictMatchState
|
||||||
|
&& ((U32)((prefixLowestIndex-1) - repIndex) >= 3 /* intentional underflow */)
|
||||||
|
&& (MEM_read32(repMatch) == MEM_read32(ip+1)) ) {
|
||||||
|
const BYTE* repMatchEnd = repIndex < prefixLowestIndex ? dictEnd : iend;
|
||||||
|
mLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixLowest) + 4;
|
||||||
|
ip++;
|
||||||
|
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH);
|
||||||
|
goto _match_stored;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* check noDict repcode */
|
||||||
|
if ( dictMode == ZSTD_noDict
|
||||||
|
&& ((offset_1 > 0) & (MEM_read32(ip+1-offset_1) == MEM_read32(ip+1)))) {
|
||||||
|
mLength = ZSTD_count(ip+1+4, ip+1+4-offset_1, iend) + 4;
|
||||||
|
ip++;
|
||||||
|
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH);
|
||||||
|
goto _match_stored;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (matchIndexL > prefixLowestIndex) {
|
||||||
|
/* check prefix long match */
|
||||||
|
if (MEM_read64(matchLong) == MEM_read64(ip)) {
|
||||||
|
mLength = ZSTD_count(ip+8, matchLong+8, iend) + 8;
|
||||||
|
offset = (U32)(ip-matchLong);
|
||||||
|
while (((ip>anchor) & (matchLong>prefixLowest)) && (ip[-1] == matchLong[-1])) { ip--; matchLong--; mLength++; } /* catch up */
|
||||||
|
goto _match_found;
|
||||||
|
}
|
||||||
|
} else if (dictMode == ZSTD_dictMatchState) {
|
||||||
|
/* check dictMatchState long match */
|
||||||
|
U32 const dictMatchIndexL = dictHashLong[dictHL];
|
||||||
|
const BYTE* dictMatchL = dictBase + dictMatchIndexL;
|
||||||
|
assert(dictMatchL < dictEnd);
|
||||||
|
|
||||||
|
if (dictMatchL > dictStart && MEM_read64(dictMatchL) == MEM_read64(ip)) {
|
||||||
|
mLength = ZSTD_count_2segments(ip+8, dictMatchL+8, iend, dictEnd, prefixLowest) + 8;
|
||||||
|
offset = (U32)(current - dictMatchIndexL - dictIndexDelta);
|
||||||
|
while (((ip>anchor) & (dictMatchL>dictStart)) && (ip[-1] == dictMatchL[-1])) { ip--; dictMatchL--; mLength++; } /* catch up */
|
||||||
|
goto _match_found;
|
||||||
|
} }
|
||||||
|
|
||||||
|
if (matchIndexS > prefixLowestIndex) {
|
||||||
|
/* check prefix short match */
|
||||||
|
if (MEM_read32(match) == MEM_read32(ip)) {
|
||||||
|
goto _search_next_long;
|
||||||
|
}
|
||||||
|
} else if (dictMode == ZSTD_dictMatchState) {
|
||||||
|
/* check dictMatchState short match */
|
||||||
|
U32 const dictMatchIndexS = dictHashSmall[dictHS];
|
||||||
|
match = dictBase + dictMatchIndexS;
|
||||||
|
matchIndexS = dictMatchIndexS + dictIndexDelta;
|
||||||
|
|
||||||
|
if (match > dictStart && MEM_read32(match) == MEM_read32(ip)) {
|
||||||
|
goto _search_next_long;
|
||||||
|
} }
|
||||||
|
|
||||||
|
ip += ((ip-anchor) >> kSearchStrength) + 1;
|
||||||
|
continue;
|
||||||
|
|
||||||
|
_search_next_long:
|
||||||
|
|
||||||
|
{ size_t const hl3 = ZSTD_hashPtr(ip+1, hBitsL, 8);
|
||||||
|
size_t const dictHLNext = ZSTD_hashPtr(ip+1, dictHBitsL, 8);
|
||||||
|
U32 const matchIndexL3 = hashLong[hl3];
|
||||||
|
const BYTE* matchL3 = base + matchIndexL3;
|
||||||
|
hashLong[hl3] = current + 1;
|
||||||
|
|
||||||
|
/* check prefix long +1 match */
|
||||||
|
if (matchIndexL3 > prefixLowestIndex) {
|
||||||
|
if (MEM_read64(matchL3) == MEM_read64(ip+1)) {
|
||||||
|
mLength = ZSTD_count(ip+9, matchL3+8, iend) + 8;
|
||||||
|
ip++;
|
||||||
|
offset = (U32)(ip-matchL3);
|
||||||
|
while (((ip>anchor) & (matchL3>prefixLowest)) && (ip[-1] == matchL3[-1])) { ip--; matchL3--; mLength++; } /* catch up */
|
||||||
|
goto _match_found;
|
||||||
|
}
|
||||||
|
} else if (dictMode == ZSTD_dictMatchState) {
|
||||||
|
/* check dict long +1 match */
|
||||||
|
U32 const dictMatchIndexL3 = dictHashLong[dictHLNext];
|
||||||
|
const BYTE* dictMatchL3 = dictBase + dictMatchIndexL3;
|
||||||
|
assert(dictMatchL3 < dictEnd);
|
||||||
|
if (dictMatchL3 > dictStart && MEM_read64(dictMatchL3) == MEM_read64(ip+1)) {
|
||||||
|
mLength = ZSTD_count_2segments(ip+1+8, dictMatchL3+8, iend, dictEnd, prefixLowest) + 8;
|
||||||
|
ip++;
|
||||||
|
offset = (U32)(current + 1 - dictMatchIndexL3 - dictIndexDelta);
|
||||||
|
while (((ip>anchor) & (dictMatchL3>dictStart)) && (ip[-1] == dictMatchL3[-1])) { ip--; dictMatchL3--; mLength++; } /* catch up */
|
||||||
|
goto _match_found;
|
||||||
|
} } }
|
||||||
|
|
||||||
|
/* if no long +1 match, explore the short match we found */
|
||||||
|
if (dictMode == ZSTD_dictMatchState && matchIndexS < prefixLowestIndex) {
|
||||||
|
mLength = ZSTD_count_2segments(ip+4, match+4, iend, dictEnd, prefixLowest) + 4;
|
||||||
|
offset = (U32)(current - matchIndexS);
|
||||||
|
while (((ip>anchor) & (match>dictStart)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */
|
||||||
|
} else {
|
||||||
|
mLength = ZSTD_count(ip+4, match+4, iend) + 4;
|
||||||
|
offset = (U32)(ip - match);
|
||||||
|
while (((ip>anchor) & (match>prefixLowest)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */
|
||||||
|
}
|
||||||
|
|
||||||
|
/* fall-through */
|
||||||
|
|
||||||
|
_match_found:
|
||||||
|
offset_2 = offset_1;
|
||||||
|
offset_1 = offset;
|
||||||
|
|
||||||
|
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
||||||
|
|
||||||
|
_match_stored:
|
||||||
|
/* match found */
|
||||||
|
ip += mLength;
|
||||||
|
anchor = ip;
|
||||||
|
|
||||||
|
if (ip <= ilimit) {
|
||||||
|
/* Complementary insertion */
|
||||||
|
/* done after iLimit test, as candidates could be > iend-8 */
|
||||||
|
{ U32 const indexToInsert = current+2;
|
||||||
|
hashLong[ZSTD_hashPtr(base+indexToInsert, hBitsL, 8)] = indexToInsert;
|
||||||
|
hashLong[ZSTD_hashPtr(ip-2, hBitsL, 8)] = (U32)(ip-2-base);
|
||||||
|
hashSmall[ZSTD_hashPtr(base+indexToInsert, hBitsS, mls)] = indexToInsert;
|
||||||
|
hashSmall[ZSTD_hashPtr(ip-1, hBitsS, mls)] = (U32)(ip-1-base);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* check immediate repcode */
|
||||||
|
if (dictMode == ZSTD_dictMatchState) {
|
||||||
|
while (ip <= ilimit) {
|
||||||
|
U32 const current2 = (U32)(ip-base);
|
||||||
|
U32 const repIndex2 = current2 - offset_2;
|
||||||
|
const BYTE* repMatch2 = dictMode == ZSTD_dictMatchState
|
||||||
|
&& repIndex2 < prefixLowestIndex ?
|
||||||
|
dictBase - dictIndexDelta + repIndex2 :
|
||||||
|
base + repIndex2;
|
||||||
|
if ( ((U32)((prefixLowestIndex-1) - (U32)repIndex2) >= 3 /* intentional overflow */)
|
||||||
|
&& (MEM_read32(repMatch2) == MEM_read32(ip)) ) {
|
||||||
|
const BYTE* const repEnd2 = repIndex2 < prefixLowestIndex ? dictEnd : iend;
|
||||||
|
size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixLowest) + 4;
|
||||||
|
U32 tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; /* swap offset_2 <=> offset_1 */
|
||||||
|
ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, repLength2-MINMATCH);
|
||||||
|
hashSmall[ZSTD_hashPtr(ip, hBitsS, mls)] = current2;
|
||||||
|
hashLong[ZSTD_hashPtr(ip, hBitsL, 8)] = current2;
|
||||||
|
ip += repLength2;
|
||||||
|
anchor = ip;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
} }
|
||||||
|
|
||||||
|
if (dictMode == ZSTD_noDict) {
|
||||||
|
while ( (ip <= ilimit)
|
||||||
|
&& ( (offset_2>0)
|
||||||
|
& (MEM_read32(ip) == MEM_read32(ip - offset_2)) )) {
|
||||||
|
/* store sequence */
|
||||||
|
size_t const rLength = ZSTD_count(ip+4, ip+4-offset_2, iend) + 4;
|
||||||
|
U32 const tmpOff = offset_2; offset_2 = offset_1; offset_1 = tmpOff; /* swap offset_2 <=> offset_1 */
|
||||||
|
hashSmall[ZSTD_hashPtr(ip, hBitsS, mls)] = (U32)(ip-base);
|
||||||
|
hashLong[ZSTD_hashPtr(ip, hBitsL, 8)] = (U32)(ip-base);
|
||||||
|
ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, rLength-MINMATCH);
|
||||||
|
ip += rLength;
|
||||||
|
anchor = ip;
|
||||||
|
continue; /* faster when present ... (?) */
|
||||||
|
} } }
|
||||||
|
} /* while (ip < ilimit) */
|
||||||
|
|
||||||
|
/* save reps for next block */
|
||||||
|
rep[0] = offset_1 ? offset_1 : offsetSaved;
|
||||||
|
rep[1] = offset_2 ? offset_2 : offsetSaved;
|
||||||
|
|
||||||
|
/* Return the last literals size */
|
||||||
|
return (size_t)(iend - anchor);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t ZSTD_compressBlock_doubleFast(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
const U32 mls = ms->cParams.minMatch;
|
||||||
|
switch(mls)
|
||||||
|
{
|
||||||
|
default: /* includes case 3 */
|
||||||
|
case 4 :
|
||||||
|
return ZSTD_compressBlock_doubleFast_generic(ms, seqStore, rep, src, srcSize, 4, ZSTD_noDict);
|
||||||
|
case 5 :
|
||||||
|
return ZSTD_compressBlock_doubleFast_generic(ms, seqStore, rep, src, srcSize, 5, ZSTD_noDict);
|
||||||
|
case 6 :
|
||||||
|
return ZSTD_compressBlock_doubleFast_generic(ms, seqStore, rep, src, srcSize, 6, ZSTD_noDict);
|
||||||
|
case 7 :
|
||||||
|
return ZSTD_compressBlock_doubleFast_generic(ms, seqStore, rep, src, srcSize, 7, ZSTD_noDict);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t ZSTD_compressBlock_doubleFast_dictMatchState(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
const U32 mls = ms->cParams.minMatch;
|
||||||
|
switch(mls)
|
||||||
|
{
|
||||||
|
default: /* includes case 3 */
|
||||||
|
case 4 :
|
||||||
|
return ZSTD_compressBlock_doubleFast_generic(ms, seqStore, rep, src, srcSize, 4, ZSTD_dictMatchState);
|
||||||
|
case 5 :
|
||||||
|
return ZSTD_compressBlock_doubleFast_generic(ms, seqStore, rep, src, srcSize, 5, ZSTD_dictMatchState);
|
||||||
|
case 6 :
|
||||||
|
return ZSTD_compressBlock_doubleFast_generic(ms, seqStore, rep, src, srcSize, 6, ZSTD_dictMatchState);
|
||||||
|
case 7 :
|
||||||
|
return ZSTD_compressBlock_doubleFast_generic(ms, seqStore, rep, src, srcSize, 7, ZSTD_dictMatchState);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static size_t ZSTD_compressBlock_doubleFast_extDict_generic(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize,
|
||||||
|
U32 const mls /* template */)
|
||||||
|
{
|
||||||
|
ZSTD_compressionParameters const* cParams = &ms->cParams;
|
||||||
|
U32* const hashLong = ms->hashTable;
|
||||||
|
U32 const hBitsL = cParams->hashLog;
|
||||||
|
U32* const hashSmall = ms->chainTable;
|
||||||
|
U32 const hBitsS = cParams->chainLog;
|
||||||
|
const BYTE* const istart = (const BYTE*)src;
|
||||||
|
const BYTE* ip = istart;
|
||||||
|
const BYTE* anchor = istart;
|
||||||
|
const BYTE* const iend = istart + srcSize;
|
||||||
|
const BYTE* const ilimit = iend - 8;
|
||||||
|
const BYTE* const base = ms->window.base;
|
||||||
|
const U32 endIndex = (U32)((size_t)(istart - base) + srcSize);
|
||||||
|
const U32 lowLimit = ZSTD_getLowestMatchIndex(ms, endIndex, cParams->windowLog);
|
||||||
|
const U32 dictStartIndex = lowLimit;
|
||||||
|
const U32 dictLimit = ms->window.dictLimit;
|
||||||
|
const U32 prefixStartIndex = (dictLimit > lowLimit) ? dictLimit : lowLimit;
|
||||||
|
const BYTE* const prefixStart = base + prefixStartIndex;
|
||||||
|
const BYTE* const dictBase = ms->window.dictBase;
|
||||||
|
const BYTE* const dictStart = dictBase + dictStartIndex;
|
||||||
|
const BYTE* const dictEnd = dictBase + prefixStartIndex;
|
||||||
|
U32 offset_1=rep[0], offset_2=rep[1];
|
||||||
|
|
||||||
|
DEBUGLOG(5, "ZSTD_compressBlock_doubleFast_extDict_generic (srcSize=%zu)", srcSize);
|
||||||
|
|
||||||
|
/* if extDict is invalidated due to maxDistance, switch to "regular" variant */
|
||||||
|
if (prefixStartIndex == dictStartIndex)
|
||||||
|
return ZSTD_compressBlock_doubleFast_generic(ms, seqStore, rep, src, srcSize, mls, ZSTD_noDict);
|
||||||
|
|
||||||
|
/* Search Loop */
|
||||||
|
while (ip < ilimit) { /* < instead of <=, because (ip+1) */
|
||||||
|
const size_t hSmall = ZSTD_hashPtr(ip, hBitsS, mls);
|
||||||
|
const U32 matchIndex = hashSmall[hSmall];
|
||||||
|
const BYTE* const matchBase = matchIndex < prefixStartIndex ? dictBase : base;
|
||||||
|
const BYTE* match = matchBase + matchIndex;
|
||||||
|
|
||||||
|
const size_t hLong = ZSTD_hashPtr(ip, hBitsL, 8);
|
||||||
|
const U32 matchLongIndex = hashLong[hLong];
|
||||||
|
const BYTE* const matchLongBase = matchLongIndex < prefixStartIndex ? dictBase : base;
|
||||||
|
const BYTE* matchLong = matchLongBase + matchLongIndex;
|
||||||
|
|
||||||
|
const U32 current = (U32)(ip-base);
|
||||||
|
const U32 repIndex = current + 1 - offset_1; /* offset_1 expected <= current +1 */
|
||||||
|
const BYTE* const repBase = repIndex < prefixStartIndex ? dictBase : base;
|
||||||
|
const BYTE* const repMatch = repBase + repIndex;
|
||||||
|
size_t mLength;
|
||||||
|
hashSmall[hSmall] = hashLong[hLong] = current; /* update hash table */
|
||||||
|
|
||||||
|
if ((((U32)((prefixStartIndex-1) - repIndex) >= 3) /* intentional underflow : ensure repIndex doesn't overlap dict + prefix */
|
||||||
|
& (repIndex > dictStartIndex))
|
||||||
|
&& (MEM_read32(repMatch) == MEM_read32(ip+1)) ) {
|
||||||
|
const BYTE* repMatchEnd = repIndex < prefixStartIndex ? dictEnd : iend;
|
||||||
|
mLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixStart) + 4;
|
||||||
|
ip++;
|
||||||
|
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH);
|
||||||
|
} else {
|
||||||
|
if ((matchLongIndex > dictStartIndex) && (MEM_read64(matchLong) == MEM_read64(ip))) {
|
||||||
|
const BYTE* const matchEnd = matchLongIndex < prefixStartIndex ? dictEnd : iend;
|
||||||
|
const BYTE* const lowMatchPtr = matchLongIndex < prefixStartIndex ? dictStart : prefixStart;
|
||||||
|
U32 offset;
|
||||||
|
mLength = ZSTD_count_2segments(ip+8, matchLong+8, iend, matchEnd, prefixStart) + 8;
|
||||||
|
offset = current - matchLongIndex;
|
||||||
|
while (((ip>anchor) & (matchLong>lowMatchPtr)) && (ip[-1] == matchLong[-1])) { ip--; matchLong--; mLength++; } /* catch up */
|
||||||
|
offset_2 = offset_1;
|
||||||
|
offset_1 = offset;
|
||||||
|
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
||||||
|
|
||||||
|
} else if ((matchIndex > dictStartIndex) && (MEM_read32(match) == MEM_read32(ip))) {
|
||||||
|
size_t const h3 = ZSTD_hashPtr(ip+1, hBitsL, 8);
|
||||||
|
U32 const matchIndex3 = hashLong[h3];
|
||||||
|
const BYTE* const match3Base = matchIndex3 < prefixStartIndex ? dictBase : base;
|
||||||
|
const BYTE* match3 = match3Base + matchIndex3;
|
||||||
|
U32 offset;
|
||||||
|
hashLong[h3] = current + 1;
|
||||||
|
if ( (matchIndex3 > dictStartIndex) && (MEM_read64(match3) == MEM_read64(ip+1)) ) {
|
||||||
|
const BYTE* const matchEnd = matchIndex3 < prefixStartIndex ? dictEnd : iend;
|
||||||
|
const BYTE* const lowMatchPtr = matchIndex3 < prefixStartIndex ? dictStart : prefixStart;
|
||||||
|
mLength = ZSTD_count_2segments(ip+9, match3+8, iend, matchEnd, prefixStart) + 8;
|
||||||
|
ip++;
|
||||||
|
offset = current+1 - matchIndex3;
|
||||||
|
while (((ip>anchor) & (match3>lowMatchPtr)) && (ip[-1] == match3[-1])) { ip--; match3--; mLength++; } /* catch up */
|
||||||
|
} else {
|
||||||
|
const BYTE* const matchEnd = matchIndex < prefixStartIndex ? dictEnd : iend;
|
||||||
|
const BYTE* const lowMatchPtr = matchIndex < prefixStartIndex ? dictStart : prefixStart;
|
||||||
|
mLength = ZSTD_count_2segments(ip+4, match+4, iend, matchEnd, prefixStart) + 4;
|
||||||
|
offset = current - matchIndex;
|
||||||
|
while (((ip>anchor) & (match>lowMatchPtr)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */
|
||||||
|
}
|
||||||
|
offset_2 = offset_1;
|
||||||
|
offset_1 = offset;
|
||||||
|
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
||||||
|
|
||||||
|
} else {
|
||||||
|
ip += ((ip-anchor) >> kSearchStrength) + 1;
|
||||||
|
continue;
|
||||||
|
} }
|
||||||
|
|
||||||
|
/* move to next sequence start */
|
||||||
|
ip += mLength;
|
||||||
|
anchor = ip;
|
||||||
|
|
||||||
|
if (ip <= ilimit) {
|
||||||
|
/* Complementary insertion */
|
||||||
|
/* done after iLimit test, as candidates could be > iend-8 */
|
||||||
|
{ U32 const indexToInsert = current+2;
|
||||||
|
hashLong[ZSTD_hashPtr(base+indexToInsert, hBitsL, 8)] = indexToInsert;
|
||||||
|
hashLong[ZSTD_hashPtr(ip-2, hBitsL, 8)] = (U32)(ip-2-base);
|
||||||
|
hashSmall[ZSTD_hashPtr(base+indexToInsert, hBitsS, mls)] = indexToInsert;
|
||||||
|
hashSmall[ZSTD_hashPtr(ip-1, hBitsS, mls)] = (U32)(ip-1-base);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* check immediate repcode */
|
||||||
|
while (ip <= ilimit) {
|
||||||
|
U32 const current2 = (U32)(ip-base);
|
||||||
|
U32 const repIndex2 = current2 - offset_2;
|
||||||
|
const BYTE* repMatch2 = repIndex2 < prefixStartIndex ? dictBase + repIndex2 : base + repIndex2;
|
||||||
|
if ( (((U32)((prefixStartIndex-1) - repIndex2) >= 3) /* intentional overflow : ensure repIndex2 doesn't overlap dict + prefix */
|
||||||
|
& (repIndex2 > dictStartIndex))
|
||||||
|
&& (MEM_read32(repMatch2) == MEM_read32(ip)) ) {
|
||||||
|
const BYTE* const repEnd2 = repIndex2 < prefixStartIndex ? dictEnd : iend;
|
||||||
|
size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixStart) + 4;
|
||||||
|
U32 const tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; /* swap offset_2 <=> offset_1 */
|
||||||
|
ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, repLength2-MINMATCH);
|
||||||
|
hashSmall[ZSTD_hashPtr(ip, hBitsS, mls)] = current2;
|
||||||
|
hashLong[ZSTD_hashPtr(ip, hBitsL, 8)] = current2;
|
||||||
|
ip += repLength2;
|
||||||
|
anchor = ip;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
} } }
|
||||||
|
|
||||||
|
/* save reps for next block */
|
||||||
|
rep[0] = offset_1;
|
||||||
|
rep[1] = offset_2;
|
||||||
|
|
||||||
|
/* Return the last literals size */
|
||||||
|
return (size_t)(iend - anchor);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t ZSTD_compressBlock_doubleFast_extDict(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
U32 const mls = ms->cParams.minMatch;
|
||||||
|
switch(mls)
|
||||||
|
{
|
||||||
|
default: /* includes case 3 */
|
||||||
|
case 4 :
|
||||||
|
return ZSTD_compressBlock_doubleFast_extDict_generic(ms, seqStore, rep, src, srcSize, 4);
|
||||||
|
case 5 :
|
||||||
|
return ZSTD_compressBlock_doubleFast_extDict_generic(ms, seqStore, rep, src, srcSize, 5);
|
||||||
|
case 6 :
|
||||||
|
return ZSTD_compressBlock_doubleFast_extDict_generic(ms, seqStore, rep, src, srcSize, 6);
|
||||||
|
case 7 :
|
||||||
|
return ZSTD_compressBlock_doubleFast_extDict_generic(ms, seqStore, rep, src, srcSize, 7);
|
||||||
|
}
|
||||||
|
}
|
||||||
38
vendor/github.com/DataDog/zstd/zstd_double_fast.h
generated
vendored
Normal file
38
vendor/github.com/DataDog/zstd/zstd_double_fast.h
generated
vendored
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_DOUBLE_FAST_H
|
||||||
|
#define ZSTD_DOUBLE_FAST_H
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#include "mem.h" /* U32 */
|
||||||
|
#include "zstd_compress_internal.h" /* ZSTD_CCtx, size_t */
|
||||||
|
|
||||||
|
void ZSTD_fillDoubleHashTable(ZSTD_matchState_t* ms,
|
||||||
|
void const* end, ZSTD_dictTableLoadMethod_e dtlm);
|
||||||
|
size_t ZSTD_compressBlock_doubleFast(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_doubleFast_dictMatchState(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_doubleFast_extDict(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_DOUBLE_FAST_H */
|
||||||
93
vendor/github.com/DataDog/zstd/zstd_errors.h
generated
vendored
Normal file
93
vendor/github.com/DataDog/zstd/zstd_errors.h
generated
vendored
Normal file
@@ -0,0 +1,93 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_ERRORS_H_398273423
|
||||||
|
#define ZSTD_ERRORS_H_398273423
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*===== dependency =====*/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
|
||||||
|
|
||||||
|
/* ===== ZSTDERRORLIB_API : control library symbols visibility ===== */
|
||||||
|
#ifndef ZSTDERRORLIB_VISIBILITY
|
||||||
|
# if defined(__GNUC__) && (__GNUC__ >= 4)
|
||||||
|
# define ZSTDERRORLIB_VISIBILITY __attribute__ ((visibility ("default")))
|
||||||
|
# else
|
||||||
|
# define ZSTDERRORLIB_VISIBILITY
|
||||||
|
# endif
|
||||||
|
#endif
|
||||||
|
#if defined(ZSTD_DLL_EXPORT) && (ZSTD_DLL_EXPORT==1)
|
||||||
|
# define ZSTDERRORLIB_API __declspec(dllexport) ZSTDERRORLIB_VISIBILITY
|
||||||
|
#elif defined(ZSTD_DLL_IMPORT) && (ZSTD_DLL_IMPORT==1)
|
||||||
|
# define ZSTDERRORLIB_API __declspec(dllimport) ZSTDERRORLIB_VISIBILITY /* It isn't required but allows to generate better code, saving a function pointer load from the IAT and an indirect jump.*/
|
||||||
|
#else
|
||||||
|
# define ZSTDERRORLIB_API ZSTDERRORLIB_VISIBILITY
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*-*********************************************
|
||||||
|
* Error codes list
|
||||||
|
*-*********************************************
|
||||||
|
* Error codes _values_ are pinned down since v1.3.1 only.
|
||||||
|
* Therefore, don't rely on values if you may link to any version < v1.3.1.
|
||||||
|
*
|
||||||
|
* Only values < 100 are considered stable.
|
||||||
|
*
|
||||||
|
* note 1 : this API shall be used with static linking only.
|
||||||
|
* dynamic linking is not yet officially supported.
|
||||||
|
* note 2 : Prefer relying on the enum than on its value whenever possible
|
||||||
|
* This is the only supported way to use the error list < v1.3.1
|
||||||
|
* note 3 : ZSTD_isError() is always correct, whatever the library version.
|
||||||
|
**********************************************/
|
||||||
|
typedef enum {
|
||||||
|
ZSTD_error_no_error = 0,
|
||||||
|
ZSTD_error_GENERIC = 1,
|
||||||
|
ZSTD_error_prefix_unknown = 10,
|
||||||
|
ZSTD_error_version_unsupported = 12,
|
||||||
|
ZSTD_error_frameParameter_unsupported = 14,
|
||||||
|
ZSTD_error_frameParameter_windowTooLarge = 16,
|
||||||
|
ZSTD_error_corruption_detected = 20,
|
||||||
|
ZSTD_error_checksum_wrong = 22,
|
||||||
|
ZSTD_error_dictionary_corrupted = 30,
|
||||||
|
ZSTD_error_dictionary_wrong = 32,
|
||||||
|
ZSTD_error_dictionaryCreation_failed = 34,
|
||||||
|
ZSTD_error_parameter_unsupported = 40,
|
||||||
|
ZSTD_error_parameter_outOfBound = 42,
|
||||||
|
ZSTD_error_tableLog_tooLarge = 44,
|
||||||
|
ZSTD_error_maxSymbolValue_tooLarge = 46,
|
||||||
|
ZSTD_error_maxSymbolValue_tooSmall = 48,
|
||||||
|
ZSTD_error_stage_wrong = 60,
|
||||||
|
ZSTD_error_init_missing = 62,
|
||||||
|
ZSTD_error_memory_allocation = 64,
|
||||||
|
ZSTD_error_workSpace_tooSmall= 66,
|
||||||
|
ZSTD_error_dstSize_tooSmall = 70,
|
||||||
|
ZSTD_error_srcSize_wrong = 72,
|
||||||
|
ZSTD_error_dstBuffer_null = 74,
|
||||||
|
/* following error codes are __NOT STABLE__, they can be removed or changed in future versions */
|
||||||
|
ZSTD_error_frameIndex_tooLarge = 100,
|
||||||
|
ZSTD_error_seekableIO = 102,
|
||||||
|
ZSTD_error_maxCode = 120 /* never EVER use this value directly, it can change in future versions! Use ZSTD_isError() instead */
|
||||||
|
} ZSTD_ErrorCode;
|
||||||
|
|
||||||
|
/*! ZSTD_getErrorCode() :
|
||||||
|
convert a `size_t` function result into a `ZSTD_ErrorCode` enum type,
|
||||||
|
which can be used to compare with enum list published above */
|
||||||
|
ZSTDERRORLIB_API ZSTD_ErrorCode ZSTD_getErrorCode(size_t functionResult);
|
||||||
|
ZSTDERRORLIB_API const char* ZSTD_getErrorString(ZSTD_ErrorCode code); /**< Same as ZSTD_getErrorName, but using a `ZSTD_ErrorCode` enum argument */
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_ERRORS_H_398273423 */
|
||||||
484
vendor/github.com/DataDog/zstd/zstd_fast.c
generated
vendored
Normal file
484
vendor/github.com/DataDog/zstd/zstd_fast.c
generated
vendored
Normal file
@@ -0,0 +1,484 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include "zstd_compress_internal.h" /* ZSTD_hashPtr, ZSTD_count, ZSTD_storeSeq */
|
||||||
|
#include "zstd_fast.h"
|
||||||
|
|
||||||
|
|
||||||
|
void ZSTD_fillHashTable(ZSTD_matchState_t* ms,
|
||||||
|
const void* const end,
|
||||||
|
ZSTD_dictTableLoadMethod_e dtlm)
|
||||||
|
{
|
||||||
|
const ZSTD_compressionParameters* const cParams = &ms->cParams;
|
||||||
|
U32* const hashTable = ms->hashTable;
|
||||||
|
U32 const hBits = cParams->hashLog;
|
||||||
|
U32 const mls = cParams->minMatch;
|
||||||
|
const BYTE* const base = ms->window.base;
|
||||||
|
const BYTE* ip = base + ms->nextToUpdate;
|
||||||
|
const BYTE* const iend = ((const BYTE*)end) - HASH_READ_SIZE;
|
||||||
|
const U32 fastHashFillStep = 3;
|
||||||
|
|
||||||
|
/* Always insert every fastHashFillStep position into the hash table.
|
||||||
|
* Insert the other positions if their hash entry is empty.
|
||||||
|
*/
|
||||||
|
for ( ; ip + fastHashFillStep < iend + 2; ip += fastHashFillStep) {
|
||||||
|
U32 const current = (U32)(ip - base);
|
||||||
|
size_t const hash0 = ZSTD_hashPtr(ip, hBits, mls);
|
||||||
|
hashTable[hash0] = current;
|
||||||
|
if (dtlm == ZSTD_dtlm_fast) continue;
|
||||||
|
/* Only load extra positions for ZSTD_dtlm_full */
|
||||||
|
{ U32 p;
|
||||||
|
for (p = 1; p < fastHashFillStep; ++p) {
|
||||||
|
size_t const hash = ZSTD_hashPtr(ip + p, hBits, mls);
|
||||||
|
if (hashTable[hash] == 0) { /* not yet filled */
|
||||||
|
hashTable[hash] = current + p;
|
||||||
|
} } } }
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE size_t
|
||||||
|
ZSTD_compressBlock_fast_generic(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize,
|
||||||
|
U32 const mls)
|
||||||
|
{
|
||||||
|
const ZSTD_compressionParameters* const cParams = &ms->cParams;
|
||||||
|
U32* const hashTable = ms->hashTable;
|
||||||
|
U32 const hlog = cParams->hashLog;
|
||||||
|
/* support stepSize of 0 */
|
||||||
|
size_t const stepSize = cParams->targetLength + !(cParams->targetLength) + 1;
|
||||||
|
const BYTE* const base = ms->window.base;
|
||||||
|
const BYTE* const istart = (const BYTE*)src;
|
||||||
|
/* We check ip0 (ip + 0) and ip1 (ip + 1) each loop */
|
||||||
|
const BYTE* ip0 = istart;
|
||||||
|
const BYTE* ip1;
|
||||||
|
const BYTE* anchor = istart;
|
||||||
|
const U32 endIndex = (U32)((size_t)(istart - base) + srcSize);
|
||||||
|
const U32 maxDistance = 1U << cParams->windowLog;
|
||||||
|
const U32 validStartIndex = ms->window.dictLimit;
|
||||||
|
const U32 prefixStartIndex = (endIndex - validStartIndex > maxDistance) ? endIndex - maxDistance : validStartIndex;
|
||||||
|
const BYTE* const prefixStart = base + prefixStartIndex;
|
||||||
|
const BYTE* const iend = istart + srcSize;
|
||||||
|
const BYTE* const ilimit = iend - HASH_READ_SIZE;
|
||||||
|
U32 offset_1=rep[0], offset_2=rep[1];
|
||||||
|
U32 offsetSaved = 0;
|
||||||
|
|
||||||
|
/* init */
|
||||||
|
DEBUGLOG(5, "ZSTD_compressBlock_fast_generic");
|
||||||
|
ip0 += (ip0 == prefixStart);
|
||||||
|
ip1 = ip0 + 1;
|
||||||
|
{ U32 const maxRep = (U32)(ip0 - prefixStart);
|
||||||
|
if (offset_2 > maxRep) offsetSaved = offset_2, offset_2 = 0;
|
||||||
|
if (offset_1 > maxRep) offsetSaved = offset_1, offset_1 = 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Main Search Loop */
|
||||||
|
while (ip1 < ilimit) { /* < instead of <=, because check at ip0+2 */
|
||||||
|
size_t mLength;
|
||||||
|
BYTE const* ip2 = ip0 + 2;
|
||||||
|
size_t const h0 = ZSTD_hashPtr(ip0, hlog, mls);
|
||||||
|
U32 const val0 = MEM_read32(ip0);
|
||||||
|
size_t const h1 = ZSTD_hashPtr(ip1, hlog, mls);
|
||||||
|
U32 const val1 = MEM_read32(ip1);
|
||||||
|
U32 const current0 = (U32)(ip0-base);
|
||||||
|
U32 const current1 = (U32)(ip1-base);
|
||||||
|
U32 const matchIndex0 = hashTable[h0];
|
||||||
|
U32 const matchIndex1 = hashTable[h1];
|
||||||
|
BYTE const* repMatch = ip2-offset_1;
|
||||||
|
const BYTE* match0 = base + matchIndex0;
|
||||||
|
const BYTE* match1 = base + matchIndex1;
|
||||||
|
U32 offcode;
|
||||||
|
hashTable[h0] = current0; /* update hash table */
|
||||||
|
hashTable[h1] = current1; /* update hash table */
|
||||||
|
|
||||||
|
assert(ip0 + 1 == ip1);
|
||||||
|
|
||||||
|
if ((offset_1 > 0) & (MEM_read32(repMatch) == MEM_read32(ip2))) {
|
||||||
|
mLength = ip2[-1] == repMatch[-1] ? 1 : 0;
|
||||||
|
ip0 = ip2 - mLength;
|
||||||
|
match0 = repMatch - mLength;
|
||||||
|
offcode = 0;
|
||||||
|
goto _match;
|
||||||
|
}
|
||||||
|
if ((matchIndex0 > prefixStartIndex) && MEM_read32(match0) == val0) {
|
||||||
|
/* found a regular match */
|
||||||
|
goto _offset;
|
||||||
|
}
|
||||||
|
if ((matchIndex1 > prefixStartIndex) && MEM_read32(match1) == val1) {
|
||||||
|
/* found a regular match after one literal */
|
||||||
|
ip0 = ip1;
|
||||||
|
match0 = match1;
|
||||||
|
goto _offset;
|
||||||
|
}
|
||||||
|
{ size_t const step = ((size_t)(ip0-anchor) >> (kSearchStrength - 1)) + stepSize;
|
||||||
|
assert(step >= 2);
|
||||||
|
ip0 += step;
|
||||||
|
ip1 += step;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
_offset: /* Requires: ip0, match0 */
|
||||||
|
/* Compute the offset code */
|
||||||
|
offset_2 = offset_1;
|
||||||
|
offset_1 = (U32)(ip0-match0);
|
||||||
|
offcode = offset_1 + ZSTD_REP_MOVE;
|
||||||
|
mLength = 0;
|
||||||
|
/* Count the backwards match length */
|
||||||
|
while (((ip0>anchor) & (match0>prefixStart))
|
||||||
|
&& (ip0[-1] == match0[-1])) { ip0--; match0--; mLength++; } /* catch up */
|
||||||
|
|
||||||
|
_match: /* Requires: ip0, match0, offcode */
|
||||||
|
/* Count the forward length */
|
||||||
|
mLength += ZSTD_count(ip0+mLength+4, match0+mLength+4, iend) + 4;
|
||||||
|
ZSTD_storeSeq(seqStore, (size_t)(ip0-anchor), anchor, iend, offcode, mLength-MINMATCH);
|
||||||
|
/* match found */
|
||||||
|
ip0 += mLength;
|
||||||
|
anchor = ip0;
|
||||||
|
ip1 = ip0 + 1;
|
||||||
|
|
||||||
|
if (ip0 <= ilimit) {
|
||||||
|
/* Fill Table */
|
||||||
|
assert(base+current0+2 > istart); /* check base overflow */
|
||||||
|
hashTable[ZSTD_hashPtr(base+current0+2, hlog, mls)] = current0+2; /* here because current+2 could be > iend-8 */
|
||||||
|
hashTable[ZSTD_hashPtr(ip0-2, hlog, mls)] = (U32)(ip0-2-base);
|
||||||
|
|
||||||
|
while ( ((ip0 <= ilimit) & (offset_2>0)) /* offset_2==0 means offset_2 is invalidated */
|
||||||
|
&& (MEM_read32(ip0) == MEM_read32(ip0 - offset_2)) ) {
|
||||||
|
/* store sequence */
|
||||||
|
size_t const rLength = ZSTD_count(ip0+4, ip0+4-offset_2, iend) + 4;
|
||||||
|
{ U32 const tmpOff = offset_2; offset_2 = offset_1; offset_1 = tmpOff; } /* swap offset_2 <=> offset_1 */
|
||||||
|
hashTable[ZSTD_hashPtr(ip0, hlog, mls)] = (U32)(ip0-base);
|
||||||
|
ip0 += rLength;
|
||||||
|
ip1 = ip0 + 1;
|
||||||
|
ZSTD_storeSeq(seqStore, 0 /*litLen*/, anchor, iend, 0 /*offCode*/, rLength-MINMATCH);
|
||||||
|
anchor = ip0;
|
||||||
|
continue; /* faster when present (confirmed on gcc-8) ... (?) */
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* save reps for next block */
|
||||||
|
rep[0] = offset_1 ? offset_1 : offsetSaved;
|
||||||
|
rep[1] = offset_2 ? offset_2 : offsetSaved;
|
||||||
|
|
||||||
|
/* Return the last literals size */
|
||||||
|
return (size_t)(iend - anchor);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t ZSTD_compressBlock_fast(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
U32 const mls = ms->cParams.minMatch;
|
||||||
|
assert(ms->dictMatchState == NULL);
|
||||||
|
switch(mls)
|
||||||
|
{
|
||||||
|
default: /* includes case 3 */
|
||||||
|
case 4 :
|
||||||
|
return ZSTD_compressBlock_fast_generic(ms, seqStore, rep, src, srcSize, 4);
|
||||||
|
case 5 :
|
||||||
|
return ZSTD_compressBlock_fast_generic(ms, seqStore, rep, src, srcSize, 5);
|
||||||
|
case 6 :
|
||||||
|
return ZSTD_compressBlock_fast_generic(ms, seqStore, rep, src, srcSize, 6);
|
||||||
|
case 7 :
|
||||||
|
return ZSTD_compressBlock_fast_generic(ms, seqStore, rep, src, srcSize, 7);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
FORCE_INLINE_TEMPLATE
|
||||||
|
size_t ZSTD_compressBlock_fast_dictMatchState_generic(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize, U32 const mls)
|
||||||
|
{
|
||||||
|
const ZSTD_compressionParameters* const cParams = &ms->cParams;
|
||||||
|
U32* const hashTable = ms->hashTable;
|
||||||
|
U32 const hlog = cParams->hashLog;
|
||||||
|
/* support stepSize of 0 */
|
||||||
|
U32 const stepSize = cParams->targetLength + !(cParams->targetLength);
|
||||||
|
const BYTE* const base = ms->window.base;
|
||||||
|
const BYTE* const istart = (const BYTE*)src;
|
||||||
|
const BYTE* ip = istart;
|
||||||
|
const BYTE* anchor = istart;
|
||||||
|
const U32 prefixStartIndex = ms->window.dictLimit;
|
||||||
|
const BYTE* const prefixStart = base + prefixStartIndex;
|
||||||
|
const BYTE* const iend = istart + srcSize;
|
||||||
|
const BYTE* const ilimit = iend - HASH_READ_SIZE;
|
||||||
|
U32 offset_1=rep[0], offset_2=rep[1];
|
||||||
|
U32 offsetSaved = 0;
|
||||||
|
|
||||||
|
const ZSTD_matchState_t* const dms = ms->dictMatchState;
|
||||||
|
const ZSTD_compressionParameters* const dictCParams = &dms->cParams ;
|
||||||
|
const U32* const dictHashTable = dms->hashTable;
|
||||||
|
const U32 dictStartIndex = dms->window.dictLimit;
|
||||||
|
const BYTE* const dictBase = dms->window.base;
|
||||||
|
const BYTE* const dictStart = dictBase + dictStartIndex;
|
||||||
|
const BYTE* const dictEnd = dms->window.nextSrc;
|
||||||
|
const U32 dictIndexDelta = prefixStartIndex - (U32)(dictEnd - dictBase);
|
||||||
|
const U32 dictAndPrefixLength = (U32)(ip - prefixStart + dictEnd - dictStart);
|
||||||
|
const U32 dictHLog = dictCParams->hashLog;
|
||||||
|
|
||||||
|
/* if a dictionary is still attached, it necessarily means that
|
||||||
|
* it is within window size. So we just check it. */
|
||||||
|
const U32 maxDistance = 1U << cParams->windowLog;
|
||||||
|
const U32 endIndex = (U32)((size_t)(ip - base) + srcSize);
|
||||||
|
assert(endIndex - prefixStartIndex <= maxDistance);
|
||||||
|
(void)maxDistance; (void)endIndex; /* these variables are not used when assert() is disabled */
|
||||||
|
|
||||||
|
/* ensure there will be no no underflow
|
||||||
|
* when translating a dict index into a local index */
|
||||||
|
assert(prefixStartIndex >= (U32)(dictEnd - dictBase));
|
||||||
|
|
||||||
|
/* init */
|
||||||
|
DEBUGLOG(5, "ZSTD_compressBlock_fast_dictMatchState_generic");
|
||||||
|
ip += (dictAndPrefixLength == 0);
|
||||||
|
/* dictMatchState repCode checks don't currently handle repCode == 0
|
||||||
|
* disabling. */
|
||||||
|
assert(offset_1 <= dictAndPrefixLength);
|
||||||
|
assert(offset_2 <= dictAndPrefixLength);
|
||||||
|
|
||||||
|
/* Main Search Loop */
|
||||||
|
while (ip < ilimit) { /* < instead of <=, because repcode check at (ip+1) */
|
||||||
|
size_t mLength;
|
||||||
|
size_t const h = ZSTD_hashPtr(ip, hlog, mls);
|
||||||
|
U32 const current = (U32)(ip-base);
|
||||||
|
U32 const matchIndex = hashTable[h];
|
||||||
|
const BYTE* match = base + matchIndex;
|
||||||
|
const U32 repIndex = current + 1 - offset_1;
|
||||||
|
const BYTE* repMatch = (repIndex < prefixStartIndex) ?
|
||||||
|
dictBase + (repIndex - dictIndexDelta) :
|
||||||
|
base + repIndex;
|
||||||
|
hashTable[h] = current; /* update hash table */
|
||||||
|
|
||||||
|
if ( ((U32)((prefixStartIndex-1) - repIndex) >= 3) /* intentional underflow : ensure repIndex isn't overlapping dict + prefix */
|
||||||
|
&& (MEM_read32(repMatch) == MEM_read32(ip+1)) ) {
|
||||||
|
const BYTE* const repMatchEnd = repIndex < prefixStartIndex ? dictEnd : iend;
|
||||||
|
mLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixStart) + 4;
|
||||||
|
ip++;
|
||||||
|
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH);
|
||||||
|
} else if ( (matchIndex <= prefixStartIndex) ) {
|
||||||
|
size_t const dictHash = ZSTD_hashPtr(ip, dictHLog, mls);
|
||||||
|
U32 const dictMatchIndex = dictHashTable[dictHash];
|
||||||
|
const BYTE* dictMatch = dictBase + dictMatchIndex;
|
||||||
|
if (dictMatchIndex <= dictStartIndex ||
|
||||||
|
MEM_read32(dictMatch) != MEM_read32(ip)) {
|
||||||
|
assert(stepSize >= 1);
|
||||||
|
ip += ((ip-anchor) >> kSearchStrength) + stepSize;
|
||||||
|
continue;
|
||||||
|
} else {
|
||||||
|
/* found a dict match */
|
||||||
|
U32 const offset = (U32)(current-dictMatchIndex-dictIndexDelta);
|
||||||
|
mLength = ZSTD_count_2segments(ip+4, dictMatch+4, iend, dictEnd, prefixStart) + 4;
|
||||||
|
while (((ip>anchor) & (dictMatch>dictStart))
|
||||||
|
&& (ip[-1] == dictMatch[-1])) {
|
||||||
|
ip--; dictMatch--; mLength++;
|
||||||
|
} /* catch up */
|
||||||
|
offset_2 = offset_1;
|
||||||
|
offset_1 = offset;
|
||||||
|
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
||||||
|
}
|
||||||
|
} else if (MEM_read32(match) != MEM_read32(ip)) {
|
||||||
|
/* it's not a match, and we're not going to check the dictionary */
|
||||||
|
assert(stepSize >= 1);
|
||||||
|
ip += ((ip-anchor) >> kSearchStrength) + stepSize;
|
||||||
|
continue;
|
||||||
|
} else {
|
||||||
|
/* found a regular match */
|
||||||
|
U32 const offset = (U32)(ip-match);
|
||||||
|
mLength = ZSTD_count(ip+4, match+4, iend) + 4;
|
||||||
|
while (((ip>anchor) & (match>prefixStart))
|
||||||
|
&& (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */
|
||||||
|
offset_2 = offset_1;
|
||||||
|
offset_1 = offset;
|
||||||
|
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* match found */
|
||||||
|
ip += mLength;
|
||||||
|
anchor = ip;
|
||||||
|
|
||||||
|
if (ip <= ilimit) {
|
||||||
|
/* Fill Table */
|
||||||
|
assert(base+current+2 > istart); /* check base overflow */
|
||||||
|
hashTable[ZSTD_hashPtr(base+current+2, hlog, mls)] = current+2; /* here because current+2 could be > iend-8 */
|
||||||
|
hashTable[ZSTD_hashPtr(ip-2, hlog, mls)] = (U32)(ip-2-base);
|
||||||
|
|
||||||
|
/* check immediate repcode */
|
||||||
|
while (ip <= ilimit) {
|
||||||
|
U32 const current2 = (U32)(ip-base);
|
||||||
|
U32 const repIndex2 = current2 - offset_2;
|
||||||
|
const BYTE* repMatch2 = repIndex2 < prefixStartIndex ?
|
||||||
|
dictBase - dictIndexDelta + repIndex2 :
|
||||||
|
base + repIndex2;
|
||||||
|
if ( ((U32)((prefixStartIndex-1) - (U32)repIndex2) >= 3 /* intentional overflow */)
|
||||||
|
&& (MEM_read32(repMatch2) == MEM_read32(ip)) ) {
|
||||||
|
const BYTE* const repEnd2 = repIndex2 < prefixStartIndex ? dictEnd : iend;
|
||||||
|
size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixStart) + 4;
|
||||||
|
U32 tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; /* swap offset_2 <=> offset_1 */
|
||||||
|
ZSTD_storeSeq(seqStore, 0, anchor, iend, 0, repLength2-MINMATCH);
|
||||||
|
hashTable[ZSTD_hashPtr(ip, hlog, mls)] = current2;
|
||||||
|
ip += repLength2;
|
||||||
|
anchor = ip;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* save reps for next block */
|
||||||
|
rep[0] = offset_1 ? offset_1 : offsetSaved;
|
||||||
|
rep[1] = offset_2 ? offset_2 : offsetSaved;
|
||||||
|
|
||||||
|
/* Return the last literals size */
|
||||||
|
return (size_t)(iend - anchor);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZSTD_compressBlock_fast_dictMatchState(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
U32 const mls = ms->cParams.minMatch;
|
||||||
|
assert(ms->dictMatchState != NULL);
|
||||||
|
switch(mls)
|
||||||
|
{
|
||||||
|
default: /* includes case 3 */
|
||||||
|
case 4 :
|
||||||
|
return ZSTD_compressBlock_fast_dictMatchState_generic(ms, seqStore, rep, src, srcSize, 4);
|
||||||
|
case 5 :
|
||||||
|
return ZSTD_compressBlock_fast_dictMatchState_generic(ms, seqStore, rep, src, srcSize, 5);
|
||||||
|
case 6 :
|
||||||
|
return ZSTD_compressBlock_fast_dictMatchState_generic(ms, seqStore, rep, src, srcSize, 6);
|
||||||
|
case 7 :
|
||||||
|
return ZSTD_compressBlock_fast_dictMatchState_generic(ms, seqStore, rep, src, srcSize, 7);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
static size_t ZSTD_compressBlock_fast_extDict_generic(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize, U32 const mls)
|
||||||
|
{
|
||||||
|
const ZSTD_compressionParameters* const cParams = &ms->cParams;
|
||||||
|
U32* const hashTable = ms->hashTable;
|
||||||
|
U32 const hlog = cParams->hashLog;
|
||||||
|
/* support stepSize of 0 */
|
||||||
|
U32 const stepSize = cParams->targetLength + !(cParams->targetLength);
|
||||||
|
const BYTE* const base = ms->window.base;
|
||||||
|
const BYTE* const dictBase = ms->window.dictBase;
|
||||||
|
const BYTE* const istart = (const BYTE*)src;
|
||||||
|
const BYTE* ip = istart;
|
||||||
|
const BYTE* anchor = istart;
|
||||||
|
const U32 endIndex = (U32)((size_t)(istart - base) + srcSize);
|
||||||
|
const U32 lowLimit = ZSTD_getLowestMatchIndex(ms, endIndex, cParams->windowLog);
|
||||||
|
const U32 dictStartIndex = lowLimit;
|
||||||
|
const BYTE* const dictStart = dictBase + dictStartIndex;
|
||||||
|
const U32 dictLimit = ms->window.dictLimit;
|
||||||
|
const U32 prefixStartIndex = dictLimit < lowLimit ? lowLimit : dictLimit;
|
||||||
|
const BYTE* const prefixStart = base + prefixStartIndex;
|
||||||
|
const BYTE* const dictEnd = dictBase + prefixStartIndex;
|
||||||
|
const BYTE* const iend = istart + srcSize;
|
||||||
|
const BYTE* const ilimit = iend - 8;
|
||||||
|
U32 offset_1=rep[0], offset_2=rep[1];
|
||||||
|
|
||||||
|
DEBUGLOG(5, "ZSTD_compressBlock_fast_extDict_generic");
|
||||||
|
|
||||||
|
/* switch to "regular" variant if extDict is invalidated due to maxDistance */
|
||||||
|
if (prefixStartIndex == dictStartIndex)
|
||||||
|
return ZSTD_compressBlock_fast_generic(ms, seqStore, rep, src, srcSize, mls);
|
||||||
|
|
||||||
|
/* Search Loop */
|
||||||
|
while (ip < ilimit) { /* < instead of <=, because (ip+1) */
|
||||||
|
const size_t h = ZSTD_hashPtr(ip, hlog, mls);
|
||||||
|
const U32 matchIndex = hashTable[h];
|
||||||
|
const BYTE* const matchBase = matchIndex < prefixStartIndex ? dictBase : base;
|
||||||
|
const BYTE* match = matchBase + matchIndex;
|
||||||
|
const U32 current = (U32)(ip-base);
|
||||||
|
const U32 repIndex = current + 1 - offset_1;
|
||||||
|
const BYTE* const repBase = repIndex < prefixStartIndex ? dictBase : base;
|
||||||
|
const BYTE* const repMatch = repBase + repIndex;
|
||||||
|
hashTable[h] = current; /* update hash table */
|
||||||
|
assert(offset_1 <= current +1); /* check repIndex */
|
||||||
|
|
||||||
|
if ( (((U32)((prefixStartIndex-1) - repIndex) >= 3) /* intentional underflow */ & (repIndex > dictStartIndex))
|
||||||
|
&& (MEM_read32(repMatch) == MEM_read32(ip+1)) ) {
|
||||||
|
const BYTE* const repMatchEnd = repIndex < prefixStartIndex ? dictEnd : iend;
|
||||||
|
size_t const rLength = ZSTD_count_2segments(ip+1 +4, repMatch +4, iend, repMatchEnd, prefixStart) + 4;
|
||||||
|
ip++;
|
||||||
|
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, rLength-MINMATCH);
|
||||||
|
ip += rLength;
|
||||||
|
anchor = ip;
|
||||||
|
} else {
|
||||||
|
if ( (matchIndex < dictStartIndex) ||
|
||||||
|
(MEM_read32(match) != MEM_read32(ip)) ) {
|
||||||
|
assert(stepSize >= 1);
|
||||||
|
ip += ((ip-anchor) >> kSearchStrength) + stepSize;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
{ const BYTE* const matchEnd = matchIndex < prefixStartIndex ? dictEnd : iend;
|
||||||
|
const BYTE* const lowMatchPtr = matchIndex < prefixStartIndex ? dictStart : prefixStart;
|
||||||
|
U32 const offset = current - matchIndex;
|
||||||
|
size_t mLength = ZSTD_count_2segments(ip+4, match+4, iend, matchEnd, prefixStart) + 4;
|
||||||
|
while (((ip>anchor) & (match>lowMatchPtr)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */
|
||||||
|
offset_2 = offset_1; offset_1 = offset; /* update offset history */
|
||||||
|
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, offset + ZSTD_REP_MOVE, mLength-MINMATCH);
|
||||||
|
ip += mLength;
|
||||||
|
anchor = ip;
|
||||||
|
} }
|
||||||
|
|
||||||
|
if (ip <= ilimit) {
|
||||||
|
/* Fill Table */
|
||||||
|
hashTable[ZSTD_hashPtr(base+current+2, hlog, mls)] = current+2;
|
||||||
|
hashTable[ZSTD_hashPtr(ip-2, hlog, mls)] = (U32)(ip-2-base);
|
||||||
|
/* check immediate repcode */
|
||||||
|
while (ip <= ilimit) {
|
||||||
|
U32 const current2 = (U32)(ip-base);
|
||||||
|
U32 const repIndex2 = current2 - offset_2;
|
||||||
|
const BYTE* const repMatch2 = repIndex2 < prefixStartIndex ? dictBase + repIndex2 : base + repIndex2;
|
||||||
|
if ( (((U32)((prefixStartIndex-1) - repIndex2) >= 3) & (repIndex2 > dictStartIndex)) /* intentional overflow */
|
||||||
|
&& (MEM_read32(repMatch2) == MEM_read32(ip)) ) {
|
||||||
|
const BYTE* const repEnd2 = repIndex2 < prefixStartIndex ? dictEnd : iend;
|
||||||
|
size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixStart) + 4;
|
||||||
|
{ U32 const tmpOffset = offset_2; offset_2 = offset_1; offset_1 = tmpOffset; } /* swap offset_2 <=> offset_1 */
|
||||||
|
ZSTD_storeSeq(seqStore, 0 /*litlen*/, anchor, iend, 0 /*offcode*/, repLength2-MINMATCH);
|
||||||
|
hashTable[ZSTD_hashPtr(ip, hlog, mls)] = current2;
|
||||||
|
ip += repLength2;
|
||||||
|
anchor = ip;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
} } }
|
||||||
|
|
||||||
|
/* save reps for next block */
|
||||||
|
rep[0] = offset_1;
|
||||||
|
rep[1] = offset_2;
|
||||||
|
|
||||||
|
/* Return the last literals size */
|
||||||
|
return (size_t)(iend - anchor);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
size_t ZSTD_compressBlock_fast_extDict(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
U32 const mls = ms->cParams.minMatch;
|
||||||
|
switch(mls)
|
||||||
|
{
|
||||||
|
default: /* includes case 3 */
|
||||||
|
case 4 :
|
||||||
|
return ZSTD_compressBlock_fast_extDict_generic(ms, seqStore, rep, src, srcSize, 4);
|
||||||
|
case 5 :
|
||||||
|
return ZSTD_compressBlock_fast_extDict_generic(ms, seqStore, rep, src, srcSize, 5);
|
||||||
|
case 6 :
|
||||||
|
return ZSTD_compressBlock_fast_extDict_generic(ms, seqStore, rep, src, srcSize, 6);
|
||||||
|
case 7 :
|
||||||
|
return ZSTD_compressBlock_fast_extDict_generic(ms, seqStore, rep, src, srcSize, 7);
|
||||||
|
}
|
||||||
|
}
|
||||||
37
vendor/github.com/DataDog/zstd/zstd_fast.h
generated
vendored
Normal file
37
vendor/github.com/DataDog/zstd/zstd_fast.h
generated
vendored
Normal file
@@ -0,0 +1,37 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_FAST_H
|
||||||
|
#define ZSTD_FAST_H
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#include "mem.h" /* U32 */
|
||||||
|
#include "zstd_compress_internal.h"
|
||||||
|
|
||||||
|
void ZSTD_fillHashTable(ZSTD_matchState_t* ms,
|
||||||
|
void const* end, ZSTD_dictTableLoadMethod_e dtlm);
|
||||||
|
size_t ZSTD_compressBlock_fast(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_fast_dictMatchState(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_fast_extDict(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_FAST_H */
|
||||||
350
vendor/github.com/DataDog/zstd/zstd_internal.h
generated
vendored
Normal file
350
vendor/github.com/DataDog/zstd/zstd_internal.h
generated
vendored
Normal file
@@ -0,0 +1,350 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_CCOMMON_H_MODULE
|
||||||
|
#define ZSTD_CCOMMON_H_MODULE
|
||||||
|
|
||||||
|
/* this module contains definitions which must be identical
|
||||||
|
* across compression, decompression and dictBuilder.
|
||||||
|
* It also contains a few functions useful to at least 2 of them
|
||||||
|
* and which benefit from being inlined */
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Dependencies
|
||||||
|
***************************************/
|
||||||
|
#include "compiler.h"
|
||||||
|
#include "mem.h"
|
||||||
|
#include "debug.h" /* assert, DEBUGLOG, RAWLOG, g_debuglevel */
|
||||||
|
#include "error_private.h"
|
||||||
|
#define ZSTD_STATIC_LINKING_ONLY
|
||||||
|
#include "zstd.h"
|
||||||
|
#define FSE_STATIC_LINKING_ONLY
|
||||||
|
#include "fse.h"
|
||||||
|
#define HUF_STATIC_LINKING_ONLY
|
||||||
|
#include "huf.h"
|
||||||
|
#ifndef XXH_STATIC_LINKING_ONLY
|
||||||
|
# define XXH_STATIC_LINKING_ONLY /* XXH64_state_t */
|
||||||
|
#endif
|
||||||
|
#include "xxhash.h" /* XXH_reset, update, digest */
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* ---- static assert (debug) --- */
|
||||||
|
#define ZSTD_STATIC_ASSERT(c) DEBUG_STATIC_ASSERT(c)
|
||||||
|
#define ZSTD_isError ERR_isError /* for inlining */
|
||||||
|
#define FSE_isError ERR_isError
|
||||||
|
#define HUF_isError ERR_isError
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* shared macros
|
||||||
|
***************************************/
|
||||||
|
#undef MIN
|
||||||
|
#undef MAX
|
||||||
|
#define MIN(a,b) ((a)<(b) ? (a) : (b))
|
||||||
|
#define MAX(a,b) ((a)>(b) ? (a) : (b))
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Return the specified error if the condition evaluates to true.
|
||||||
|
*
|
||||||
|
* In debug modes, prints additional information.
|
||||||
|
* In order to do that (particularly, printing the conditional that failed),
|
||||||
|
* this can't just wrap RETURN_ERROR().
|
||||||
|
*/
|
||||||
|
#define RETURN_ERROR_IF(cond, err, ...) \
|
||||||
|
if (cond) { \
|
||||||
|
RAWLOG(3, "%s:%d: ERROR!: check %s failed, returning %s", __FILE__, __LINE__, ZSTD_QUOTE(cond), ZSTD_QUOTE(ERROR(err))); \
|
||||||
|
RAWLOG(3, ": " __VA_ARGS__); \
|
||||||
|
RAWLOG(3, "\n"); \
|
||||||
|
return ERROR(err); \
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Unconditionally return the specified error.
|
||||||
|
*
|
||||||
|
* In debug modes, prints additional information.
|
||||||
|
*/
|
||||||
|
#define RETURN_ERROR(err, ...) \
|
||||||
|
do { \
|
||||||
|
RAWLOG(3, "%s:%d: ERROR!: unconditional check failed, returning %s", __FILE__, __LINE__, ZSTD_QUOTE(ERROR(err))); \
|
||||||
|
RAWLOG(3, ": " __VA_ARGS__); \
|
||||||
|
RAWLOG(3, "\n"); \
|
||||||
|
return ERROR(err); \
|
||||||
|
} while(0);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* If the provided expression evaluates to an error code, returns that error code.
|
||||||
|
*
|
||||||
|
* In debug modes, prints additional information.
|
||||||
|
*/
|
||||||
|
#define FORWARD_IF_ERROR(err, ...) \
|
||||||
|
do { \
|
||||||
|
size_t const err_code = (err); \
|
||||||
|
if (ERR_isError(err_code)) { \
|
||||||
|
RAWLOG(3, "%s:%d: ERROR!: forwarding error in %s: %s", __FILE__, __LINE__, ZSTD_QUOTE(err), ERR_getErrorName(err_code)); \
|
||||||
|
RAWLOG(3, ": " __VA_ARGS__); \
|
||||||
|
RAWLOG(3, "\n"); \
|
||||||
|
return err_code; \
|
||||||
|
} \
|
||||||
|
} while(0);
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Common constants
|
||||||
|
***************************************/
|
||||||
|
#define ZSTD_OPT_NUM (1<<12)
|
||||||
|
|
||||||
|
#define ZSTD_REP_NUM 3 /* number of repcodes */
|
||||||
|
#define ZSTD_REP_MOVE (ZSTD_REP_NUM-1)
|
||||||
|
static const U32 repStartValue[ZSTD_REP_NUM] = { 1, 4, 8 };
|
||||||
|
|
||||||
|
#define KB *(1 <<10)
|
||||||
|
#define MB *(1 <<20)
|
||||||
|
#define GB *(1U<<30)
|
||||||
|
|
||||||
|
#define BIT7 128
|
||||||
|
#define BIT6 64
|
||||||
|
#define BIT5 32
|
||||||
|
#define BIT4 16
|
||||||
|
#define BIT1 2
|
||||||
|
#define BIT0 1
|
||||||
|
|
||||||
|
#define ZSTD_WINDOWLOG_ABSOLUTEMIN 10
|
||||||
|
static const size_t ZSTD_fcs_fieldSize[4] = { 0, 2, 4, 8 };
|
||||||
|
static const size_t ZSTD_did_fieldSize[4] = { 0, 1, 2, 4 };
|
||||||
|
|
||||||
|
#define ZSTD_FRAMEIDSIZE 4 /* magic number size */
|
||||||
|
|
||||||
|
#define ZSTD_BLOCKHEADERSIZE 3 /* C standard doesn't allow `static const` variable to be init using another `static const` variable */
|
||||||
|
static const size_t ZSTD_blockHeaderSize = ZSTD_BLOCKHEADERSIZE;
|
||||||
|
typedef enum { bt_raw, bt_rle, bt_compressed, bt_reserved } blockType_e;
|
||||||
|
|
||||||
|
#define MIN_SEQUENCES_SIZE 1 /* nbSeq==0 */
|
||||||
|
#define MIN_CBLOCK_SIZE (1 /*litCSize*/ + 1 /* RLE or RAW */ + MIN_SEQUENCES_SIZE /* nbSeq==0 */) /* for a non-null block */
|
||||||
|
|
||||||
|
#define HufLog 12
|
||||||
|
typedef enum { set_basic, set_rle, set_compressed, set_repeat } symbolEncodingType_e;
|
||||||
|
|
||||||
|
#define LONGNBSEQ 0x7F00
|
||||||
|
|
||||||
|
#define MINMATCH 3
|
||||||
|
|
||||||
|
#define Litbits 8
|
||||||
|
#define MaxLit ((1<<Litbits) - 1)
|
||||||
|
#define MaxML 52
|
||||||
|
#define MaxLL 35
|
||||||
|
#define DefaultMaxOff 28
|
||||||
|
#define MaxOff 31
|
||||||
|
#define MaxSeq MAX(MaxLL, MaxML) /* Assumption : MaxOff < MaxLL,MaxML */
|
||||||
|
#define MLFSELog 9
|
||||||
|
#define LLFSELog 9
|
||||||
|
#define OffFSELog 8
|
||||||
|
#define MaxFSELog MAX(MAX(MLFSELog, LLFSELog), OffFSELog)
|
||||||
|
|
||||||
|
static const U32 LL_bits[MaxLL+1] = { 0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
|
0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
|
1, 1, 1, 1, 2, 2, 3, 3,
|
||||||
|
4, 6, 7, 8, 9,10,11,12,
|
||||||
|
13,14,15,16 };
|
||||||
|
static const S16 LL_defaultNorm[MaxLL+1] = { 4, 3, 2, 2, 2, 2, 2, 2,
|
||||||
|
2, 2, 2, 2, 2, 1, 1, 1,
|
||||||
|
2, 2, 2, 2, 2, 2, 2, 2,
|
||||||
|
2, 3, 2, 1, 1, 1, 1, 1,
|
||||||
|
-1,-1,-1,-1 };
|
||||||
|
#define LL_DEFAULTNORMLOG 6 /* for static allocation */
|
||||||
|
static const U32 LL_defaultNormLog = LL_DEFAULTNORMLOG;
|
||||||
|
|
||||||
|
static const U32 ML_bits[MaxML+1] = { 0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
|
0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
|
0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
|
0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
|
1, 1, 1, 1, 2, 2, 3, 3,
|
||||||
|
4, 4, 5, 7, 8, 9,10,11,
|
||||||
|
12,13,14,15,16 };
|
||||||
|
static const S16 ML_defaultNorm[MaxML+1] = { 1, 4, 3, 2, 2, 2, 2, 2,
|
||||||
|
2, 1, 1, 1, 1, 1, 1, 1,
|
||||||
|
1, 1, 1, 1, 1, 1, 1, 1,
|
||||||
|
1, 1, 1, 1, 1, 1, 1, 1,
|
||||||
|
1, 1, 1, 1, 1, 1, 1, 1,
|
||||||
|
1, 1, 1, 1, 1, 1,-1,-1,
|
||||||
|
-1,-1,-1,-1,-1 };
|
||||||
|
#define ML_DEFAULTNORMLOG 6 /* for static allocation */
|
||||||
|
static const U32 ML_defaultNormLog = ML_DEFAULTNORMLOG;
|
||||||
|
|
||||||
|
static const S16 OF_defaultNorm[DefaultMaxOff+1] = { 1, 1, 1, 1, 1, 1, 2, 2,
|
||||||
|
2, 1, 1, 1, 1, 1, 1, 1,
|
||||||
|
1, 1, 1, 1, 1, 1, 1, 1,
|
||||||
|
-1,-1,-1,-1,-1 };
|
||||||
|
#define OF_DEFAULTNORMLOG 5 /* for static allocation */
|
||||||
|
static const U32 OF_defaultNormLog = OF_DEFAULTNORMLOG;
|
||||||
|
|
||||||
|
|
||||||
|
/*-*******************************************
|
||||||
|
* Shared functions to include for inlining
|
||||||
|
*********************************************/
|
||||||
|
static void ZSTD_copy8(void* dst, const void* src) { memcpy(dst, src, 8); }
|
||||||
|
|
||||||
|
#define COPY8(d,s) { ZSTD_copy8(d,s); d+=8; s+=8; }
|
||||||
|
static void ZSTD_copy16(void* dst, const void* src) { memcpy(dst, src, 16); }
|
||||||
|
#define COPY16(d,s) { ZSTD_copy16(d,s); d+=16; s+=16; }
|
||||||
|
|
||||||
|
#define WILDCOPY_OVERLENGTH 32
|
||||||
|
#define WILDCOPY_VECLEN 16
|
||||||
|
|
||||||
|
typedef enum {
|
||||||
|
ZSTD_no_overlap,
|
||||||
|
ZSTD_overlap_src_before_dst
|
||||||
|
/* ZSTD_overlap_dst_before_src, */
|
||||||
|
} ZSTD_overlap_e;
|
||||||
|
|
||||||
|
/*! ZSTD_wildcopy() :
|
||||||
|
* Custom version of memcpy(), can over read/write up to WILDCOPY_OVERLENGTH bytes (if length==0)
|
||||||
|
* @param ovtype controls the overlap detection
|
||||||
|
* - ZSTD_no_overlap: The source and destination are guaranteed to be at least WILDCOPY_VECLEN bytes apart.
|
||||||
|
* - ZSTD_overlap_src_before_dst: The src and dst may overlap, but they MUST be at least 8 bytes apart.
|
||||||
|
* The src buffer must be before the dst buffer.
|
||||||
|
*/
|
||||||
|
MEM_STATIC FORCE_INLINE_ATTR DONT_VECTORIZE
|
||||||
|
void ZSTD_wildcopy(void* dst, const void* src, ptrdiff_t length, ZSTD_overlap_e const ovtype)
|
||||||
|
{
|
||||||
|
ptrdiff_t diff = (BYTE*)dst - (const BYTE*)src;
|
||||||
|
const BYTE* ip = (const BYTE*)src;
|
||||||
|
BYTE* op = (BYTE*)dst;
|
||||||
|
BYTE* const oend = op + length;
|
||||||
|
|
||||||
|
assert(diff >= 8 || (ovtype == ZSTD_no_overlap && diff <= -WILDCOPY_VECLEN));
|
||||||
|
|
||||||
|
if (ovtype == ZSTD_overlap_src_before_dst && diff < WILDCOPY_VECLEN) {
|
||||||
|
/* Handle short offset copies. */
|
||||||
|
do {
|
||||||
|
COPY8(op, ip)
|
||||||
|
} while (op < oend);
|
||||||
|
} else {
|
||||||
|
assert(diff >= WILDCOPY_VECLEN || diff <= -WILDCOPY_VECLEN);
|
||||||
|
/* Separate out the first two COPY16() calls because the copy length is
|
||||||
|
* almost certain to be short, so the branches have different
|
||||||
|
* probabilities.
|
||||||
|
* On gcc-9 unrolling once is +1.6%, twice is +2%, thrice is +1.8%.
|
||||||
|
* On clang-8 unrolling once is +1.4%, twice is +3.3%, thrice is +3%.
|
||||||
|
*/
|
||||||
|
COPY16(op, ip);
|
||||||
|
COPY16(op, ip);
|
||||||
|
if (op >= oend) return;
|
||||||
|
do {
|
||||||
|
COPY16(op, ip);
|
||||||
|
COPY16(op, ip);
|
||||||
|
}
|
||||||
|
while (op < oend);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/*-*******************************************
|
||||||
|
* Private declarations
|
||||||
|
*********************************************/
|
||||||
|
typedef struct seqDef_s {
|
||||||
|
U32 offset;
|
||||||
|
U16 litLength;
|
||||||
|
U16 matchLength;
|
||||||
|
} seqDef;
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
seqDef* sequencesStart;
|
||||||
|
seqDef* sequences;
|
||||||
|
BYTE* litStart;
|
||||||
|
BYTE* lit;
|
||||||
|
BYTE* llCode;
|
||||||
|
BYTE* mlCode;
|
||||||
|
BYTE* ofCode;
|
||||||
|
size_t maxNbSeq;
|
||||||
|
size_t maxNbLit;
|
||||||
|
U32 longLengthID; /* 0 == no longLength; 1 == Lit.longLength; 2 == Match.longLength; */
|
||||||
|
U32 longLengthPos;
|
||||||
|
} seqStore_t;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Contains the compressed frame size and an upper-bound for the decompressed frame size.
|
||||||
|
* Note: before using `compressedSize`, check for errors using ZSTD_isError().
|
||||||
|
* similarly, before using `decompressedBound`, check for errors using:
|
||||||
|
* `decompressedBound != ZSTD_CONTENTSIZE_ERROR`
|
||||||
|
*/
|
||||||
|
typedef struct {
|
||||||
|
size_t compressedSize;
|
||||||
|
unsigned long long decompressedBound;
|
||||||
|
} ZSTD_frameSizeInfo; /* decompress & legacy */
|
||||||
|
|
||||||
|
const seqStore_t* ZSTD_getSeqStore(const ZSTD_CCtx* ctx); /* compress & dictBuilder */
|
||||||
|
void ZSTD_seqToCodes(const seqStore_t* seqStorePtr); /* compress, dictBuilder, decodeCorpus (shouldn't get its definition from here) */
|
||||||
|
|
||||||
|
/* custom memory allocation functions */
|
||||||
|
void* ZSTD_malloc(size_t size, ZSTD_customMem customMem);
|
||||||
|
void* ZSTD_calloc(size_t size, ZSTD_customMem customMem);
|
||||||
|
void ZSTD_free(void* ptr, ZSTD_customMem customMem);
|
||||||
|
|
||||||
|
|
||||||
|
MEM_STATIC U32 ZSTD_highbit32(U32 val) /* compress, dictBuilder, decodeCorpus */
|
||||||
|
{
|
||||||
|
assert(val != 0);
|
||||||
|
{
|
||||||
|
# if defined(_MSC_VER) /* Visual */
|
||||||
|
unsigned long r=0;
|
||||||
|
_BitScanReverse(&r, val);
|
||||||
|
return (unsigned)r;
|
||||||
|
# elif defined(__GNUC__) && (__GNUC__ >= 3) /* GCC Intrinsic */
|
||||||
|
return __builtin_clz (val) ^ 31;
|
||||||
|
# elif defined(__ICCARM__) /* IAR Intrinsic */
|
||||||
|
return 31 - __CLZ(val);
|
||||||
|
# else /* Software version */
|
||||||
|
static const U32 DeBruijnClz[32] = { 0, 9, 1, 10, 13, 21, 2, 29, 11, 14, 16, 18, 22, 25, 3, 30, 8, 12, 20, 28, 15, 17, 24, 7, 19, 27, 23, 6, 26, 5, 4, 31 };
|
||||||
|
U32 v = val;
|
||||||
|
v |= v >> 1;
|
||||||
|
v |= v >> 2;
|
||||||
|
v |= v >> 4;
|
||||||
|
v |= v >> 8;
|
||||||
|
v |= v >> 16;
|
||||||
|
return DeBruijnClz[(v * 0x07C4ACDDU) >> 27];
|
||||||
|
# endif
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/* ZSTD_invalidateRepCodes() :
|
||||||
|
* ensures next compression will not use repcodes from previous block.
|
||||||
|
* Note : only works with regular variant;
|
||||||
|
* do not use with extDict variant ! */
|
||||||
|
void ZSTD_invalidateRepCodes(ZSTD_CCtx* cctx); /* zstdmt, adaptive_compression (shouldn't get this definition from here) */
|
||||||
|
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
blockType_e blockType;
|
||||||
|
U32 lastBlock;
|
||||||
|
U32 origSize;
|
||||||
|
} blockProperties_t; /* declared here for decompress and fullbench */
|
||||||
|
|
||||||
|
/*! ZSTD_getcBlockSize() :
|
||||||
|
* Provides the size of compressed block from block header `src` */
|
||||||
|
/* Used by: decompress, fullbench (does not get its definition from here) */
|
||||||
|
size_t ZSTD_getcBlockSize(const void* src, size_t srcSize,
|
||||||
|
blockProperties_t* bpPtr);
|
||||||
|
|
||||||
|
/*! ZSTD_decodeSeqHeaders() :
|
||||||
|
* decode sequence header from src */
|
||||||
|
/* Used by: decompress, fullbench (does not get its definition from here) */
|
||||||
|
size_t ZSTD_decodeSeqHeaders(ZSTD_DCtx* dctx, int* nbSeqPtr,
|
||||||
|
const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_CCOMMON_H_MODULE */
|
||||||
1115
vendor/github.com/DataDog/zstd/zstd_lazy.c
generated
vendored
Normal file
1115
vendor/github.com/DataDog/zstd/zstd_lazy.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
67
vendor/github.com/DataDog/zstd/zstd_lazy.h
generated
vendored
Normal file
67
vendor/github.com/DataDog/zstd/zstd_lazy.h
generated
vendored
Normal file
@@ -0,0 +1,67 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_LAZY_H
|
||||||
|
#define ZSTD_LAZY_H
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#include "zstd_compress_internal.h"
|
||||||
|
|
||||||
|
U32 ZSTD_insertAndFindFirstIndex(ZSTD_matchState_t* ms, const BYTE* ip);
|
||||||
|
|
||||||
|
void ZSTD_preserveUnsortedMark (U32* const table, U32 const size, U32 const reducerValue); /*! used in ZSTD_reduceIndex(). preemptively increase value of ZSTD_DUBT_UNSORTED_MARK */
|
||||||
|
|
||||||
|
size_t ZSTD_compressBlock_btlazy2(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_lazy2(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_lazy(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_greedy(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
|
||||||
|
size_t ZSTD_compressBlock_btlazy2_dictMatchState(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_lazy2_dictMatchState(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_lazy_dictMatchState(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_greedy_dictMatchState(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
|
||||||
|
size_t ZSTD_compressBlock_greedy_extDict(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_lazy_extDict(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_lazy2_extDict(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_btlazy2_extDict(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_LAZY_H */
|
||||||
597
vendor/github.com/DataDog/zstd/zstd_ldm.c
generated
vendored
Normal file
597
vendor/github.com/DataDog/zstd/zstd_ldm.c
generated
vendored
Normal file
@@ -0,0 +1,597 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include "zstd_ldm.h"
|
||||||
|
|
||||||
|
#include "debug.h"
|
||||||
|
#include "zstd_fast.h" /* ZSTD_fillHashTable() */
|
||||||
|
#include "zstd_double_fast.h" /* ZSTD_fillDoubleHashTable() */
|
||||||
|
|
||||||
|
#define LDM_BUCKET_SIZE_LOG 3
|
||||||
|
#define LDM_MIN_MATCH_LENGTH 64
|
||||||
|
#define LDM_HASH_RLOG 7
|
||||||
|
#define LDM_HASH_CHAR_OFFSET 10
|
||||||
|
|
||||||
|
void ZSTD_ldm_adjustParameters(ldmParams_t* params,
|
||||||
|
ZSTD_compressionParameters const* cParams)
|
||||||
|
{
|
||||||
|
params->windowLog = cParams->windowLog;
|
||||||
|
ZSTD_STATIC_ASSERT(LDM_BUCKET_SIZE_LOG <= ZSTD_LDM_BUCKETSIZELOG_MAX);
|
||||||
|
DEBUGLOG(4, "ZSTD_ldm_adjustParameters");
|
||||||
|
if (!params->bucketSizeLog) params->bucketSizeLog = LDM_BUCKET_SIZE_LOG;
|
||||||
|
if (!params->minMatchLength) params->minMatchLength = LDM_MIN_MATCH_LENGTH;
|
||||||
|
if (cParams->strategy >= ZSTD_btopt) {
|
||||||
|
/* Get out of the way of the optimal parser */
|
||||||
|
U32 const minMatch = MAX(cParams->targetLength, params->minMatchLength);
|
||||||
|
assert(minMatch >= ZSTD_LDM_MINMATCH_MIN);
|
||||||
|
assert(minMatch <= ZSTD_LDM_MINMATCH_MAX);
|
||||||
|
params->minMatchLength = minMatch;
|
||||||
|
}
|
||||||
|
if (params->hashLog == 0) {
|
||||||
|
params->hashLog = MAX(ZSTD_HASHLOG_MIN, params->windowLog - LDM_HASH_RLOG);
|
||||||
|
assert(params->hashLog <= ZSTD_HASHLOG_MAX);
|
||||||
|
}
|
||||||
|
if (params->hashRateLog == 0) {
|
||||||
|
params->hashRateLog = params->windowLog < params->hashLog
|
||||||
|
? 0
|
||||||
|
: params->windowLog - params->hashLog;
|
||||||
|
}
|
||||||
|
params->bucketSizeLog = MIN(params->bucketSizeLog, params->hashLog);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZSTD_ldm_getTableSize(ldmParams_t params)
|
||||||
|
{
|
||||||
|
size_t const ldmHSize = ((size_t)1) << params.hashLog;
|
||||||
|
size_t const ldmBucketSizeLog = MIN(params.bucketSizeLog, params.hashLog);
|
||||||
|
size_t const ldmBucketSize = ((size_t)1) << (params.hashLog - ldmBucketSizeLog);
|
||||||
|
size_t const totalSize = ZSTD_cwksp_alloc_size(ldmBucketSize)
|
||||||
|
+ ZSTD_cwksp_alloc_size(ldmHSize * sizeof(ldmEntry_t));
|
||||||
|
return params.enableLdm ? totalSize : 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZSTD_ldm_getMaxNbSeq(ldmParams_t params, size_t maxChunkSize)
|
||||||
|
{
|
||||||
|
return params.enableLdm ? (maxChunkSize / params.minMatchLength) : 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** ZSTD_ldm_getSmallHash() :
|
||||||
|
* numBits should be <= 32
|
||||||
|
* If numBits==0, returns 0.
|
||||||
|
* @return : the most significant numBits of value. */
|
||||||
|
static U32 ZSTD_ldm_getSmallHash(U64 value, U32 numBits)
|
||||||
|
{
|
||||||
|
assert(numBits <= 32);
|
||||||
|
return numBits == 0 ? 0 : (U32)(value >> (64 - numBits));
|
||||||
|
}
|
||||||
|
|
||||||
|
/** ZSTD_ldm_getChecksum() :
|
||||||
|
* numBitsToDiscard should be <= 32
|
||||||
|
* @return : the next most significant 32 bits after numBitsToDiscard */
|
||||||
|
static U32 ZSTD_ldm_getChecksum(U64 hash, U32 numBitsToDiscard)
|
||||||
|
{
|
||||||
|
assert(numBitsToDiscard <= 32);
|
||||||
|
return (hash >> (64 - 32 - numBitsToDiscard)) & 0xFFFFFFFF;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** ZSTD_ldm_getTag() ;
|
||||||
|
* Given the hash, returns the most significant numTagBits bits
|
||||||
|
* after (32 + hbits) bits.
|
||||||
|
*
|
||||||
|
* If there are not enough bits remaining, return the last
|
||||||
|
* numTagBits bits. */
|
||||||
|
static U32 ZSTD_ldm_getTag(U64 hash, U32 hbits, U32 numTagBits)
|
||||||
|
{
|
||||||
|
assert(numTagBits < 32 && hbits <= 32);
|
||||||
|
if (32 - hbits < numTagBits) {
|
||||||
|
return hash & (((U32)1 << numTagBits) - 1);
|
||||||
|
} else {
|
||||||
|
return (hash >> (32 - hbits - numTagBits)) & (((U32)1 << numTagBits) - 1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/** ZSTD_ldm_getBucket() :
|
||||||
|
* Returns a pointer to the start of the bucket associated with hash. */
|
||||||
|
static ldmEntry_t* ZSTD_ldm_getBucket(
|
||||||
|
ldmState_t* ldmState, size_t hash, ldmParams_t const ldmParams)
|
||||||
|
{
|
||||||
|
return ldmState->hashTable + (hash << ldmParams.bucketSizeLog);
|
||||||
|
}
|
||||||
|
|
||||||
|
/** ZSTD_ldm_insertEntry() :
|
||||||
|
* Insert the entry with corresponding hash into the hash table */
|
||||||
|
static void ZSTD_ldm_insertEntry(ldmState_t* ldmState,
|
||||||
|
size_t const hash, const ldmEntry_t entry,
|
||||||
|
ldmParams_t const ldmParams)
|
||||||
|
{
|
||||||
|
BYTE* const bucketOffsets = ldmState->bucketOffsets;
|
||||||
|
*(ZSTD_ldm_getBucket(ldmState, hash, ldmParams) + bucketOffsets[hash]) = entry;
|
||||||
|
bucketOffsets[hash]++;
|
||||||
|
bucketOffsets[hash] &= ((U32)1 << ldmParams.bucketSizeLog) - 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** ZSTD_ldm_makeEntryAndInsertByTag() :
|
||||||
|
*
|
||||||
|
* Gets the small hash, checksum, and tag from the rollingHash.
|
||||||
|
*
|
||||||
|
* If the tag matches (1 << ldmParams.hashRateLog)-1, then
|
||||||
|
* creates an ldmEntry from the offset, and inserts it into the hash table.
|
||||||
|
*
|
||||||
|
* hBits is the length of the small hash, which is the most significant hBits
|
||||||
|
* of rollingHash. The checksum is the next 32 most significant bits, followed
|
||||||
|
* by ldmParams.hashRateLog bits that make up the tag. */
|
||||||
|
static void ZSTD_ldm_makeEntryAndInsertByTag(ldmState_t* ldmState,
|
||||||
|
U64 const rollingHash,
|
||||||
|
U32 const hBits,
|
||||||
|
U32 const offset,
|
||||||
|
ldmParams_t const ldmParams)
|
||||||
|
{
|
||||||
|
U32 const tag = ZSTD_ldm_getTag(rollingHash, hBits, ldmParams.hashRateLog);
|
||||||
|
U32 const tagMask = ((U32)1 << ldmParams.hashRateLog) - 1;
|
||||||
|
if (tag == tagMask) {
|
||||||
|
U32 const hash = ZSTD_ldm_getSmallHash(rollingHash, hBits);
|
||||||
|
U32 const checksum = ZSTD_ldm_getChecksum(rollingHash, hBits);
|
||||||
|
ldmEntry_t entry;
|
||||||
|
entry.offset = offset;
|
||||||
|
entry.checksum = checksum;
|
||||||
|
ZSTD_ldm_insertEntry(ldmState, hash, entry, ldmParams);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/** ZSTD_ldm_countBackwardsMatch() :
|
||||||
|
* Returns the number of bytes that match backwards before pIn and pMatch.
|
||||||
|
*
|
||||||
|
* We count only bytes where pMatch >= pBase and pIn >= pAnchor. */
|
||||||
|
static size_t ZSTD_ldm_countBackwardsMatch(
|
||||||
|
const BYTE* pIn, const BYTE* pAnchor,
|
||||||
|
const BYTE* pMatch, const BYTE* pBase)
|
||||||
|
{
|
||||||
|
size_t matchLength = 0;
|
||||||
|
while (pIn > pAnchor && pMatch > pBase && pIn[-1] == pMatch[-1]) {
|
||||||
|
pIn--;
|
||||||
|
pMatch--;
|
||||||
|
matchLength++;
|
||||||
|
}
|
||||||
|
return matchLength;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** ZSTD_ldm_fillFastTables() :
|
||||||
|
*
|
||||||
|
* Fills the relevant tables for the ZSTD_fast and ZSTD_dfast strategies.
|
||||||
|
* This is similar to ZSTD_loadDictionaryContent.
|
||||||
|
*
|
||||||
|
* The tables for the other strategies are filled within their
|
||||||
|
* block compressors. */
|
||||||
|
static size_t ZSTD_ldm_fillFastTables(ZSTD_matchState_t* ms,
|
||||||
|
void const* end)
|
||||||
|
{
|
||||||
|
const BYTE* const iend = (const BYTE*)end;
|
||||||
|
|
||||||
|
switch(ms->cParams.strategy)
|
||||||
|
{
|
||||||
|
case ZSTD_fast:
|
||||||
|
ZSTD_fillHashTable(ms, iend, ZSTD_dtlm_fast);
|
||||||
|
break;
|
||||||
|
|
||||||
|
case ZSTD_dfast:
|
||||||
|
ZSTD_fillDoubleHashTable(ms, iend, ZSTD_dtlm_fast);
|
||||||
|
break;
|
||||||
|
|
||||||
|
case ZSTD_greedy:
|
||||||
|
case ZSTD_lazy:
|
||||||
|
case ZSTD_lazy2:
|
||||||
|
case ZSTD_btlazy2:
|
||||||
|
case ZSTD_btopt:
|
||||||
|
case ZSTD_btultra:
|
||||||
|
case ZSTD_btultra2:
|
||||||
|
break;
|
||||||
|
default:
|
||||||
|
assert(0); /* not possible : not a valid strategy id */
|
||||||
|
}
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** ZSTD_ldm_fillLdmHashTable() :
|
||||||
|
*
|
||||||
|
* Fills hashTable from (lastHashed + 1) to iend (non-inclusive).
|
||||||
|
* lastHash is the rolling hash that corresponds to lastHashed.
|
||||||
|
*
|
||||||
|
* Returns the rolling hash corresponding to position iend-1. */
|
||||||
|
static U64 ZSTD_ldm_fillLdmHashTable(ldmState_t* state,
|
||||||
|
U64 lastHash, const BYTE* lastHashed,
|
||||||
|
const BYTE* iend, const BYTE* base,
|
||||||
|
U32 hBits, ldmParams_t const ldmParams)
|
||||||
|
{
|
||||||
|
U64 rollingHash = lastHash;
|
||||||
|
const BYTE* cur = lastHashed + 1;
|
||||||
|
|
||||||
|
while (cur < iend) {
|
||||||
|
rollingHash = ZSTD_rollingHash_rotate(rollingHash, cur[-1],
|
||||||
|
cur[ldmParams.minMatchLength-1],
|
||||||
|
state->hashPower);
|
||||||
|
ZSTD_ldm_makeEntryAndInsertByTag(state,
|
||||||
|
rollingHash, hBits,
|
||||||
|
(U32)(cur - base), ldmParams);
|
||||||
|
++cur;
|
||||||
|
}
|
||||||
|
return rollingHash;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/** ZSTD_ldm_limitTableUpdate() :
|
||||||
|
*
|
||||||
|
* Sets cctx->nextToUpdate to a position corresponding closer to anchor
|
||||||
|
* if it is far way
|
||||||
|
* (after a long match, only update tables a limited amount). */
|
||||||
|
static void ZSTD_ldm_limitTableUpdate(ZSTD_matchState_t* ms, const BYTE* anchor)
|
||||||
|
{
|
||||||
|
U32 const current = (U32)(anchor - ms->window.base);
|
||||||
|
if (current > ms->nextToUpdate + 1024) {
|
||||||
|
ms->nextToUpdate =
|
||||||
|
current - MIN(512, current - ms->nextToUpdate - 1024);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static size_t ZSTD_ldm_generateSequences_internal(
|
||||||
|
ldmState_t* ldmState, rawSeqStore_t* rawSeqStore,
|
||||||
|
ldmParams_t const* params, void const* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
/* LDM parameters */
|
||||||
|
int const extDict = ZSTD_window_hasExtDict(ldmState->window);
|
||||||
|
U32 const minMatchLength = params->minMatchLength;
|
||||||
|
U64 const hashPower = ldmState->hashPower;
|
||||||
|
U32 const hBits = params->hashLog - params->bucketSizeLog;
|
||||||
|
U32 const ldmBucketSize = 1U << params->bucketSizeLog;
|
||||||
|
U32 const hashRateLog = params->hashRateLog;
|
||||||
|
U32 const ldmTagMask = (1U << params->hashRateLog) - 1;
|
||||||
|
/* Prefix and extDict parameters */
|
||||||
|
U32 const dictLimit = ldmState->window.dictLimit;
|
||||||
|
U32 const lowestIndex = extDict ? ldmState->window.lowLimit : dictLimit;
|
||||||
|
BYTE const* const base = ldmState->window.base;
|
||||||
|
BYTE const* const dictBase = extDict ? ldmState->window.dictBase : NULL;
|
||||||
|
BYTE const* const dictStart = extDict ? dictBase + lowestIndex : NULL;
|
||||||
|
BYTE const* const dictEnd = extDict ? dictBase + dictLimit : NULL;
|
||||||
|
BYTE const* const lowPrefixPtr = base + dictLimit;
|
||||||
|
/* Input bounds */
|
||||||
|
BYTE const* const istart = (BYTE const*)src;
|
||||||
|
BYTE const* const iend = istart + srcSize;
|
||||||
|
BYTE const* const ilimit = iend - MAX(minMatchLength, HASH_READ_SIZE);
|
||||||
|
/* Input positions */
|
||||||
|
BYTE const* anchor = istart;
|
||||||
|
BYTE const* ip = istart;
|
||||||
|
/* Rolling hash */
|
||||||
|
BYTE const* lastHashed = NULL;
|
||||||
|
U64 rollingHash = 0;
|
||||||
|
|
||||||
|
while (ip <= ilimit) {
|
||||||
|
size_t mLength;
|
||||||
|
U32 const current = (U32)(ip - base);
|
||||||
|
size_t forwardMatchLength = 0, backwardMatchLength = 0;
|
||||||
|
ldmEntry_t* bestEntry = NULL;
|
||||||
|
if (ip != istart) {
|
||||||
|
rollingHash = ZSTD_rollingHash_rotate(rollingHash, lastHashed[0],
|
||||||
|
lastHashed[minMatchLength],
|
||||||
|
hashPower);
|
||||||
|
} else {
|
||||||
|
rollingHash = ZSTD_rollingHash_compute(ip, minMatchLength);
|
||||||
|
}
|
||||||
|
lastHashed = ip;
|
||||||
|
|
||||||
|
/* Do not insert and do not look for a match */
|
||||||
|
if (ZSTD_ldm_getTag(rollingHash, hBits, hashRateLog) != ldmTagMask) {
|
||||||
|
ip++;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Get the best entry and compute the match lengths */
|
||||||
|
{
|
||||||
|
ldmEntry_t* const bucket =
|
||||||
|
ZSTD_ldm_getBucket(ldmState,
|
||||||
|
ZSTD_ldm_getSmallHash(rollingHash, hBits),
|
||||||
|
*params);
|
||||||
|
ldmEntry_t* cur;
|
||||||
|
size_t bestMatchLength = 0;
|
||||||
|
U32 const checksum = ZSTD_ldm_getChecksum(rollingHash, hBits);
|
||||||
|
|
||||||
|
for (cur = bucket; cur < bucket + ldmBucketSize; ++cur) {
|
||||||
|
size_t curForwardMatchLength, curBackwardMatchLength,
|
||||||
|
curTotalMatchLength;
|
||||||
|
if (cur->checksum != checksum || cur->offset <= lowestIndex) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
if (extDict) {
|
||||||
|
BYTE const* const curMatchBase =
|
||||||
|
cur->offset < dictLimit ? dictBase : base;
|
||||||
|
BYTE const* const pMatch = curMatchBase + cur->offset;
|
||||||
|
BYTE const* const matchEnd =
|
||||||
|
cur->offset < dictLimit ? dictEnd : iend;
|
||||||
|
BYTE const* const lowMatchPtr =
|
||||||
|
cur->offset < dictLimit ? dictStart : lowPrefixPtr;
|
||||||
|
|
||||||
|
curForwardMatchLength = ZSTD_count_2segments(
|
||||||
|
ip, pMatch, iend,
|
||||||
|
matchEnd, lowPrefixPtr);
|
||||||
|
if (curForwardMatchLength < minMatchLength) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
curBackwardMatchLength =
|
||||||
|
ZSTD_ldm_countBackwardsMatch(ip, anchor, pMatch,
|
||||||
|
lowMatchPtr);
|
||||||
|
curTotalMatchLength = curForwardMatchLength +
|
||||||
|
curBackwardMatchLength;
|
||||||
|
} else { /* !extDict */
|
||||||
|
BYTE const* const pMatch = base + cur->offset;
|
||||||
|
curForwardMatchLength = ZSTD_count(ip, pMatch, iend);
|
||||||
|
if (curForwardMatchLength < minMatchLength) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
curBackwardMatchLength =
|
||||||
|
ZSTD_ldm_countBackwardsMatch(ip, anchor, pMatch,
|
||||||
|
lowPrefixPtr);
|
||||||
|
curTotalMatchLength = curForwardMatchLength +
|
||||||
|
curBackwardMatchLength;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (curTotalMatchLength > bestMatchLength) {
|
||||||
|
bestMatchLength = curTotalMatchLength;
|
||||||
|
forwardMatchLength = curForwardMatchLength;
|
||||||
|
backwardMatchLength = curBackwardMatchLength;
|
||||||
|
bestEntry = cur;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* No match found -- continue searching */
|
||||||
|
if (bestEntry == NULL) {
|
||||||
|
ZSTD_ldm_makeEntryAndInsertByTag(ldmState, rollingHash,
|
||||||
|
hBits, current,
|
||||||
|
*params);
|
||||||
|
ip++;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Match found */
|
||||||
|
mLength = forwardMatchLength + backwardMatchLength;
|
||||||
|
ip -= backwardMatchLength;
|
||||||
|
|
||||||
|
{
|
||||||
|
/* Store the sequence:
|
||||||
|
* ip = current - backwardMatchLength
|
||||||
|
* The match is at (bestEntry->offset - backwardMatchLength)
|
||||||
|
*/
|
||||||
|
U32 const matchIndex = bestEntry->offset;
|
||||||
|
U32 const offset = current - matchIndex;
|
||||||
|
rawSeq* const seq = rawSeqStore->seq + rawSeqStore->size;
|
||||||
|
|
||||||
|
/* Out of sequence storage */
|
||||||
|
if (rawSeqStore->size == rawSeqStore->capacity)
|
||||||
|
return ERROR(dstSize_tooSmall);
|
||||||
|
seq->litLength = (U32)(ip - anchor);
|
||||||
|
seq->matchLength = (U32)mLength;
|
||||||
|
seq->offset = offset;
|
||||||
|
rawSeqStore->size++;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Insert the current entry into the hash table */
|
||||||
|
ZSTD_ldm_makeEntryAndInsertByTag(ldmState, rollingHash, hBits,
|
||||||
|
(U32)(lastHashed - base),
|
||||||
|
*params);
|
||||||
|
|
||||||
|
assert(ip + backwardMatchLength == lastHashed);
|
||||||
|
|
||||||
|
/* Fill the hash table from lastHashed+1 to ip+mLength*/
|
||||||
|
/* Heuristic: don't need to fill the entire table at end of block */
|
||||||
|
if (ip + mLength <= ilimit) {
|
||||||
|
rollingHash = ZSTD_ldm_fillLdmHashTable(
|
||||||
|
ldmState, rollingHash, lastHashed,
|
||||||
|
ip + mLength, base, hBits, *params);
|
||||||
|
lastHashed = ip + mLength - 1;
|
||||||
|
}
|
||||||
|
ip += mLength;
|
||||||
|
anchor = ip;
|
||||||
|
}
|
||||||
|
return iend - anchor;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*! ZSTD_ldm_reduceTable() :
|
||||||
|
* reduce table indexes by `reducerValue` */
|
||||||
|
static void ZSTD_ldm_reduceTable(ldmEntry_t* const table, U32 const size,
|
||||||
|
U32 const reducerValue)
|
||||||
|
{
|
||||||
|
U32 u;
|
||||||
|
for (u = 0; u < size; u++) {
|
||||||
|
if (table[u].offset < reducerValue) table[u].offset = 0;
|
||||||
|
else table[u].offset -= reducerValue;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZSTD_ldm_generateSequences(
|
||||||
|
ldmState_t* ldmState, rawSeqStore_t* sequences,
|
||||||
|
ldmParams_t const* params, void const* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
U32 const maxDist = 1U << params->windowLog;
|
||||||
|
BYTE const* const istart = (BYTE const*)src;
|
||||||
|
BYTE const* const iend = istart + srcSize;
|
||||||
|
size_t const kMaxChunkSize = 1 << 20;
|
||||||
|
size_t const nbChunks = (srcSize / kMaxChunkSize) + ((srcSize % kMaxChunkSize) != 0);
|
||||||
|
size_t chunk;
|
||||||
|
size_t leftoverSize = 0;
|
||||||
|
|
||||||
|
assert(ZSTD_CHUNKSIZE_MAX >= kMaxChunkSize);
|
||||||
|
/* Check that ZSTD_window_update() has been called for this chunk prior
|
||||||
|
* to passing it to this function.
|
||||||
|
*/
|
||||||
|
assert(ldmState->window.nextSrc >= (BYTE const*)src + srcSize);
|
||||||
|
/* The input could be very large (in zstdmt), so it must be broken up into
|
||||||
|
* chunks to enforce the maximum distance and handle overflow correction.
|
||||||
|
*/
|
||||||
|
assert(sequences->pos <= sequences->size);
|
||||||
|
assert(sequences->size <= sequences->capacity);
|
||||||
|
for (chunk = 0; chunk < nbChunks && sequences->size < sequences->capacity; ++chunk) {
|
||||||
|
BYTE const* const chunkStart = istart + chunk * kMaxChunkSize;
|
||||||
|
size_t const remaining = (size_t)(iend - chunkStart);
|
||||||
|
BYTE const *const chunkEnd =
|
||||||
|
(remaining < kMaxChunkSize) ? iend : chunkStart + kMaxChunkSize;
|
||||||
|
size_t const chunkSize = chunkEnd - chunkStart;
|
||||||
|
size_t newLeftoverSize;
|
||||||
|
size_t const prevSize = sequences->size;
|
||||||
|
|
||||||
|
assert(chunkStart < iend);
|
||||||
|
/* 1. Perform overflow correction if necessary. */
|
||||||
|
if (ZSTD_window_needOverflowCorrection(ldmState->window, chunkEnd)) {
|
||||||
|
U32 const ldmHSize = 1U << params->hashLog;
|
||||||
|
U32 const correction = ZSTD_window_correctOverflow(
|
||||||
|
&ldmState->window, /* cycleLog */ 0, maxDist, chunkStart);
|
||||||
|
ZSTD_ldm_reduceTable(ldmState->hashTable, ldmHSize, correction);
|
||||||
|
}
|
||||||
|
/* 2. We enforce the maximum offset allowed.
|
||||||
|
*
|
||||||
|
* kMaxChunkSize should be small enough that we don't lose too much of
|
||||||
|
* the window through early invalidation.
|
||||||
|
* TODO: * Test the chunk size.
|
||||||
|
* * Try invalidation after the sequence generation and test the
|
||||||
|
* the offset against maxDist directly.
|
||||||
|
*/
|
||||||
|
ZSTD_window_enforceMaxDist(&ldmState->window, chunkEnd, maxDist, NULL, NULL);
|
||||||
|
/* 3. Generate the sequences for the chunk, and get newLeftoverSize. */
|
||||||
|
newLeftoverSize = ZSTD_ldm_generateSequences_internal(
|
||||||
|
ldmState, sequences, params, chunkStart, chunkSize);
|
||||||
|
if (ZSTD_isError(newLeftoverSize))
|
||||||
|
return newLeftoverSize;
|
||||||
|
/* 4. We add the leftover literals from previous iterations to the first
|
||||||
|
* newly generated sequence, or add the `newLeftoverSize` if none are
|
||||||
|
* generated.
|
||||||
|
*/
|
||||||
|
/* Prepend the leftover literals from the last call */
|
||||||
|
if (prevSize < sequences->size) {
|
||||||
|
sequences->seq[prevSize].litLength += (U32)leftoverSize;
|
||||||
|
leftoverSize = newLeftoverSize;
|
||||||
|
} else {
|
||||||
|
assert(newLeftoverSize == chunkSize);
|
||||||
|
leftoverSize += chunkSize;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
void ZSTD_ldm_skipSequences(rawSeqStore_t* rawSeqStore, size_t srcSize, U32 const minMatch) {
|
||||||
|
while (srcSize > 0 && rawSeqStore->pos < rawSeqStore->size) {
|
||||||
|
rawSeq* seq = rawSeqStore->seq + rawSeqStore->pos;
|
||||||
|
if (srcSize <= seq->litLength) {
|
||||||
|
/* Skip past srcSize literals */
|
||||||
|
seq->litLength -= (U32)srcSize;
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
srcSize -= seq->litLength;
|
||||||
|
seq->litLength = 0;
|
||||||
|
if (srcSize < seq->matchLength) {
|
||||||
|
/* Skip past the first srcSize of the match */
|
||||||
|
seq->matchLength -= (U32)srcSize;
|
||||||
|
if (seq->matchLength < minMatch) {
|
||||||
|
/* The match is too short, omit it */
|
||||||
|
if (rawSeqStore->pos + 1 < rawSeqStore->size) {
|
||||||
|
seq[1].litLength += seq[0].matchLength;
|
||||||
|
}
|
||||||
|
rawSeqStore->pos++;
|
||||||
|
}
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
srcSize -= seq->matchLength;
|
||||||
|
seq->matchLength = 0;
|
||||||
|
rawSeqStore->pos++;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* If the sequence length is longer than remaining then the sequence is split
|
||||||
|
* between this block and the next.
|
||||||
|
*
|
||||||
|
* Returns the current sequence to handle, or if the rest of the block should
|
||||||
|
* be literals, it returns a sequence with offset == 0.
|
||||||
|
*/
|
||||||
|
static rawSeq maybeSplitSequence(rawSeqStore_t* rawSeqStore,
|
||||||
|
U32 const remaining, U32 const minMatch)
|
||||||
|
{
|
||||||
|
rawSeq sequence = rawSeqStore->seq[rawSeqStore->pos];
|
||||||
|
assert(sequence.offset > 0);
|
||||||
|
/* Likely: No partial sequence */
|
||||||
|
if (remaining >= sequence.litLength + sequence.matchLength) {
|
||||||
|
rawSeqStore->pos++;
|
||||||
|
return sequence;
|
||||||
|
}
|
||||||
|
/* Cut the sequence short (offset == 0 ==> rest is literals). */
|
||||||
|
if (remaining <= sequence.litLength) {
|
||||||
|
sequence.offset = 0;
|
||||||
|
} else if (remaining < sequence.litLength + sequence.matchLength) {
|
||||||
|
sequence.matchLength = remaining - sequence.litLength;
|
||||||
|
if (sequence.matchLength < minMatch) {
|
||||||
|
sequence.offset = 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
/* Skip past `remaining` bytes for the future sequences. */
|
||||||
|
ZSTD_ldm_skipSequences(rawSeqStore, remaining, minMatch);
|
||||||
|
return sequence;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t ZSTD_ldm_blockCompress(rawSeqStore_t* rawSeqStore,
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
const ZSTD_compressionParameters* const cParams = &ms->cParams;
|
||||||
|
unsigned const minMatch = cParams->minMatch;
|
||||||
|
ZSTD_blockCompressor const blockCompressor =
|
||||||
|
ZSTD_selectBlockCompressor(cParams->strategy, ZSTD_matchState_dictMode(ms));
|
||||||
|
/* Input bounds */
|
||||||
|
BYTE const* const istart = (BYTE const*)src;
|
||||||
|
BYTE const* const iend = istart + srcSize;
|
||||||
|
/* Input positions */
|
||||||
|
BYTE const* ip = istart;
|
||||||
|
|
||||||
|
DEBUGLOG(5, "ZSTD_ldm_blockCompress: srcSize=%zu", srcSize);
|
||||||
|
assert(rawSeqStore->pos <= rawSeqStore->size);
|
||||||
|
assert(rawSeqStore->size <= rawSeqStore->capacity);
|
||||||
|
/* Loop through each sequence and apply the block compressor to the lits */
|
||||||
|
while (rawSeqStore->pos < rawSeqStore->size && ip < iend) {
|
||||||
|
/* maybeSplitSequence updates rawSeqStore->pos */
|
||||||
|
rawSeq const sequence = maybeSplitSequence(rawSeqStore,
|
||||||
|
(U32)(iend - ip), minMatch);
|
||||||
|
int i;
|
||||||
|
/* End signal */
|
||||||
|
if (sequence.offset == 0)
|
||||||
|
break;
|
||||||
|
|
||||||
|
assert(sequence.offset <= (1U << cParams->windowLog));
|
||||||
|
assert(ip + sequence.litLength + sequence.matchLength <= iend);
|
||||||
|
|
||||||
|
/* Fill tables for block compressor */
|
||||||
|
ZSTD_ldm_limitTableUpdate(ms, ip);
|
||||||
|
ZSTD_ldm_fillFastTables(ms, ip);
|
||||||
|
/* Run the block compressor */
|
||||||
|
DEBUGLOG(5, "calling block compressor on segment of size %u", sequence.litLength);
|
||||||
|
{
|
||||||
|
size_t const newLitLength =
|
||||||
|
blockCompressor(ms, seqStore, rep, ip, sequence.litLength);
|
||||||
|
ip += sequence.litLength;
|
||||||
|
/* Update the repcodes */
|
||||||
|
for (i = ZSTD_REP_NUM - 1; i > 0; i--)
|
||||||
|
rep[i] = rep[i-1];
|
||||||
|
rep[0] = sequence.offset;
|
||||||
|
/* Store the sequence */
|
||||||
|
ZSTD_storeSeq(seqStore, newLitLength, ip - newLitLength, iend,
|
||||||
|
sequence.offset + ZSTD_REP_MOVE,
|
||||||
|
sequence.matchLength - MINMATCH);
|
||||||
|
ip += sequence.matchLength;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
/* Fill the tables for the block compressor */
|
||||||
|
ZSTD_ldm_limitTableUpdate(ms, ip);
|
||||||
|
ZSTD_ldm_fillFastTables(ms, ip);
|
||||||
|
/* Compress the last literals */
|
||||||
|
return blockCompressor(ms, seqStore, rep, ip, iend - ip);
|
||||||
|
}
|
||||||
105
vendor/github.com/DataDog/zstd/zstd_ldm.h
generated
vendored
Normal file
105
vendor/github.com/DataDog/zstd/zstd_ldm.h
generated
vendored
Normal file
@@ -0,0 +1,105 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_LDM_H
|
||||||
|
#define ZSTD_LDM_H
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#include "zstd_compress_internal.h" /* ldmParams_t, U32 */
|
||||||
|
#include "zstd.h" /* ZSTD_CCtx, size_t */
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Long distance matching
|
||||||
|
***************************************/
|
||||||
|
|
||||||
|
#define ZSTD_LDM_DEFAULT_WINDOW_LOG ZSTD_WINDOWLOG_LIMIT_DEFAULT
|
||||||
|
|
||||||
|
/**
|
||||||
|
* ZSTD_ldm_generateSequences():
|
||||||
|
*
|
||||||
|
* Generates the sequences using the long distance match finder.
|
||||||
|
* Generates long range matching sequences in `sequences`, which parse a prefix
|
||||||
|
* of the source. `sequences` must be large enough to store every sequence,
|
||||||
|
* which can be checked with `ZSTD_ldm_getMaxNbSeq()`.
|
||||||
|
* @returns 0 or an error code.
|
||||||
|
*
|
||||||
|
* NOTE: The user must have called ZSTD_window_update() for all of the input
|
||||||
|
* they have, even if they pass it to ZSTD_ldm_generateSequences() in chunks.
|
||||||
|
* NOTE: This function returns an error if it runs out of space to store
|
||||||
|
* sequences.
|
||||||
|
*/
|
||||||
|
size_t ZSTD_ldm_generateSequences(
|
||||||
|
ldmState_t* ldms, rawSeqStore_t* sequences,
|
||||||
|
ldmParams_t const* params, void const* src, size_t srcSize);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* ZSTD_ldm_blockCompress():
|
||||||
|
*
|
||||||
|
* Compresses a block using the predefined sequences, along with a secondary
|
||||||
|
* block compressor. The literals section of every sequence is passed to the
|
||||||
|
* secondary block compressor, and those sequences are interspersed with the
|
||||||
|
* predefined sequences. Returns the length of the last literals.
|
||||||
|
* Updates `rawSeqStore.pos` to indicate how many sequences have been consumed.
|
||||||
|
* `rawSeqStore.seq` may also be updated to split the last sequence between two
|
||||||
|
* blocks.
|
||||||
|
* @return The length of the last literals.
|
||||||
|
*
|
||||||
|
* NOTE: The source must be at most the maximum block size, but the predefined
|
||||||
|
* sequences can be any size, and may be longer than the block. In the case that
|
||||||
|
* they are longer than the block, the last sequences may need to be split into
|
||||||
|
* two. We handle that case correctly, and update `rawSeqStore` appropriately.
|
||||||
|
* NOTE: This function does not return any errors.
|
||||||
|
*/
|
||||||
|
size_t ZSTD_ldm_blockCompress(rawSeqStore_t* rawSeqStore,
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* ZSTD_ldm_skipSequences():
|
||||||
|
*
|
||||||
|
* Skip past `srcSize` bytes worth of sequences in `rawSeqStore`.
|
||||||
|
* Avoids emitting matches less than `minMatch` bytes.
|
||||||
|
* Must be called for data with is not passed to ZSTD_ldm_blockCompress().
|
||||||
|
*/
|
||||||
|
void ZSTD_ldm_skipSequences(rawSeqStore_t* rawSeqStore, size_t srcSize,
|
||||||
|
U32 const minMatch);
|
||||||
|
|
||||||
|
|
||||||
|
/** ZSTD_ldm_getTableSize() :
|
||||||
|
* Estimate the space needed for long distance matching tables or 0 if LDM is
|
||||||
|
* disabled.
|
||||||
|
*/
|
||||||
|
size_t ZSTD_ldm_getTableSize(ldmParams_t params);
|
||||||
|
|
||||||
|
/** ZSTD_ldm_getSeqSpace() :
|
||||||
|
* Return an upper bound on the number of sequences that can be produced by
|
||||||
|
* the long distance matcher, or 0 if LDM is disabled.
|
||||||
|
*/
|
||||||
|
size_t ZSTD_ldm_getMaxNbSeq(ldmParams_t params, size_t maxChunkSize);
|
||||||
|
|
||||||
|
/** ZSTD_ldm_adjustParameters() :
|
||||||
|
* If the params->hashRateLog is not set, set it to its default value based on
|
||||||
|
* windowLog and params->hashLog.
|
||||||
|
*
|
||||||
|
* Ensures that params->bucketSizeLog is <= params->hashLog (setting it to
|
||||||
|
* params->hashLog if it is not).
|
||||||
|
*
|
||||||
|
* Ensures that the minMatchLength >= targetLength during optimal parsing.
|
||||||
|
*/
|
||||||
|
void ZSTD_ldm_adjustParameters(ldmParams_t* params,
|
||||||
|
ZSTD_compressionParameters const* cParams);
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_FAST_H */
|
||||||
415
vendor/github.com/DataDog/zstd/zstd_legacy.h
generated
vendored
Normal file
415
vendor/github.com/DataDog/zstd/zstd_legacy.h
generated
vendored
Normal file
@@ -0,0 +1,415 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_LEGACY_H
|
||||||
|
#define ZSTD_LEGACY_H
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Includes
|
||||||
|
***************************************/
|
||||||
|
#include "mem.h" /* MEM_STATIC */
|
||||||
|
#include "error_private.h" /* ERROR */
|
||||||
|
#include "zstd_internal.h" /* ZSTD_inBuffer, ZSTD_outBuffer, ZSTD_frameSizeInfo */
|
||||||
|
|
||||||
|
#if !defined (ZSTD_LEGACY_SUPPORT) || (ZSTD_LEGACY_SUPPORT == 0)
|
||||||
|
# undef ZSTD_LEGACY_SUPPORT
|
||||||
|
# define ZSTD_LEGACY_SUPPORT 8
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 1)
|
||||||
|
# include "zstd_v01.h"
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 2)
|
||||||
|
# include "zstd_v02.h"
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 3)
|
||||||
|
# include "zstd_v03.h"
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 4)
|
||||||
|
# include "zstd_v04.h"
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 5)
|
||||||
|
# include "zstd_v05.h"
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 6)
|
||||||
|
# include "zstd_v06.h"
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 7)
|
||||||
|
# include "zstd_v07.h"
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/** ZSTD_isLegacy() :
|
||||||
|
@return : > 0 if supported by legacy decoder. 0 otherwise.
|
||||||
|
return value is the version.
|
||||||
|
*/
|
||||||
|
MEM_STATIC unsigned ZSTD_isLegacy(const void* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
U32 magicNumberLE;
|
||||||
|
if (srcSize<4) return 0;
|
||||||
|
magicNumberLE = MEM_readLE32(src);
|
||||||
|
switch(magicNumberLE)
|
||||||
|
{
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 1)
|
||||||
|
case ZSTDv01_magicNumberLE:return 1;
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 2)
|
||||||
|
case ZSTDv02_magicNumber : return 2;
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 3)
|
||||||
|
case ZSTDv03_magicNumber : return 3;
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 4)
|
||||||
|
case ZSTDv04_magicNumber : return 4;
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 5)
|
||||||
|
case ZSTDv05_MAGICNUMBER : return 5;
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 6)
|
||||||
|
case ZSTDv06_MAGICNUMBER : return 6;
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 7)
|
||||||
|
case ZSTDv07_MAGICNUMBER : return 7;
|
||||||
|
#endif
|
||||||
|
default : return 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
MEM_STATIC unsigned long long ZSTD_getDecompressedSize_legacy(const void* src, size_t srcSize)
|
||||||
|
{
|
||||||
|
U32 const version = ZSTD_isLegacy(src, srcSize);
|
||||||
|
if (version < 5) return 0; /* no decompressed size in frame header, or not a legacy format */
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 5)
|
||||||
|
if (version==5) {
|
||||||
|
ZSTDv05_parameters fParams;
|
||||||
|
size_t const frResult = ZSTDv05_getFrameParams(&fParams, src, srcSize);
|
||||||
|
if (frResult != 0) return 0;
|
||||||
|
return fParams.srcSize;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 6)
|
||||||
|
if (version==6) {
|
||||||
|
ZSTDv06_frameParams fParams;
|
||||||
|
size_t const frResult = ZSTDv06_getFrameParams(&fParams, src, srcSize);
|
||||||
|
if (frResult != 0) return 0;
|
||||||
|
return fParams.frameContentSize;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 7)
|
||||||
|
if (version==7) {
|
||||||
|
ZSTDv07_frameParams fParams;
|
||||||
|
size_t const frResult = ZSTDv07_getFrameParams(&fParams, src, srcSize);
|
||||||
|
if (frResult != 0) return 0;
|
||||||
|
return fParams.frameContentSize;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
return 0; /* should not be possible */
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
MEM_STATIC size_t ZSTD_decompressLegacy(
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t compressedSize,
|
||||||
|
const void* dict,size_t dictSize)
|
||||||
|
{
|
||||||
|
U32 const version = ZSTD_isLegacy(src, compressedSize);
|
||||||
|
(void)dst; (void)dstCapacity; (void)dict; (void)dictSize; /* unused when ZSTD_LEGACY_SUPPORT >= 8 */
|
||||||
|
switch(version)
|
||||||
|
{
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 1)
|
||||||
|
case 1 :
|
||||||
|
return ZSTDv01_decompress(dst, dstCapacity, src, compressedSize);
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 2)
|
||||||
|
case 2 :
|
||||||
|
return ZSTDv02_decompress(dst, dstCapacity, src, compressedSize);
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 3)
|
||||||
|
case 3 :
|
||||||
|
return ZSTDv03_decompress(dst, dstCapacity, src, compressedSize);
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 4)
|
||||||
|
case 4 :
|
||||||
|
return ZSTDv04_decompress(dst, dstCapacity, src, compressedSize);
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 5)
|
||||||
|
case 5 :
|
||||||
|
{ size_t result;
|
||||||
|
ZSTDv05_DCtx* const zd = ZSTDv05_createDCtx();
|
||||||
|
if (zd==NULL) return ERROR(memory_allocation);
|
||||||
|
result = ZSTDv05_decompress_usingDict(zd, dst, dstCapacity, src, compressedSize, dict, dictSize);
|
||||||
|
ZSTDv05_freeDCtx(zd);
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 6)
|
||||||
|
case 6 :
|
||||||
|
{ size_t result;
|
||||||
|
ZSTDv06_DCtx* const zd = ZSTDv06_createDCtx();
|
||||||
|
if (zd==NULL) return ERROR(memory_allocation);
|
||||||
|
result = ZSTDv06_decompress_usingDict(zd, dst, dstCapacity, src, compressedSize, dict, dictSize);
|
||||||
|
ZSTDv06_freeDCtx(zd);
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 7)
|
||||||
|
case 7 :
|
||||||
|
{ size_t result;
|
||||||
|
ZSTDv07_DCtx* const zd = ZSTDv07_createDCtx();
|
||||||
|
if (zd==NULL) return ERROR(memory_allocation);
|
||||||
|
result = ZSTDv07_decompress_usingDict(zd, dst, dstCapacity, src, compressedSize, dict, dictSize);
|
||||||
|
ZSTDv07_freeDCtx(zd);
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
default :
|
||||||
|
return ERROR(prefix_unknown);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC ZSTD_frameSizeInfo ZSTD_findFrameSizeInfoLegacy(const void *src, size_t srcSize)
|
||||||
|
{
|
||||||
|
ZSTD_frameSizeInfo frameSizeInfo;
|
||||||
|
U32 const version = ZSTD_isLegacy(src, srcSize);
|
||||||
|
switch(version)
|
||||||
|
{
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 1)
|
||||||
|
case 1 :
|
||||||
|
ZSTDv01_findFrameSizeInfoLegacy(src, srcSize,
|
||||||
|
&frameSizeInfo.compressedSize,
|
||||||
|
&frameSizeInfo.decompressedBound);
|
||||||
|
break;
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 2)
|
||||||
|
case 2 :
|
||||||
|
ZSTDv02_findFrameSizeInfoLegacy(src, srcSize,
|
||||||
|
&frameSizeInfo.compressedSize,
|
||||||
|
&frameSizeInfo.decompressedBound);
|
||||||
|
break;
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 3)
|
||||||
|
case 3 :
|
||||||
|
ZSTDv03_findFrameSizeInfoLegacy(src, srcSize,
|
||||||
|
&frameSizeInfo.compressedSize,
|
||||||
|
&frameSizeInfo.decompressedBound);
|
||||||
|
break;
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 4)
|
||||||
|
case 4 :
|
||||||
|
ZSTDv04_findFrameSizeInfoLegacy(src, srcSize,
|
||||||
|
&frameSizeInfo.compressedSize,
|
||||||
|
&frameSizeInfo.decompressedBound);
|
||||||
|
break;
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 5)
|
||||||
|
case 5 :
|
||||||
|
ZSTDv05_findFrameSizeInfoLegacy(src, srcSize,
|
||||||
|
&frameSizeInfo.compressedSize,
|
||||||
|
&frameSizeInfo.decompressedBound);
|
||||||
|
break;
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 6)
|
||||||
|
case 6 :
|
||||||
|
ZSTDv06_findFrameSizeInfoLegacy(src, srcSize,
|
||||||
|
&frameSizeInfo.compressedSize,
|
||||||
|
&frameSizeInfo.decompressedBound);
|
||||||
|
break;
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 7)
|
||||||
|
case 7 :
|
||||||
|
ZSTDv07_findFrameSizeInfoLegacy(src, srcSize,
|
||||||
|
&frameSizeInfo.compressedSize,
|
||||||
|
&frameSizeInfo.decompressedBound);
|
||||||
|
break;
|
||||||
|
#endif
|
||||||
|
default :
|
||||||
|
frameSizeInfo.compressedSize = ERROR(prefix_unknown);
|
||||||
|
frameSizeInfo.decompressedBound = ZSTD_CONTENTSIZE_ERROR;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
if (!ZSTD_isError(frameSizeInfo.compressedSize) && frameSizeInfo.compressedSize > srcSize) {
|
||||||
|
frameSizeInfo.compressedSize = ERROR(srcSize_wrong);
|
||||||
|
frameSizeInfo.decompressedBound = ZSTD_CONTENTSIZE_ERROR;
|
||||||
|
}
|
||||||
|
return frameSizeInfo;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC size_t ZSTD_findFrameCompressedSizeLegacy(const void *src, size_t srcSize)
|
||||||
|
{
|
||||||
|
ZSTD_frameSizeInfo frameSizeInfo = ZSTD_findFrameSizeInfoLegacy(src, srcSize);
|
||||||
|
return frameSizeInfo.compressedSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
MEM_STATIC size_t ZSTD_freeLegacyStreamContext(void* legacyContext, U32 version)
|
||||||
|
{
|
||||||
|
switch(version)
|
||||||
|
{
|
||||||
|
default :
|
||||||
|
case 1 :
|
||||||
|
case 2 :
|
||||||
|
case 3 :
|
||||||
|
(void)legacyContext;
|
||||||
|
return ERROR(version_unsupported);
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 4)
|
||||||
|
case 4 : return ZBUFFv04_freeDCtx((ZBUFFv04_DCtx*)legacyContext);
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 5)
|
||||||
|
case 5 : return ZBUFFv05_freeDCtx((ZBUFFv05_DCtx*)legacyContext);
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 6)
|
||||||
|
case 6 : return ZBUFFv06_freeDCtx((ZBUFFv06_DCtx*)legacyContext);
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 7)
|
||||||
|
case 7 : return ZBUFFv07_freeDCtx((ZBUFFv07_DCtx*)legacyContext);
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
MEM_STATIC size_t ZSTD_initLegacyStream(void** legacyContext, U32 prevVersion, U32 newVersion,
|
||||||
|
const void* dict, size_t dictSize)
|
||||||
|
{
|
||||||
|
DEBUGLOG(5, "ZSTD_initLegacyStream for v0.%u", newVersion);
|
||||||
|
if (prevVersion != newVersion) ZSTD_freeLegacyStreamContext(*legacyContext, prevVersion);
|
||||||
|
switch(newVersion)
|
||||||
|
{
|
||||||
|
default :
|
||||||
|
case 1 :
|
||||||
|
case 2 :
|
||||||
|
case 3 :
|
||||||
|
(void)dict; (void)dictSize;
|
||||||
|
return 0;
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 4)
|
||||||
|
case 4 :
|
||||||
|
{
|
||||||
|
ZBUFFv04_DCtx* dctx = (prevVersion != newVersion) ? ZBUFFv04_createDCtx() : (ZBUFFv04_DCtx*)*legacyContext;
|
||||||
|
if (dctx==NULL) return ERROR(memory_allocation);
|
||||||
|
ZBUFFv04_decompressInit(dctx);
|
||||||
|
ZBUFFv04_decompressWithDictionary(dctx, dict, dictSize);
|
||||||
|
*legacyContext = dctx;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 5)
|
||||||
|
case 5 :
|
||||||
|
{
|
||||||
|
ZBUFFv05_DCtx* dctx = (prevVersion != newVersion) ? ZBUFFv05_createDCtx() : (ZBUFFv05_DCtx*)*legacyContext;
|
||||||
|
if (dctx==NULL) return ERROR(memory_allocation);
|
||||||
|
ZBUFFv05_decompressInitDictionary(dctx, dict, dictSize);
|
||||||
|
*legacyContext = dctx;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 6)
|
||||||
|
case 6 :
|
||||||
|
{
|
||||||
|
ZBUFFv06_DCtx* dctx = (prevVersion != newVersion) ? ZBUFFv06_createDCtx() : (ZBUFFv06_DCtx*)*legacyContext;
|
||||||
|
if (dctx==NULL) return ERROR(memory_allocation);
|
||||||
|
ZBUFFv06_decompressInitDictionary(dctx, dict, dictSize);
|
||||||
|
*legacyContext = dctx;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 7)
|
||||||
|
case 7 :
|
||||||
|
{
|
||||||
|
ZBUFFv07_DCtx* dctx = (prevVersion != newVersion) ? ZBUFFv07_createDCtx() : (ZBUFFv07_DCtx*)*legacyContext;
|
||||||
|
if (dctx==NULL) return ERROR(memory_allocation);
|
||||||
|
ZBUFFv07_decompressInitDictionary(dctx, dict, dictSize);
|
||||||
|
*legacyContext = dctx;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
MEM_STATIC size_t ZSTD_decompressLegacyStream(void* legacyContext, U32 version,
|
||||||
|
ZSTD_outBuffer* output, ZSTD_inBuffer* input)
|
||||||
|
{
|
||||||
|
DEBUGLOG(5, "ZSTD_decompressLegacyStream for v0.%u", version);
|
||||||
|
switch(version)
|
||||||
|
{
|
||||||
|
default :
|
||||||
|
case 1 :
|
||||||
|
case 2 :
|
||||||
|
case 3 :
|
||||||
|
(void)legacyContext; (void)output; (void)input;
|
||||||
|
return ERROR(version_unsupported);
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 4)
|
||||||
|
case 4 :
|
||||||
|
{
|
||||||
|
ZBUFFv04_DCtx* dctx = (ZBUFFv04_DCtx*) legacyContext;
|
||||||
|
const void* src = (const char*)input->src + input->pos;
|
||||||
|
size_t readSize = input->size - input->pos;
|
||||||
|
void* dst = (char*)output->dst + output->pos;
|
||||||
|
size_t decodedSize = output->size - output->pos;
|
||||||
|
size_t const hintSize = ZBUFFv04_decompressContinue(dctx, dst, &decodedSize, src, &readSize);
|
||||||
|
output->pos += decodedSize;
|
||||||
|
input->pos += readSize;
|
||||||
|
return hintSize;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 5)
|
||||||
|
case 5 :
|
||||||
|
{
|
||||||
|
ZBUFFv05_DCtx* dctx = (ZBUFFv05_DCtx*) legacyContext;
|
||||||
|
const void* src = (const char*)input->src + input->pos;
|
||||||
|
size_t readSize = input->size - input->pos;
|
||||||
|
void* dst = (char*)output->dst + output->pos;
|
||||||
|
size_t decodedSize = output->size - output->pos;
|
||||||
|
size_t const hintSize = ZBUFFv05_decompressContinue(dctx, dst, &decodedSize, src, &readSize);
|
||||||
|
output->pos += decodedSize;
|
||||||
|
input->pos += readSize;
|
||||||
|
return hintSize;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 6)
|
||||||
|
case 6 :
|
||||||
|
{
|
||||||
|
ZBUFFv06_DCtx* dctx = (ZBUFFv06_DCtx*) legacyContext;
|
||||||
|
const void* src = (const char*)input->src + input->pos;
|
||||||
|
size_t readSize = input->size - input->pos;
|
||||||
|
void* dst = (char*)output->dst + output->pos;
|
||||||
|
size_t decodedSize = output->size - output->pos;
|
||||||
|
size_t const hintSize = ZBUFFv06_decompressContinue(dctx, dst, &decodedSize, src, &readSize);
|
||||||
|
output->pos += decodedSize;
|
||||||
|
input->pos += readSize;
|
||||||
|
return hintSize;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
#if (ZSTD_LEGACY_SUPPORT <= 7)
|
||||||
|
case 7 :
|
||||||
|
{
|
||||||
|
ZBUFFv07_DCtx* dctx = (ZBUFFv07_DCtx*) legacyContext;
|
||||||
|
const void* src = (const char*)input->src + input->pos;
|
||||||
|
size_t readSize = input->size - input->pos;
|
||||||
|
void* dst = (char*)output->dst + output->pos;
|
||||||
|
size_t decodedSize = output->size - output->pos;
|
||||||
|
size_t const hintSize = ZBUFFv07_decompressContinue(dctx, dst, &decodedSize, src, &readSize);
|
||||||
|
output->pos += decodedSize;
|
||||||
|
input->pos += readSize;
|
||||||
|
return hintSize;
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_LEGACY_H */
|
||||||
1246
vendor/github.com/DataDog/zstd/zstd_opt.c
generated
vendored
Normal file
1246
vendor/github.com/DataDog/zstd/zstd_opt.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
56
vendor/github.com/DataDog/zstd/zstd_opt.h
generated
vendored
Normal file
56
vendor/github.com/DataDog/zstd/zstd_opt.h
generated
vendored
Normal file
@@ -0,0 +1,56 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_OPT_H
|
||||||
|
#define ZSTD_OPT_H
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#include "zstd_compress_internal.h"
|
||||||
|
|
||||||
|
/* used in ZSTD_loadDictionaryContent() */
|
||||||
|
void ZSTD_updateTree(ZSTD_matchState_t* ms, const BYTE* ip, const BYTE* iend);
|
||||||
|
|
||||||
|
size_t ZSTD_compressBlock_btopt(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_btultra(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_btultra2(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
|
||||||
|
|
||||||
|
size_t ZSTD_compressBlock_btopt_dictMatchState(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_btultra_dictMatchState(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
|
||||||
|
size_t ZSTD_compressBlock_btopt_extDict(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
size_t ZSTD_compressBlock_btultra_extDict(
|
||||||
|
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
|
||||||
|
void const* src, size_t srcSize);
|
||||||
|
|
||||||
|
/* note : no btultra2 variant for extDict nor dictMatchState,
|
||||||
|
* because btultra2 is not meant to work with dictionaries
|
||||||
|
* and is only specific for the first block (no prefix) */
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_OPT_H */
|
||||||
332
vendor/github.com/DataDog/zstd/zstd_stream.go
generated
vendored
Normal file
332
vendor/github.com/DataDog/zstd/zstd_stream.go
generated
vendored
Normal file
@@ -0,0 +1,332 @@
|
|||||||
|
package zstd
|
||||||
|
|
||||||
|
/*
|
||||||
|
#define ZSTD_STATIC_LINKING_ONLY
|
||||||
|
#define ZBUFF_DISABLE_DEPRECATE_WARNINGS
|
||||||
|
#include "zstd.h"
|
||||||
|
#include "zbuff.h"
|
||||||
|
*/
|
||||||
|
import "C"
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"runtime"
|
||||||
|
"sync"
|
||||||
|
"unsafe"
|
||||||
|
)
|
||||||
|
|
||||||
|
var errShortRead = errors.New("short read")
|
||||||
|
|
||||||
|
// Writer is an io.WriteCloser that zstd-compresses its input.
|
||||||
|
type Writer struct {
|
||||||
|
CompressionLevel int
|
||||||
|
|
||||||
|
ctx *C.ZSTD_CCtx
|
||||||
|
dict []byte
|
||||||
|
dstBuffer []byte
|
||||||
|
firstError error
|
||||||
|
underlyingWriter io.Writer
|
||||||
|
}
|
||||||
|
|
||||||
|
func resize(in []byte, newSize int) []byte {
|
||||||
|
if in == nil {
|
||||||
|
return make([]byte, newSize)
|
||||||
|
}
|
||||||
|
if newSize <= cap(in) {
|
||||||
|
return in[:newSize]
|
||||||
|
}
|
||||||
|
toAdd := newSize - len(in)
|
||||||
|
return append(in, make([]byte, toAdd)...)
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewWriter creates a new Writer with default compression options. Writes to
|
||||||
|
// the writer will be written in compressed form to w.
|
||||||
|
func NewWriter(w io.Writer) *Writer {
|
||||||
|
return NewWriterLevelDict(w, DefaultCompression, nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewWriterLevel is like NewWriter but specifies the compression level instead
|
||||||
|
// of assuming default compression.
|
||||||
|
//
|
||||||
|
// The level can be DefaultCompression or any integer value between BestSpeed
|
||||||
|
// and BestCompression inclusive.
|
||||||
|
func NewWriterLevel(w io.Writer, level int) *Writer {
|
||||||
|
return NewWriterLevelDict(w, level, nil)
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewWriterLevelDict is like NewWriterLevel but specifies a dictionary to
|
||||||
|
// compress with. If the dictionary is empty or nil it is ignored. The dictionary
|
||||||
|
// should not be modified until the writer is closed.
|
||||||
|
func NewWriterLevelDict(w io.Writer, level int, dict []byte) *Writer {
|
||||||
|
var err error
|
||||||
|
ctx := C.ZSTD_createCCtx()
|
||||||
|
|
||||||
|
if dict == nil {
|
||||||
|
err = getError(int(C.ZSTD_compressBegin(ctx,
|
||||||
|
C.int(level))))
|
||||||
|
} else {
|
||||||
|
err = getError(int(C.ZSTD_compressBegin_usingDict(
|
||||||
|
ctx,
|
||||||
|
unsafe.Pointer(&dict[0]),
|
||||||
|
C.size_t(len(dict)),
|
||||||
|
C.int(level))))
|
||||||
|
}
|
||||||
|
|
||||||
|
return &Writer{
|
||||||
|
CompressionLevel: level,
|
||||||
|
ctx: ctx,
|
||||||
|
dict: dict,
|
||||||
|
dstBuffer: make([]byte, CompressBound(1024)),
|
||||||
|
firstError: err,
|
||||||
|
underlyingWriter: w,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Write writes a compressed form of p to the underlying io.Writer.
|
||||||
|
func (w *Writer) Write(p []byte) (int, error) {
|
||||||
|
if w.firstError != nil {
|
||||||
|
return 0, w.firstError
|
||||||
|
}
|
||||||
|
if len(p) == 0 {
|
||||||
|
return 0, nil
|
||||||
|
}
|
||||||
|
// Check if dstBuffer is enough
|
||||||
|
if len(w.dstBuffer) < CompressBound(len(p)) {
|
||||||
|
w.dstBuffer = make([]byte, CompressBound(len(p)))
|
||||||
|
}
|
||||||
|
|
||||||
|
retCode := C.ZSTD_compressContinue(
|
||||||
|
w.ctx,
|
||||||
|
unsafe.Pointer(&w.dstBuffer[0]),
|
||||||
|
C.size_t(len(w.dstBuffer)),
|
||||||
|
unsafe.Pointer(&p[0]),
|
||||||
|
C.size_t(len(p)))
|
||||||
|
|
||||||
|
if err := getError(int(retCode)); err != nil {
|
||||||
|
return 0, err
|
||||||
|
}
|
||||||
|
written := int(retCode)
|
||||||
|
|
||||||
|
// Write to underlying buffer
|
||||||
|
_, err := w.underlyingWriter.Write(w.dstBuffer[:written])
|
||||||
|
|
||||||
|
// Same behaviour as zlib, we can't know how much data we wrote, only
|
||||||
|
// if there was an error
|
||||||
|
if err != nil {
|
||||||
|
return 0, err
|
||||||
|
}
|
||||||
|
return len(p), err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close closes the Writer, flushing any unwritten data to the underlying
|
||||||
|
// io.Writer and freeing objects, but does not close the underlying io.Writer.
|
||||||
|
func (w *Writer) Close() error {
|
||||||
|
retCode := C.ZSTD_compressEnd(
|
||||||
|
w.ctx,
|
||||||
|
unsafe.Pointer(&w.dstBuffer[0]),
|
||||||
|
C.size_t(len(w.dstBuffer)),
|
||||||
|
unsafe.Pointer(nil),
|
||||||
|
C.size_t(0))
|
||||||
|
|
||||||
|
if err := getError(int(retCode)); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
written := int(retCode)
|
||||||
|
retCode = C.ZSTD_freeCCtx(w.ctx) // Safely close buffer before writing the end
|
||||||
|
|
||||||
|
if err := getError(int(retCode)); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
_, err := w.underlyingWriter.Write(w.dstBuffer[:written])
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// cSize is the recommended size of reader.compressionBuffer. This func and
|
||||||
|
// invocation allow for a one-time check for validity.
|
||||||
|
var cSize = func() int {
|
||||||
|
v := int(C.ZBUFF_recommendedDInSize())
|
||||||
|
if v <= 0 {
|
||||||
|
panic(fmt.Errorf("ZBUFF_recommendedDInSize() returned invalid size: %v", v))
|
||||||
|
}
|
||||||
|
return v
|
||||||
|
}()
|
||||||
|
|
||||||
|
// dSize is the recommended size of reader.decompressionBuffer. This func and
|
||||||
|
// invocation allow for a one-time check for validity.
|
||||||
|
var dSize = func() int {
|
||||||
|
v := int(C.ZBUFF_recommendedDOutSize())
|
||||||
|
if v <= 0 {
|
||||||
|
panic(fmt.Errorf("ZBUFF_recommendedDOutSize() returned invalid size: %v", v))
|
||||||
|
}
|
||||||
|
return v
|
||||||
|
}()
|
||||||
|
|
||||||
|
// cPool is a pool of buffers for use in reader.compressionBuffer. Buffers are
|
||||||
|
// taken from the pool in NewReaderDict, returned in reader.Close(). Returns a
|
||||||
|
// pointer to a slice to avoid the extra allocation of returning the slice as a
|
||||||
|
// value.
|
||||||
|
var cPool = sync.Pool{
|
||||||
|
New: func() interface{} {
|
||||||
|
buff := make([]byte, cSize)
|
||||||
|
return &buff
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
// dPool is a pool of buffers for use in reader.decompressionBuffer. Buffers are
|
||||||
|
// taken from the pool in NewReaderDict, returned in reader.Close(). Returns a
|
||||||
|
// pointer to a slice to avoid the extra allocation of returning the slice as a
|
||||||
|
// value.
|
||||||
|
var dPool = sync.Pool{
|
||||||
|
New: func() interface{} {
|
||||||
|
buff := make([]byte, dSize)
|
||||||
|
return &buff
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
// reader is an io.ReadCloser that decompresses when read from.
|
||||||
|
type reader struct {
|
||||||
|
ctx *C.ZBUFF_DCtx
|
||||||
|
compressionBuffer []byte
|
||||||
|
compressionLeft int
|
||||||
|
decompressionBuffer []byte
|
||||||
|
decompOff int
|
||||||
|
decompSize int
|
||||||
|
dict []byte
|
||||||
|
firstError error
|
||||||
|
recommendedSrcSize int
|
||||||
|
underlyingReader io.Reader
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewReader creates a new io.ReadCloser. Reads from the returned ReadCloser
|
||||||
|
// read and decompress data from r. It is the caller's responsibility to call
|
||||||
|
// Close on the ReadCloser when done. If this is not done, underlying objects
|
||||||
|
// in the zstd library will not be freed.
|
||||||
|
func NewReader(r io.Reader) io.ReadCloser {
|
||||||
|
return NewReaderDict(r, nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewReaderDict is like NewReader but uses a preset dictionary. NewReaderDict
|
||||||
|
// ignores the dictionary if it is nil.
|
||||||
|
func NewReaderDict(r io.Reader, dict []byte) io.ReadCloser {
|
||||||
|
var err error
|
||||||
|
ctx := C.ZBUFF_createDCtx()
|
||||||
|
if len(dict) == 0 {
|
||||||
|
err = getError(int(C.ZBUFF_decompressInit(ctx)))
|
||||||
|
} else {
|
||||||
|
err = getError(int(C.ZBUFF_decompressInitDictionary(
|
||||||
|
ctx,
|
||||||
|
unsafe.Pointer(&dict[0]),
|
||||||
|
C.size_t(len(dict)))))
|
||||||
|
}
|
||||||
|
compressionBufferP := cPool.Get().(*[]byte)
|
||||||
|
decompressionBufferP := dPool.Get().(*[]byte)
|
||||||
|
return &reader{
|
||||||
|
ctx: ctx,
|
||||||
|
dict: dict,
|
||||||
|
compressionBuffer: *compressionBufferP,
|
||||||
|
decompressionBuffer: *decompressionBufferP,
|
||||||
|
firstError: err,
|
||||||
|
recommendedSrcSize: cSize,
|
||||||
|
underlyingReader: r,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close frees the allocated C objects
|
||||||
|
func (r *reader) Close() error {
|
||||||
|
cb := r.compressionBuffer
|
||||||
|
db := r.decompressionBuffer
|
||||||
|
cPool.Put(&cb)
|
||||||
|
dPool.Put(&db)
|
||||||
|
return getError(int(C.ZBUFF_freeDCtx(r.ctx)))
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *reader) Read(p []byte) (int, error) {
|
||||||
|
|
||||||
|
// If we already have enough bytes, return
|
||||||
|
if r.decompSize-r.decompOff >= len(p) {
|
||||||
|
copy(p, r.decompressionBuffer[r.decompOff:])
|
||||||
|
r.decompOff += len(p)
|
||||||
|
return len(p), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
copy(p, r.decompressionBuffer[r.decompOff:r.decompSize])
|
||||||
|
got := r.decompSize - r.decompOff
|
||||||
|
r.decompSize = 0
|
||||||
|
r.decompOff = 0
|
||||||
|
|
||||||
|
for got < len(p) {
|
||||||
|
// Populate src
|
||||||
|
src := r.compressionBuffer
|
||||||
|
reader := r.underlyingReader
|
||||||
|
n, err := TryReadFull(reader, src[r.compressionLeft:])
|
||||||
|
if err != nil && err != errShortRead { // Handle underlying reader errors first
|
||||||
|
return 0, fmt.Errorf("failed to read from underlying reader: %s", err)
|
||||||
|
} else if n == 0 && r.compressionLeft == 0 {
|
||||||
|
return got, io.EOF
|
||||||
|
}
|
||||||
|
src = src[:r.compressionLeft+n]
|
||||||
|
|
||||||
|
// C code
|
||||||
|
cSrcSize := C.size_t(len(src))
|
||||||
|
cDstSize := C.size_t(len(r.decompressionBuffer))
|
||||||
|
retCode := int(C.ZBUFF_decompressContinue(
|
||||||
|
r.ctx,
|
||||||
|
unsafe.Pointer(&r.decompressionBuffer[0]),
|
||||||
|
&cDstSize,
|
||||||
|
unsafe.Pointer(&src[0]),
|
||||||
|
&cSrcSize))
|
||||||
|
|
||||||
|
// Keep src here eventhough, we reuse later, the code might be deleted at some point
|
||||||
|
runtime.KeepAlive(src)
|
||||||
|
if err = getError(retCode); err != nil {
|
||||||
|
return 0, fmt.Errorf("failed to decompress: %s", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Put everything in buffer
|
||||||
|
if int(cSrcSize) < len(src) {
|
||||||
|
left := src[int(cSrcSize):]
|
||||||
|
copy(r.compressionBuffer, left)
|
||||||
|
}
|
||||||
|
r.compressionLeft = len(src) - int(cSrcSize)
|
||||||
|
r.decompSize = int(cDstSize)
|
||||||
|
r.decompOff = copy(p[got:], r.decompressionBuffer[:r.decompSize])
|
||||||
|
got += r.decompOff
|
||||||
|
|
||||||
|
// Resize buffers
|
||||||
|
nsize := retCode // Hint for next src buffer size
|
||||||
|
if nsize <= 0 {
|
||||||
|
// Reset to recommended size
|
||||||
|
nsize = r.recommendedSrcSize
|
||||||
|
}
|
||||||
|
if nsize < r.compressionLeft {
|
||||||
|
nsize = r.compressionLeft
|
||||||
|
}
|
||||||
|
r.compressionBuffer = resize(r.compressionBuffer, nsize)
|
||||||
|
}
|
||||||
|
return got, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// TryReadFull reads buffer just as ReadFull does
|
||||||
|
// Here we expect that buffer may end and we do not return ErrUnexpectedEOF as ReadAtLeast does.
|
||||||
|
// We return errShortRead instead to distinguish short reads and failures.
|
||||||
|
// We cannot use ReadFull/ReadAtLeast because it masks Reader errors, such as network failures
|
||||||
|
// and causes panic instead of error.
|
||||||
|
func TryReadFull(r io.Reader, buf []byte) (n int, err error) {
|
||||||
|
for n < len(buf) && err == nil {
|
||||||
|
var nn int
|
||||||
|
nn, err = r.Read(buf[n:])
|
||||||
|
n += nn
|
||||||
|
}
|
||||||
|
if n == len(buf) && err == io.EOF {
|
||||||
|
err = nil // EOF at the end is somewhat expected
|
||||||
|
} else if err == io.EOF {
|
||||||
|
err = errShortRead
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
2152
vendor/github.com/DataDog/zstd/zstd_v01.c
generated
vendored
Normal file
2152
vendor/github.com/DataDog/zstd/zstd_v01.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
94
vendor/github.com/DataDog/zstd/zstd_v01.h
generated
vendored
Normal file
94
vendor/github.com/DataDog/zstd/zstd_v01.h
generated
vendored
Normal file
@@ -0,0 +1,94 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_V01_H_28739879432
|
||||||
|
#define ZSTD_V01_H_28739879432
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Includes
|
||||||
|
***************************************/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Simple one-step function
|
||||||
|
***************************************/
|
||||||
|
/**
|
||||||
|
ZSTDv01_decompress() : decompress ZSTD frames compliant with v0.1.x format
|
||||||
|
compressedSize : is the exact source size
|
||||||
|
maxOriginalSize : is the size of the 'dst' buffer, which must be already allocated.
|
||||||
|
It must be equal or larger than originalSize, otherwise decompression will fail.
|
||||||
|
return : the number of bytes decompressed into destination buffer (originalSize)
|
||||||
|
or an errorCode if it fails (which can be tested using ZSTDv01_isError())
|
||||||
|
*/
|
||||||
|
size_t ZSTDv01_decompress( void* dst, size_t maxOriginalSize,
|
||||||
|
const void* src, size_t compressedSize);
|
||||||
|
|
||||||
|
/**
|
||||||
|
ZSTDv01_findFrameSizeInfoLegacy() : get the source length and decompressed bound of a ZSTD frame compliant with v0.1.x format
|
||||||
|
srcSize : The size of the 'src' buffer, at least as large as the frame pointed to by 'src'
|
||||||
|
cSize (output parameter) : the number of bytes that would be read to decompress this frame
|
||||||
|
or an error code if it fails (which can be tested using ZSTDv01_isError())
|
||||||
|
dBound (output parameter) : an upper-bound for the decompressed size of the data in the frame
|
||||||
|
or ZSTD_CONTENTSIZE_ERROR if an error occurs
|
||||||
|
|
||||||
|
note : assumes `cSize` and `dBound` are _not_ NULL.
|
||||||
|
*/
|
||||||
|
void ZSTDv01_findFrameSizeInfoLegacy(const void *src, size_t srcSize,
|
||||||
|
size_t* cSize, unsigned long long* dBound);
|
||||||
|
|
||||||
|
/**
|
||||||
|
ZSTDv01_isError() : tells if the result of ZSTDv01_decompress() is an error
|
||||||
|
*/
|
||||||
|
unsigned ZSTDv01_isError(size_t code);
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Advanced functions
|
||||||
|
***************************************/
|
||||||
|
typedef struct ZSTDv01_Dctx_s ZSTDv01_Dctx;
|
||||||
|
ZSTDv01_Dctx* ZSTDv01_createDCtx(void);
|
||||||
|
size_t ZSTDv01_freeDCtx(ZSTDv01_Dctx* dctx);
|
||||||
|
|
||||||
|
size_t ZSTDv01_decompressDCtx(void* ctx,
|
||||||
|
void* dst, size_t maxOriginalSize,
|
||||||
|
const void* src, size_t compressedSize);
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Streaming functions
|
||||||
|
***************************************/
|
||||||
|
size_t ZSTDv01_resetDCtx(ZSTDv01_Dctx* dctx);
|
||||||
|
|
||||||
|
size_t ZSTDv01_nextSrcSizeToDecompress(ZSTDv01_Dctx* dctx);
|
||||||
|
size_t ZSTDv01_decompressContinue(ZSTDv01_Dctx* dctx, void* dst, size_t maxDstSize, const void* src, size_t srcSize);
|
||||||
|
/**
|
||||||
|
Use above functions alternatively.
|
||||||
|
ZSTD_nextSrcSizeToDecompress() tells how much bytes to provide as 'srcSize' to ZSTD_decompressContinue().
|
||||||
|
ZSTD_decompressContinue() will use previous data blocks to improve compression if they are located prior to current block.
|
||||||
|
Result is the number of bytes regenerated within 'dst'.
|
||||||
|
It can be zero, which is not an error; it just means ZSTD_decompressContinue() has decoded some header.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Prefix - version detection
|
||||||
|
***************************************/
|
||||||
|
#define ZSTDv01_magicNumber 0xFD2FB51E /* Big Endian version */
|
||||||
|
#define ZSTDv01_magicNumberLE 0x1EB52FFD /* Little Endian version */
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_V01_H_28739879432 */
|
||||||
3514
vendor/github.com/DataDog/zstd/zstd_v02.c
generated
vendored
Normal file
3514
vendor/github.com/DataDog/zstd/zstd_v02.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
93
vendor/github.com/DataDog/zstd/zstd_v02.h
generated
vendored
Normal file
93
vendor/github.com/DataDog/zstd/zstd_v02.h
generated
vendored
Normal file
@@ -0,0 +1,93 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_V02_H_4174539423
|
||||||
|
#define ZSTD_V02_H_4174539423
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Includes
|
||||||
|
***************************************/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Simple one-step function
|
||||||
|
***************************************/
|
||||||
|
/**
|
||||||
|
ZSTDv02_decompress() : decompress ZSTD frames compliant with v0.2.x format
|
||||||
|
compressedSize : is the exact source size
|
||||||
|
maxOriginalSize : is the size of the 'dst' buffer, which must be already allocated.
|
||||||
|
It must be equal or larger than originalSize, otherwise decompression will fail.
|
||||||
|
return : the number of bytes decompressed into destination buffer (originalSize)
|
||||||
|
or an errorCode if it fails (which can be tested using ZSTDv01_isError())
|
||||||
|
*/
|
||||||
|
size_t ZSTDv02_decompress( void* dst, size_t maxOriginalSize,
|
||||||
|
const void* src, size_t compressedSize);
|
||||||
|
|
||||||
|
/**
|
||||||
|
ZSTDv02_findFrameSizeInfoLegacy() : get the source length and decompressed bound of a ZSTD frame compliant with v0.2.x format
|
||||||
|
srcSize : The size of the 'src' buffer, at least as large as the frame pointed to by 'src'
|
||||||
|
cSize (output parameter) : the number of bytes that would be read to decompress this frame
|
||||||
|
or an error code if it fails (which can be tested using ZSTDv01_isError())
|
||||||
|
dBound (output parameter) : an upper-bound for the decompressed size of the data in the frame
|
||||||
|
or ZSTD_CONTENTSIZE_ERROR if an error occurs
|
||||||
|
|
||||||
|
note : assumes `cSize` and `dBound` are _not_ NULL.
|
||||||
|
*/
|
||||||
|
void ZSTDv02_findFrameSizeInfoLegacy(const void *src, size_t srcSize,
|
||||||
|
size_t* cSize, unsigned long long* dBound);
|
||||||
|
|
||||||
|
/**
|
||||||
|
ZSTDv02_isError() : tells if the result of ZSTDv02_decompress() is an error
|
||||||
|
*/
|
||||||
|
unsigned ZSTDv02_isError(size_t code);
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Advanced functions
|
||||||
|
***************************************/
|
||||||
|
typedef struct ZSTDv02_Dctx_s ZSTDv02_Dctx;
|
||||||
|
ZSTDv02_Dctx* ZSTDv02_createDCtx(void);
|
||||||
|
size_t ZSTDv02_freeDCtx(ZSTDv02_Dctx* dctx);
|
||||||
|
|
||||||
|
size_t ZSTDv02_decompressDCtx(void* ctx,
|
||||||
|
void* dst, size_t maxOriginalSize,
|
||||||
|
const void* src, size_t compressedSize);
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Streaming functions
|
||||||
|
***************************************/
|
||||||
|
size_t ZSTDv02_resetDCtx(ZSTDv02_Dctx* dctx);
|
||||||
|
|
||||||
|
size_t ZSTDv02_nextSrcSizeToDecompress(ZSTDv02_Dctx* dctx);
|
||||||
|
size_t ZSTDv02_decompressContinue(ZSTDv02_Dctx* dctx, void* dst, size_t maxDstSize, const void* src, size_t srcSize);
|
||||||
|
/**
|
||||||
|
Use above functions alternatively.
|
||||||
|
ZSTD_nextSrcSizeToDecompress() tells how much bytes to provide as 'srcSize' to ZSTD_decompressContinue().
|
||||||
|
ZSTD_decompressContinue() will use previous data blocks to improve compression if they are located prior to current block.
|
||||||
|
Result is the number of bytes regenerated within 'dst'.
|
||||||
|
It can be zero, which is not an error; it just means ZSTD_decompressContinue() has decoded some header.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Prefix - version detection
|
||||||
|
***************************************/
|
||||||
|
#define ZSTDv02_magicNumber 0xFD2FB522 /* v0.2 */
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_V02_H_4174539423 */
|
||||||
3156
vendor/github.com/DataDog/zstd/zstd_v03.c
generated
vendored
Normal file
3156
vendor/github.com/DataDog/zstd/zstd_v03.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
93
vendor/github.com/DataDog/zstd/zstd_v03.h
generated
vendored
Normal file
93
vendor/github.com/DataDog/zstd/zstd_v03.h
generated
vendored
Normal file
@@ -0,0 +1,93 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_V03_H_298734209782
|
||||||
|
#define ZSTD_V03_H_298734209782
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Includes
|
||||||
|
***************************************/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Simple one-step function
|
||||||
|
***************************************/
|
||||||
|
/**
|
||||||
|
ZSTDv03_decompress() : decompress ZSTD frames compliant with v0.3.x format
|
||||||
|
compressedSize : is the exact source size
|
||||||
|
maxOriginalSize : is the size of the 'dst' buffer, which must be already allocated.
|
||||||
|
It must be equal or larger than originalSize, otherwise decompression will fail.
|
||||||
|
return : the number of bytes decompressed into destination buffer (originalSize)
|
||||||
|
or an errorCode if it fails (which can be tested using ZSTDv01_isError())
|
||||||
|
*/
|
||||||
|
size_t ZSTDv03_decompress( void* dst, size_t maxOriginalSize,
|
||||||
|
const void* src, size_t compressedSize);
|
||||||
|
|
||||||
|
/**
|
||||||
|
ZSTDv03_findFrameSizeInfoLegacy() : get the source length and decompressed bound of a ZSTD frame compliant with v0.3.x format
|
||||||
|
srcSize : The size of the 'src' buffer, at least as large as the frame pointed to by 'src'
|
||||||
|
cSize (output parameter) : the number of bytes that would be read to decompress this frame
|
||||||
|
or an error code if it fails (which can be tested using ZSTDv01_isError())
|
||||||
|
dBound (output parameter) : an upper-bound for the decompressed size of the data in the frame
|
||||||
|
or ZSTD_CONTENTSIZE_ERROR if an error occurs
|
||||||
|
|
||||||
|
note : assumes `cSize` and `dBound` are _not_ NULL.
|
||||||
|
*/
|
||||||
|
void ZSTDv03_findFrameSizeInfoLegacy(const void *src, size_t srcSize,
|
||||||
|
size_t* cSize, unsigned long long* dBound);
|
||||||
|
|
||||||
|
/**
|
||||||
|
ZSTDv03_isError() : tells if the result of ZSTDv03_decompress() is an error
|
||||||
|
*/
|
||||||
|
unsigned ZSTDv03_isError(size_t code);
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Advanced functions
|
||||||
|
***************************************/
|
||||||
|
typedef struct ZSTDv03_Dctx_s ZSTDv03_Dctx;
|
||||||
|
ZSTDv03_Dctx* ZSTDv03_createDCtx(void);
|
||||||
|
size_t ZSTDv03_freeDCtx(ZSTDv03_Dctx* dctx);
|
||||||
|
|
||||||
|
size_t ZSTDv03_decompressDCtx(void* ctx,
|
||||||
|
void* dst, size_t maxOriginalSize,
|
||||||
|
const void* src, size_t compressedSize);
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Streaming functions
|
||||||
|
***************************************/
|
||||||
|
size_t ZSTDv03_resetDCtx(ZSTDv03_Dctx* dctx);
|
||||||
|
|
||||||
|
size_t ZSTDv03_nextSrcSizeToDecompress(ZSTDv03_Dctx* dctx);
|
||||||
|
size_t ZSTDv03_decompressContinue(ZSTDv03_Dctx* dctx, void* dst, size_t maxDstSize, const void* src, size_t srcSize);
|
||||||
|
/**
|
||||||
|
Use above functions alternatively.
|
||||||
|
ZSTD_nextSrcSizeToDecompress() tells how much bytes to provide as 'srcSize' to ZSTD_decompressContinue().
|
||||||
|
ZSTD_decompressContinue() will use previous data blocks to improve compression if they are located prior to current block.
|
||||||
|
Result is the number of bytes regenerated within 'dst'.
|
||||||
|
It can be zero, which is not an error; it just means ZSTD_decompressContinue() has decoded some header.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Prefix - version detection
|
||||||
|
***************************************/
|
||||||
|
#define ZSTDv03_magicNumber 0xFD2FB523 /* v0.3 */
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_V03_H_298734209782 */
|
||||||
3641
vendor/github.com/DataDog/zstd/zstd_v04.c
generated
vendored
Normal file
3641
vendor/github.com/DataDog/zstd/zstd_v04.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
142
vendor/github.com/DataDog/zstd/zstd_v04.h
generated
vendored
Normal file
142
vendor/github.com/DataDog/zstd/zstd_v04.h
generated
vendored
Normal file
@@ -0,0 +1,142 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTD_V04_H_91868324769238
|
||||||
|
#define ZSTD_V04_H_91868324769238
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Includes
|
||||||
|
***************************************/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Simple one-step function
|
||||||
|
***************************************/
|
||||||
|
/**
|
||||||
|
ZSTDv04_decompress() : decompress ZSTD frames compliant with v0.4.x format
|
||||||
|
compressedSize : is the exact source size
|
||||||
|
maxOriginalSize : is the size of the 'dst' buffer, which must be already allocated.
|
||||||
|
It must be equal or larger than originalSize, otherwise decompression will fail.
|
||||||
|
return : the number of bytes decompressed into destination buffer (originalSize)
|
||||||
|
or an errorCode if it fails (which can be tested using ZSTDv01_isError())
|
||||||
|
*/
|
||||||
|
size_t ZSTDv04_decompress( void* dst, size_t maxOriginalSize,
|
||||||
|
const void* src, size_t compressedSize);
|
||||||
|
|
||||||
|
/**
|
||||||
|
ZSTDv04_findFrameSizeInfoLegacy() : get the source length and decompressed bound of a ZSTD frame compliant with v0.4.x format
|
||||||
|
srcSize : The size of the 'src' buffer, at least as large as the frame pointed to by 'src'
|
||||||
|
cSize (output parameter) : the number of bytes that would be read to decompress this frame
|
||||||
|
or an error code if it fails (which can be tested using ZSTDv01_isError())
|
||||||
|
dBound (output parameter) : an upper-bound for the decompressed size of the data in the frame
|
||||||
|
or ZSTD_CONTENTSIZE_ERROR if an error occurs
|
||||||
|
|
||||||
|
note : assumes `cSize` and `dBound` are _not_ NULL.
|
||||||
|
*/
|
||||||
|
void ZSTDv04_findFrameSizeInfoLegacy(const void *src, size_t srcSize,
|
||||||
|
size_t* cSize, unsigned long long* dBound);
|
||||||
|
|
||||||
|
/**
|
||||||
|
ZSTDv04_isError() : tells if the result of ZSTDv04_decompress() is an error
|
||||||
|
*/
|
||||||
|
unsigned ZSTDv04_isError(size_t code);
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Advanced functions
|
||||||
|
***************************************/
|
||||||
|
typedef struct ZSTDv04_Dctx_s ZSTDv04_Dctx;
|
||||||
|
ZSTDv04_Dctx* ZSTDv04_createDCtx(void);
|
||||||
|
size_t ZSTDv04_freeDCtx(ZSTDv04_Dctx* dctx);
|
||||||
|
|
||||||
|
size_t ZSTDv04_decompressDCtx(ZSTDv04_Dctx* dctx,
|
||||||
|
void* dst, size_t maxOriginalSize,
|
||||||
|
const void* src, size_t compressedSize);
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Direct Streaming
|
||||||
|
***************************************/
|
||||||
|
size_t ZSTDv04_resetDCtx(ZSTDv04_Dctx* dctx);
|
||||||
|
|
||||||
|
size_t ZSTDv04_nextSrcSizeToDecompress(ZSTDv04_Dctx* dctx);
|
||||||
|
size_t ZSTDv04_decompressContinue(ZSTDv04_Dctx* dctx, void* dst, size_t maxDstSize, const void* src, size_t srcSize);
|
||||||
|
/**
|
||||||
|
Use above functions alternatively.
|
||||||
|
ZSTD_nextSrcSizeToDecompress() tells how much bytes to provide as 'srcSize' to ZSTD_decompressContinue().
|
||||||
|
ZSTD_decompressContinue() will use previous data blocks to improve compression if they are located prior to current block.
|
||||||
|
Result is the number of bytes regenerated within 'dst'.
|
||||||
|
It can be zero, which is not an error; it just means ZSTD_decompressContinue() has decoded some header.
|
||||||
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Buffered Streaming
|
||||||
|
***************************************/
|
||||||
|
typedef struct ZBUFFv04_DCtx_s ZBUFFv04_DCtx;
|
||||||
|
ZBUFFv04_DCtx* ZBUFFv04_createDCtx(void);
|
||||||
|
size_t ZBUFFv04_freeDCtx(ZBUFFv04_DCtx* dctx);
|
||||||
|
|
||||||
|
size_t ZBUFFv04_decompressInit(ZBUFFv04_DCtx* dctx);
|
||||||
|
size_t ZBUFFv04_decompressWithDictionary(ZBUFFv04_DCtx* dctx, const void* dict, size_t dictSize);
|
||||||
|
|
||||||
|
size_t ZBUFFv04_decompressContinue(ZBUFFv04_DCtx* dctx, void* dst, size_t* maxDstSizePtr, const void* src, size_t* srcSizePtr);
|
||||||
|
|
||||||
|
/** ************************************************
|
||||||
|
* Streaming decompression
|
||||||
|
*
|
||||||
|
* A ZBUFF_DCtx object is required to track streaming operation.
|
||||||
|
* Use ZBUFF_createDCtx() and ZBUFF_freeDCtx() to create/release resources.
|
||||||
|
* Use ZBUFF_decompressInit() to start a new decompression operation.
|
||||||
|
* ZBUFF_DCtx objects can be reused multiple times.
|
||||||
|
*
|
||||||
|
* Optionally, a reference to a static dictionary can be set, using ZBUFF_decompressWithDictionary()
|
||||||
|
* It must be the same content as the one set during compression phase.
|
||||||
|
* Dictionary content must remain accessible during the decompression process.
|
||||||
|
*
|
||||||
|
* Use ZBUFF_decompressContinue() repetitively to consume your input.
|
||||||
|
* *srcSizePtr and *maxDstSizePtr can be any size.
|
||||||
|
* The function will report how many bytes were read or written by modifying *srcSizePtr and *maxDstSizePtr.
|
||||||
|
* Note that it may not consume the entire input, in which case it's up to the caller to present remaining input again.
|
||||||
|
* The content of dst will be overwritten (up to *maxDstSizePtr) at each function call, so save its content if it matters or change dst.
|
||||||
|
* @return : a hint to preferred nb of bytes to use as input for next function call (it's only a hint, to improve latency)
|
||||||
|
* or 0 when a frame is completely decoded
|
||||||
|
* or an error code, which can be tested using ZBUFF_isError().
|
||||||
|
*
|
||||||
|
* Hint : recommended buffer sizes (not compulsory) : ZBUFF_recommendedDInSize / ZBUFF_recommendedDOutSize
|
||||||
|
* output : ZBUFF_recommendedDOutSize==128 KB block size is the internal unit, it ensures it's always possible to write a full block when it's decoded.
|
||||||
|
* input : ZBUFF_recommendedDInSize==128Kb+3; just follow indications from ZBUFF_decompressContinue() to minimize latency. It should always be <= 128 KB + 3 .
|
||||||
|
* **************************************************/
|
||||||
|
unsigned ZBUFFv04_isError(size_t errorCode);
|
||||||
|
const char* ZBUFFv04_getErrorName(size_t errorCode);
|
||||||
|
|
||||||
|
|
||||||
|
/** The below functions provide recommended buffer sizes for Compression or Decompression operations.
|
||||||
|
* These sizes are not compulsory, they just tend to offer better latency */
|
||||||
|
size_t ZBUFFv04_recommendedDInSize(void);
|
||||||
|
size_t ZBUFFv04_recommendedDOutSize(void);
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Prefix - version detection
|
||||||
|
***************************************/
|
||||||
|
#define ZSTDv04_magicNumber 0xFD2FB524 /* v0.4 */
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTD_V04_H_91868324769238 */
|
||||||
4046
vendor/github.com/DataDog/zstd/zstd_v05.c
generated
vendored
Normal file
4046
vendor/github.com/DataDog/zstd/zstd_v05.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
162
vendor/github.com/DataDog/zstd/zstd_v05.h
generated
vendored
Normal file
162
vendor/github.com/DataDog/zstd/zstd_v05.h
generated
vendored
Normal file
@@ -0,0 +1,162 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTDv05_H
|
||||||
|
#define ZSTDv05_H
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Dependencies
|
||||||
|
***************************************/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
#include "mem.h" /* U64, U32 */
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Simple functions
|
||||||
|
***************************************/
|
||||||
|
/*! ZSTDv05_decompress() :
|
||||||
|
`compressedSize` : is the _exact_ size of the compressed blob, otherwise decompression will fail.
|
||||||
|
`dstCapacity` must be large enough, equal or larger than originalSize.
|
||||||
|
@return : the number of bytes decompressed into `dst` (<= `dstCapacity`),
|
||||||
|
or an errorCode if it fails (which can be tested using ZSTDv05_isError()) */
|
||||||
|
size_t ZSTDv05_decompress( void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t compressedSize);
|
||||||
|
|
||||||
|
/**
|
||||||
|
ZSTDv05_findFrameSizeInfoLegacy() : get the source length and decompressed bound of a ZSTD frame compliant with v0.5.x format
|
||||||
|
srcSize : The size of the 'src' buffer, at least as large as the frame pointed to by 'src'
|
||||||
|
cSize (output parameter) : the number of bytes that would be read to decompress this frame
|
||||||
|
or an error code if it fails (which can be tested using ZSTDv01_isError())
|
||||||
|
dBound (output parameter) : an upper-bound for the decompressed size of the data in the frame
|
||||||
|
or ZSTD_CONTENTSIZE_ERROR if an error occurs
|
||||||
|
|
||||||
|
note : assumes `cSize` and `dBound` are _not_ NULL.
|
||||||
|
*/
|
||||||
|
void ZSTDv05_findFrameSizeInfoLegacy(const void *src, size_t srcSize,
|
||||||
|
size_t* cSize, unsigned long long* dBound);
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Helper functions
|
||||||
|
***************************************/
|
||||||
|
/* Error Management */
|
||||||
|
unsigned ZSTDv05_isError(size_t code); /*!< tells if a `size_t` function result is an error code */
|
||||||
|
const char* ZSTDv05_getErrorName(size_t code); /*!< provides readable string for an error code */
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Explicit memory management
|
||||||
|
***************************************/
|
||||||
|
/** Decompression context */
|
||||||
|
typedef struct ZSTDv05_DCtx_s ZSTDv05_DCtx;
|
||||||
|
ZSTDv05_DCtx* ZSTDv05_createDCtx(void);
|
||||||
|
size_t ZSTDv05_freeDCtx(ZSTDv05_DCtx* dctx); /*!< @return : errorCode */
|
||||||
|
|
||||||
|
/** ZSTDv05_decompressDCtx() :
|
||||||
|
* Same as ZSTDv05_decompress(), but requires an already allocated ZSTDv05_DCtx (see ZSTDv05_createDCtx()) */
|
||||||
|
size_t ZSTDv05_decompressDCtx(ZSTDv05_DCtx* ctx, void* dst, size_t dstCapacity, const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
|
||||||
|
/*-***********************
|
||||||
|
* Simple Dictionary API
|
||||||
|
*************************/
|
||||||
|
/*! ZSTDv05_decompress_usingDict() :
|
||||||
|
* Decompression using a pre-defined Dictionary content (see dictBuilder).
|
||||||
|
* Dictionary must be identical to the one used during compression, otherwise regenerated data will be corrupted.
|
||||||
|
* Note : dict can be NULL, in which case, it's equivalent to ZSTDv05_decompressDCtx() */
|
||||||
|
size_t ZSTDv05_decompress_usingDict(ZSTDv05_DCtx* dctx,
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const void* dict,size_t dictSize);
|
||||||
|
|
||||||
|
/*-************************
|
||||||
|
* Advanced Streaming API
|
||||||
|
***************************/
|
||||||
|
typedef enum { ZSTDv05_fast, ZSTDv05_greedy, ZSTDv05_lazy, ZSTDv05_lazy2, ZSTDv05_btlazy2, ZSTDv05_opt, ZSTDv05_btopt } ZSTDv05_strategy;
|
||||||
|
typedef struct {
|
||||||
|
U64 srcSize;
|
||||||
|
U32 windowLog; /* the only useful information to retrieve */
|
||||||
|
U32 contentLog; U32 hashLog; U32 searchLog; U32 searchLength; U32 targetLength; ZSTDv05_strategy strategy;
|
||||||
|
} ZSTDv05_parameters;
|
||||||
|
size_t ZSTDv05_getFrameParams(ZSTDv05_parameters* params, const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
size_t ZSTDv05_decompressBegin_usingDict(ZSTDv05_DCtx* dctx, const void* dict, size_t dictSize);
|
||||||
|
void ZSTDv05_copyDCtx(ZSTDv05_DCtx* dstDCtx, const ZSTDv05_DCtx* srcDCtx);
|
||||||
|
size_t ZSTDv05_nextSrcSizeToDecompress(ZSTDv05_DCtx* dctx);
|
||||||
|
size_t ZSTDv05_decompressContinue(ZSTDv05_DCtx* dctx, void* dst, size_t dstCapacity, const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
|
||||||
|
/*-***********************
|
||||||
|
* ZBUFF API
|
||||||
|
*************************/
|
||||||
|
typedef struct ZBUFFv05_DCtx_s ZBUFFv05_DCtx;
|
||||||
|
ZBUFFv05_DCtx* ZBUFFv05_createDCtx(void);
|
||||||
|
size_t ZBUFFv05_freeDCtx(ZBUFFv05_DCtx* dctx);
|
||||||
|
|
||||||
|
size_t ZBUFFv05_decompressInit(ZBUFFv05_DCtx* dctx);
|
||||||
|
size_t ZBUFFv05_decompressInitDictionary(ZBUFFv05_DCtx* dctx, const void* dict, size_t dictSize);
|
||||||
|
|
||||||
|
size_t ZBUFFv05_decompressContinue(ZBUFFv05_DCtx* dctx,
|
||||||
|
void* dst, size_t* dstCapacityPtr,
|
||||||
|
const void* src, size_t* srcSizePtr);
|
||||||
|
|
||||||
|
/*-***************************************************************************
|
||||||
|
* Streaming decompression
|
||||||
|
*
|
||||||
|
* A ZBUFFv05_DCtx object is required to track streaming operations.
|
||||||
|
* Use ZBUFFv05_createDCtx() and ZBUFFv05_freeDCtx() to create/release resources.
|
||||||
|
* Use ZBUFFv05_decompressInit() to start a new decompression operation,
|
||||||
|
* or ZBUFFv05_decompressInitDictionary() if decompression requires a dictionary.
|
||||||
|
* Note that ZBUFFv05_DCtx objects can be reused multiple times.
|
||||||
|
*
|
||||||
|
* Use ZBUFFv05_decompressContinue() repetitively to consume your input.
|
||||||
|
* *srcSizePtr and *dstCapacityPtr can be any size.
|
||||||
|
* The function will report how many bytes were read or written by modifying *srcSizePtr and *dstCapacityPtr.
|
||||||
|
* Note that it may not consume the entire input, in which case it's up to the caller to present remaining input again.
|
||||||
|
* The content of @dst will be overwritten (up to *dstCapacityPtr) at each function call, so save its content if it matters or change @dst.
|
||||||
|
* @return : a hint to preferred nb of bytes to use as input for next function call (it's only a hint, to help latency)
|
||||||
|
* or 0 when a frame is completely decoded
|
||||||
|
* or an error code, which can be tested using ZBUFFv05_isError().
|
||||||
|
*
|
||||||
|
* Hint : recommended buffer sizes (not compulsory) : ZBUFFv05_recommendedDInSize() / ZBUFFv05_recommendedDOutSize()
|
||||||
|
* output : ZBUFFv05_recommendedDOutSize==128 KB block size is the internal unit, it ensures it's always possible to write a full block when decoded.
|
||||||
|
* input : ZBUFFv05_recommendedDInSize==128Kb+3; just follow indications from ZBUFFv05_decompressContinue() to minimize latency. It should always be <= 128 KB + 3 .
|
||||||
|
* *******************************************************************************/
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Tool functions
|
||||||
|
***************************************/
|
||||||
|
unsigned ZBUFFv05_isError(size_t errorCode);
|
||||||
|
const char* ZBUFFv05_getErrorName(size_t errorCode);
|
||||||
|
|
||||||
|
/** Functions below provide recommended buffer sizes for Compression or Decompression operations.
|
||||||
|
* These sizes are just hints, and tend to offer better latency */
|
||||||
|
size_t ZBUFFv05_recommendedDInSize(void);
|
||||||
|
size_t ZBUFFv05_recommendedDOutSize(void);
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Constants
|
||||||
|
***************************************/
|
||||||
|
#define ZSTDv05_MAGICNUMBER 0xFD2FB525 /* v0.5 */
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTDv0505_H */
|
||||||
4150
vendor/github.com/DataDog/zstd/zstd_v06.c
generated
vendored
Normal file
4150
vendor/github.com/DataDog/zstd/zstd_v06.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
172
vendor/github.com/DataDog/zstd/zstd_v06.h
generated
vendored
Normal file
172
vendor/github.com/DataDog/zstd/zstd_v06.h
generated
vendored
Normal file
@@ -0,0 +1,172 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTDv06_H
|
||||||
|
#define ZSTDv06_H
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*====== Dependency ======*/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
|
||||||
|
|
||||||
|
/*====== Export for Windows ======*/
|
||||||
|
/*!
|
||||||
|
* ZSTDv06_DLL_EXPORT :
|
||||||
|
* Enable exporting of functions when building a Windows DLL
|
||||||
|
*/
|
||||||
|
#if defined(_WIN32) && defined(ZSTDv06_DLL_EXPORT) && (ZSTDv06_DLL_EXPORT==1)
|
||||||
|
# define ZSTDLIBv06_API __declspec(dllexport)
|
||||||
|
#else
|
||||||
|
# define ZSTDLIBv06_API
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Simple functions
|
||||||
|
***************************************/
|
||||||
|
/*! ZSTDv06_decompress() :
|
||||||
|
`compressedSize` : is the _exact_ size of the compressed blob, otherwise decompression will fail.
|
||||||
|
`dstCapacity` must be large enough, equal or larger than originalSize.
|
||||||
|
@return : the number of bytes decompressed into `dst` (<= `dstCapacity`),
|
||||||
|
or an errorCode if it fails (which can be tested using ZSTDv06_isError()) */
|
||||||
|
ZSTDLIBv06_API size_t ZSTDv06_decompress( void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t compressedSize);
|
||||||
|
|
||||||
|
/**
|
||||||
|
ZSTDv06_findFrameSizeInfoLegacy() : get the source length and decompressed bound of a ZSTD frame compliant with v0.6.x format
|
||||||
|
srcSize : The size of the 'src' buffer, at least as large as the frame pointed to by 'src'
|
||||||
|
cSize (output parameter) : the number of bytes that would be read to decompress this frame
|
||||||
|
or an error code if it fails (which can be tested using ZSTDv01_isError())
|
||||||
|
dBound (output parameter) : an upper-bound for the decompressed size of the data in the frame
|
||||||
|
or ZSTD_CONTENTSIZE_ERROR if an error occurs
|
||||||
|
|
||||||
|
note : assumes `cSize` and `dBound` are _not_ NULL.
|
||||||
|
*/
|
||||||
|
void ZSTDv06_findFrameSizeInfoLegacy(const void *src, size_t srcSize,
|
||||||
|
size_t* cSize, unsigned long long* dBound);
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Helper functions
|
||||||
|
***************************************/
|
||||||
|
ZSTDLIBv06_API size_t ZSTDv06_compressBound(size_t srcSize); /*!< maximum compressed size (worst case scenario) */
|
||||||
|
|
||||||
|
/* Error Management */
|
||||||
|
ZSTDLIBv06_API unsigned ZSTDv06_isError(size_t code); /*!< tells if a `size_t` function result is an error code */
|
||||||
|
ZSTDLIBv06_API const char* ZSTDv06_getErrorName(size_t code); /*!< provides readable string for an error code */
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Explicit memory management
|
||||||
|
***************************************/
|
||||||
|
/** Decompression context */
|
||||||
|
typedef struct ZSTDv06_DCtx_s ZSTDv06_DCtx;
|
||||||
|
ZSTDLIBv06_API ZSTDv06_DCtx* ZSTDv06_createDCtx(void);
|
||||||
|
ZSTDLIBv06_API size_t ZSTDv06_freeDCtx(ZSTDv06_DCtx* dctx); /*!< @return : errorCode */
|
||||||
|
|
||||||
|
/** ZSTDv06_decompressDCtx() :
|
||||||
|
* Same as ZSTDv06_decompress(), but requires an already allocated ZSTDv06_DCtx (see ZSTDv06_createDCtx()) */
|
||||||
|
ZSTDLIBv06_API size_t ZSTDv06_decompressDCtx(ZSTDv06_DCtx* ctx, void* dst, size_t dstCapacity, const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
|
||||||
|
/*-***********************
|
||||||
|
* Dictionary API
|
||||||
|
*************************/
|
||||||
|
/*! ZSTDv06_decompress_usingDict() :
|
||||||
|
* Decompression using a pre-defined Dictionary content (see dictBuilder).
|
||||||
|
* Dictionary must be identical to the one used during compression, otherwise regenerated data will be corrupted.
|
||||||
|
* Note : dict can be NULL, in which case, it's equivalent to ZSTDv06_decompressDCtx() */
|
||||||
|
ZSTDLIBv06_API size_t ZSTDv06_decompress_usingDict(ZSTDv06_DCtx* dctx,
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const void* dict,size_t dictSize);
|
||||||
|
|
||||||
|
|
||||||
|
/*-************************
|
||||||
|
* Advanced Streaming API
|
||||||
|
***************************/
|
||||||
|
struct ZSTDv06_frameParams_s { unsigned long long frameContentSize; unsigned windowLog; };
|
||||||
|
typedef struct ZSTDv06_frameParams_s ZSTDv06_frameParams;
|
||||||
|
|
||||||
|
ZSTDLIBv06_API size_t ZSTDv06_getFrameParams(ZSTDv06_frameParams* fparamsPtr, const void* src, size_t srcSize); /**< doesn't consume input */
|
||||||
|
ZSTDLIBv06_API size_t ZSTDv06_decompressBegin_usingDict(ZSTDv06_DCtx* dctx, const void* dict, size_t dictSize);
|
||||||
|
ZSTDLIBv06_API void ZSTDv06_copyDCtx(ZSTDv06_DCtx* dctx, const ZSTDv06_DCtx* preparedDCtx);
|
||||||
|
|
||||||
|
ZSTDLIBv06_API size_t ZSTDv06_nextSrcSizeToDecompress(ZSTDv06_DCtx* dctx);
|
||||||
|
ZSTDLIBv06_API size_t ZSTDv06_decompressContinue(ZSTDv06_DCtx* dctx, void* dst, size_t dstCapacity, const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* ZBUFF API
|
||||||
|
***************************************/
|
||||||
|
|
||||||
|
typedef struct ZBUFFv06_DCtx_s ZBUFFv06_DCtx;
|
||||||
|
ZSTDLIBv06_API ZBUFFv06_DCtx* ZBUFFv06_createDCtx(void);
|
||||||
|
ZSTDLIBv06_API size_t ZBUFFv06_freeDCtx(ZBUFFv06_DCtx* dctx);
|
||||||
|
|
||||||
|
ZSTDLIBv06_API size_t ZBUFFv06_decompressInit(ZBUFFv06_DCtx* dctx);
|
||||||
|
ZSTDLIBv06_API size_t ZBUFFv06_decompressInitDictionary(ZBUFFv06_DCtx* dctx, const void* dict, size_t dictSize);
|
||||||
|
|
||||||
|
ZSTDLIBv06_API size_t ZBUFFv06_decompressContinue(ZBUFFv06_DCtx* dctx,
|
||||||
|
void* dst, size_t* dstCapacityPtr,
|
||||||
|
const void* src, size_t* srcSizePtr);
|
||||||
|
|
||||||
|
/*-***************************************************************************
|
||||||
|
* Streaming decompression howto
|
||||||
|
*
|
||||||
|
* A ZBUFFv06_DCtx object is required to track streaming operations.
|
||||||
|
* Use ZBUFFv06_createDCtx() and ZBUFFv06_freeDCtx() to create/release resources.
|
||||||
|
* Use ZBUFFv06_decompressInit() to start a new decompression operation,
|
||||||
|
* or ZBUFFv06_decompressInitDictionary() if decompression requires a dictionary.
|
||||||
|
* Note that ZBUFFv06_DCtx objects can be re-init multiple times.
|
||||||
|
*
|
||||||
|
* Use ZBUFFv06_decompressContinue() repetitively to consume your input.
|
||||||
|
* *srcSizePtr and *dstCapacityPtr can be any size.
|
||||||
|
* The function will report how many bytes were read or written by modifying *srcSizePtr and *dstCapacityPtr.
|
||||||
|
* Note that it may not consume the entire input, in which case it's up to the caller to present remaining input again.
|
||||||
|
* The content of `dst` will be overwritten (up to *dstCapacityPtr) at each function call, so save its content if it matters, or change `dst`.
|
||||||
|
* @return : a hint to preferred nb of bytes to use as input for next function call (it's only a hint, to help latency),
|
||||||
|
* or 0 when a frame is completely decoded,
|
||||||
|
* or an error code, which can be tested using ZBUFFv06_isError().
|
||||||
|
*
|
||||||
|
* Hint : recommended buffer sizes (not compulsory) : ZBUFFv06_recommendedDInSize() and ZBUFFv06_recommendedDOutSize()
|
||||||
|
* output : ZBUFFv06_recommendedDOutSize== 128 KB block size is the internal unit, it ensures it's always possible to write a full block when decoded.
|
||||||
|
* input : ZBUFFv06_recommendedDInSize == 128KB + 3;
|
||||||
|
* just follow indications from ZBUFFv06_decompressContinue() to minimize latency. It should always be <= 128 KB + 3 .
|
||||||
|
* *******************************************************************************/
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Tool functions
|
||||||
|
***************************************/
|
||||||
|
ZSTDLIBv06_API unsigned ZBUFFv06_isError(size_t errorCode);
|
||||||
|
ZSTDLIBv06_API const char* ZBUFFv06_getErrorName(size_t errorCode);
|
||||||
|
|
||||||
|
/** Functions below provide recommended buffer sizes for Compression or Decompression operations.
|
||||||
|
* These sizes are just hints, they tend to offer better latency */
|
||||||
|
ZSTDLIBv06_API size_t ZBUFFv06_recommendedDInSize(void);
|
||||||
|
ZSTDLIBv06_API size_t ZBUFFv06_recommendedDOutSize(void);
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Constants
|
||||||
|
***************************************/
|
||||||
|
#define ZSTDv06_MAGICNUMBER 0xFD2FB526 /* v0.6 */
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTDv06_BUFFERED_H */
|
||||||
4533
vendor/github.com/DataDog/zstd/zstd_v07.c
generated
vendored
Normal file
4533
vendor/github.com/DataDog/zstd/zstd_v07.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
187
vendor/github.com/DataDog/zstd/zstd_v07.h
generated
vendored
Normal file
187
vendor/github.com/DataDog/zstd/zstd_v07.h
generated
vendored
Normal file
@@ -0,0 +1,187 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTDv07_H_235446
|
||||||
|
#define ZSTDv07_H_235446
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*====== Dependency ======*/
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
|
||||||
|
|
||||||
|
/*====== Export for Windows ======*/
|
||||||
|
/*!
|
||||||
|
* ZSTDv07_DLL_EXPORT :
|
||||||
|
* Enable exporting of functions when building a Windows DLL
|
||||||
|
*/
|
||||||
|
#if defined(_WIN32) && defined(ZSTDv07_DLL_EXPORT) && (ZSTDv07_DLL_EXPORT==1)
|
||||||
|
# define ZSTDLIBv07_API __declspec(dllexport)
|
||||||
|
#else
|
||||||
|
# define ZSTDLIBv07_API
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Simple API
|
||||||
|
***************************************/
|
||||||
|
/*! ZSTDv07_getDecompressedSize() :
|
||||||
|
* @return : decompressed size if known, 0 otherwise.
|
||||||
|
note 1 : if `0`, follow up with ZSTDv07_getFrameParams() to know precise failure cause.
|
||||||
|
note 2 : decompressed size could be wrong or intentionally modified !
|
||||||
|
always ensure results fit within application's authorized limits */
|
||||||
|
unsigned long long ZSTDv07_getDecompressedSize(const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
/*! ZSTDv07_decompress() :
|
||||||
|
`compressedSize` : must be _exact_ size of compressed input, otherwise decompression will fail.
|
||||||
|
`dstCapacity` must be equal or larger than originalSize.
|
||||||
|
@return : the number of bytes decompressed into `dst` (<= `dstCapacity`),
|
||||||
|
or an errorCode if it fails (which can be tested using ZSTDv07_isError()) */
|
||||||
|
ZSTDLIBv07_API size_t ZSTDv07_decompress( void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t compressedSize);
|
||||||
|
|
||||||
|
/**
|
||||||
|
ZSTDv07_findFrameSizeInfoLegacy() : get the source length and decompressed bound of a ZSTD frame compliant with v0.7.x format
|
||||||
|
srcSize : The size of the 'src' buffer, at least as large as the frame pointed to by 'src'
|
||||||
|
cSize (output parameter) : the number of bytes that would be read to decompress this frame
|
||||||
|
or an error code if it fails (which can be tested using ZSTDv01_isError())
|
||||||
|
dBound (output parameter) : an upper-bound for the decompressed size of the data in the frame
|
||||||
|
or ZSTD_CONTENTSIZE_ERROR if an error occurs
|
||||||
|
|
||||||
|
note : assumes `cSize` and `dBound` are _not_ NULL.
|
||||||
|
*/
|
||||||
|
void ZSTDv07_findFrameSizeInfoLegacy(const void *src, size_t srcSize,
|
||||||
|
size_t* cSize, unsigned long long* dBound);
|
||||||
|
|
||||||
|
/*====== Helper functions ======*/
|
||||||
|
ZSTDLIBv07_API unsigned ZSTDv07_isError(size_t code); /*!< tells if a `size_t` function result is an error code */
|
||||||
|
ZSTDLIBv07_API const char* ZSTDv07_getErrorName(size_t code); /*!< provides readable string from an error code */
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Explicit memory management
|
||||||
|
***************************************/
|
||||||
|
/** Decompression context */
|
||||||
|
typedef struct ZSTDv07_DCtx_s ZSTDv07_DCtx;
|
||||||
|
ZSTDLIBv07_API ZSTDv07_DCtx* ZSTDv07_createDCtx(void);
|
||||||
|
ZSTDLIBv07_API size_t ZSTDv07_freeDCtx(ZSTDv07_DCtx* dctx); /*!< @return : errorCode */
|
||||||
|
|
||||||
|
/** ZSTDv07_decompressDCtx() :
|
||||||
|
* Same as ZSTDv07_decompress(), requires an allocated ZSTDv07_DCtx (see ZSTDv07_createDCtx()) */
|
||||||
|
ZSTDLIBv07_API size_t ZSTDv07_decompressDCtx(ZSTDv07_DCtx* ctx, void* dst, size_t dstCapacity, const void* src, size_t srcSize);
|
||||||
|
|
||||||
|
|
||||||
|
/*-************************
|
||||||
|
* Simple dictionary API
|
||||||
|
***************************/
|
||||||
|
/*! ZSTDv07_decompress_usingDict() :
|
||||||
|
* Decompression using a pre-defined Dictionary content (see dictBuilder).
|
||||||
|
* Dictionary must be identical to the one used during compression.
|
||||||
|
* Note : This function load the dictionary, resulting in a significant startup time */
|
||||||
|
ZSTDLIBv07_API size_t ZSTDv07_decompress_usingDict(ZSTDv07_DCtx* dctx,
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const void* dict,size_t dictSize);
|
||||||
|
|
||||||
|
|
||||||
|
/*-**************************
|
||||||
|
* Advanced Dictionary API
|
||||||
|
****************************/
|
||||||
|
/*! ZSTDv07_createDDict() :
|
||||||
|
* Create a digested dictionary, ready to start decompression operation without startup delay.
|
||||||
|
* `dict` can be released after creation */
|
||||||
|
typedef struct ZSTDv07_DDict_s ZSTDv07_DDict;
|
||||||
|
ZSTDLIBv07_API ZSTDv07_DDict* ZSTDv07_createDDict(const void* dict, size_t dictSize);
|
||||||
|
ZSTDLIBv07_API size_t ZSTDv07_freeDDict(ZSTDv07_DDict* ddict);
|
||||||
|
|
||||||
|
/*! ZSTDv07_decompress_usingDDict() :
|
||||||
|
* Decompression using a pre-digested Dictionary
|
||||||
|
* Faster startup than ZSTDv07_decompress_usingDict(), recommended when same dictionary is used multiple times. */
|
||||||
|
ZSTDLIBv07_API size_t ZSTDv07_decompress_usingDDict(ZSTDv07_DCtx* dctx,
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const ZSTDv07_DDict* ddict);
|
||||||
|
|
||||||
|
typedef struct {
|
||||||
|
unsigned long long frameContentSize;
|
||||||
|
unsigned windowSize;
|
||||||
|
unsigned dictID;
|
||||||
|
unsigned checksumFlag;
|
||||||
|
} ZSTDv07_frameParams;
|
||||||
|
|
||||||
|
ZSTDLIBv07_API size_t ZSTDv07_getFrameParams(ZSTDv07_frameParams* fparamsPtr, const void* src, size_t srcSize); /**< doesn't consume input */
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Streaming functions
|
||||||
|
***************************************/
|
||||||
|
typedef struct ZBUFFv07_DCtx_s ZBUFFv07_DCtx;
|
||||||
|
ZSTDLIBv07_API ZBUFFv07_DCtx* ZBUFFv07_createDCtx(void);
|
||||||
|
ZSTDLIBv07_API size_t ZBUFFv07_freeDCtx(ZBUFFv07_DCtx* dctx);
|
||||||
|
|
||||||
|
ZSTDLIBv07_API size_t ZBUFFv07_decompressInit(ZBUFFv07_DCtx* dctx);
|
||||||
|
ZSTDLIBv07_API size_t ZBUFFv07_decompressInitDictionary(ZBUFFv07_DCtx* dctx, const void* dict, size_t dictSize);
|
||||||
|
|
||||||
|
ZSTDLIBv07_API size_t ZBUFFv07_decompressContinue(ZBUFFv07_DCtx* dctx,
|
||||||
|
void* dst, size_t* dstCapacityPtr,
|
||||||
|
const void* src, size_t* srcSizePtr);
|
||||||
|
|
||||||
|
/*-***************************************************************************
|
||||||
|
* Streaming decompression howto
|
||||||
|
*
|
||||||
|
* A ZBUFFv07_DCtx object is required to track streaming operations.
|
||||||
|
* Use ZBUFFv07_createDCtx() and ZBUFFv07_freeDCtx() to create/release resources.
|
||||||
|
* Use ZBUFFv07_decompressInit() to start a new decompression operation,
|
||||||
|
* or ZBUFFv07_decompressInitDictionary() if decompression requires a dictionary.
|
||||||
|
* Note that ZBUFFv07_DCtx objects can be re-init multiple times.
|
||||||
|
*
|
||||||
|
* Use ZBUFFv07_decompressContinue() repetitively to consume your input.
|
||||||
|
* *srcSizePtr and *dstCapacityPtr can be any size.
|
||||||
|
* The function will report how many bytes were read or written by modifying *srcSizePtr and *dstCapacityPtr.
|
||||||
|
* Note that it may not consume the entire input, in which case it's up to the caller to present remaining input again.
|
||||||
|
* The content of `dst` will be overwritten (up to *dstCapacityPtr) at each function call, so save its content if it matters, or change `dst`.
|
||||||
|
* @return : a hint to preferred nb of bytes to use as input for next function call (it's only a hint, to help latency),
|
||||||
|
* or 0 when a frame is completely decoded,
|
||||||
|
* or an error code, which can be tested using ZBUFFv07_isError().
|
||||||
|
*
|
||||||
|
* Hint : recommended buffer sizes (not compulsory) : ZBUFFv07_recommendedDInSize() and ZBUFFv07_recommendedDOutSize()
|
||||||
|
* output : ZBUFFv07_recommendedDOutSize== 128 KB block size is the internal unit, it ensures it's always possible to write a full block when decoded.
|
||||||
|
* input : ZBUFFv07_recommendedDInSize == 128KB + 3;
|
||||||
|
* just follow indications from ZBUFFv07_decompressContinue() to minimize latency. It should always be <= 128 KB + 3 .
|
||||||
|
* *******************************************************************************/
|
||||||
|
|
||||||
|
|
||||||
|
/* *************************************
|
||||||
|
* Tool functions
|
||||||
|
***************************************/
|
||||||
|
ZSTDLIBv07_API unsigned ZBUFFv07_isError(size_t errorCode);
|
||||||
|
ZSTDLIBv07_API const char* ZBUFFv07_getErrorName(size_t errorCode);
|
||||||
|
|
||||||
|
/** Functions below provide recommended buffer sizes for Compression or Decompression operations.
|
||||||
|
* These sizes are just hints, they tend to offer better latency */
|
||||||
|
ZSTDLIBv07_API size_t ZBUFFv07_recommendedDInSize(void);
|
||||||
|
ZSTDLIBv07_API size_t ZBUFFv07_recommendedDOutSize(void);
|
||||||
|
|
||||||
|
|
||||||
|
/*-*************************************
|
||||||
|
* Constants
|
||||||
|
***************************************/
|
||||||
|
#define ZSTDv07_MAGICNUMBER 0xFD2FB527 /* v0.7 */
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTDv07_H_235446 */
|
||||||
2116
vendor/github.com/DataDog/zstd/zstdmt_compress.c
generated
vendored
Normal file
2116
vendor/github.com/DataDog/zstd/zstdmt_compress.c
generated
vendored
Normal file
File diff suppressed because it is too large
Load Diff
192
vendor/github.com/DataDog/zstd/zstdmt_compress.h
generated
vendored
Normal file
192
vendor/github.com/DataDog/zstd/zstdmt_compress.h
generated
vendored
Normal file
@@ -0,0 +1,192 @@
|
|||||||
|
/*
|
||||||
|
* Copyright (c) 2016-present, Yann Collet, Facebook, Inc.
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* This source code is licensed under both the BSD-style license (found in the
|
||||||
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
||||||
|
* in the COPYING file in the root directory of this source tree).
|
||||||
|
* You may select, at your option, one of the above-listed licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef ZSTDMT_COMPRESS_H
|
||||||
|
#define ZSTDMT_COMPRESS_H
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
|
||||||
|
|
||||||
|
/* Note : This is an internal API.
|
||||||
|
* These APIs used to be exposed with ZSTDLIB_API,
|
||||||
|
* because it used to be the only way to invoke MT compression.
|
||||||
|
* Now, it's recommended to use ZSTD_compress2 and ZSTD_compressStream2()
|
||||||
|
* instead.
|
||||||
|
*
|
||||||
|
* If you depend on these APIs and can't switch, then define
|
||||||
|
* ZSTD_LEGACY_MULTITHREADED_API when making the dynamic library.
|
||||||
|
* However, we may completely remove these functions in a future
|
||||||
|
* release, so please switch soon.
|
||||||
|
*
|
||||||
|
* This API requires ZSTD_MULTITHREAD to be defined during compilation,
|
||||||
|
* otherwise ZSTDMT_createCCtx*() will fail.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifdef ZSTD_LEGACY_MULTITHREADED_API
|
||||||
|
# define ZSTDMT_API ZSTDLIB_API
|
||||||
|
#else
|
||||||
|
# define ZSTDMT_API
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/* === Dependencies === */
|
||||||
|
#include <stddef.h> /* size_t */
|
||||||
|
#define ZSTD_STATIC_LINKING_ONLY /* ZSTD_parameters */
|
||||||
|
#include "zstd.h" /* ZSTD_inBuffer, ZSTD_outBuffer, ZSTDLIB_API */
|
||||||
|
|
||||||
|
|
||||||
|
/* === Constants === */
|
||||||
|
#ifndef ZSTDMT_NBWORKERS_MAX
|
||||||
|
# define ZSTDMT_NBWORKERS_MAX 200
|
||||||
|
#endif
|
||||||
|
#ifndef ZSTDMT_JOBSIZE_MIN
|
||||||
|
# define ZSTDMT_JOBSIZE_MIN (1 MB)
|
||||||
|
#endif
|
||||||
|
#define ZSTDMT_JOBLOG_MAX (MEM_32bits() ? 29 : 30)
|
||||||
|
#define ZSTDMT_JOBSIZE_MAX (MEM_32bits() ? (512 MB) : (1024 MB))
|
||||||
|
|
||||||
|
|
||||||
|
/* === Memory management === */
|
||||||
|
typedef struct ZSTDMT_CCtx_s ZSTDMT_CCtx;
|
||||||
|
/* Requires ZSTD_MULTITHREAD to be defined during compilation, otherwise it will return NULL. */
|
||||||
|
ZSTDMT_API ZSTDMT_CCtx* ZSTDMT_createCCtx(unsigned nbWorkers);
|
||||||
|
/* Requires ZSTD_MULTITHREAD to be defined during compilation, otherwise it will return NULL. */
|
||||||
|
ZSTDMT_API ZSTDMT_CCtx* ZSTDMT_createCCtx_advanced(unsigned nbWorkers,
|
||||||
|
ZSTD_customMem cMem);
|
||||||
|
ZSTDMT_API size_t ZSTDMT_freeCCtx(ZSTDMT_CCtx* mtctx);
|
||||||
|
|
||||||
|
ZSTDMT_API size_t ZSTDMT_sizeof_CCtx(ZSTDMT_CCtx* mtctx);
|
||||||
|
|
||||||
|
|
||||||
|
/* === Simple one-pass compression function === */
|
||||||
|
|
||||||
|
ZSTDMT_API size_t ZSTDMT_compressCCtx(ZSTDMT_CCtx* mtctx,
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
int compressionLevel);
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
/* === Streaming functions === */
|
||||||
|
|
||||||
|
ZSTDMT_API size_t ZSTDMT_initCStream(ZSTDMT_CCtx* mtctx, int compressionLevel);
|
||||||
|
ZSTDMT_API size_t ZSTDMT_resetCStream(ZSTDMT_CCtx* mtctx, unsigned long long pledgedSrcSize); /**< if srcSize is not known at reset time, use ZSTD_CONTENTSIZE_UNKNOWN. Note: for compatibility with older programs, 0 means the same as ZSTD_CONTENTSIZE_UNKNOWN, but it will change in the future to mean "empty" */
|
||||||
|
|
||||||
|
ZSTDMT_API size_t ZSTDMT_nextInputSizeHint(const ZSTDMT_CCtx* mtctx);
|
||||||
|
ZSTDMT_API size_t ZSTDMT_compressStream(ZSTDMT_CCtx* mtctx, ZSTD_outBuffer* output, ZSTD_inBuffer* input);
|
||||||
|
|
||||||
|
ZSTDMT_API size_t ZSTDMT_flushStream(ZSTDMT_CCtx* mtctx, ZSTD_outBuffer* output); /**< @return : 0 == all flushed; >0 : still some data to be flushed; or an error code (ZSTD_isError()) */
|
||||||
|
ZSTDMT_API size_t ZSTDMT_endStream(ZSTDMT_CCtx* mtctx, ZSTD_outBuffer* output); /**< @return : 0 == all flushed; >0 : still some data to be flushed; or an error code (ZSTD_isError()) */
|
||||||
|
|
||||||
|
|
||||||
|
/* === Advanced functions and parameters === */
|
||||||
|
|
||||||
|
ZSTDMT_API size_t ZSTDMT_compress_advanced(ZSTDMT_CCtx* mtctx,
|
||||||
|
void* dst, size_t dstCapacity,
|
||||||
|
const void* src, size_t srcSize,
|
||||||
|
const ZSTD_CDict* cdict,
|
||||||
|
ZSTD_parameters params,
|
||||||
|
int overlapLog);
|
||||||
|
|
||||||
|
ZSTDMT_API size_t ZSTDMT_initCStream_advanced(ZSTDMT_CCtx* mtctx,
|
||||||
|
const void* dict, size_t dictSize, /* dict can be released after init, a local copy is preserved within zcs */
|
||||||
|
ZSTD_parameters params,
|
||||||
|
unsigned long long pledgedSrcSize); /* pledgedSrcSize is optional and can be zero == unknown */
|
||||||
|
|
||||||
|
ZSTDMT_API size_t ZSTDMT_initCStream_usingCDict(ZSTDMT_CCtx* mtctx,
|
||||||
|
const ZSTD_CDict* cdict,
|
||||||
|
ZSTD_frameParameters fparams,
|
||||||
|
unsigned long long pledgedSrcSize); /* note : zero means empty */
|
||||||
|
|
||||||
|
/* ZSTDMT_parameter :
|
||||||
|
* List of parameters that can be set using ZSTDMT_setMTCtxParameter() */
|
||||||
|
typedef enum {
|
||||||
|
ZSTDMT_p_jobSize, /* Each job is compressed in parallel. By default, this value is dynamically determined depending on compression parameters. Can be set explicitly here. */
|
||||||
|
ZSTDMT_p_overlapLog, /* Each job may reload a part of previous job to enhance compression ratio; 0 == no overlap, 6(default) == use 1/8th of window, >=9 == use full window. This is a "sticky" parameter : its value will be re-used on next compression job */
|
||||||
|
ZSTDMT_p_rsyncable /* Enables rsyncable mode. */
|
||||||
|
} ZSTDMT_parameter;
|
||||||
|
|
||||||
|
/* ZSTDMT_setMTCtxParameter() :
|
||||||
|
* allow setting individual parameters, one at a time, among a list of enums defined in ZSTDMT_parameter.
|
||||||
|
* The function must be called typically after ZSTD_createCCtx() but __before ZSTDMT_init*() !__
|
||||||
|
* Parameters not explicitly reset by ZSTDMT_init*() remain the same in consecutive compression sessions.
|
||||||
|
* @return : 0, or an error code (which can be tested using ZSTD_isError()) */
|
||||||
|
ZSTDMT_API size_t ZSTDMT_setMTCtxParameter(ZSTDMT_CCtx* mtctx, ZSTDMT_parameter parameter, int value);
|
||||||
|
|
||||||
|
/* ZSTDMT_getMTCtxParameter() :
|
||||||
|
* Query the ZSTDMT_CCtx for a parameter value.
|
||||||
|
* @return : 0, or an error code (which can be tested using ZSTD_isError()) */
|
||||||
|
ZSTDMT_API size_t ZSTDMT_getMTCtxParameter(ZSTDMT_CCtx* mtctx, ZSTDMT_parameter parameter, int* value);
|
||||||
|
|
||||||
|
|
||||||
|
/*! ZSTDMT_compressStream_generic() :
|
||||||
|
* Combines ZSTDMT_compressStream() with optional ZSTDMT_flushStream() or ZSTDMT_endStream()
|
||||||
|
* depending on flush directive.
|
||||||
|
* @return : minimum amount of data still to be flushed
|
||||||
|
* 0 if fully flushed
|
||||||
|
* or an error code
|
||||||
|
* note : needs to be init using any ZSTD_initCStream*() variant */
|
||||||
|
ZSTDMT_API size_t ZSTDMT_compressStream_generic(ZSTDMT_CCtx* mtctx,
|
||||||
|
ZSTD_outBuffer* output,
|
||||||
|
ZSTD_inBuffer* input,
|
||||||
|
ZSTD_EndDirective endOp);
|
||||||
|
|
||||||
|
|
||||||
|
/* ========================================================
|
||||||
|
* === Private interface, for use by ZSTD_compress.c ===
|
||||||
|
* === Not exposed in libzstd. Never invoke directly ===
|
||||||
|
* ======================================================== */
|
||||||
|
|
||||||
|
/*! ZSTDMT_toFlushNow()
|
||||||
|
* Tell how many bytes are ready to be flushed immediately.
|
||||||
|
* Probe the oldest active job (not yet entirely flushed) and check its output buffer.
|
||||||
|
* If return 0, it means there is no active job,
|
||||||
|
* or, it means oldest job is still active, but everything produced has been flushed so far,
|
||||||
|
* therefore flushing is limited by speed of oldest job. */
|
||||||
|
size_t ZSTDMT_toFlushNow(ZSTDMT_CCtx* mtctx);
|
||||||
|
|
||||||
|
/*! ZSTDMT_CCtxParam_setMTCtxParameter()
|
||||||
|
* like ZSTDMT_setMTCtxParameter(), but into a ZSTD_CCtx_Params */
|
||||||
|
size_t ZSTDMT_CCtxParam_setMTCtxParameter(ZSTD_CCtx_params* params, ZSTDMT_parameter parameter, int value);
|
||||||
|
|
||||||
|
/*! ZSTDMT_CCtxParam_setNbWorkers()
|
||||||
|
* Set nbWorkers, and clamp it.
|
||||||
|
* Also reset jobSize and overlapLog */
|
||||||
|
size_t ZSTDMT_CCtxParam_setNbWorkers(ZSTD_CCtx_params* params, unsigned nbWorkers);
|
||||||
|
|
||||||
|
/*! ZSTDMT_updateCParams_whileCompressing() :
|
||||||
|
* Updates only a selected set of compression parameters, to remain compatible with current frame.
|
||||||
|
* New parameters will be applied to next compression job. */
|
||||||
|
void ZSTDMT_updateCParams_whileCompressing(ZSTDMT_CCtx* mtctx, const ZSTD_CCtx_params* cctxParams);
|
||||||
|
|
||||||
|
/*! ZSTDMT_getFrameProgression():
|
||||||
|
* tells how much data has been consumed (input) and produced (output) for current frame.
|
||||||
|
* able to count progression inside worker threads.
|
||||||
|
*/
|
||||||
|
ZSTD_frameProgression ZSTDMT_getFrameProgression(ZSTDMT_CCtx* mtctx);
|
||||||
|
|
||||||
|
|
||||||
|
/*! ZSTDMT_initCStream_internal() :
|
||||||
|
* Private use only. Init streaming operation.
|
||||||
|
* expects params to be valid.
|
||||||
|
* must receive dict, or cdict, or none, but not both.
|
||||||
|
* @return : 0, or an error code */
|
||||||
|
size_t ZSTDMT_initCStream_internal(ZSTDMT_CCtx* zcs,
|
||||||
|
const void* dict, size_t dictSize, ZSTD_dictContentType_e dictContentType,
|
||||||
|
const ZSTD_CDict* cdict,
|
||||||
|
ZSTD_CCtx_params params, unsigned long long pledgedSrcSize);
|
||||||
|
|
||||||
|
|
||||||
|
#if defined (__cplusplus)
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#endif /* ZSTDMT_COMPRESS_H */
|
||||||
15
vendor/github.com/go-stack/stack/.travis.yml
generated
vendored
Normal file
15
vendor/github.com/go-stack/stack/.travis.yml
generated
vendored
Normal file
@@ -0,0 +1,15 @@
|
|||||||
|
language: go
|
||||||
|
sudo: false
|
||||||
|
go:
|
||||||
|
- 1.7.x
|
||||||
|
- 1.8.x
|
||||||
|
- 1.9.x
|
||||||
|
- 1.10.x
|
||||||
|
- 1.11.x
|
||||||
|
- tip
|
||||||
|
|
||||||
|
before_install:
|
||||||
|
- go get github.com/mattn/goveralls
|
||||||
|
|
||||||
|
script:
|
||||||
|
- goveralls -service=travis-ci
|
||||||
21
vendor/github.com/go-stack/stack/LICENSE.md
generated
vendored
Normal file
21
vendor/github.com/go-stack/stack/LICENSE.md
generated
vendored
Normal file
@@ -0,0 +1,21 @@
|
|||||||
|
The MIT License (MIT)
|
||||||
|
|
||||||
|
Copyright (c) 2014 Chris Hines
|
||||||
|
|
||||||
|
Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||||
|
of this software and associated documentation files (the "Software"), to deal
|
||||||
|
in the Software without restriction, including without limitation the rights
|
||||||
|
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
||||||
|
copies of the Software, and to permit persons to whom the Software is
|
||||||
|
furnished to do so, subject to the following conditions:
|
||||||
|
|
||||||
|
The above copyright notice and this permission notice shall be included in all
|
||||||
|
copies or substantial portions of the Software.
|
||||||
|
|
||||||
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||||
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||||
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
||||||
|
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||||
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
||||||
|
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
||||||
|
SOFTWARE.
|
||||||
38
vendor/github.com/go-stack/stack/README.md
generated
vendored
Normal file
38
vendor/github.com/go-stack/stack/README.md
generated
vendored
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
[](https://godoc.org/github.com/go-stack/stack)
|
||||||
|
[](https://goreportcard.com/report/go-stack/stack)
|
||||||
|
[](https://travis-ci.org/go-stack/stack)
|
||||||
|
[](https://coveralls.io/github/go-stack/stack?branch=master)
|
||||||
|
|
||||||
|
# stack
|
||||||
|
|
||||||
|
Package stack implements utilities to capture, manipulate, and format call
|
||||||
|
stacks. It provides a simpler API than package runtime.
|
||||||
|
|
||||||
|
The implementation takes care of the minutia and special cases of interpreting
|
||||||
|
the program counter (pc) values returned by runtime.Callers.
|
||||||
|
|
||||||
|
## Versioning
|
||||||
|
|
||||||
|
Package stack publishes releases via [semver](http://semver.org/) compatible Git
|
||||||
|
tags prefixed with a single 'v'. The master branch always contains the latest
|
||||||
|
release. The develop branch contains unreleased commits.
|
||||||
|
|
||||||
|
## Formatting
|
||||||
|
|
||||||
|
Package stack's types implement fmt.Formatter, which provides a simple and
|
||||||
|
flexible way to declaratively configure formatting when used with logging or
|
||||||
|
error tracking packages.
|
||||||
|
|
||||||
|
```go
|
||||||
|
func DoTheThing() {
|
||||||
|
c := stack.Caller(0)
|
||||||
|
log.Print(c) // "source.go:10"
|
||||||
|
log.Printf("%+v", c) // "pkg/path/source.go:10"
|
||||||
|
log.Printf("%n", c) // "DoTheThing"
|
||||||
|
|
||||||
|
s := stack.Trace().TrimRuntime()
|
||||||
|
log.Print(s) // "[source.go:15 caller.go:42 main.go:14]"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
See the docs for all of the supported formatting options.
|
||||||
1
vendor/github.com/go-stack/stack/go.mod
generated
vendored
Normal file
1
vendor/github.com/go-stack/stack/go.mod
generated
vendored
Normal file
@@ -0,0 +1 @@
|
|||||||
|
module github.com/go-stack/stack
|
||||||
400
vendor/github.com/go-stack/stack/stack.go
generated
vendored
Normal file
400
vendor/github.com/go-stack/stack/stack.go
generated
vendored
Normal file
@@ -0,0 +1,400 @@
|
|||||||
|
// +build go1.7
|
||||||
|
|
||||||
|
// Package stack implements utilities to capture, manipulate, and format call
|
||||||
|
// stacks. It provides a simpler API than package runtime.
|
||||||
|
//
|
||||||
|
// The implementation takes care of the minutia and special cases of
|
||||||
|
// interpreting the program counter (pc) values returned by runtime.Callers.
|
||||||
|
//
|
||||||
|
// Package stack's types implement fmt.Formatter, which provides a simple and
|
||||||
|
// flexible way to declaratively configure formatting when used with logging
|
||||||
|
// or error tracking packages.
|
||||||
|
package stack
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"runtime"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Call records a single function invocation from a goroutine stack.
|
||||||
|
type Call struct {
|
||||||
|
frame runtime.Frame
|
||||||
|
}
|
||||||
|
|
||||||
|
// Caller returns a Call from the stack of the current goroutine. The argument
|
||||||
|
// skip is the number of stack frames to ascend, with 0 identifying the
|
||||||
|
// calling function.
|
||||||
|
func Caller(skip int) Call {
|
||||||
|
// As of Go 1.9 we need room for up to three PC entries.
|
||||||
|
//
|
||||||
|
// 0. An entry for the stack frame prior to the target to check for
|
||||||
|
// special handling needed if that prior entry is runtime.sigpanic.
|
||||||
|
// 1. A possible second entry to hold metadata about skipped inlined
|
||||||
|
// functions. If inline functions were not skipped the target frame
|
||||||
|
// PC will be here.
|
||||||
|
// 2. A third entry for the target frame PC when the second entry
|
||||||
|
// is used for skipped inline functions.
|
||||||
|
var pcs [3]uintptr
|
||||||
|
n := runtime.Callers(skip+1, pcs[:])
|
||||||
|
frames := runtime.CallersFrames(pcs[:n])
|
||||||
|
frame, _ := frames.Next()
|
||||||
|
frame, _ = frames.Next()
|
||||||
|
|
||||||
|
return Call{
|
||||||
|
frame: frame,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// String implements fmt.Stinger. It is equivalent to fmt.Sprintf("%v", c).
|
||||||
|
func (c Call) String() string {
|
||||||
|
return fmt.Sprint(c)
|
||||||
|
}
|
||||||
|
|
||||||
|
// MarshalText implements encoding.TextMarshaler. It formats the Call the same
|
||||||
|
// as fmt.Sprintf("%v", c).
|
||||||
|
func (c Call) MarshalText() ([]byte, error) {
|
||||||
|
if c.frame == (runtime.Frame{}) {
|
||||||
|
return nil, ErrNoFunc
|
||||||
|
}
|
||||||
|
|
||||||
|
buf := bytes.Buffer{}
|
||||||
|
fmt.Fprint(&buf, c)
|
||||||
|
return buf.Bytes(), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// ErrNoFunc means that the Call has a nil *runtime.Func. The most likely
|
||||||
|
// cause is a Call with the zero value.
|
||||||
|
var ErrNoFunc = errors.New("no call stack information")
|
||||||
|
|
||||||
|
// Format implements fmt.Formatter with support for the following verbs.
|
||||||
|
//
|
||||||
|
// %s source file
|
||||||
|
// %d line number
|
||||||
|
// %n function name
|
||||||
|
// %k last segment of the package path
|
||||||
|
// %v equivalent to %s:%d
|
||||||
|
//
|
||||||
|
// It accepts the '+' and '#' flags for most of the verbs as follows.
|
||||||
|
//
|
||||||
|
// %+s path of source file relative to the compile time GOPATH,
|
||||||
|
// or the module path joined to the path of source file relative
|
||||||
|
// to module root
|
||||||
|
// %#s full path of source file
|
||||||
|
// %+n import path qualified function name
|
||||||
|
// %+k full package path
|
||||||
|
// %+v equivalent to %+s:%d
|
||||||
|
// %#v equivalent to %#s:%d
|
||||||
|
func (c Call) Format(s fmt.State, verb rune) {
|
||||||
|
if c.frame == (runtime.Frame{}) {
|
||||||
|
fmt.Fprintf(s, "%%!%c(NOFUNC)", verb)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
switch verb {
|
||||||
|
case 's', 'v':
|
||||||
|
file := c.frame.File
|
||||||
|
switch {
|
||||||
|
case s.Flag('#'):
|
||||||
|
// done
|
||||||
|
case s.Flag('+'):
|
||||||
|
file = pkgFilePath(&c.frame)
|
||||||
|
default:
|
||||||
|
const sep = "/"
|
||||||
|
if i := strings.LastIndex(file, sep); i != -1 {
|
||||||
|
file = file[i+len(sep):]
|
||||||
|
}
|
||||||
|
}
|
||||||
|
io.WriteString(s, file)
|
||||||
|
if verb == 'v' {
|
||||||
|
buf := [7]byte{':'}
|
||||||
|
s.Write(strconv.AppendInt(buf[:1], int64(c.frame.Line), 10))
|
||||||
|
}
|
||||||
|
|
||||||
|
case 'd':
|
||||||
|
buf := [6]byte{}
|
||||||
|
s.Write(strconv.AppendInt(buf[:0], int64(c.frame.Line), 10))
|
||||||
|
|
||||||
|
case 'k':
|
||||||
|
name := c.frame.Function
|
||||||
|
const pathSep = "/"
|
||||||
|
start, end := 0, len(name)
|
||||||
|
if i := strings.LastIndex(name, pathSep); i != -1 {
|
||||||
|
start = i + len(pathSep)
|
||||||
|
}
|
||||||
|
const pkgSep = "."
|
||||||
|
if i := strings.Index(name[start:], pkgSep); i != -1 {
|
||||||
|
end = start + i
|
||||||
|
}
|
||||||
|
if s.Flag('+') {
|
||||||
|
start = 0
|
||||||
|
}
|
||||||
|
io.WriteString(s, name[start:end])
|
||||||
|
|
||||||
|
case 'n':
|
||||||
|
name := c.frame.Function
|
||||||
|
if !s.Flag('+') {
|
||||||
|
const pathSep = "/"
|
||||||
|
if i := strings.LastIndex(name, pathSep); i != -1 {
|
||||||
|
name = name[i+len(pathSep):]
|
||||||
|
}
|
||||||
|
const pkgSep = "."
|
||||||
|
if i := strings.Index(name, pkgSep); i != -1 {
|
||||||
|
name = name[i+len(pkgSep):]
|
||||||
|
}
|
||||||
|
}
|
||||||
|
io.WriteString(s, name)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Frame returns the call frame infomation for the Call.
|
||||||
|
func (c Call) Frame() runtime.Frame {
|
||||||
|
return c.frame
|
||||||
|
}
|
||||||
|
|
||||||
|
// PC returns the program counter for this call frame; multiple frames may
|
||||||
|
// have the same PC value.
|
||||||
|
//
|
||||||
|
// Deprecated: Use Call.Frame instead.
|
||||||
|
func (c Call) PC() uintptr {
|
||||||
|
return c.frame.PC
|
||||||
|
}
|
||||||
|
|
||||||
|
// CallStack records a sequence of function invocations from a goroutine
|
||||||
|
// stack.
|
||||||
|
type CallStack []Call
|
||||||
|
|
||||||
|
// String implements fmt.Stinger. It is equivalent to fmt.Sprintf("%v", cs).
|
||||||
|
func (cs CallStack) String() string {
|
||||||
|
return fmt.Sprint(cs)
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
openBracketBytes = []byte("[")
|
||||||
|
closeBracketBytes = []byte("]")
|
||||||
|
spaceBytes = []byte(" ")
|
||||||
|
)
|
||||||
|
|
||||||
|
// MarshalText implements encoding.TextMarshaler. It formats the CallStack the
|
||||||
|
// same as fmt.Sprintf("%v", cs).
|
||||||
|
func (cs CallStack) MarshalText() ([]byte, error) {
|
||||||
|
buf := bytes.Buffer{}
|
||||||
|
buf.Write(openBracketBytes)
|
||||||
|
for i, pc := range cs {
|
||||||
|
if i > 0 {
|
||||||
|
buf.Write(spaceBytes)
|
||||||
|
}
|
||||||
|
fmt.Fprint(&buf, pc)
|
||||||
|
}
|
||||||
|
buf.Write(closeBracketBytes)
|
||||||
|
return buf.Bytes(), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Format implements fmt.Formatter by printing the CallStack as square brackets
|
||||||
|
// ([, ]) surrounding a space separated list of Calls each formatted with the
|
||||||
|
// supplied verb and options.
|
||||||
|
func (cs CallStack) Format(s fmt.State, verb rune) {
|
||||||
|
s.Write(openBracketBytes)
|
||||||
|
for i, pc := range cs {
|
||||||
|
if i > 0 {
|
||||||
|
s.Write(spaceBytes)
|
||||||
|
}
|
||||||
|
pc.Format(s, verb)
|
||||||
|
}
|
||||||
|
s.Write(closeBracketBytes)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Trace returns a CallStack for the current goroutine with element 0
|
||||||
|
// identifying the calling function.
|
||||||
|
func Trace() CallStack {
|
||||||
|
var pcs [512]uintptr
|
||||||
|
n := runtime.Callers(1, pcs[:])
|
||||||
|
|
||||||
|
frames := runtime.CallersFrames(pcs[:n])
|
||||||
|
cs := make(CallStack, 0, n)
|
||||||
|
|
||||||
|
// Skip extra frame retrieved just to make sure the runtime.sigpanic
|
||||||
|
// special case is handled.
|
||||||
|
frame, more := frames.Next()
|
||||||
|
|
||||||
|
for more {
|
||||||
|
frame, more = frames.Next()
|
||||||
|
cs = append(cs, Call{frame: frame})
|
||||||
|
}
|
||||||
|
|
||||||
|
return cs
|
||||||
|
}
|
||||||
|
|
||||||
|
// TrimBelow returns a slice of the CallStack with all entries below c
|
||||||
|
// removed.
|
||||||
|
func (cs CallStack) TrimBelow(c Call) CallStack {
|
||||||
|
for len(cs) > 0 && cs[0] != c {
|
||||||
|
cs = cs[1:]
|
||||||
|
}
|
||||||
|
return cs
|
||||||
|
}
|
||||||
|
|
||||||
|
// TrimAbove returns a slice of the CallStack with all entries above c
|
||||||
|
// removed.
|
||||||
|
func (cs CallStack) TrimAbove(c Call) CallStack {
|
||||||
|
for len(cs) > 0 && cs[len(cs)-1] != c {
|
||||||
|
cs = cs[:len(cs)-1]
|
||||||
|
}
|
||||||
|
return cs
|
||||||
|
}
|
||||||
|
|
||||||
|
// pkgIndex returns the index that results in file[index:] being the path of
|
||||||
|
// file relative to the compile time GOPATH, and file[:index] being the
|
||||||
|
// $GOPATH/src/ portion of file. funcName must be the name of a function in
|
||||||
|
// file as returned by runtime.Func.Name.
|
||||||
|
func pkgIndex(file, funcName string) int {
|
||||||
|
// As of Go 1.6.2 there is no direct way to know the compile time GOPATH
|
||||||
|
// at runtime, but we can infer the number of path segments in the GOPATH.
|
||||||
|
// We note that runtime.Func.Name() returns the function name qualified by
|
||||||
|
// the import path, which does not include the GOPATH. Thus we can trim
|
||||||
|
// segments from the beginning of the file path until the number of path
|
||||||
|
// separators remaining is one more than the number of path separators in
|
||||||
|
// the function name. For example, given:
|
||||||
|
//
|
||||||
|
// GOPATH /home/user
|
||||||
|
// file /home/user/src/pkg/sub/file.go
|
||||||
|
// fn.Name() pkg/sub.Type.Method
|
||||||
|
//
|
||||||
|
// We want to produce:
|
||||||
|
//
|
||||||
|
// file[:idx] == /home/user/src/
|
||||||
|
// file[idx:] == pkg/sub/file.go
|
||||||
|
//
|
||||||
|
// From this we can easily see that fn.Name() has one less path separator
|
||||||
|
// than our desired result for file[idx:]. We count separators from the
|
||||||
|
// end of the file path until it finds two more than in the function name
|
||||||
|
// and then move one character forward to preserve the initial path
|
||||||
|
// segment without a leading separator.
|
||||||
|
const sep = "/"
|
||||||
|
i := len(file)
|
||||||
|
for n := strings.Count(funcName, sep) + 2; n > 0; n-- {
|
||||||
|
i = strings.LastIndex(file[:i], sep)
|
||||||
|
if i == -1 {
|
||||||
|
i = -len(sep)
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// get back to 0 or trim the leading separator
|
||||||
|
return i + len(sep)
|
||||||
|
}
|
||||||
|
|
||||||
|
// pkgFilePath returns the frame's filepath relative to the compile-time GOPATH,
|
||||||
|
// or its module path joined to its path relative to the module root.
|
||||||
|
//
|
||||||
|
// As of Go 1.11 there is no direct way to know the compile time GOPATH or
|
||||||
|
// module paths at runtime, but we can piece together the desired information
|
||||||
|
// from available information. We note that runtime.Frame.Function contains the
|
||||||
|
// function name qualified by the package path, which includes the module path
|
||||||
|
// but not the GOPATH. We can extract the package path from that and append the
|
||||||
|
// last segments of the file path to arrive at the desired package qualified
|
||||||
|
// file path. For example, given:
|
||||||
|
//
|
||||||
|
// GOPATH /home/user
|
||||||
|
// import path pkg/sub
|
||||||
|
// frame.File /home/user/src/pkg/sub/file.go
|
||||||
|
// frame.Function pkg/sub.Type.Method
|
||||||
|
// Desired return pkg/sub/file.go
|
||||||
|
//
|
||||||
|
// It appears that we simply need to trim ".Type.Method" from frame.Function and
|
||||||
|
// append "/" + path.Base(file).
|
||||||
|
//
|
||||||
|
// But there are other wrinkles. Although it is idiomatic to do so, the internal
|
||||||
|
// name of a package is not required to match the last segment of its import
|
||||||
|
// path. In addition, the introduction of modules in Go 1.11 allows working
|
||||||
|
// without a GOPATH. So we also must make these work right:
|
||||||
|
//
|
||||||
|
// GOPATH /home/user
|
||||||
|
// import path pkg/go-sub
|
||||||
|
// package name sub
|
||||||
|
// frame.File /home/user/src/pkg/go-sub/file.go
|
||||||
|
// frame.Function pkg/sub.Type.Method
|
||||||
|
// Desired return pkg/go-sub/file.go
|
||||||
|
//
|
||||||
|
// Module path pkg/v2
|
||||||
|
// import path pkg/v2/go-sub
|
||||||
|
// package name sub
|
||||||
|
// frame.File /home/user/cloned-pkg/go-sub/file.go
|
||||||
|
// frame.Function pkg/v2/sub.Type.Method
|
||||||
|
// Desired return pkg/v2/go-sub/file.go
|
||||||
|
//
|
||||||
|
// We can handle all of these situations by using the package path extracted
|
||||||
|
// from frame.Function up to, but not including, the last segment as the prefix
|
||||||
|
// and the last two segments of frame.File as the suffix of the returned path.
|
||||||
|
// This preserves the existing behavior when working in a GOPATH without modules
|
||||||
|
// and a semantically equivalent behavior when used in module aware project.
|
||||||
|
func pkgFilePath(frame *runtime.Frame) string {
|
||||||
|
pre := pkgPrefix(frame.Function)
|
||||||
|
post := pathSuffix(frame.File)
|
||||||
|
if pre == "" {
|
||||||
|
return post
|
||||||
|
}
|
||||||
|
return pre + "/" + post
|
||||||
|
}
|
||||||
|
|
||||||
|
// pkgPrefix returns the import path of the function's package with the final
|
||||||
|
// segment removed.
|
||||||
|
func pkgPrefix(funcName string) string {
|
||||||
|
const pathSep = "/"
|
||||||
|
end := strings.LastIndex(funcName, pathSep)
|
||||||
|
if end == -1 {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
return funcName[:end]
|
||||||
|
}
|
||||||
|
|
||||||
|
// pathSuffix returns the last two segments of path.
|
||||||
|
func pathSuffix(path string) string {
|
||||||
|
const pathSep = "/"
|
||||||
|
lastSep := strings.LastIndex(path, pathSep)
|
||||||
|
if lastSep == -1 {
|
||||||
|
return path
|
||||||
|
}
|
||||||
|
return path[strings.LastIndex(path[:lastSep], pathSep)+1:]
|
||||||
|
}
|
||||||
|
|
||||||
|
var runtimePath string
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
var pcs [3]uintptr
|
||||||
|
runtime.Callers(0, pcs[:])
|
||||||
|
frames := runtime.CallersFrames(pcs[:])
|
||||||
|
frame, _ := frames.Next()
|
||||||
|
file := frame.File
|
||||||
|
|
||||||
|
idx := pkgIndex(frame.File, frame.Function)
|
||||||
|
|
||||||
|
runtimePath = file[:idx]
|
||||||
|
if runtime.GOOS == "windows" {
|
||||||
|
runtimePath = strings.ToLower(runtimePath)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func inGoroot(c Call) bool {
|
||||||
|
file := c.frame.File
|
||||||
|
if len(file) == 0 || file[0] == '?' {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
if runtime.GOOS == "windows" {
|
||||||
|
file = strings.ToLower(file)
|
||||||
|
}
|
||||||
|
return strings.HasPrefix(file, runtimePath) || strings.HasSuffix(file, "/_testmain.go")
|
||||||
|
}
|
||||||
|
|
||||||
|
// TrimRuntime returns a slice of the CallStack with the topmost entries from
|
||||||
|
// the go runtime removed. It considers any calls originating from unknown
|
||||||
|
// files, files under GOROOT, or _testmain.go as part of the runtime.
|
||||||
|
func (cs CallStack) TrimRuntime() CallStack {
|
||||||
|
for len(cs) > 0 && inGoroot(cs[len(cs)-1]) {
|
||||||
|
cs = cs[:len(cs)-1]
|
||||||
|
}
|
||||||
|
return cs
|
||||||
|
}
|
||||||
0
vendor/github.com/xdg/scram/.gitignore
generated
vendored
Normal file
0
vendor/github.com/xdg/scram/.gitignore
generated
vendored
Normal file
11
vendor/github.com/xdg/scram/.travis.yml
generated
vendored
Normal file
11
vendor/github.com/xdg/scram/.travis.yml
generated
vendored
Normal file
@@ -0,0 +1,11 @@
|
|||||||
|
language: go
|
||||||
|
sudo: false
|
||||||
|
go:
|
||||||
|
- "1.7"
|
||||||
|
- "1.8"
|
||||||
|
- "1.9"
|
||||||
|
- "1.10"
|
||||||
|
- master
|
||||||
|
matrix:
|
||||||
|
allow_failures:
|
||||||
|
- go: master
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user