mirror of
https://github.com/go-gitea/gitea
synced 2024-12-21 17:17:56 +01:00
chore: update github.com/couchbase/gomemcached and github.com/couchbase/go-couchbase (#9419)
This commit is contained in:
parent
8873a80276
commit
559fb6ccf0
2
go.mod
2
go.mod
@ -24,6 +24,8 @@ require (
|
||||
github.com/blevesearch/go-porterstemmer v1.0.2 // indirect
|
||||
github.com/blevesearch/segment v0.0.0-20160915185041-762005e7a34f // indirect
|
||||
github.com/boombuler/barcode v0.0.0-20161226211916-fe0f26ff6d26 // indirect
|
||||
github.com/couchbase/gomemcached v0.0.0-20191004160342-7b5da2ec40b2 // indirect
|
||||
github.com/couchbase/goutils v0.0.0-20191018232750-b49639060d85 // indirect
|
||||
github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd // indirect
|
||||
github.com/cznic/b v0.0.0-20181122101859-a26611c4d92d // indirect
|
||||
github.com/cznic/mathutil v0.0.0-20181122101859-297441e03548 // indirect
|
||||
|
4
go.sum
4
go.sum
@ -97,8 +97,12 @@ github.com/coreos/go-systemd v0.0.0-20190321100706-95778dfbb74e/go.mod h1:F5haX7
|
||||
github.com/coreos/pkg v0.0.0-20180928190104-399ea9e2e55f/go.mod h1:E3G3o1h8I7cfcXa63jLwjI0eiQQMgzzUDFVpN/nH/eA=
|
||||
github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d h1:XMf4E1U+b9E3ElF0mjvfXZdflBRZz4gLp16nQ/QSHQM=
|
||||
github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d/go.mod h1:srVSlQLB8iXBVXHgnqemxUXqN6FCvClgCMPCsjBDR7c=
|
||||
github.com/couchbase/gomemcached v0.0.0-20191004160342-7b5da2ec40b2 h1:vZryARwW4PSFXd9arwegEywvMTvPuXL3/oa+4L5NTe8=
|
||||
github.com/couchbase/gomemcached v0.0.0-20191004160342-7b5da2ec40b2/go.mod h1:srVSlQLB8iXBVXHgnqemxUXqN6FCvClgCMPCsjBDR7c=
|
||||
github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b h1:bZ9rKU2/V8sY+NulSfxDOnXTWcs1rySqdF1sVepihvo=
|
||||
github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b/go.mod h1:BQwMFlJzDjFDG3DJUdU0KORxn88UlsOULuxLExMh3Hs=
|
||||
github.com/couchbase/goutils v0.0.0-20191018232750-b49639060d85 h1:0WMIDtuXCKEm4wtAJgAAXa/qtM5O9MariLwgHaRlYmk=
|
||||
github.com/couchbase/goutils v0.0.0-20191018232750-b49639060d85/go.mod h1:BQwMFlJzDjFDG3DJUdU0KORxn88UlsOULuxLExMh3Hs=
|
||||
github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd h1:zeuJhcG3f8eePshH3KxkNE+Xtl53pVln9MOUPMyr/1w=
|
||||
github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd/go.mod h1:xbc8Ff/oG7h2ejd7AlwOpfd+6QZntc92ygpAOfGwcKY=
|
||||
github.com/couchbaselabs/go-couchbase v0.0.0-20190708161019-23e7ca2ce2b7 h1:1XjEY/gnjQ+AfXef2U6dxCquhiRzkEpxZuWqs+QxTL8=
|
||||
|
123
vendor/github.com/couchbase/gomemcached/client/collections_filter.go
generated
vendored
Normal file
123
vendor/github.com/couchbase/gomemcached/client/collections_filter.go
generated
vendored
Normal file
@ -0,0 +1,123 @@
|
||||
package memcached
|
||||
|
||||
import (
|
||||
"encoding/json"
|
||||
"fmt"
|
||||
)
|
||||
|
||||
// Collection based filter
|
||||
type CollectionsFilter struct {
|
||||
ManifestUid uint64
|
||||
UseManifestUid bool
|
||||
StreamId uint16
|
||||
UseStreamId bool
|
||||
|
||||
// Use either ScopeId OR CollectionsList, not both
|
||||
CollectionsList []uint32
|
||||
ScopeId uint32
|
||||
}
|
||||
|
||||
type nonStreamIdNonResumeScopeMeta struct {
|
||||
ScopeId string `json:"scope"`
|
||||
}
|
||||
|
||||
type nonStreamIdResumeScopeMeta struct {
|
||||
ManifestId string `json:"uid"`
|
||||
}
|
||||
|
||||
type nonStreamIdNonResumeCollectionsMeta struct {
|
||||
CollectionsList []string `json:"collections"`
|
||||
}
|
||||
|
||||
type nonStreamIdResumeCollectionsMeta struct {
|
||||
ManifestId string `json:"uid"`
|
||||
CollectionsList []string `json:"collections"`
|
||||
}
|
||||
|
||||
type streamIdNonResumeCollectionsMeta struct {
|
||||
CollectionsList []string `json:"collections"`
|
||||
StreamId uint16 `json:"sid"`
|
||||
}
|
||||
|
||||
type streamIdNonResumeScopeMeta struct {
|
||||
ScopeId string `json:"scope"`
|
||||
StreamId uint16 `json:"sid"`
|
||||
}
|
||||
|
||||
func (c *CollectionsFilter) IsValid() error {
|
||||
if c.UseManifestUid {
|
||||
return fmt.Errorf("Not implemented yet")
|
||||
}
|
||||
|
||||
if len(c.CollectionsList) > 0 && c.ScopeId > 0 {
|
||||
return fmt.Errorf("Collection list is specified but scope ID is also specified")
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func (c *CollectionsFilter) outputCollectionsFilterColList() (outputList []string) {
|
||||
for _, collectionUint := range c.CollectionsList {
|
||||
outputList = append(outputList, fmt.Sprintf("%x", collectionUint))
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func (c *CollectionsFilter) outputScopeId() string {
|
||||
return fmt.Sprintf("%x", c.ScopeId)
|
||||
}
|
||||
|
||||
func (c *CollectionsFilter) ToStreamReqBody() ([]byte, error) {
|
||||
if err := c.IsValid(); err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
var output interface{}
|
||||
|
||||
switch c.UseStreamId {
|
||||
case true:
|
||||
switch c.UseManifestUid {
|
||||
case true:
|
||||
// TODO
|
||||
return nil, fmt.Errorf("NotImplemented0")
|
||||
case false:
|
||||
switch len(c.CollectionsList) > 0 {
|
||||
case true:
|
||||
filter := &streamIdNonResumeCollectionsMeta{
|
||||
StreamId: c.StreamId,
|
||||
CollectionsList: c.outputCollectionsFilterColList(),
|
||||
}
|
||||
output = *filter
|
||||
case false:
|
||||
filter := &streamIdNonResumeScopeMeta{
|
||||
StreamId: c.StreamId,
|
||||
ScopeId: c.outputScopeId(),
|
||||
}
|
||||
output = *filter
|
||||
}
|
||||
}
|
||||
case false:
|
||||
switch c.UseManifestUid {
|
||||
case true:
|
||||
// TODO
|
||||
return nil, fmt.Errorf("NotImplemented1")
|
||||
case false:
|
||||
switch len(c.CollectionsList) > 0 {
|
||||
case true:
|
||||
filter := &nonStreamIdNonResumeCollectionsMeta{
|
||||
CollectionsList: c.outputCollectionsFilterColList(),
|
||||
}
|
||||
output = *filter
|
||||
case false:
|
||||
output = nonStreamIdNonResumeScopeMeta{ScopeId: c.outputScopeId()}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
data, err := json.Marshal(output)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
} else {
|
||||
return data, nil
|
||||
}
|
||||
}
|
30
vendor/github.com/couchbase/gomemcached/client/mc.go
generated
vendored
30
vendor/github.com/couchbase/gomemcached/client/mc.go
generated
vendored
@ -28,10 +28,12 @@ type ClientIface interface {
|
||||
CASNext(vb uint16, k string, exp int, state *CASState) bool
|
||||
CAS(vb uint16, k string, f CasFunc, initexp int) (*gomemcached.MCResponse, error)
|
||||
CollectionsGetCID(scope string, collection string) (*gomemcached.MCResponse, error)
|
||||
CollectionEnabled() bool
|
||||
Close() error
|
||||
Decr(vb uint16, key string, amt, def uint64, exp int) (uint64, error)
|
||||
Del(vb uint16, key string) (*gomemcached.MCResponse, error)
|
||||
EnableMutationToken() (*gomemcached.MCResponse, error)
|
||||
EnableFeatures(features Features) (*gomemcached.MCResponse, error)
|
||||
Get(vb uint16, key string) (*gomemcached.MCResponse, error)
|
||||
GetCollectionsManifest() (*gomemcached.MCResponse, error)
|
||||
GetFromCollection(vb uint16, cid uint32, key string) (*gomemcached.MCResponse, error)
|
||||
@ -76,9 +78,12 @@ var Healthy uint32 = 1
|
||||
type Features []Feature
|
||||
type Feature uint16
|
||||
|
||||
const FeatureMutationToken = Feature(0x04)
|
||||
const FeatureTcpNoDelay = Feature(0x03)
|
||||
const FeatureMutationToken = Feature(0x04) // XATTR bit in data type field with dcp mutations
|
||||
const FeatureXattr = Feature(0x06)
|
||||
const FeatureXerror = Feature(0x07)
|
||||
const FeatureCollections = Feature(0x12)
|
||||
const FeatureSnappyCompression = Feature(0x0a)
|
||||
const FeatureDataType = Feature(0x0b)
|
||||
|
||||
type memcachedConnection interface {
|
||||
@ -96,6 +101,9 @@ type Client struct {
|
||||
opaque uint32
|
||||
|
||||
hdrBuf []byte
|
||||
|
||||
featureMtx sync.RWMutex
|
||||
sentHeloFeatures Features
|
||||
}
|
||||
|
||||
var (
|
||||
@ -285,6 +293,10 @@ func (c *Client) EnableFeatures(features Features) (*gomemcached.MCResponse, err
|
||||
binary.BigEndian.PutUint16(payload[len(payload)-2:], uint16(feature))
|
||||
}
|
||||
|
||||
c.featureMtx.Lock()
|
||||
c.sentHeloFeatures = features
|
||||
c.featureMtx.Unlock()
|
||||
|
||||
return c.Send(&gomemcached.MCRequest{
|
||||
Opcode: gomemcached.HELLO,
|
||||
Key: []byte("GoMemcached"),
|
||||
@ -363,6 +375,18 @@ func (c *Client) CollectionsGetCID(scope string, collection string) (*gomemcache
|
||||
return res, nil
|
||||
}
|
||||
|
||||
func (c *Client) CollectionEnabled() bool {
|
||||
c.featureMtx.RLock()
|
||||
defer c.featureMtx.RUnlock()
|
||||
|
||||
for _, feature := range c.sentHeloFeatures {
|
||||
if feature == FeatureCollections {
|
||||
return true
|
||||
}
|
||||
}
|
||||
return false
|
||||
}
|
||||
|
||||
// Get the value for a key, and update expiry
|
||||
func (c *Client) GetAndTouch(vb uint16, key string, exp int) (*gomemcached.MCResponse, error) {
|
||||
extraBuf := make([]byte, 4)
|
||||
@ -1138,3 +1162,7 @@ func IfResStatusError(response *gomemcached.MCResponse) bool {
|
||||
response.Status != gomemcached.SUBDOC_PATH_NOT_FOUND &&
|
||||
response.Status != gomemcached.SUBDOC_MULTI_PATH_FAILURE_DELETED)
|
||||
}
|
||||
|
||||
func (c *Client) Conn() io.ReadWriteCloser {
|
||||
return c.conn
|
||||
}
|
||||
|
346
vendor/github.com/couchbase/gomemcached/client/upr_event.go
generated
vendored
Normal file
346
vendor/github.com/couchbase/gomemcached/client/upr_event.go
generated
vendored
Normal file
@ -0,0 +1,346 @@
|
||||
package memcached
|
||||
|
||||
import (
|
||||
"encoding/binary"
|
||||
"fmt"
|
||||
"github.com/couchbase/gomemcached"
|
||||
"math"
|
||||
)
|
||||
|
||||
type SystemEventType int
|
||||
|
||||
const InvalidSysEvent SystemEventType = -1
|
||||
|
||||
const (
|
||||
CollectionCreate SystemEventType = 0
|
||||
CollectionDrop SystemEventType = iota
|
||||
CollectionFlush SystemEventType = iota // KV did not implement
|
||||
ScopeCreate SystemEventType = iota
|
||||
ScopeDrop SystemEventType = iota
|
||||
CollectionChanged SystemEventType = iota
|
||||
)
|
||||
|
||||
type ScopeCreateEvent interface {
|
||||
GetSystemEventName() (string, error)
|
||||
GetScopeId() (uint32, error)
|
||||
GetManifestId() (uint64, error)
|
||||
}
|
||||
|
||||
type CollectionCreateEvent interface {
|
||||
GetSystemEventName() (string, error)
|
||||
GetScopeId() (uint32, error)
|
||||
GetCollectionId() (uint32, error)
|
||||
GetManifestId() (uint64, error)
|
||||
GetMaxTTL() (uint32, error)
|
||||
}
|
||||
|
||||
type CollectionDropEvent interface {
|
||||
GetScopeId() (uint32, error)
|
||||
GetCollectionId() (uint32, error)
|
||||
GetManifestId() (uint64, error)
|
||||
}
|
||||
|
||||
type ScopeDropEvent interface {
|
||||
GetScopeId() (uint32, error)
|
||||
GetManifestId() (uint64, error)
|
||||
}
|
||||
|
||||
type CollectionChangedEvent interface {
|
||||
GetCollectionId() (uint32, error)
|
||||
GetManifestId() (uint64, error)
|
||||
GetMaxTTL() (uint32, error)
|
||||
}
|
||||
|
||||
var ErrorInvalidOp error = fmt.Errorf("Invalid Operation")
|
||||
var ErrorInvalidVersion error = fmt.Errorf("Invalid version for parsing")
|
||||
var ErrorValueTooShort error = fmt.Errorf("Value length is too short")
|
||||
var ErrorNoMaxTTL error = fmt.Errorf("This event has no max TTL")
|
||||
|
||||
// UprEvent memcached events for UPR streams.
|
||||
type UprEvent struct {
|
||||
Opcode gomemcached.CommandCode // Type of event
|
||||
Status gomemcached.Status // Response status
|
||||
VBucket uint16 // VBucket this event applies to
|
||||
DataType uint8 // data type
|
||||
Opaque uint16 // 16 MSB of opaque
|
||||
VBuuid uint64 // This field is set by downstream
|
||||
Flags uint32 // Item flags
|
||||
Expiry uint32 // Item expiration time
|
||||
Key, Value []byte // Item key/value
|
||||
OldValue []byte // TODO: TBD: old document value
|
||||
Cas uint64 // CAS value of the item
|
||||
Seqno uint64 // sequence number of the mutation
|
||||
RevSeqno uint64 // rev sequence number : deletions
|
||||
LockTime uint32 // Lock time
|
||||
MetadataSize uint16 // Metadata size
|
||||
SnapstartSeq uint64 // start sequence number of this snapshot
|
||||
SnapendSeq uint64 // End sequence number of the snapshot
|
||||
SnapshotType uint32 // 0: disk 1: memory
|
||||
FailoverLog *FailoverLog // Failover log containing vvuid and sequnce number
|
||||
Error error // Error value in case of a failure
|
||||
ExtMeta []byte // Extended Metadata
|
||||
AckSize uint32 // The number of bytes that can be Acked to DCP
|
||||
SystemEvent SystemEventType // Only valid if IsSystemEvent() is true
|
||||
SysEventVersion uint8 // Based on the version, the way Extra bytes is parsed is different
|
||||
ValueLen int // Cache it to avoid len() calls for performance
|
||||
CollectionId uint64 // Valid if Collection is in use
|
||||
}
|
||||
|
||||
// FailoverLog containing vvuid and sequnce number
|
||||
type FailoverLog [][2]uint64
|
||||
|
||||
func makeUprEvent(rq gomemcached.MCRequest, stream *UprStream, bytesReceivedFromDCP int) *UprEvent {
|
||||
event := &UprEvent{
|
||||
Opcode: rq.Opcode,
|
||||
VBucket: stream.Vbucket,
|
||||
VBuuid: stream.Vbuuid,
|
||||
Value: rq.Body,
|
||||
Cas: rq.Cas,
|
||||
ExtMeta: rq.ExtMeta,
|
||||
DataType: rq.DataType,
|
||||
ValueLen: len(rq.Body),
|
||||
SystemEvent: InvalidSysEvent,
|
||||
CollectionId: math.MaxUint64,
|
||||
}
|
||||
|
||||
event.PopulateFieldsBasedOnStreamType(rq, stream.StreamType)
|
||||
|
||||
// set AckSize for events that need to be acked to DCP,
|
||||
// i.e., events with CommandCodes that need to be buffered in DCP
|
||||
if _, ok := gomemcached.BufferedCommandCodeMap[rq.Opcode]; ok {
|
||||
event.AckSize = uint32(bytesReceivedFromDCP)
|
||||
}
|
||||
|
||||
// 16 LSBits are used by client library to encode vbucket number.
|
||||
// 16 MSBits are left for application to multiplex on opaque value.
|
||||
event.Opaque = appOpaque(rq.Opaque)
|
||||
|
||||
if len(rq.Extras) >= uprMutationExtraLen &&
|
||||
event.Opcode == gomemcached.UPR_MUTATION {
|
||||
|
||||
event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8])
|
||||
event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16])
|
||||
event.Flags = binary.BigEndian.Uint32(rq.Extras[16:20])
|
||||
event.Expiry = binary.BigEndian.Uint32(rq.Extras[20:24])
|
||||
event.LockTime = binary.BigEndian.Uint32(rq.Extras[24:28])
|
||||
event.MetadataSize = binary.BigEndian.Uint16(rq.Extras[28:30])
|
||||
|
||||
} else if len(rq.Extras) >= uprDeletetionWithDeletionTimeExtraLen &&
|
||||
event.Opcode == gomemcached.UPR_DELETION {
|
||||
|
||||
event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8])
|
||||
event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16])
|
||||
event.Expiry = binary.BigEndian.Uint32(rq.Extras[16:20])
|
||||
|
||||
} else if len(rq.Extras) >= uprDeletetionExtraLen &&
|
||||
event.Opcode == gomemcached.UPR_DELETION ||
|
||||
event.Opcode == gomemcached.UPR_EXPIRATION {
|
||||
|
||||
event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8])
|
||||
event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16])
|
||||
event.MetadataSize = binary.BigEndian.Uint16(rq.Extras[16:18])
|
||||
|
||||
} else if len(rq.Extras) >= uprSnapshotExtraLen &&
|
||||
event.Opcode == gomemcached.UPR_SNAPSHOT {
|
||||
|
||||
event.SnapstartSeq = binary.BigEndian.Uint64(rq.Extras[:8])
|
||||
event.SnapendSeq = binary.BigEndian.Uint64(rq.Extras[8:16])
|
||||
event.SnapshotType = binary.BigEndian.Uint32(rq.Extras[16:20])
|
||||
} else if event.IsSystemEvent() {
|
||||
event.PopulateEvent(rq.Extras)
|
||||
}
|
||||
|
||||
return event
|
||||
}
|
||||
|
||||
func (event *UprEvent) PopulateFieldsBasedOnStreamType(rq gomemcached.MCRequest, streamType DcpStreamType) {
|
||||
switch streamType {
|
||||
case CollectionsNonStreamId:
|
||||
switch rq.Opcode {
|
||||
// Only these will have CID encoded within the key
|
||||
case gomemcached.UPR_MUTATION,
|
||||
gomemcached.UPR_DELETION,
|
||||
gomemcached.UPR_EXPIRATION:
|
||||
uleb128 := Uleb128(rq.Key)
|
||||
result, bytesShifted := uleb128.ToUint64(rq.Keylen)
|
||||
event.CollectionId = result
|
||||
event.Key = rq.Key[bytesShifted:]
|
||||
default:
|
||||
event.Key = rq.Key
|
||||
}
|
||||
case CollectionsStreamId:
|
||||
// TODO - not implemented
|
||||
fallthrough
|
||||
case NonCollectionStream:
|
||||
// Let default behavior be legacy stream type
|
||||
fallthrough
|
||||
default:
|
||||
event.Key = rq.Key
|
||||
}
|
||||
}
|
||||
|
||||
func (event *UprEvent) String() string {
|
||||
name := gomemcached.CommandNames[event.Opcode]
|
||||
if name == "" {
|
||||
name = fmt.Sprintf("#%d", event.Opcode)
|
||||
}
|
||||
return name
|
||||
}
|
||||
|
||||
func (event *UprEvent) IsSnappyDataType() bool {
|
||||
return event.Opcode == gomemcached.UPR_MUTATION && (event.DataType&SnappyDataType > 0)
|
||||
}
|
||||
|
||||
func (event *UprEvent) IsCollectionType() bool {
|
||||
return event.IsSystemEvent() || event.CollectionId <= math.MaxUint32
|
||||
}
|
||||
|
||||
func (event *UprEvent) IsSystemEvent() bool {
|
||||
return event.Opcode == gomemcached.DCP_SYSTEM_EVENT
|
||||
}
|
||||
|
||||
func (event *UprEvent) PopulateEvent(extras []byte) {
|
||||
if len(extras) < dcpSystemEventExtraLen {
|
||||
// Wrong length, don't parse
|
||||
return
|
||||
}
|
||||
event.Seqno = binary.BigEndian.Uint64(extras[:8])
|
||||
event.SystemEvent = SystemEventType(binary.BigEndian.Uint32(extras[8:12]))
|
||||
var versionTemp uint16 = binary.BigEndian.Uint16(extras[12:14])
|
||||
event.SysEventVersion = uint8(versionTemp >> 8)
|
||||
}
|
||||
|
||||
func (event *UprEvent) GetSystemEventName() (string, error) {
|
||||
switch event.SystemEvent {
|
||||
case CollectionCreate:
|
||||
fallthrough
|
||||
case ScopeCreate:
|
||||
return string(event.Key), nil
|
||||
default:
|
||||
return "", ErrorInvalidOp
|
||||
}
|
||||
}
|
||||
|
||||
func (event *UprEvent) GetManifestId() (uint64, error) {
|
||||
switch event.SystemEvent {
|
||||
// Version 0 only checks
|
||||
case CollectionChanged:
|
||||
fallthrough
|
||||
case ScopeDrop:
|
||||
fallthrough
|
||||
case ScopeCreate:
|
||||
fallthrough
|
||||
case CollectionDrop:
|
||||
if event.SysEventVersion > 0 {
|
||||
return 0, ErrorInvalidVersion
|
||||
}
|
||||
fallthrough
|
||||
case CollectionCreate:
|
||||
// CollectionCreate supports version 1
|
||||
if event.SysEventVersion > 1 {
|
||||
return 0, ErrorInvalidVersion
|
||||
}
|
||||
if event.ValueLen < 8 {
|
||||
return 0, ErrorValueTooShort
|
||||
}
|
||||
return binary.BigEndian.Uint64(event.Value[0:8]), nil
|
||||
default:
|
||||
return 0, ErrorInvalidOp
|
||||
}
|
||||
}
|
||||
|
||||
func (event *UprEvent) GetCollectionId() (uint32, error) {
|
||||
switch event.SystemEvent {
|
||||
case CollectionDrop:
|
||||
if event.SysEventVersion > 0 {
|
||||
return 0, ErrorInvalidVersion
|
||||
}
|
||||
fallthrough
|
||||
case CollectionCreate:
|
||||
if event.SysEventVersion > 1 {
|
||||
return 0, ErrorInvalidVersion
|
||||
}
|
||||
if event.ValueLen < 16 {
|
||||
return 0, ErrorValueTooShort
|
||||
}
|
||||
return binary.BigEndian.Uint32(event.Value[12:16]), nil
|
||||
case CollectionChanged:
|
||||
if event.SysEventVersion > 0 {
|
||||
return 0, ErrorInvalidVersion
|
||||
}
|
||||
if event.ValueLen < 12 {
|
||||
return 0, ErrorValueTooShort
|
||||
}
|
||||
return binary.BigEndian.Uint32(event.Value[8:12]), nil
|
||||
default:
|
||||
return 0, ErrorInvalidOp
|
||||
}
|
||||
}
|
||||
|
||||
func (event *UprEvent) GetScopeId() (uint32, error) {
|
||||
switch event.SystemEvent {
|
||||
// version 0 checks
|
||||
case ScopeCreate:
|
||||
fallthrough
|
||||
case ScopeDrop:
|
||||
fallthrough
|
||||
case CollectionDrop:
|
||||
if event.SysEventVersion > 0 {
|
||||
return 0, ErrorInvalidVersion
|
||||
}
|
||||
fallthrough
|
||||
case CollectionCreate:
|
||||
// CollectionCreate could be either 0 or 1
|
||||
if event.SysEventVersion > 1 {
|
||||
return 0, ErrorInvalidVersion
|
||||
}
|
||||
if event.ValueLen < 12 {
|
||||
return 0, ErrorValueTooShort
|
||||
}
|
||||
return binary.BigEndian.Uint32(event.Value[8:12]), nil
|
||||
default:
|
||||
return 0, ErrorInvalidOp
|
||||
}
|
||||
}
|
||||
|
||||
func (event *UprEvent) GetMaxTTL() (uint32, error) {
|
||||
switch event.SystemEvent {
|
||||
case CollectionCreate:
|
||||
if event.SysEventVersion < 1 {
|
||||
return 0, ErrorNoMaxTTL
|
||||
}
|
||||
if event.ValueLen < 20 {
|
||||
return 0, ErrorValueTooShort
|
||||
}
|
||||
return binary.BigEndian.Uint32(event.Value[16:20]), nil
|
||||
case CollectionChanged:
|
||||
if event.SysEventVersion > 0 {
|
||||
return 0, ErrorInvalidVersion
|
||||
}
|
||||
if event.ValueLen < 16 {
|
||||
return 0, ErrorValueTooShort
|
||||
}
|
||||
return binary.BigEndian.Uint32(event.Value[12:16]), nil
|
||||
default:
|
||||
return 0, ErrorInvalidOp
|
||||
}
|
||||
}
|
||||
|
||||
type Uleb128 []byte
|
||||
|
||||
func (u Uleb128) ToUint64(cachedLen int) (result uint64, bytesShifted int) {
|
||||
var shift uint = 0
|
||||
|
||||
for curByte := 0; curByte < cachedLen; curByte++ {
|
||||
oneByte := u[curByte]
|
||||
last7Bits := 0x7f & oneByte
|
||||
result |= uint64(last7Bits) << shift
|
||||
bytesShifted++
|
||||
if oneByte&0x80 == 0 {
|
||||
break
|
||||
}
|
||||
shift += 7
|
||||
}
|
||||
|
||||
return
|
||||
}
|
248
vendor/github.com/couchbase/gomemcached/client/upr_feed.go
generated
vendored
248
vendor/github.com/couchbase/gomemcached/client/upr_feed.go
generated
vendored
@ -19,6 +19,7 @@ const uprMutationExtraLen = 30
|
||||
const uprDeletetionExtraLen = 18
|
||||
const uprDeletetionWithDeletionTimeExtraLen = 21
|
||||
const uprSnapshotExtraLen = 20
|
||||
const dcpSystemEventExtraLen = 13
|
||||
const bufferAckThreshold = 0.2
|
||||
const opaqueOpen = 0xBEAF0001
|
||||
const opaqueFailover = 0xDEADBEEF
|
||||
@ -27,32 +28,6 @@ const uprDefaultNoopInterval = 120
|
||||
// Counter on top of opaqueOpen that others can draw from for open and control msgs
|
||||
var opaqueOpenCtrlWell uint32 = opaqueOpen
|
||||
|
||||
// UprEvent memcached events for UPR streams.
|
||||
type UprEvent struct {
|
||||
Opcode gomemcached.CommandCode // Type of event
|
||||
Status gomemcached.Status // Response status
|
||||
VBucket uint16 // VBucket this event applies to
|
||||
DataType uint8 // data type
|
||||
Opaque uint16 // 16 MSB of opaque
|
||||
VBuuid uint64 // This field is set by downstream
|
||||
Flags uint32 // Item flags
|
||||
Expiry uint32 // Item expiration time
|
||||
Key, Value []byte // Item key/value
|
||||
OldValue []byte // TODO: TBD: old document value
|
||||
Cas uint64 // CAS value of the item
|
||||
Seqno uint64 // sequence number of the mutation
|
||||
RevSeqno uint64 // rev sequence number : deletions
|
||||
LockTime uint32 // Lock time
|
||||
MetadataSize uint16 // Metadata size
|
||||
SnapstartSeq uint64 // start sequence number of this snapshot
|
||||
SnapendSeq uint64 // End sequence number of the snapshot
|
||||
SnapshotType uint32 // 0: disk 1: memory
|
||||
FailoverLog *FailoverLog // Failover log containing vvuid and sequnce number
|
||||
Error error // Error value in case of a failure
|
||||
ExtMeta []byte
|
||||
AckSize uint32 // The number of bytes that can be Acked to DCP
|
||||
}
|
||||
|
||||
type PriorityType string
|
||||
|
||||
// high > medium > disabled > low
|
||||
@ -63,13 +38,39 @@ const (
|
||||
PriorityHigh PriorityType = "high"
|
||||
)
|
||||
|
||||
type DcpStreamType int32
|
||||
|
||||
var UninitializedStream DcpStreamType = -1
|
||||
|
||||
const (
|
||||
NonCollectionStream DcpStreamType = 0
|
||||
CollectionsNonStreamId DcpStreamType = iota
|
||||
CollectionsStreamId DcpStreamType = iota
|
||||
)
|
||||
|
||||
func (t DcpStreamType) String() string {
|
||||
switch t {
|
||||
case UninitializedStream:
|
||||
return "Un-Initialized Stream"
|
||||
case NonCollectionStream:
|
||||
return "Traditional Non-Collection Stream"
|
||||
case CollectionsNonStreamId:
|
||||
return "Collections Stream without StreamID"
|
||||
case CollectionsStreamId:
|
||||
return "Collection Stream with StreamID"
|
||||
default:
|
||||
return "Unknown Stream Type"
|
||||
}
|
||||
}
|
||||
|
||||
// UprStream is per stream data structure over an UPR Connection.
|
||||
type UprStream struct {
|
||||
Vbucket uint16 // Vbucket id
|
||||
Vbuuid uint64 // vbucket uuid
|
||||
StartSeq uint64 // start sequence number
|
||||
EndSeq uint64 // end sequence number
|
||||
connected bool
|
||||
Vbucket uint16 // Vbucket id
|
||||
Vbuuid uint64 // vbucket uuid
|
||||
StartSeq uint64 // start sequence number
|
||||
EndSeq uint64 // end sequence number
|
||||
connected bool
|
||||
StreamType DcpStreamType
|
||||
}
|
||||
|
||||
type FeedState int
|
||||
@ -113,6 +114,7 @@ type UprFeatures struct {
|
||||
IncludeDeletionTime bool
|
||||
DcpPriority PriorityType
|
||||
EnableExpiry bool
|
||||
EnableStreamId bool
|
||||
}
|
||||
|
||||
/**
|
||||
@ -274,9 +276,15 @@ type UprFeed struct {
|
||||
// if flag is true, upr feed will use ack from client to determine whether/when to send ack to DCP
|
||||
// if flag is false, upr feed will track how many bytes it has sent to client
|
||||
// and use that to determine whether/when to send ack to DCP
|
||||
ackByClient bool
|
||||
feedState FeedState
|
||||
muFeedState sync.RWMutex
|
||||
ackByClient bool
|
||||
feedState FeedState
|
||||
muFeedState sync.RWMutex
|
||||
activatedFeatures UprFeatures
|
||||
collectionEnabled bool // This is needed separately because parsing depends on this
|
||||
// DCP StreamID allows multiple filtered collection streams to share a single DCP Stream
|
||||
// It is not allowed once a regular/legacy stream was started originally
|
||||
streamsType DcpStreamType
|
||||
initStreamTypeOnce sync.Once
|
||||
}
|
||||
|
||||
// Exported interface - to allow for mocking
|
||||
@ -296,6 +304,9 @@ type UprFeedIface interface {
|
||||
UprRequestStream(vbno, opaqueMSB uint16, flags uint32, vuuid, startSequence, endSequence, snapStart, snapEnd uint64) error
|
||||
// Set DCP priority on an existing DCP connection. The command is sent asynchronously without waiting for a response
|
||||
SetPriorityAsync(p PriorityType) error
|
||||
|
||||
// Various Collection-Type RequestStreams
|
||||
UprRequestCollectionsStream(vbno, opaqueMSB uint16, flags uint32, vbuuid, startSeq, endSeq, snapStart, snapEnd uint64, filter *CollectionsFilter) error
|
||||
}
|
||||
|
||||
type UprStats struct {
|
||||
@ -305,9 +316,6 @@ type UprStats struct {
|
||||
TotalSnapShot uint64
|
||||
}
|
||||
|
||||
// FailoverLog containing vvuid and sequnce number
|
||||
type FailoverLog [][2]uint64
|
||||
|
||||
// error codes
|
||||
var ErrorInvalidLog = errors.New("couchbase.errorInvalidLog")
|
||||
|
||||
@ -320,76 +328,6 @@ func (flogp *FailoverLog) Latest() (vbuuid, seqno uint64, err error) {
|
||||
return vbuuid, seqno, ErrorInvalidLog
|
||||
}
|
||||
|
||||
func makeUprEvent(rq gomemcached.MCRequest, stream *UprStream, bytesReceivedFromDCP int) *UprEvent {
|
||||
event := &UprEvent{
|
||||
Opcode: rq.Opcode,
|
||||
VBucket: stream.Vbucket,
|
||||
VBuuid: stream.Vbuuid,
|
||||
Key: rq.Key,
|
||||
Value: rq.Body,
|
||||
Cas: rq.Cas,
|
||||
ExtMeta: rq.ExtMeta,
|
||||
DataType: rq.DataType,
|
||||
}
|
||||
|
||||
// set AckSize for events that need to be acked to DCP,
|
||||
// i.e., events with CommandCodes that need to be buffered in DCP
|
||||
if _, ok := gomemcached.BufferedCommandCodeMap[rq.Opcode]; ok {
|
||||
event.AckSize = uint32(bytesReceivedFromDCP)
|
||||
}
|
||||
|
||||
// 16 LSBits are used by client library to encode vbucket number.
|
||||
// 16 MSBits are left for application to multiplex on opaque value.
|
||||
event.Opaque = appOpaque(rq.Opaque)
|
||||
|
||||
if len(rq.Extras) >= uprMutationExtraLen &&
|
||||
event.Opcode == gomemcached.UPR_MUTATION {
|
||||
|
||||
event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8])
|
||||
event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16])
|
||||
event.Flags = binary.BigEndian.Uint32(rq.Extras[16:20])
|
||||
event.Expiry = binary.BigEndian.Uint32(rq.Extras[20:24])
|
||||
event.LockTime = binary.BigEndian.Uint32(rq.Extras[24:28])
|
||||
event.MetadataSize = binary.BigEndian.Uint16(rq.Extras[28:30])
|
||||
|
||||
} else if len(rq.Extras) >= uprDeletetionWithDeletionTimeExtraLen &&
|
||||
event.Opcode == gomemcached.UPR_DELETION {
|
||||
|
||||
event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8])
|
||||
event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16])
|
||||
event.Expiry = binary.BigEndian.Uint32(rq.Extras[16:20])
|
||||
|
||||
} else if len(rq.Extras) >= uprDeletetionExtraLen &&
|
||||
event.Opcode == gomemcached.UPR_DELETION ||
|
||||
event.Opcode == gomemcached.UPR_EXPIRATION {
|
||||
|
||||
event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8])
|
||||
event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16])
|
||||
event.MetadataSize = binary.BigEndian.Uint16(rq.Extras[16:18])
|
||||
|
||||
} else if len(rq.Extras) >= uprSnapshotExtraLen &&
|
||||
event.Opcode == gomemcached.UPR_SNAPSHOT {
|
||||
|
||||
event.SnapstartSeq = binary.BigEndian.Uint64(rq.Extras[:8])
|
||||
event.SnapendSeq = binary.BigEndian.Uint64(rq.Extras[8:16])
|
||||
event.SnapshotType = binary.BigEndian.Uint32(rq.Extras[16:20])
|
||||
}
|
||||
|
||||
return event
|
||||
}
|
||||
|
||||
func (event *UprEvent) String() string {
|
||||
name := gomemcached.CommandNames[event.Opcode]
|
||||
if name == "" {
|
||||
name = fmt.Sprintf("#%d", event.Opcode)
|
||||
}
|
||||
return name
|
||||
}
|
||||
|
||||
func (event *UprEvent) IsSnappyDataType() bool {
|
||||
return event.Opcode == gomemcached.UPR_MUTATION && (event.DataType&SnappyDataType > 0)
|
||||
}
|
||||
|
||||
func (feed *UprFeed) sendCommands(mc *Client) {
|
||||
transmitCh := feed.transmitCh
|
||||
transmitCl := feed.transmitCl
|
||||
@ -420,6 +358,10 @@ func (feed *UprFeed) activateStream(vbno, opaque uint16, stream *UprStream) erro
|
||||
feed.muVbstreams.Lock()
|
||||
defer feed.muVbstreams.Unlock()
|
||||
|
||||
if feed.collectionEnabled {
|
||||
stream.StreamType = feed.streamsType
|
||||
}
|
||||
|
||||
// Set this stream as the officially connected stream for this vb
|
||||
stream.connected = true
|
||||
feed.vbstreams[vbno] = stream
|
||||
@ -440,14 +382,15 @@ func (mc *Client) NewUprFeed() (*UprFeed, error) {
|
||||
}
|
||||
|
||||
func (mc *Client) NewUprFeedWithConfig(ackByClient bool) (*UprFeed, error) {
|
||||
|
||||
feed := &UprFeed{
|
||||
conn: mc,
|
||||
closer: make(chan bool, 1),
|
||||
vbstreams: make(map[uint16]*UprStream),
|
||||
transmitCh: make(chan *gomemcached.MCRequest),
|
||||
transmitCl: make(chan bool),
|
||||
ackByClient: ackByClient,
|
||||
conn: mc,
|
||||
closer: make(chan bool, 1),
|
||||
vbstreams: make(map[uint16]*UprStream),
|
||||
transmitCh: make(chan *gomemcached.MCRequest),
|
||||
transmitCl: make(chan bool),
|
||||
ackByClient: ackByClient,
|
||||
collectionEnabled: mc.CollectionEnabled(),
|
||||
streamsType: UninitializedStream,
|
||||
}
|
||||
|
||||
feed.negotiator.initialize()
|
||||
@ -642,7 +585,22 @@ func (feed *UprFeed) uprOpen(name string, sequence uint32, bufSize uint32, featu
|
||||
activatedFeatures.EnableExpiry = true
|
||||
}
|
||||
|
||||
if features.EnableStreamId {
|
||||
rq := &gomemcached.MCRequest{
|
||||
Opcode: gomemcached.UPR_CONTROL,
|
||||
Key: []byte("enable_stream_id"),
|
||||
Body: []byte("true"),
|
||||
Opaque: getUprOpenCtrlOpaque(),
|
||||
}
|
||||
err = sendMcRequestSync(feed.conn, rq)
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
activatedFeatures.EnableStreamId = true
|
||||
}
|
||||
|
||||
// everything is ok so far, set upr feed to open state
|
||||
feed.activatedFeatures = activatedFeatures
|
||||
feed.setOpen()
|
||||
return
|
||||
}
|
||||
@ -689,10 +647,60 @@ func (mc *Client) UprGetFailoverLog(
|
||||
func (feed *UprFeed) UprRequestStream(vbno, opaqueMSB uint16, flags uint32,
|
||||
vuuid, startSequence, endSequence, snapStart, snapEnd uint64) error {
|
||||
|
||||
return feed.UprRequestCollectionsStream(vbno, opaqueMSB, flags, vuuid, startSequence, endSequence, snapStart, snapEnd, nil)
|
||||
}
|
||||
|
||||
func (feed *UprFeed) initStreamType(filter *CollectionsFilter) (err error) {
|
||||
if filter != nil && filter.UseStreamId && !feed.activatedFeatures.EnableStreamId {
|
||||
err = fmt.Errorf("Cannot use streamID based filter if the feed was not started with the streamID feature")
|
||||
return
|
||||
}
|
||||
|
||||
streamInitFunc := func() {
|
||||
if feed.streamsType != UninitializedStream {
|
||||
// Shouldn't happen
|
||||
err = fmt.Errorf("The current feed has already been started in %v mode", feed.streamsType.String())
|
||||
} else {
|
||||
if !feed.collectionEnabled {
|
||||
feed.streamsType = NonCollectionStream
|
||||
} else {
|
||||
if filter != nil && filter.UseStreamId {
|
||||
feed.streamsType = CollectionsStreamId
|
||||
} else {
|
||||
feed.streamsType = CollectionsNonStreamId
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
feed.initStreamTypeOnce.Do(streamInitFunc)
|
||||
return
|
||||
}
|
||||
|
||||
func (feed *UprFeed) UprRequestCollectionsStream(vbno, opaqueMSB uint16, flags uint32,
|
||||
vbuuid, startSequence, endSequence, snapStart, snapEnd uint64, filter *CollectionsFilter) error {
|
||||
|
||||
err := feed.initStreamType(filter)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
var mcRequestBody []byte
|
||||
if filter != nil {
|
||||
err = filter.IsValid()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
mcRequestBody, err = filter.ToStreamReqBody()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
|
||||
rq := &gomemcached.MCRequest{
|
||||
Opcode: gomemcached.UPR_STREAMREQ,
|
||||
VBucket: vbno,
|
||||
Opaque: composeOpaque(vbno, opaqueMSB),
|
||||
Body: mcRequestBody,
|
||||
}
|
||||
|
||||
rq.Extras = make([]byte, 48) // #Extras
|
||||
@ -700,15 +708,15 @@ func (feed *UprFeed) UprRequestStream(vbno, opaqueMSB uint16, flags uint32,
|
||||
binary.BigEndian.PutUint32(rq.Extras[4:8], uint32(0))
|
||||
binary.BigEndian.PutUint64(rq.Extras[8:16], startSequence)
|
||||
binary.BigEndian.PutUint64(rq.Extras[16:24], endSequence)
|
||||
binary.BigEndian.PutUint64(rq.Extras[24:32], vuuid)
|
||||
binary.BigEndian.PutUint64(rq.Extras[24:32], vbuuid)
|
||||
binary.BigEndian.PutUint64(rq.Extras[32:40], snapStart)
|
||||
binary.BigEndian.PutUint64(rq.Extras[40:48], snapEnd)
|
||||
|
||||
feed.negotiator.registerRequest(vbno, opaqueMSB, vuuid, startSequence, endSequence)
|
||||
feed.negotiator.registerRequest(vbno, opaqueMSB, vbuuid, startSequence, endSequence)
|
||||
// Any client that has ever called this method, regardless of return code,
|
||||
// should expect a potential UPR_CLOSESTREAM message due to this new map entry prior to Transmit.
|
||||
|
||||
if err := feed.conn.Transmit(rq); err != nil {
|
||||
if err = feed.conn.Transmit(rq); err != nil {
|
||||
logging.Errorf("Error in StreamRequest %s", err.Error())
|
||||
// If an error occurs during transmit, then the UPRFeed will keep the stream
|
||||
// in the vbstreams map. This is to prevent nil lookup from any previously
|
||||
@ -973,6 +981,12 @@ loop:
|
||||
if err := feed.conn.TransmitResponse(noop); err != nil {
|
||||
logging.Warnf("failed to transmit command %s. Error %s", noop.Opcode.String(), err.Error())
|
||||
}
|
||||
case gomemcached.DCP_SYSTEM_EVENT:
|
||||
if stream == nil {
|
||||
logging.Infof("Stream not found for vb %d: %#v", vb, pkt)
|
||||
break loop
|
||||
}
|
||||
event = makeUprEvent(pkt, stream, bytes)
|
||||
default:
|
||||
logging.Infof("Recived an unknown response for vbucket %d", vb)
|
||||
}
|
||||
|
381
vendor/github.com/couchbase/gomemcached/flexibleFraming.go
generated
vendored
Normal file
381
vendor/github.com/couchbase/gomemcached/flexibleFraming.go
generated
vendored
Normal file
@ -0,0 +1,381 @@
|
||||
package gomemcached
|
||||
|
||||
import (
|
||||
"encoding/binary"
|
||||
"fmt"
|
||||
)
|
||||
|
||||
type FrameObjType int
|
||||
|
||||
const (
|
||||
FrameBarrier FrameObjType = iota
|
||||
FrameDurability FrameObjType = iota
|
||||
FrameDcpStreamId FrameObjType = iota
|
||||
FrameOpenTracing FrameObjType = iota
|
||||
)
|
||||
|
||||
type FrameInfo struct {
|
||||
ObjId FrameObjType
|
||||
ObjLen int
|
||||
ObjData []byte
|
||||
}
|
||||
|
||||
var ErrorInvalidOp error = fmt.Errorf("Specified method is not applicable")
|
||||
var ErrorObjLenNotMatch error = fmt.Errorf("Object length does not match data")
|
||||
|
||||
func (f *FrameInfo) Validate() error {
|
||||
switch f.ObjId {
|
||||
case FrameBarrier:
|
||||
if f.ObjLen != 0 {
|
||||
return fmt.Errorf("Invalid FrameBarrier - length is %v\n", f.ObjLen)
|
||||
} else if f.ObjLen != len(f.ObjData) {
|
||||
return ErrorObjLenNotMatch
|
||||
}
|
||||
case FrameDurability:
|
||||
if f.ObjLen != 1 && f.ObjLen != 3 {
|
||||
return fmt.Errorf("Invalid FrameDurability - length is %v\n", f.ObjLen)
|
||||
} else if f.ObjLen != len(f.ObjData) {
|
||||
return ErrorObjLenNotMatch
|
||||
}
|
||||
case FrameDcpStreamId:
|
||||
if f.ObjLen != 2 {
|
||||
return fmt.Errorf("Invalid FrameDcpStreamId - length is %v\n", f.ObjLen)
|
||||
} else if f.ObjLen != len(f.ObjData) {
|
||||
return ErrorObjLenNotMatch
|
||||
}
|
||||
case FrameOpenTracing:
|
||||
if f.ObjLen == 0 {
|
||||
return fmt.Errorf("Invalid FrameOpenTracing - length must be > 0")
|
||||
} else if f.ObjLen != len(f.ObjData) {
|
||||
return ErrorObjLenNotMatch
|
||||
}
|
||||
default:
|
||||
return fmt.Errorf("Unknown FrameInfo type")
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (f *FrameInfo) GetStreamId() (uint16, error) {
|
||||
if f.ObjId != FrameDcpStreamId {
|
||||
return 0, ErrorInvalidOp
|
||||
}
|
||||
|
||||
var output uint16
|
||||
output = uint16(f.ObjData[0])
|
||||
output = output << 8
|
||||
output |= uint16(f.ObjData[1])
|
||||
return output, nil
|
||||
}
|
||||
|
||||
type DurabilityLvl uint8
|
||||
|
||||
const (
|
||||
DuraInvalid DurabilityLvl = iota // Not used (0x0)
|
||||
DuraMajority DurabilityLvl = iota // (0x01)
|
||||
DuraMajorityAndPersistOnMaster DurabilityLvl = iota // (0x02)
|
||||
DuraPersistToMajority DurabilityLvl = iota // (0x03)
|
||||
)
|
||||
|
||||
func (f *FrameInfo) GetDurabilityRequirements() (lvl DurabilityLvl, timeoutProvided bool, timeoutMs uint16, err error) {
|
||||
if f.ObjId != FrameDurability {
|
||||
err = ErrorInvalidOp
|
||||
return
|
||||
}
|
||||
if f.ObjLen != 1 && f.ObjLen != 3 {
|
||||
err = ErrorObjLenNotMatch
|
||||
return
|
||||
}
|
||||
|
||||
lvl = DurabilityLvl(uint8(f.ObjData[0]))
|
||||
|
||||
if f.ObjLen == 3 {
|
||||
timeoutProvided = true
|
||||
timeoutMs = binary.BigEndian.Uint16(f.ObjData[1:2])
|
||||
}
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
func incrementMarker(bitsToBeIncremented, byteIncrementCnt *int, framingElen, curObjIdx int) (int, error) {
|
||||
for *bitsToBeIncremented >= 8 {
|
||||
*byteIncrementCnt++
|
||||
*bitsToBeIncremented -= 8
|
||||
}
|
||||
marker := curObjIdx + *byteIncrementCnt
|
||||
if marker > framingElen {
|
||||
return -1, fmt.Errorf("Out of bounds")
|
||||
}
|
||||
return marker, nil
|
||||
}
|
||||
|
||||
// Right now, halfByteRemaining will always be false, because ObjID and Len haven't gotten that large yet
|
||||
func (f *FrameInfo) Bytes() (output []byte, halfByteRemaining bool) {
|
||||
// ObjIdentifier - 4 bits + ObjLength - 4 bits
|
||||
var idAndLen uint8
|
||||
idAndLen |= uint8(f.ObjId) << 4
|
||||
idAndLen |= uint8(f.ObjLen)
|
||||
output = append(output, byte(idAndLen))
|
||||
|
||||
// Rest is Data
|
||||
output = append(output, f.ObjData...)
|
||||
return
|
||||
}
|
||||
|
||||
func parseFrameInfoObjects(buf []byte, framingElen int) (objs []FrameInfo, err error, halfByteRemaining bool) {
|
||||
var curObjIdx int
|
||||
var byteIncrementCnt int
|
||||
var bitsToBeIncremented int
|
||||
var marker int
|
||||
|
||||
// Parse frameInfo objects
|
||||
for curObjIdx = 0; curObjIdx < framingElen; curObjIdx += byteIncrementCnt {
|
||||
byteIncrementCnt = 0
|
||||
var oneFrameObj FrameInfo
|
||||
|
||||
// First get the objId
|
||||
// -------------------------
|
||||
var objId int
|
||||
var objHeader uint8 = buf[curObjIdx]
|
||||
var objIdentifierRaw uint8
|
||||
if bitsToBeIncremented == 0 {
|
||||
// ObjHeader
|
||||
// 0 1 2 3 4 5 6 7
|
||||
// ^-----^
|
||||
// ObjIdentifierRaw
|
||||
objIdentifierRaw = (objHeader & 0xf0) >> 4
|
||||
} else {
|
||||
// ObjHeader
|
||||
// 0 1 2 3 4 5 6 7
|
||||
// ^-----^
|
||||
// ObjIdentifierRaw
|
||||
objIdentifierRaw = (objHeader & 0x0f)
|
||||
}
|
||||
bitsToBeIncremented += 4
|
||||
|
||||
marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx)
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
|
||||
// Value is 0-14
|
||||
objId = int(objIdentifierRaw & 0xe)
|
||||
// If bit 15 is set, ID is 15 + value of next byte
|
||||
if objIdentifierRaw&0x1 > 0 {
|
||||
if bitsToBeIncremented > 0 {
|
||||
// ObjHeader
|
||||
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
||||
// ^-----^ ^---------------^
|
||||
// ObjId1 Extension
|
||||
// ^ marker
|
||||
buffer := uint16(buf[marker])
|
||||
buffer = buffer << 8
|
||||
buffer |= uint16(buf[marker+1])
|
||||
var extension uint8 = uint8(buffer & 0xff0 >> 4)
|
||||
objId += int(extension)
|
||||
} else {
|
||||
// ObjHeader
|
||||
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
||||
// ^-----^ ^-------------------^
|
||||
// ObjId1 extension
|
||||
// ^ marker
|
||||
var extension uint8 = uint8(buf[marker])
|
||||
objId += int(extension)
|
||||
}
|
||||
bitsToBeIncremented += 8
|
||||
}
|
||||
|
||||
marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx)
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
oneFrameObj.ObjId = FrameObjType(objId)
|
||||
|
||||
// Then get the obj length
|
||||
// -------------------------
|
||||
var objLenRaw uint8
|
||||
var objLen int
|
||||
if bitsToBeIncremented > 0 {
|
||||
// ObjHeader
|
||||
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
||||
// ^ ^---------^
|
||||
// marker objLen
|
||||
objLenRaw = uint8(buf[marker]) & 0x0f
|
||||
} else {
|
||||
// ObjHeader
|
||||
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
|
||||
// ^--------^
|
||||
// objLen
|
||||
// ^ marker
|
||||
objLenRaw = uint8(buf[marker]) & 0xf0 >> 4
|
||||
}
|
||||
bitsToBeIncremented += 4
|
||||
|
||||
marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx)
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
|
||||
// Length is 0-14
|
||||
objLen = int(objLenRaw & 0xe)
|
||||
// If bit 15 is set, lenghth is 15 + value of next byte
|
||||
if objLenRaw&0x1 > 0 {
|
||||
if bitsToBeIncremented == 0 {
|
||||
// ObjHeader
|
||||
// 12 13 14 15 16 17 18 19 20 21 22 23
|
||||
// ^---------^ ^--------------------^
|
||||
// objLen extension
|
||||
// ^ marker
|
||||
var extension uint8 = uint8(buf[marker])
|
||||
objLen += int(extension)
|
||||
} else {
|
||||
// ObjHeader
|
||||
// 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31
|
||||
// ^--------^ ^---------------------^
|
||||
// objLen extension
|
||||
// ^ marker var buffer uint16
|
||||
buffer := uint16(buf[marker])
|
||||
buffer = buffer << 8
|
||||
buffer |= uint16(buf[marker+1])
|
||||
var extension uint8 = uint8(buffer & 0xff0 >> 4)
|
||||
objLen += int(extension)
|
||||
}
|
||||
bitsToBeIncremented += 8
|
||||
}
|
||||
|
||||
marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx)
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
oneFrameObj.ObjLen = objLen
|
||||
|
||||
// The rest is N-bytes of data based on the length
|
||||
if bitsToBeIncremented == 0 {
|
||||
// No weird alignment needed
|
||||
oneFrameObj.ObjData = buf[marker : marker+objLen]
|
||||
} else {
|
||||
// 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31
|
||||
// ^--------^ ^---------------------^ ^--------->
|
||||
// objLen extension data
|
||||
// ^ marker
|
||||
oneFrameObj.ObjData = ShiftByteSliceLeft4Bits(buf[marker : marker+objLen+1])
|
||||
}
|
||||
err = oneFrameObj.Validate()
|
||||
if err != nil {
|
||||
return
|
||||
}
|
||||
objs = append(objs, oneFrameObj)
|
||||
|
||||
bitsToBeIncremented += 8 * objLen
|
||||
marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx)
|
||||
}
|
||||
|
||||
if bitsToBeIncremented > 0 {
|
||||
halfByteRemaining = true
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func ShiftByteSliceLeft4Bits(slice []byte) (replacement []byte) {
|
||||
var buffer uint16
|
||||
var i int
|
||||
sliceLen := len(slice)
|
||||
|
||||
if sliceLen < 2 {
|
||||
// Let's not shift less than 16 bits
|
||||
return
|
||||
}
|
||||
|
||||
replacement = make([]byte, sliceLen, cap(slice))
|
||||
|
||||
for i = 0; i < sliceLen-1; i++ {
|
||||
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
||||
// ^-----^ ^---------------^ ^-----------
|
||||
// garbage data byte 0 data byte 1
|
||||
buffer = uint16(slice[i])
|
||||
buffer = buffer << 8
|
||||
buffer |= uint16(slice[i+1])
|
||||
replacement[i] = uint8(buffer & 0xff0 >> 4)
|
||||
}
|
||||
|
||||
if i < sliceLen {
|
||||
lastByte := slice[sliceLen-1]
|
||||
lastByte = lastByte << 4
|
||||
replacement[i] = lastByte
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
// The following is used to theoretically support frameInfo ObjID extensions
|
||||
// for completeness, but they are not very efficient though
|
||||
func ShiftByteSliceRight4Bits(slice []byte) (replacement []byte) {
|
||||
var buffer uint16
|
||||
var i int
|
||||
var leftovers uint8 // 4 bits only
|
||||
var replacementUnit uint16
|
||||
var first bool = true
|
||||
var firstLeftovers uint8
|
||||
var lastLeftovers uint8
|
||||
sliceLen := len(slice)
|
||||
|
||||
if sliceLen < 2 {
|
||||
// Let's not shift less than 16 bits
|
||||
return
|
||||
}
|
||||
|
||||
if slice[sliceLen-1]&0xf == 0 {
|
||||
replacement = make([]byte, sliceLen, cap(slice))
|
||||
} else {
|
||||
replacement = make([]byte, sliceLen+1, cap(slice)+1)
|
||||
}
|
||||
|
||||
for i = 0; i < sliceLen-1; i++ {
|
||||
buffer = binary.BigEndian.Uint16(slice[i : i+2])
|
||||
// (buffer)
|
||||
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
||||
// ^-------------^ ^-------------------^
|
||||
// data byte 0 data byte 1
|
||||
//
|
||||
// into
|
||||
//
|
||||
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
||||
// ^-----^ ^---------------^ ^--------------------^ ^----------^
|
||||
// zeroes data byte 0 data byte 1 zeroes
|
||||
|
||||
if first {
|
||||
// The leftover OR'ing will overwrite the first 4 bits of data byte 0. Save them
|
||||
firstLeftovers = uint8(buffer & 0xf000 >> 12)
|
||||
first = false
|
||||
}
|
||||
replacementUnit = 0
|
||||
replacementUnit |= uint16(leftovers) << 12
|
||||
replacementUnit |= (buffer & 0xff00) >> 4 // data byte 0
|
||||
replacementUnit |= buffer & 0xff >> 4 // data byte 1 first 4 bits
|
||||
lastLeftovers = uint8(buffer&0xf) << 4
|
||||
|
||||
replacement[i+1] = byte(replacementUnit)
|
||||
|
||||
leftovers = uint8((buffer & 0x000f) << 4)
|
||||
}
|
||||
|
||||
replacement[0] = byte(uint8(replacement[0]) | firstLeftovers)
|
||||
if lastLeftovers > 0 {
|
||||
replacement[sliceLen] = byte(lastLeftovers)
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func Merge2HalfByteSlices(src1, src2 []byte) (output []byte) {
|
||||
src1Len := len(src1)
|
||||
src2Len := len(src2)
|
||||
output = make([]byte, src1Len+src2Len-1)
|
||||
|
||||
var mergeByte uint8 = src1[src1Len-1]
|
||||
mergeByte |= uint8(src2[0])
|
||||
|
||||
copy(output, src1)
|
||||
copy(output[src1Len:], src2[1:])
|
||||
|
||||
output[src1Len-1] = byte(mergeByte)
|
||||
|
||||
return
|
||||
}
|
8
vendor/github.com/couchbase/gomemcached/mc_constants.go
generated
vendored
8
vendor/github.com/couchbase/gomemcached/mc_constants.go
generated
vendored
@ -6,8 +6,10 @@ import (
|
||||
)
|
||||
|
||||
const (
|
||||
REQ_MAGIC = 0x80
|
||||
RES_MAGIC = 0x81
|
||||
REQ_MAGIC = 0x80
|
||||
RES_MAGIC = 0x81
|
||||
FLEX_MAGIC = 0x08
|
||||
FLEX_RES_MAGIC = 0x18
|
||||
)
|
||||
|
||||
// CommandCode for memcached packets.
|
||||
@ -99,6 +101,8 @@ const (
|
||||
SUBDOC_GET = CommandCode(0xc5) // Get subdoc. Returns with xattrs
|
||||
SUBDOC_MULTI_LOOKUP = CommandCode(0xd0) // Multi lookup. Doc xattrs and meta.
|
||||
|
||||
DCP_SYSTEM_EVENT = CommandCode(0x5f) // A system event has occurred
|
||||
|
||||
)
|
||||
|
||||
// command codes that are counted toward DCP control buffer
|
||||
|
359
vendor/github.com/couchbase/gomemcached/mc_req.go
generated
vendored
359
vendor/github.com/couchbase/gomemcached/mc_req.go
generated
vendored
@ -25,11 +25,17 @@ type MCRequest struct {
|
||||
Extras, Key, Body, ExtMeta []byte
|
||||
// Datatype identifier
|
||||
DataType uint8
|
||||
// len() calls are expensive - cache this in case for collection
|
||||
Keylen int
|
||||
// Flexible Framing Extras
|
||||
FramingExtras []FrameInfo
|
||||
// Stored length of incoming framing extras
|
||||
FramingElen int
|
||||
}
|
||||
|
||||
// Size gives the number of bytes this request requires.
|
||||
func (req *MCRequest) Size() int {
|
||||
return HDR_LEN + len(req.Extras) + len(req.Key) + len(req.Body) + len(req.ExtMeta)
|
||||
return HDR_LEN + len(req.Extras) + len(req.Key) + len(req.Body) + len(req.ExtMeta) + req.FramingElen
|
||||
}
|
||||
|
||||
// A debugging string representation of this request
|
||||
@ -38,7 +44,23 @@ func (req MCRequest) String() string {
|
||||
req.Opcode, len(req.Body), req.Key)
|
||||
}
|
||||
|
||||
func (req *MCRequest) fillHeaderBytes(data []byte) int {
|
||||
func (req *MCRequest) fillRegularHeaderBytes(data []byte) int {
|
||||
// Byte/ 0 | 1 | 2 | 3 |
|
||||
// / | | | |
|
||||
// |0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|
|
||||
// +---------------+---------------+---------------+---------------+
|
||||
// 0| Magic | Opcode | Key length |
|
||||
// +---------------+---------------+---------------+---------------+
|
||||
// 4| Extras length | Data type | vbucket id |
|
||||
// +---------------+---------------+---------------+---------------+
|
||||
// 8| Total body length |
|
||||
// +---------------+---------------+---------------+---------------+
|
||||
// 12| Opaque |
|
||||
// +---------------+---------------+---------------+---------------+
|
||||
// 16| CAS |
|
||||
// | |
|
||||
// +---------------+---------------+---------------+---------------+
|
||||
// Total 24 bytes
|
||||
|
||||
pos := 0
|
||||
data[pos] = REQ_MAGIC
|
||||
@ -84,16 +106,130 @@ func (req *MCRequest) fillHeaderBytes(data []byte) int {
|
||||
copy(data[pos:pos+len(req.Key)], req.Key)
|
||||
pos += len(req.Key)
|
||||
}
|
||||
|
||||
return pos
|
||||
}
|
||||
|
||||
// Returns pos and if trailing by half byte
|
||||
func (req *MCRequest) fillFlexHeaderBytes(data []byte) (int, bool) {
|
||||
|
||||
// Byte/ 0 | 1 | 2 | 3 |
|
||||
// / | | | |
|
||||
// |0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|
|
||||
// +---------------+---------------+---------------+---------------+
|
||||
// 0| Magic (0x08) | Opcode | Framing extras| Key Length |
|
||||
// +---------------+---------------+---------------+---------------+
|
||||
// 4| Extras length | Data type | vbucket id |
|
||||
// +---------------+---------------+---------------+---------------+
|
||||
// 8| Total body length |
|
||||
// +---------------+---------------+---------------+---------------+
|
||||
// 12| Opaque |
|
||||
// +---------------+---------------+---------------+---------------+
|
||||
// 16| CAS |
|
||||
// | |
|
||||
// +---------------+---------------+---------------+---------------+
|
||||
// Total 24 bytes
|
||||
|
||||
data[0] = FLEX_MAGIC
|
||||
data[1] = byte(req.Opcode)
|
||||
data[2] = byte(req.FramingElen)
|
||||
data[3] = byte(req.Keylen)
|
||||
elen := len(req.Extras)
|
||||
data[4] = byte(elen)
|
||||
if req.DataType != 0 {
|
||||
data[5] = byte(req.DataType)
|
||||
}
|
||||
binary.BigEndian.PutUint16(data[6:8], req.VBucket)
|
||||
binary.BigEndian.PutUint32(data[8:12],
|
||||
uint32(len(req.Body)+req.Keylen+elen+len(req.ExtMeta)+req.FramingElen))
|
||||
binary.BigEndian.PutUint32(data[12:16], req.Opaque)
|
||||
if req.Cas != 0 {
|
||||
binary.BigEndian.PutUint64(data[16:24], req.Cas)
|
||||
}
|
||||
pos := HDR_LEN
|
||||
|
||||
// Add framing infos
|
||||
var framingExtras []byte
|
||||
var outputBytes []byte
|
||||
var mergeModeSrc []byte
|
||||
var frameBytes int
|
||||
var halfByteMode bool
|
||||
var mergeMode bool
|
||||
for _, frameInfo := range req.FramingExtras {
|
||||
if !mergeMode {
|
||||
outputBytes, halfByteMode = frameInfo.Bytes()
|
||||
if !halfByteMode {
|
||||
framingExtras = append(framingExtras, outputBytes...)
|
||||
frameBytes += len(outputBytes)
|
||||
} else {
|
||||
mergeMode = true
|
||||
mergeModeSrc = outputBytes
|
||||
}
|
||||
} else {
|
||||
outputBytes, halfByteMode = frameInfo.Bytes()
|
||||
outputBytes := ShiftByteSliceRight4Bits(outputBytes)
|
||||
if halfByteMode {
|
||||
// Previous halfbyte merge with this halfbyte will result in a complete byte
|
||||
mergeMode = false
|
||||
outputBytes = Merge2HalfByteSlices(mergeModeSrc, outputBytes)
|
||||
framingExtras = append(framingExtras, outputBytes...)
|
||||
frameBytes += len(outputBytes)
|
||||
} else {
|
||||
// Merge half byte with a non-half byte will result in a combined half-byte that will
|
||||
// become the source for the next iteration
|
||||
mergeModeSrc = Merge2HalfByteSlices(mergeModeSrc, outputBytes)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if mergeMode {
|
||||
// Commit the temporary merge area into framingExtras
|
||||
framingExtras = append(framingExtras, mergeModeSrc...)
|
||||
frameBytes += len(mergeModeSrc)
|
||||
}
|
||||
|
||||
copy(data[pos:pos+frameBytes], framingExtras)
|
||||
|
||||
pos += frameBytes
|
||||
|
||||
// Add Extras
|
||||
if len(req.Extras) > 0 {
|
||||
if mergeMode {
|
||||
outputBytes = ShiftByteSliceRight4Bits(req.Extras)
|
||||
data = Merge2HalfByteSlices(data, outputBytes)
|
||||
} else {
|
||||
copy(data[pos:pos+elen], req.Extras)
|
||||
}
|
||||
pos += elen
|
||||
}
|
||||
|
||||
// Add keys
|
||||
if req.Keylen > 0 {
|
||||
if mergeMode {
|
||||
outputBytes = ShiftByteSliceRight4Bits(req.Key)
|
||||
data = Merge2HalfByteSlices(data, outputBytes)
|
||||
} else {
|
||||
copy(data[pos:pos+req.Keylen], req.Key)
|
||||
}
|
||||
pos += req.Keylen
|
||||
}
|
||||
|
||||
return pos, mergeMode
|
||||
}
|
||||
|
||||
func (req *MCRequest) FillHeaderBytes(data []byte) (int, bool) {
|
||||
if req.FramingElen == 0 {
|
||||
return req.fillRegularHeaderBytes(data), false
|
||||
} else {
|
||||
return req.fillFlexHeaderBytes(data)
|
||||
}
|
||||
}
|
||||
|
||||
// HeaderBytes will return the wire representation of the request header
|
||||
// (with the extras and key).
|
||||
func (req *MCRequest) HeaderBytes() []byte {
|
||||
data := make([]byte, HDR_LEN+len(req.Extras)+len(req.Key))
|
||||
data := make([]byte, HDR_LEN+len(req.Extras)+len(req.Key)+req.FramingElen)
|
||||
|
||||
req.fillHeaderBytes(data)
|
||||
req.FillHeaderBytes(data)
|
||||
|
||||
return data
|
||||
}
|
||||
@ -102,16 +238,27 @@ func (req *MCRequest) HeaderBytes() []byte {
|
||||
func (req *MCRequest) Bytes() []byte {
|
||||
data := make([]byte, req.Size())
|
||||
|
||||
pos := req.fillHeaderBytes(data)
|
||||
pos, halfByteMode := req.FillHeaderBytes(data)
|
||||
// TODO - the halfByteMode should be revisited for a more efficient
|
||||
// way of doing things
|
||||
|
||||
if len(req.Body) > 0 {
|
||||
copy(data[pos:pos+len(req.Body)], req.Body)
|
||||
if halfByteMode {
|
||||
shifted := ShiftByteSliceRight4Bits(req.Body)
|
||||
data = Merge2HalfByteSlices(data, shifted)
|
||||
} else {
|
||||
copy(data[pos:pos+len(req.Body)], req.Body)
|
||||
}
|
||||
}
|
||||
|
||||
if len(req.ExtMeta) > 0 {
|
||||
copy(data[pos+len(req.Body):pos+len(req.Body)+len(req.ExtMeta)], req.ExtMeta)
|
||||
if halfByteMode {
|
||||
shifted := ShiftByteSliceRight4Bits(req.ExtMeta)
|
||||
data = Merge2HalfByteSlices(data, shifted)
|
||||
} else {
|
||||
copy(data[pos+len(req.Body):pos+len(req.Body)+len(req.ExtMeta)], req.ExtMeta)
|
||||
}
|
||||
}
|
||||
|
||||
return data
|
||||
}
|
||||
|
||||
@ -130,40 +277,44 @@ func (req *MCRequest) Transmit(w io.Writer) (n int, err error) {
|
||||
return
|
||||
}
|
||||
|
||||
// Receive will fill this MCRequest with the data from a reader.
|
||||
func (req *MCRequest) Receive(r io.Reader, hdrBytes []byte) (int, error) {
|
||||
if len(hdrBytes) < HDR_LEN {
|
||||
hdrBytes = []byte{
|
||||
0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0}
|
||||
}
|
||||
n, err := io.ReadFull(r, hdrBytes)
|
||||
if err != nil {
|
||||
return n, err
|
||||
}
|
||||
|
||||
if hdrBytes[0] != RES_MAGIC && hdrBytes[0] != REQ_MAGIC {
|
||||
return n, fmt.Errorf("bad magic: 0x%02x", hdrBytes[0])
|
||||
}
|
||||
|
||||
klen := int(binary.BigEndian.Uint16(hdrBytes[2:]))
|
||||
elen := int(hdrBytes[4])
|
||||
func (req *MCRequest) receiveHeaderCommon(hdrBytes []byte) (elen, totalBodyLen int) {
|
||||
elen = int(hdrBytes[4])
|
||||
// Data type at 5
|
||||
req.DataType = uint8(hdrBytes[5])
|
||||
|
||||
req.Opcode = CommandCode(hdrBytes[1])
|
||||
// Vbucket at 6:7
|
||||
req.VBucket = binary.BigEndian.Uint16(hdrBytes[6:])
|
||||
totalBodyLen := int(binary.BigEndian.Uint32(hdrBytes[8:]))
|
||||
totalBodyLen = int(binary.BigEndian.Uint32(hdrBytes[8:]))
|
||||
|
||||
req.Opaque = binary.BigEndian.Uint32(hdrBytes[12:])
|
||||
req.Cas = binary.BigEndian.Uint64(hdrBytes[16:])
|
||||
return
|
||||
}
|
||||
|
||||
func (req *MCRequest) receiveRegHeader(hdrBytes []byte) (elen, totalBodyLen int) {
|
||||
elen, totalBodyLen = req.receiveHeaderCommon(hdrBytes)
|
||||
req.Keylen = int(binary.BigEndian.Uint16(hdrBytes[2:]))
|
||||
return
|
||||
}
|
||||
|
||||
func (req *MCRequest) receiveFlexibleFramingHeader(hdrBytes []byte) (elen, totalBodyLen, framingElen int) {
|
||||
elen, totalBodyLen = req.receiveHeaderCommon(hdrBytes)
|
||||
|
||||
// For flexible framing header, key length is a single byte at byte index 3
|
||||
req.Keylen = int(binary.BigEndian.Uint16(hdrBytes[2:]) & 0x0ff)
|
||||
// Flexible framing lengh is a single byte at index 2
|
||||
framingElen = int(binary.BigEndian.Uint16(hdrBytes[2:]) >> 8)
|
||||
req.FramingElen = framingElen
|
||||
return
|
||||
}
|
||||
|
||||
func (req *MCRequest) populateRegularBody(r io.Reader, totalBodyLen, elen int) (int, error) {
|
||||
var m int
|
||||
var err error
|
||||
if totalBodyLen > 0 {
|
||||
buf := make([]byte, totalBodyLen)
|
||||
m, err := io.ReadFull(r, buf)
|
||||
n += m
|
||||
m, err = io.ReadFull(r, buf)
|
||||
if err == nil {
|
||||
if req.Opcode >= TAP_MUTATION &&
|
||||
req.Opcode <= TAP_CHECKPOINT_END &&
|
||||
@ -175,7 +326,7 @@ func (req *MCRequest) Receive(r io.Reader, hdrBytes []byte) (int, error) {
|
||||
}
|
||||
|
||||
req.Extras = buf[0:elen]
|
||||
req.Key = buf[elen : klen+elen]
|
||||
req.Key = buf[elen : req.Keylen+elen]
|
||||
|
||||
// get the length of extended metadata
|
||||
extMetaLen := 0
|
||||
@ -183,15 +334,149 @@ func (req *MCRequest) Receive(r io.Reader, hdrBytes []byte) (int, error) {
|
||||
extMetaLen = int(binary.BigEndian.Uint16(req.Extras[28:30]))
|
||||
}
|
||||
|
||||
bodyLen := totalBodyLen - klen - elen - extMetaLen
|
||||
bodyLen := totalBodyLen - req.Keylen - elen - extMetaLen
|
||||
if bodyLen > MaxBodyLen {
|
||||
return n, fmt.Errorf("%d is too big (max %d)",
|
||||
return m, fmt.Errorf("%d is too big (max %d)",
|
||||
bodyLen, MaxBodyLen)
|
||||
}
|
||||
|
||||
req.Body = buf[klen+elen : klen+elen+bodyLen]
|
||||
req.ExtMeta = buf[klen+elen+bodyLen:]
|
||||
req.Body = buf[req.Keylen+elen : req.Keylen+elen+bodyLen]
|
||||
req.ExtMeta = buf[req.Keylen+elen+bodyLen:]
|
||||
}
|
||||
}
|
||||
return n, err
|
||||
return m, err
|
||||
}
|
||||
|
||||
func (req *MCRequest) populateFlexBody(r io.Reader, totalBodyLen, elen, framingElen int) (int, error) {
|
||||
var m int
|
||||
var err error
|
||||
if totalBodyLen > 0 {
|
||||
buf := make([]byte, totalBodyLen)
|
||||
m, err = io.ReadFull(r, buf)
|
||||
if err != nil {
|
||||
return m, err
|
||||
}
|
||||
err = req.populateFlexBodyInternal(buf, totalBodyLen, elen, framingElen)
|
||||
}
|
||||
return m, err
|
||||
}
|
||||
|
||||
func (req *MCRequest) populateFlexBodyInternal(buf []byte, totalBodyLen, elen, framingElen int) error {
|
||||
var halfByteOffset bool
|
||||
var err error
|
||||
if framingElen > 0 {
|
||||
var objs []FrameInfo
|
||||
objs, err, halfByteOffset = parseFrameInfoObjects(buf, framingElen)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
req.FramingExtras = objs
|
||||
}
|
||||
|
||||
err = req.populateFlexBodyAfterFrames(buf, totalBodyLen, elen, framingElen, halfByteOffset)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func (req *MCRequest) populateFlexBodyAfterFrames(buf []byte, totalBodyLen, elen, framingElen int, halfByteOffset bool) error {
|
||||
var idxCursor int = framingElen
|
||||
if req.Opcode >= TAP_MUTATION && req.Opcode <= TAP_CHECKPOINT_END && len(buf[idxCursor:]) > 1 {
|
||||
// In these commands there is "engine private"
|
||||
// data at the end of the extras. The first 2
|
||||
// bytes of extra data give its length.
|
||||
if !halfByteOffset {
|
||||
elen += int(binary.BigEndian.Uint16(buf[idxCursor:]))
|
||||
} else {
|
||||
// 0 1 2 3 4 .... 19 20 21 22 ... 32
|
||||
// ^-----^ ^-------^ ^------------^
|
||||
// offset data do not care
|
||||
var buffer uint32 = binary.BigEndian.Uint32(buf[idxCursor:])
|
||||
buffer &= 0xffff000
|
||||
elen += int(buffer >> 12)
|
||||
}
|
||||
}
|
||||
|
||||
// Get the extras
|
||||
if !halfByteOffset {
|
||||
req.Extras = buf[idxCursor : idxCursor+elen]
|
||||
} else {
|
||||
preShift := buf[idxCursor : idxCursor+elen+1]
|
||||
req.Extras = ShiftByteSliceLeft4Bits(preShift)
|
||||
}
|
||||
idxCursor += elen
|
||||
|
||||
// Get the Key
|
||||
if !halfByteOffset {
|
||||
req.Key = buf[idxCursor : idxCursor+req.Keylen]
|
||||
} else {
|
||||
preShift := buf[idxCursor : idxCursor+req.Keylen+1]
|
||||
req.Key = ShiftByteSliceLeft4Bits(preShift)
|
||||
}
|
||||
idxCursor += req.Keylen
|
||||
|
||||
// get the length of extended metadata
|
||||
extMetaLen := 0
|
||||
if elen > 29 {
|
||||
extMetaLen = int(binary.BigEndian.Uint16(req.Extras[28:30]))
|
||||
}
|
||||
idxCursor += extMetaLen
|
||||
|
||||
bodyLen := totalBodyLen - req.Keylen - elen - extMetaLen - framingElen
|
||||
if bodyLen > MaxBodyLen {
|
||||
return fmt.Errorf("%d is too big (max %d)",
|
||||
bodyLen, MaxBodyLen)
|
||||
}
|
||||
|
||||
if !halfByteOffset {
|
||||
req.Body = buf[idxCursor : idxCursor+bodyLen]
|
||||
idxCursor += bodyLen
|
||||
} else {
|
||||
preShift := buf[idxCursor : idxCursor+bodyLen+1]
|
||||
req.Body = ShiftByteSliceLeft4Bits(preShift)
|
||||
idxCursor += bodyLen
|
||||
}
|
||||
|
||||
if extMetaLen > 0 {
|
||||
if !halfByteOffset {
|
||||
req.ExtMeta = buf[idxCursor:]
|
||||
} else {
|
||||
preShift := buf[idxCursor:]
|
||||
req.ExtMeta = ShiftByteSliceLeft4Bits(preShift)
|
||||
}
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// Receive will fill this MCRequest with the data from a reader.
|
||||
func (req *MCRequest) Receive(r io.Reader, hdrBytes []byte) (int, error) {
|
||||
if len(hdrBytes) < HDR_LEN {
|
||||
hdrBytes = []byte{
|
||||
0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0}
|
||||
}
|
||||
n, err := io.ReadFull(r, hdrBytes)
|
||||
if err != nil {
|
||||
fmt.Printf("Err %v\n", err)
|
||||
return n, err
|
||||
}
|
||||
|
||||
switch hdrBytes[0] {
|
||||
case RES_MAGIC:
|
||||
fallthrough
|
||||
case REQ_MAGIC:
|
||||
elen, totalBodyLen := req.receiveRegHeader(hdrBytes)
|
||||
bodyRead, err := req.populateRegularBody(r, totalBodyLen, elen)
|
||||
return n + bodyRead, err
|
||||
case FLEX_MAGIC:
|
||||
elen, totalBodyLen, framingElen := req.receiveFlexibleFramingHeader(hdrBytes)
|
||||
bodyRead, err := req.populateFlexBody(r, totalBodyLen, elen, framingElen)
|
||||
return n + bodyRead, err
|
||||
default:
|
||||
return n, fmt.Errorf("bad magic: 0x%02x", hdrBytes[0])
|
||||
}
|
||||
}
|
||||
|
15
vendor/github.com/couchbase/gomemcached/mc_res.go
generated
vendored
15
vendor/github.com/couchbase/gomemcached/mc_res.go
generated
vendored
@ -153,6 +153,13 @@ func (res *MCResponse) Transmit(w io.Writer) (n int, err error) {
|
||||
|
||||
// Receive will fill this MCResponse with the data from this reader.
|
||||
func (res *MCResponse) Receive(r io.Reader, hdrBytes []byte) (n int, err error) {
|
||||
return res.ReceiveWithBuf(r, hdrBytes, nil)
|
||||
}
|
||||
|
||||
// ReceiveWithBuf takes an optional pre-allocated []byte buf which
|
||||
// will be used if its capacity is large enough, otherwise a new
|
||||
// []byte slice is allocated.
|
||||
func (res *MCResponse) ReceiveWithBuf(r io.Reader, hdrBytes, buf []byte) (n int, err error) {
|
||||
if len(hdrBytes) < HDR_LEN {
|
||||
hdrBytes = []byte{
|
||||
0, 0, 0, 0, 0, 0, 0, 0,
|
||||
@ -187,7 +194,13 @@ func (res *MCResponse) Receive(r io.Reader, hdrBytes []byte) (n int, err error)
|
||||
}
|
||||
}()
|
||||
|
||||
buf := make([]byte, klen+elen+bodyLen)
|
||||
bufNeed := klen + elen + bodyLen
|
||||
if buf != nil && cap(buf) >= bufNeed {
|
||||
buf = buf[0:bufNeed]
|
||||
} else {
|
||||
buf = make([]byte, bufNeed)
|
||||
}
|
||||
|
||||
m, err := io.ReadFull(r, buf)
|
||||
if err == nil {
|
||||
res.Extras = buf[0:elen]
|
||||
|
235
vendor/github.com/couchbase/goutils/LICENSE.md
generated
vendored
235
vendor/github.com/couchbase/goutils/LICENSE.md
generated
vendored
@ -1,47 +1,202 @@
|
||||
COUCHBASE INC. COMMUNITY EDITION LICENSE AGREEMENT
|
||||
Apache License
|
||||
Version 2.0, January 2004
|
||||
http://www.apache.org/licenses/
|
||||
|
||||
IMPORTANT-READ CAREFULLY: BY CLICKING THE "I ACCEPT" BOX OR INSTALLING,
|
||||
DOWNLOADING OR OTHERWISE USING THIS SOFTWARE AND ANY ASSOCIATED
|
||||
DOCUMENTATION, YOU, ON BEHALF OF YOURSELF OR AS AN AUTHORIZED
|
||||
REPRESENTATIVE ON BEHALF OF AN ENTITY ("LICENSEE") AGREE TO ALL THE
|
||||
TERMS OF THIS COMMUNITY EDITION LICENSE AGREEMENT (THE "AGREEMENT")
|
||||
REGARDING YOUR USE OF THE SOFTWARE. YOU REPRESENT AND WARRANT THAT YOU
|
||||
HAVE FULL LEGAL AUTHORITY TO BIND THE LICENSEE TO THIS AGREEMENT. IF YOU
|
||||
DO NOT AGREE WITH ALL OF THESE TERMS, DO NOT SELECT THE "I ACCEPT" BOX
|
||||
AND DO NOT INSTALL, DOWNLOAD OR OTHERWISE USE THE SOFTWARE. THE
|
||||
EFFECTIVE DATE OF THIS AGREEMENT IS THE DATE ON WHICH YOU CLICK "I
|
||||
ACCEPT" OR OTHERWISE INSTALL, DOWNLOAD OR USE THE SOFTWARE.
|
||||
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
||||
|
||||
1. License Grant. Couchbase Inc. hereby grants Licensee, free of charge,
|
||||
the non-exclusive right to use, copy, merge, publish, distribute,
|
||||
sublicense, and/or sell copies of the Software, and to permit persons to
|
||||
whom the Software is furnished to do so, subject to Licensee including
|
||||
the following copyright notice in all copies or substantial portions of
|
||||
the Software:
|
||||
1. Definitions.
|
||||
|
||||
Couchbase (r) http://www.Couchbase.com Copyright 2016 Couchbase, Inc.
|
||||
"License" shall mean the terms and conditions for use, reproduction,
|
||||
and distribution as defined by Sections 1 through 9 of this document.
|
||||
|
||||
As used in this Agreement, "Software" means the object code version of
|
||||
the applicable elastic data management server software provided by
|
||||
Couchbase Inc.
|
||||
"Licensor" shall mean the copyright owner or entity authorized by
|
||||
the copyright owner that is granting the License.
|
||||
|
||||
2. Restrictions. Licensee will not reverse engineer, disassemble, or
|
||||
decompile the Software (except to the extent such restrictions are
|
||||
prohibited by law).
|
||||
"Legal Entity" shall mean the union of the acting entity and all
|
||||
other entities that control, are controlled by, or are under common
|
||||
control with that entity. For the purposes of this definition,
|
||||
"control" means (i) the power, direct or indirect, to cause the
|
||||
direction or management of such entity, whether by contract or
|
||||
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
||||
outstanding shares, or (iii) beneficial ownership of such entity.
|
||||
|
||||
3. Support. Couchbase, Inc. will provide Licensee with access to, and
|
||||
use of, the Couchbase, Inc. support forum available at the following
|
||||
URL: http://www.couchbase.org/forums/. Couchbase, Inc. may, at its
|
||||
discretion, modify, suspend or terminate support at any time upon notice
|
||||
to Licensee.
|
||||
"You" (or "Your") shall mean an individual or Legal Entity
|
||||
exercising permissions granted by this License.
|
||||
|
||||
"Source" form shall mean the preferred form for making modifications,
|
||||
including but not limited to software source code, documentation
|
||||
source, and configuration files.
|
||||
|
||||
"Object" form shall mean any form resulting from mechanical
|
||||
transformation or translation of a Source form, including but
|
||||
not limited to compiled object code, generated documentation,
|
||||
and conversions to other media types.
|
||||
|
||||
"Work" shall mean the work of authorship, whether in Source or
|
||||
Object form, made available under the License, as indicated by a
|
||||
copyright notice that is included in or attached to the work
|
||||
(an example is provided in the Appendix below).
|
||||
|
||||
"Derivative Works" shall mean any work, whether in Source or Object
|
||||
form, that is based on (or derived from) the Work and for which the
|
||||
editorial revisions, annotations, elaborations, or other modifications
|
||||
represent, as a whole, an original work of authorship. For the purposes
|
||||
of this License, Derivative Works shall not include works that remain
|
||||
separable from, or merely link (or bind by name) to the interfaces of,
|
||||
the Work and Derivative Works thereof.
|
||||
|
||||
"Contribution" shall mean any work of authorship, including
|
||||
the original version of the Work and any modifications or additions
|
||||
to that Work or Derivative Works thereof, that is intentionally
|
||||
submitted to Licensor for inclusion in the Work by the copyright owner
|
||||
or by an individual or Legal Entity authorized to submit on behalf of
|
||||
the copyright owner. For the purposes of this definition, "submitted"
|
||||
means any form of electronic, verbal, or written communication sent
|
||||
to the Licensor or its representatives, including but not limited to
|
||||
communication on electronic mailing lists, source code control systems,
|
||||
and issue tracking systems that are managed by, or on behalf of, the
|
||||
Licensor for the purpose of discussing and improving the Work, but
|
||||
excluding communication that is conspicuously marked or otherwise
|
||||
designated in writing by the copyright owner as "Not a Contribution."
|
||||
|
||||
"Contributor" shall mean Licensor and any individual or Legal Entity
|
||||
on behalf of whom a Contribution has been received by Licensor and
|
||||
subsequently incorporated within the Work.
|
||||
|
||||
2. Grant of Copyright License. Subject to the terms and conditions of
|
||||
this License, each Contributor hereby grants to You a perpetual,
|
||||
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||
copyright license to reproduce, prepare Derivative Works of,
|
||||
publicly display, publicly perform, sublicense, and distribute the
|
||||
Work and such Derivative Works in Source or Object form.
|
||||
|
||||
3. Grant of Patent License. Subject to the terms and conditions of
|
||||
this License, each Contributor hereby grants to You a perpetual,
|
||||
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||
(except as stated in this section) patent license to make, have made,
|
||||
use, offer to sell, sell, import, and otherwise transfer the Work,
|
||||
where such license applies only to those patent claims licensable
|
||||
by such Contributor that are necessarily infringed by their
|
||||
Contribution(s) alone or by combination of their Contribution(s)
|
||||
with the Work to which such Contribution(s) was submitted. If You
|
||||
institute patent litigation against any entity (including a
|
||||
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
||||
or a Contribution incorporated within the Work constitutes direct
|
||||
or contributory patent infringement, then any patent licenses
|
||||
granted to You under this License for that Work shall terminate
|
||||
as of the date such litigation is filed.
|
||||
|
||||
4. Redistribution. You may reproduce and distribute copies of the
|
||||
Work or Derivative Works thereof in any medium, with or without
|
||||
modifications, and in Source or Object form, provided that You
|
||||
meet the following conditions:
|
||||
|
||||
(a) You must give any other recipients of the Work or
|
||||
Derivative Works a copy of this License; and
|
||||
|
||||
(b) You must cause any modified files to carry prominent notices
|
||||
stating that You changed the files; and
|
||||
|
||||
(c) You must retain, in the Source form of any Derivative Works
|
||||
that You distribute, all copyright, patent, trademark, and
|
||||
attribution notices from the Source form of the Work,
|
||||
excluding those notices that do not pertain to any part of
|
||||
the Derivative Works; and
|
||||
|
||||
(d) If the Work includes a "NOTICE" text file as part of its
|
||||
distribution, then any Derivative Works that You distribute must
|
||||
include a readable copy of the attribution notices contained
|
||||
within such NOTICE file, excluding those notices that do not
|
||||
pertain to any part of the Derivative Works, in at least one
|
||||
of the following places: within a NOTICE text file distributed
|
||||
as part of the Derivative Works; within the Source form or
|
||||
documentation, if provided along with the Derivative Works; or,
|
||||
within a display generated by the Derivative Works, if and
|
||||
wherever such third-party notices normally appear. The contents
|
||||
of the NOTICE file are for informational purposes only and
|
||||
do not modify the License. You may add Your own attribution
|
||||
notices within Derivative Works that You distribute, alongside
|
||||
or as an addendum to the NOTICE text from the Work, provided
|
||||
that such additional attribution notices cannot be construed
|
||||
as modifying the License.
|
||||
|
||||
You may add Your own copyright statement to Your modifications and
|
||||
may provide additional or different license terms and conditions
|
||||
for use, reproduction, or distribution of Your modifications, or
|
||||
for any such Derivative Works as a whole, provided Your use,
|
||||
reproduction, and distribution of the Work otherwise complies with
|
||||
the conditions stated in this License.
|
||||
|
||||
5. Submission of Contributions. Unless You explicitly state otherwise,
|
||||
any Contribution intentionally submitted for inclusion in the Work
|
||||
by You to the Licensor shall be under the terms and conditions of
|
||||
this License, without any additional terms or conditions.
|
||||
Notwithstanding the above, nothing herein shall supersede or modify
|
||||
the terms of any separate license agreement you may have executed
|
||||
with Licensor regarding such Contributions.
|
||||
|
||||
6. Trademarks. This License does not grant permission to use the trade
|
||||
names, trademarks, service marks, or product names of the Licensor,
|
||||
except as required for reasonable and customary use in describing the
|
||||
origin of the Work and reproducing the content of the NOTICE file.
|
||||
|
||||
7. Disclaimer of Warranty. Unless required by applicable law or
|
||||
agreed to in writing, Licensor provides the Work (and each
|
||||
Contributor provides its Contributions) on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
implied, including, without limitation, any warranties or conditions
|
||||
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
||||
PARTICULAR PURPOSE. You are solely responsible for determining the
|
||||
appropriateness of using or redistributing the Work and assume any
|
||||
risks associated with Your exercise of permissions under this License.
|
||||
|
||||
8. Limitation of Liability. In no event and under no legal theory,
|
||||
whether in tort (including negligence), contract, or otherwise,
|
||||
unless required by applicable law (such as deliberate and grossly
|
||||
negligent acts) or agreed to in writing, shall any Contributor be
|
||||
liable to You for damages, including any direct, indirect, special,
|
||||
incidental, or consequential damages of any character arising as a
|
||||
result of this License or out of the use or inability to use the
|
||||
Work (including but not limited to damages for loss of goodwill,
|
||||
work stoppage, computer failure or malfunction, or any and all
|
||||
other commercial damages or losses), even if such Contributor
|
||||
has been advised of the possibility of such damages.
|
||||
|
||||
9. Accepting Warranty or Additional Liability. While redistributing
|
||||
the Work or Derivative Works thereof, You may choose to offer,
|
||||
and charge a fee for, acceptance of support, warranty, indemnity,
|
||||
or other liability obligations and/or rights consistent with this
|
||||
License. However, in accepting such obligations, You may act only
|
||||
on Your own behalf and on Your sole responsibility, not on behalf
|
||||
of any other Contributor, and only if You agree to indemnify,
|
||||
defend, and hold each Contributor harmless for any liability
|
||||
incurred by, or claims asserted against, such Contributor by reason
|
||||
of your accepting any such warranty or additional liability.
|
||||
|
||||
END OF TERMS AND CONDITIONS
|
||||
|
||||
APPENDIX: How to apply the Apache License to your work.
|
||||
|
||||
To apply the Apache License to your work, attach the following
|
||||
boilerplate notice, with the fields enclosed by brackets "{}"
|
||||
replaced with your own identifying information. (Don't include
|
||||
the brackets!) The text should be enclosed in the appropriate
|
||||
comment syntax for the file format. We also recommend that a
|
||||
file or class name and description of purpose be included on the
|
||||
same "printed page" as the copyright notice for easier
|
||||
identification within third-party archives.
|
||||
|
||||
Copyright {yyyy} {name of copyright owner}
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
|
||||
4. Warranty Disclaimer and Limitation of Liability. THE SOFTWARE IS
|
||||
PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED,
|
||||
INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS
|
||||
FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||||
COUCHBASE INC. OR THE AUTHORS OR COPYRIGHT HOLDERS IN THE SOFTWARE BE
|
||||
LIABLE FOR ANY CLAIM, DAMAGES (IINCLUDING, WITHOUT LIMITATION, DIRECT,
|
||||
INDIRECT OR CONSEQUENTIAL DAMAGES) OR OTHER LIABILITY, WHETHER IN AN
|
||||
ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
||||
CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
||||
SOFTWARE.
|
||||
|
4
vendor/modules.txt
vendored
4
vendor/modules.txt
vendored
@ -99,10 +99,10 @@ github.com/boombuler/barcode/qr
|
||||
github.com/boombuler/barcode/utils
|
||||
# github.com/bradfitz/gomemcache v0.0.0-20190329173943-551aad21a668
|
||||
github.com/bradfitz/gomemcache/memcache
|
||||
# github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d
|
||||
# github.com/couchbase/gomemcached v0.0.0-20191004160342-7b5da2ec40b2
|
||||
github.com/couchbase/gomemcached
|
||||
github.com/couchbase/gomemcached/client
|
||||
# github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b
|
||||
# github.com/couchbase/goutils v0.0.0-20191018232750-b49639060d85
|
||||
github.com/couchbase/goutils/logging
|
||||
github.com/couchbase/goutils/scramsha
|
||||
# github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd
|
||||
|
Loading…
Reference in New Issue
Block a user