mirror of https://github.com/go-gitea/gitea.git
chore: update github.com/couchbase/gomemcached and github.com/couchbase/go-couchbase (#9419)
This commit is contained in:
parent
8873a80276
commit
559fb6ccf0
2
go.mod
2
go.mod
|
@ -24,6 +24,8 @@ require (
|
||||||
github.com/blevesearch/go-porterstemmer v1.0.2 // indirect
|
github.com/blevesearch/go-porterstemmer v1.0.2 // indirect
|
||||||
github.com/blevesearch/segment v0.0.0-20160915185041-762005e7a34f // indirect
|
github.com/blevesearch/segment v0.0.0-20160915185041-762005e7a34f // indirect
|
||||||
github.com/boombuler/barcode v0.0.0-20161226211916-fe0f26ff6d26 // indirect
|
github.com/boombuler/barcode v0.0.0-20161226211916-fe0f26ff6d26 // indirect
|
||||||
|
github.com/couchbase/gomemcached v0.0.0-20191004160342-7b5da2ec40b2 // indirect
|
||||||
|
github.com/couchbase/goutils v0.0.0-20191018232750-b49639060d85 // indirect
|
||||||
github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd // indirect
|
github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd // indirect
|
||||||
github.com/cznic/b v0.0.0-20181122101859-a26611c4d92d // indirect
|
github.com/cznic/b v0.0.0-20181122101859-a26611c4d92d // indirect
|
||||||
github.com/cznic/mathutil v0.0.0-20181122101859-297441e03548 // indirect
|
github.com/cznic/mathutil v0.0.0-20181122101859-297441e03548 // indirect
|
||||||
|
|
4
go.sum
4
go.sum
|
@ -97,8 +97,12 @@ github.com/coreos/go-systemd v0.0.0-20190321100706-95778dfbb74e/go.mod h1:F5haX7
|
||||||
github.com/coreos/pkg v0.0.0-20180928190104-399ea9e2e55f/go.mod h1:E3G3o1h8I7cfcXa63jLwjI0eiQQMgzzUDFVpN/nH/eA=
|
github.com/coreos/pkg v0.0.0-20180928190104-399ea9e2e55f/go.mod h1:E3G3o1h8I7cfcXa63jLwjI0eiQQMgzzUDFVpN/nH/eA=
|
||||||
github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d h1:XMf4E1U+b9E3ElF0mjvfXZdflBRZz4gLp16nQ/QSHQM=
|
github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d h1:XMf4E1U+b9E3ElF0mjvfXZdflBRZz4gLp16nQ/QSHQM=
|
||||||
github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d/go.mod h1:srVSlQLB8iXBVXHgnqemxUXqN6FCvClgCMPCsjBDR7c=
|
github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d/go.mod h1:srVSlQLB8iXBVXHgnqemxUXqN6FCvClgCMPCsjBDR7c=
|
||||||
|
github.com/couchbase/gomemcached v0.0.0-20191004160342-7b5da2ec40b2 h1:vZryARwW4PSFXd9arwegEywvMTvPuXL3/oa+4L5NTe8=
|
||||||
|
github.com/couchbase/gomemcached v0.0.0-20191004160342-7b5da2ec40b2/go.mod h1:srVSlQLB8iXBVXHgnqemxUXqN6FCvClgCMPCsjBDR7c=
|
||||||
github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b h1:bZ9rKU2/V8sY+NulSfxDOnXTWcs1rySqdF1sVepihvo=
|
github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b h1:bZ9rKU2/V8sY+NulSfxDOnXTWcs1rySqdF1sVepihvo=
|
||||||
github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b/go.mod h1:BQwMFlJzDjFDG3DJUdU0KORxn88UlsOULuxLExMh3Hs=
|
github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b/go.mod h1:BQwMFlJzDjFDG3DJUdU0KORxn88UlsOULuxLExMh3Hs=
|
||||||
|
github.com/couchbase/goutils v0.0.0-20191018232750-b49639060d85 h1:0WMIDtuXCKEm4wtAJgAAXa/qtM5O9MariLwgHaRlYmk=
|
||||||
|
github.com/couchbase/goutils v0.0.0-20191018232750-b49639060d85/go.mod h1:BQwMFlJzDjFDG3DJUdU0KORxn88UlsOULuxLExMh3Hs=
|
||||||
github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd h1:zeuJhcG3f8eePshH3KxkNE+Xtl53pVln9MOUPMyr/1w=
|
github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd h1:zeuJhcG3f8eePshH3KxkNE+Xtl53pVln9MOUPMyr/1w=
|
||||||
github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd/go.mod h1:xbc8Ff/oG7h2ejd7AlwOpfd+6QZntc92ygpAOfGwcKY=
|
github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd/go.mod h1:xbc8Ff/oG7h2ejd7AlwOpfd+6QZntc92ygpAOfGwcKY=
|
||||||
github.com/couchbaselabs/go-couchbase v0.0.0-20190708161019-23e7ca2ce2b7 h1:1XjEY/gnjQ+AfXef2U6dxCquhiRzkEpxZuWqs+QxTL8=
|
github.com/couchbaselabs/go-couchbase v0.0.0-20190708161019-23e7ca2ce2b7 h1:1XjEY/gnjQ+AfXef2U6dxCquhiRzkEpxZuWqs+QxTL8=
|
||||||
|
|
123
vendor/github.com/couchbase/gomemcached/client/collections_filter.go
generated
vendored
Normal file
123
vendor/github.com/couchbase/gomemcached/client/collections_filter.go
generated
vendored
Normal file
|
@ -0,0 +1,123 @@
|
||||||
|
package memcached
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Collection based filter
|
||||||
|
type CollectionsFilter struct {
|
||||||
|
ManifestUid uint64
|
||||||
|
UseManifestUid bool
|
||||||
|
StreamId uint16
|
||||||
|
UseStreamId bool
|
||||||
|
|
||||||
|
// Use either ScopeId OR CollectionsList, not both
|
||||||
|
CollectionsList []uint32
|
||||||
|
ScopeId uint32
|
||||||
|
}
|
||||||
|
|
||||||
|
type nonStreamIdNonResumeScopeMeta struct {
|
||||||
|
ScopeId string `json:"scope"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type nonStreamIdResumeScopeMeta struct {
|
||||||
|
ManifestId string `json:"uid"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type nonStreamIdNonResumeCollectionsMeta struct {
|
||||||
|
CollectionsList []string `json:"collections"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type nonStreamIdResumeCollectionsMeta struct {
|
||||||
|
ManifestId string `json:"uid"`
|
||||||
|
CollectionsList []string `json:"collections"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type streamIdNonResumeCollectionsMeta struct {
|
||||||
|
CollectionsList []string `json:"collections"`
|
||||||
|
StreamId uint16 `json:"sid"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type streamIdNonResumeScopeMeta struct {
|
||||||
|
ScopeId string `json:"scope"`
|
||||||
|
StreamId uint16 `json:"sid"`
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *CollectionsFilter) IsValid() error {
|
||||||
|
if c.UseManifestUid {
|
||||||
|
return fmt.Errorf("Not implemented yet")
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(c.CollectionsList) > 0 && c.ScopeId > 0 {
|
||||||
|
return fmt.Errorf("Collection list is specified but scope ID is also specified")
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *CollectionsFilter) outputCollectionsFilterColList() (outputList []string) {
|
||||||
|
for _, collectionUint := range c.CollectionsList {
|
||||||
|
outputList = append(outputList, fmt.Sprintf("%x", collectionUint))
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *CollectionsFilter) outputScopeId() string {
|
||||||
|
return fmt.Sprintf("%x", c.ScopeId)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *CollectionsFilter) ToStreamReqBody() ([]byte, error) {
|
||||||
|
if err := c.IsValid(); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
var output interface{}
|
||||||
|
|
||||||
|
switch c.UseStreamId {
|
||||||
|
case true:
|
||||||
|
switch c.UseManifestUid {
|
||||||
|
case true:
|
||||||
|
// TODO
|
||||||
|
return nil, fmt.Errorf("NotImplemented0")
|
||||||
|
case false:
|
||||||
|
switch len(c.CollectionsList) > 0 {
|
||||||
|
case true:
|
||||||
|
filter := &streamIdNonResumeCollectionsMeta{
|
||||||
|
StreamId: c.StreamId,
|
||||||
|
CollectionsList: c.outputCollectionsFilterColList(),
|
||||||
|
}
|
||||||
|
output = *filter
|
||||||
|
case false:
|
||||||
|
filter := &streamIdNonResumeScopeMeta{
|
||||||
|
StreamId: c.StreamId,
|
||||||
|
ScopeId: c.outputScopeId(),
|
||||||
|
}
|
||||||
|
output = *filter
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case false:
|
||||||
|
switch c.UseManifestUid {
|
||||||
|
case true:
|
||||||
|
// TODO
|
||||||
|
return nil, fmt.Errorf("NotImplemented1")
|
||||||
|
case false:
|
||||||
|
switch len(c.CollectionsList) > 0 {
|
||||||
|
case true:
|
||||||
|
filter := &nonStreamIdNonResumeCollectionsMeta{
|
||||||
|
CollectionsList: c.outputCollectionsFilterColList(),
|
||||||
|
}
|
||||||
|
output = *filter
|
||||||
|
case false:
|
||||||
|
output = nonStreamIdNonResumeScopeMeta{ScopeId: c.outputScopeId()}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
data, err := json.Marshal(output)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
} else {
|
||||||
|
return data, nil
|
||||||
|
}
|
||||||
|
}
|
|
@ -28,10 +28,12 @@ type ClientIface interface {
|
||||||
CASNext(vb uint16, k string, exp int, state *CASState) bool
|
CASNext(vb uint16, k string, exp int, state *CASState) bool
|
||||||
CAS(vb uint16, k string, f CasFunc, initexp int) (*gomemcached.MCResponse, error)
|
CAS(vb uint16, k string, f CasFunc, initexp int) (*gomemcached.MCResponse, error)
|
||||||
CollectionsGetCID(scope string, collection string) (*gomemcached.MCResponse, error)
|
CollectionsGetCID(scope string, collection string) (*gomemcached.MCResponse, error)
|
||||||
|
CollectionEnabled() bool
|
||||||
Close() error
|
Close() error
|
||||||
Decr(vb uint16, key string, amt, def uint64, exp int) (uint64, error)
|
Decr(vb uint16, key string, amt, def uint64, exp int) (uint64, error)
|
||||||
Del(vb uint16, key string) (*gomemcached.MCResponse, error)
|
Del(vb uint16, key string) (*gomemcached.MCResponse, error)
|
||||||
EnableMutationToken() (*gomemcached.MCResponse, error)
|
EnableMutationToken() (*gomemcached.MCResponse, error)
|
||||||
|
EnableFeatures(features Features) (*gomemcached.MCResponse, error)
|
||||||
Get(vb uint16, key string) (*gomemcached.MCResponse, error)
|
Get(vb uint16, key string) (*gomemcached.MCResponse, error)
|
||||||
GetCollectionsManifest() (*gomemcached.MCResponse, error)
|
GetCollectionsManifest() (*gomemcached.MCResponse, error)
|
||||||
GetFromCollection(vb uint16, cid uint32, key string) (*gomemcached.MCResponse, error)
|
GetFromCollection(vb uint16, cid uint32, key string) (*gomemcached.MCResponse, error)
|
||||||
|
@ -76,9 +78,12 @@ var Healthy uint32 = 1
|
||||||
type Features []Feature
|
type Features []Feature
|
||||||
type Feature uint16
|
type Feature uint16
|
||||||
|
|
||||||
const FeatureMutationToken = Feature(0x04)
|
const FeatureTcpNoDelay = Feature(0x03)
|
||||||
|
const FeatureMutationToken = Feature(0x04) // XATTR bit in data type field with dcp mutations
|
||||||
const FeatureXattr = Feature(0x06)
|
const FeatureXattr = Feature(0x06)
|
||||||
|
const FeatureXerror = Feature(0x07)
|
||||||
const FeatureCollections = Feature(0x12)
|
const FeatureCollections = Feature(0x12)
|
||||||
|
const FeatureSnappyCompression = Feature(0x0a)
|
||||||
const FeatureDataType = Feature(0x0b)
|
const FeatureDataType = Feature(0x0b)
|
||||||
|
|
||||||
type memcachedConnection interface {
|
type memcachedConnection interface {
|
||||||
|
@ -96,6 +101,9 @@ type Client struct {
|
||||||
opaque uint32
|
opaque uint32
|
||||||
|
|
||||||
hdrBuf []byte
|
hdrBuf []byte
|
||||||
|
|
||||||
|
featureMtx sync.RWMutex
|
||||||
|
sentHeloFeatures Features
|
||||||
}
|
}
|
||||||
|
|
||||||
var (
|
var (
|
||||||
|
@ -285,6 +293,10 @@ func (c *Client) EnableFeatures(features Features) (*gomemcached.MCResponse, err
|
||||||
binary.BigEndian.PutUint16(payload[len(payload)-2:], uint16(feature))
|
binary.BigEndian.PutUint16(payload[len(payload)-2:], uint16(feature))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
c.featureMtx.Lock()
|
||||||
|
c.sentHeloFeatures = features
|
||||||
|
c.featureMtx.Unlock()
|
||||||
|
|
||||||
return c.Send(&gomemcached.MCRequest{
|
return c.Send(&gomemcached.MCRequest{
|
||||||
Opcode: gomemcached.HELLO,
|
Opcode: gomemcached.HELLO,
|
||||||
Key: []byte("GoMemcached"),
|
Key: []byte("GoMemcached"),
|
||||||
|
@ -363,6 +375,18 @@ func (c *Client) CollectionsGetCID(scope string, collection string) (*gomemcache
|
||||||
return res, nil
|
return res, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (c *Client) CollectionEnabled() bool {
|
||||||
|
c.featureMtx.RLock()
|
||||||
|
defer c.featureMtx.RUnlock()
|
||||||
|
|
||||||
|
for _, feature := range c.sentHeloFeatures {
|
||||||
|
if feature == FeatureCollections {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
// Get the value for a key, and update expiry
|
// Get the value for a key, and update expiry
|
||||||
func (c *Client) GetAndTouch(vb uint16, key string, exp int) (*gomemcached.MCResponse, error) {
|
func (c *Client) GetAndTouch(vb uint16, key string, exp int) (*gomemcached.MCResponse, error) {
|
||||||
extraBuf := make([]byte, 4)
|
extraBuf := make([]byte, 4)
|
||||||
|
@ -1138,3 +1162,7 @@ func IfResStatusError(response *gomemcached.MCResponse) bool {
|
||||||
response.Status != gomemcached.SUBDOC_PATH_NOT_FOUND &&
|
response.Status != gomemcached.SUBDOC_PATH_NOT_FOUND &&
|
||||||
response.Status != gomemcached.SUBDOC_MULTI_PATH_FAILURE_DELETED)
|
response.Status != gomemcached.SUBDOC_MULTI_PATH_FAILURE_DELETED)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (c *Client) Conn() io.ReadWriteCloser {
|
||||||
|
return c.conn
|
||||||
|
}
|
||||||
|
|
|
@ -0,0 +1,346 @@
|
||||||
|
package memcached
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/binary"
|
||||||
|
"fmt"
|
||||||
|
"github.com/couchbase/gomemcached"
|
||||||
|
"math"
|
||||||
|
)
|
||||||
|
|
||||||
|
type SystemEventType int
|
||||||
|
|
||||||
|
const InvalidSysEvent SystemEventType = -1
|
||||||
|
|
||||||
|
const (
|
||||||
|
CollectionCreate SystemEventType = 0
|
||||||
|
CollectionDrop SystemEventType = iota
|
||||||
|
CollectionFlush SystemEventType = iota // KV did not implement
|
||||||
|
ScopeCreate SystemEventType = iota
|
||||||
|
ScopeDrop SystemEventType = iota
|
||||||
|
CollectionChanged SystemEventType = iota
|
||||||
|
)
|
||||||
|
|
||||||
|
type ScopeCreateEvent interface {
|
||||||
|
GetSystemEventName() (string, error)
|
||||||
|
GetScopeId() (uint32, error)
|
||||||
|
GetManifestId() (uint64, error)
|
||||||
|
}
|
||||||
|
|
||||||
|
type CollectionCreateEvent interface {
|
||||||
|
GetSystemEventName() (string, error)
|
||||||
|
GetScopeId() (uint32, error)
|
||||||
|
GetCollectionId() (uint32, error)
|
||||||
|
GetManifestId() (uint64, error)
|
||||||
|
GetMaxTTL() (uint32, error)
|
||||||
|
}
|
||||||
|
|
||||||
|
type CollectionDropEvent interface {
|
||||||
|
GetScopeId() (uint32, error)
|
||||||
|
GetCollectionId() (uint32, error)
|
||||||
|
GetManifestId() (uint64, error)
|
||||||
|
}
|
||||||
|
|
||||||
|
type ScopeDropEvent interface {
|
||||||
|
GetScopeId() (uint32, error)
|
||||||
|
GetManifestId() (uint64, error)
|
||||||
|
}
|
||||||
|
|
||||||
|
type CollectionChangedEvent interface {
|
||||||
|
GetCollectionId() (uint32, error)
|
||||||
|
GetManifestId() (uint64, error)
|
||||||
|
GetMaxTTL() (uint32, error)
|
||||||
|
}
|
||||||
|
|
||||||
|
var ErrorInvalidOp error = fmt.Errorf("Invalid Operation")
|
||||||
|
var ErrorInvalidVersion error = fmt.Errorf("Invalid version for parsing")
|
||||||
|
var ErrorValueTooShort error = fmt.Errorf("Value length is too short")
|
||||||
|
var ErrorNoMaxTTL error = fmt.Errorf("This event has no max TTL")
|
||||||
|
|
||||||
|
// UprEvent memcached events for UPR streams.
|
||||||
|
type UprEvent struct {
|
||||||
|
Opcode gomemcached.CommandCode // Type of event
|
||||||
|
Status gomemcached.Status // Response status
|
||||||
|
VBucket uint16 // VBucket this event applies to
|
||||||
|
DataType uint8 // data type
|
||||||
|
Opaque uint16 // 16 MSB of opaque
|
||||||
|
VBuuid uint64 // This field is set by downstream
|
||||||
|
Flags uint32 // Item flags
|
||||||
|
Expiry uint32 // Item expiration time
|
||||||
|
Key, Value []byte // Item key/value
|
||||||
|
OldValue []byte // TODO: TBD: old document value
|
||||||
|
Cas uint64 // CAS value of the item
|
||||||
|
Seqno uint64 // sequence number of the mutation
|
||||||
|
RevSeqno uint64 // rev sequence number : deletions
|
||||||
|
LockTime uint32 // Lock time
|
||||||
|
MetadataSize uint16 // Metadata size
|
||||||
|
SnapstartSeq uint64 // start sequence number of this snapshot
|
||||||
|
SnapendSeq uint64 // End sequence number of the snapshot
|
||||||
|
SnapshotType uint32 // 0: disk 1: memory
|
||||||
|
FailoverLog *FailoverLog // Failover log containing vvuid and sequnce number
|
||||||
|
Error error // Error value in case of a failure
|
||||||
|
ExtMeta []byte // Extended Metadata
|
||||||
|
AckSize uint32 // The number of bytes that can be Acked to DCP
|
||||||
|
SystemEvent SystemEventType // Only valid if IsSystemEvent() is true
|
||||||
|
SysEventVersion uint8 // Based on the version, the way Extra bytes is parsed is different
|
||||||
|
ValueLen int // Cache it to avoid len() calls for performance
|
||||||
|
CollectionId uint64 // Valid if Collection is in use
|
||||||
|
}
|
||||||
|
|
||||||
|
// FailoverLog containing vvuid and sequnce number
|
||||||
|
type FailoverLog [][2]uint64
|
||||||
|
|
||||||
|
func makeUprEvent(rq gomemcached.MCRequest, stream *UprStream, bytesReceivedFromDCP int) *UprEvent {
|
||||||
|
event := &UprEvent{
|
||||||
|
Opcode: rq.Opcode,
|
||||||
|
VBucket: stream.Vbucket,
|
||||||
|
VBuuid: stream.Vbuuid,
|
||||||
|
Value: rq.Body,
|
||||||
|
Cas: rq.Cas,
|
||||||
|
ExtMeta: rq.ExtMeta,
|
||||||
|
DataType: rq.DataType,
|
||||||
|
ValueLen: len(rq.Body),
|
||||||
|
SystemEvent: InvalidSysEvent,
|
||||||
|
CollectionId: math.MaxUint64,
|
||||||
|
}
|
||||||
|
|
||||||
|
event.PopulateFieldsBasedOnStreamType(rq, stream.StreamType)
|
||||||
|
|
||||||
|
// set AckSize for events that need to be acked to DCP,
|
||||||
|
// i.e., events with CommandCodes that need to be buffered in DCP
|
||||||
|
if _, ok := gomemcached.BufferedCommandCodeMap[rq.Opcode]; ok {
|
||||||
|
event.AckSize = uint32(bytesReceivedFromDCP)
|
||||||
|
}
|
||||||
|
|
||||||
|
// 16 LSBits are used by client library to encode vbucket number.
|
||||||
|
// 16 MSBits are left for application to multiplex on opaque value.
|
||||||
|
event.Opaque = appOpaque(rq.Opaque)
|
||||||
|
|
||||||
|
if len(rq.Extras) >= uprMutationExtraLen &&
|
||||||
|
event.Opcode == gomemcached.UPR_MUTATION {
|
||||||
|
|
||||||
|
event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8])
|
||||||
|
event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16])
|
||||||
|
event.Flags = binary.BigEndian.Uint32(rq.Extras[16:20])
|
||||||
|
event.Expiry = binary.BigEndian.Uint32(rq.Extras[20:24])
|
||||||
|
event.LockTime = binary.BigEndian.Uint32(rq.Extras[24:28])
|
||||||
|
event.MetadataSize = binary.BigEndian.Uint16(rq.Extras[28:30])
|
||||||
|
|
||||||
|
} else if len(rq.Extras) >= uprDeletetionWithDeletionTimeExtraLen &&
|
||||||
|
event.Opcode == gomemcached.UPR_DELETION {
|
||||||
|
|
||||||
|
event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8])
|
||||||
|
event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16])
|
||||||
|
event.Expiry = binary.BigEndian.Uint32(rq.Extras[16:20])
|
||||||
|
|
||||||
|
} else if len(rq.Extras) >= uprDeletetionExtraLen &&
|
||||||
|
event.Opcode == gomemcached.UPR_DELETION ||
|
||||||
|
event.Opcode == gomemcached.UPR_EXPIRATION {
|
||||||
|
|
||||||
|
event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8])
|
||||||
|
event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16])
|
||||||
|
event.MetadataSize = binary.BigEndian.Uint16(rq.Extras[16:18])
|
||||||
|
|
||||||
|
} else if len(rq.Extras) >= uprSnapshotExtraLen &&
|
||||||
|
event.Opcode == gomemcached.UPR_SNAPSHOT {
|
||||||
|
|
||||||
|
event.SnapstartSeq = binary.BigEndian.Uint64(rq.Extras[:8])
|
||||||
|
event.SnapendSeq = binary.BigEndian.Uint64(rq.Extras[8:16])
|
||||||
|
event.SnapshotType = binary.BigEndian.Uint32(rq.Extras[16:20])
|
||||||
|
} else if event.IsSystemEvent() {
|
||||||
|
event.PopulateEvent(rq.Extras)
|
||||||
|
}
|
||||||
|
|
||||||
|
return event
|
||||||
|
}
|
||||||
|
|
||||||
|
func (event *UprEvent) PopulateFieldsBasedOnStreamType(rq gomemcached.MCRequest, streamType DcpStreamType) {
|
||||||
|
switch streamType {
|
||||||
|
case CollectionsNonStreamId:
|
||||||
|
switch rq.Opcode {
|
||||||
|
// Only these will have CID encoded within the key
|
||||||
|
case gomemcached.UPR_MUTATION,
|
||||||
|
gomemcached.UPR_DELETION,
|
||||||
|
gomemcached.UPR_EXPIRATION:
|
||||||
|
uleb128 := Uleb128(rq.Key)
|
||||||
|
result, bytesShifted := uleb128.ToUint64(rq.Keylen)
|
||||||
|
event.CollectionId = result
|
||||||
|
event.Key = rq.Key[bytesShifted:]
|
||||||
|
default:
|
||||||
|
event.Key = rq.Key
|
||||||
|
}
|
||||||
|
case CollectionsStreamId:
|
||||||
|
// TODO - not implemented
|
||||||
|
fallthrough
|
||||||
|
case NonCollectionStream:
|
||||||
|
// Let default behavior be legacy stream type
|
||||||
|
fallthrough
|
||||||
|
default:
|
||||||
|
event.Key = rq.Key
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (event *UprEvent) String() string {
|
||||||
|
name := gomemcached.CommandNames[event.Opcode]
|
||||||
|
if name == "" {
|
||||||
|
name = fmt.Sprintf("#%d", event.Opcode)
|
||||||
|
}
|
||||||
|
return name
|
||||||
|
}
|
||||||
|
|
||||||
|
func (event *UprEvent) IsSnappyDataType() bool {
|
||||||
|
return event.Opcode == gomemcached.UPR_MUTATION && (event.DataType&SnappyDataType > 0)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (event *UprEvent) IsCollectionType() bool {
|
||||||
|
return event.IsSystemEvent() || event.CollectionId <= math.MaxUint32
|
||||||
|
}
|
||||||
|
|
||||||
|
func (event *UprEvent) IsSystemEvent() bool {
|
||||||
|
return event.Opcode == gomemcached.DCP_SYSTEM_EVENT
|
||||||
|
}
|
||||||
|
|
||||||
|
func (event *UprEvent) PopulateEvent(extras []byte) {
|
||||||
|
if len(extras) < dcpSystemEventExtraLen {
|
||||||
|
// Wrong length, don't parse
|
||||||
|
return
|
||||||
|
}
|
||||||
|
event.Seqno = binary.BigEndian.Uint64(extras[:8])
|
||||||
|
event.SystemEvent = SystemEventType(binary.BigEndian.Uint32(extras[8:12]))
|
||||||
|
var versionTemp uint16 = binary.BigEndian.Uint16(extras[12:14])
|
||||||
|
event.SysEventVersion = uint8(versionTemp >> 8)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (event *UprEvent) GetSystemEventName() (string, error) {
|
||||||
|
switch event.SystemEvent {
|
||||||
|
case CollectionCreate:
|
||||||
|
fallthrough
|
||||||
|
case ScopeCreate:
|
||||||
|
return string(event.Key), nil
|
||||||
|
default:
|
||||||
|
return "", ErrorInvalidOp
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (event *UprEvent) GetManifestId() (uint64, error) {
|
||||||
|
switch event.SystemEvent {
|
||||||
|
// Version 0 only checks
|
||||||
|
case CollectionChanged:
|
||||||
|
fallthrough
|
||||||
|
case ScopeDrop:
|
||||||
|
fallthrough
|
||||||
|
case ScopeCreate:
|
||||||
|
fallthrough
|
||||||
|
case CollectionDrop:
|
||||||
|
if event.SysEventVersion > 0 {
|
||||||
|
return 0, ErrorInvalidVersion
|
||||||
|
}
|
||||||
|
fallthrough
|
||||||
|
case CollectionCreate:
|
||||||
|
// CollectionCreate supports version 1
|
||||||
|
if event.SysEventVersion > 1 {
|
||||||
|
return 0, ErrorInvalidVersion
|
||||||
|
}
|
||||||
|
if event.ValueLen < 8 {
|
||||||
|
return 0, ErrorValueTooShort
|
||||||
|
}
|
||||||
|
return binary.BigEndian.Uint64(event.Value[0:8]), nil
|
||||||
|
default:
|
||||||
|
return 0, ErrorInvalidOp
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (event *UprEvent) GetCollectionId() (uint32, error) {
|
||||||
|
switch event.SystemEvent {
|
||||||
|
case CollectionDrop:
|
||||||
|
if event.SysEventVersion > 0 {
|
||||||
|
return 0, ErrorInvalidVersion
|
||||||
|
}
|
||||||
|
fallthrough
|
||||||
|
case CollectionCreate:
|
||||||
|
if event.SysEventVersion > 1 {
|
||||||
|
return 0, ErrorInvalidVersion
|
||||||
|
}
|
||||||
|
if event.ValueLen < 16 {
|
||||||
|
return 0, ErrorValueTooShort
|
||||||
|
}
|
||||||
|
return binary.BigEndian.Uint32(event.Value[12:16]), nil
|
||||||
|
case CollectionChanged:
|
||||||
|
if event.SysEventVersion > 0 {
|
||||||
|
return 0, ErrorInvalidVersion
|
||||||
|
}
|
||||||
|
if event.ValueLen < 12 {
|
||||||
|
return 0, ErrorValueTooShort
|
||||||
|
}
|
||||||
|
return binary.BigEndian.Uint32(event.Value[8:12]), nil
|
||||||
|
default:
|
||||||
|
return 0, ErrorInvalidOp
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (event *UprEvent) GetScopeId() (uint32, error) {
|
||||||
|
switch event.SystemEvent {
|
||||||
|
// version 0 checks
|
||||||
|
case ScopeCreate:
|
||||||
|
fallthrough
|
||||||
|
case ScopeDrop:
|
||||||
|
fallthrough
|
||||||
|
case CollectionDrop:
|
||||||
|
if event.SysEventVersion > 0 {
|
||||||
|
return 0, ErrorInvalidVersion
|
||||||
|
}
|
||||||
|
fallthrough
|
||||||
|
case CollectionCreate:
|
||||||
|
// CollectionCreate could be either 0 or 1
|
||||||
|
if event.SysEventVersion > 1 {
|
||||||
|
return 0, ErrorInvalidVersion
|
||||||
|
}
|
||||||
|
if event.ValueLen < 12 {
|
||||||
|
return 0, ErrorValueTooShort
|
||||||
|
}
|
||||||
|
return binary.BigEndian.Uint32(event.Value[8:12]), nil
|
||||||
|
default:
|
||||||
|
return 0, ErrorInvalidOp
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (event *UprEvent) GetMaxTTL() (uint32, error) {
|
||||||
|
switch event.SystemEvent {
|
||||||
|
case CollectionCreate:
|
||||||
|
if event.SysEventVersion < 1 {
|
||||||
|
return 0, ErrorNoMaxTTL
|
||||||
|
}
|
||||||
|
if event.ValueLen < 20 {
|
||||||
|
return 0, ErrorValueTooShort
|
||||||
|
}
|
||||||
|
return binary.BigEndian.Uint32(event.Value[16:20]), nil
|
||||||
|
case CollectionChanged:
|
||||||
|
if event.SysEventVersion > 0 {
|
||||||
|
return 0, ErrorInvalidVersion
|
||||||
|
}
|
||||||
|
if event.ValueLen < 16 {
|
||||||
|
return 0, ErrorValueTooShort
|
||||||
|
}
|
||||||
|
return binary.BigEndian.Uint32(event.Value[12:16]), nil
|
||||||
|
default:
|
||||||
|
return 0, ErrorInvalidOp
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type Uleb128 []byte
|
||||||
|
|
||||||
|
func (u Uleb128) ToUint64(cachedLen int) (result uint64, bytesShifted int) {
|
||||||
|
var shift uint = 0
|
||||||
|
|
||||||
|
for curByte := 0; curByte < cachedLen; curByte++ {
|
||||||
|
oneByte := u[curByte]
|
||||||
|
last7Bits := 0x7f & oneByte
|
||||||
|
result |= uint64(last7Bits) << shift
|
||||||
|
bytesShifted++
|
||||||
|
if oneByte&0x80 == 0 {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
shift += 7
|
||||||
|
}
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
|
@ -19,6 +19,7 @@ const uprMutationExtraLen = 30
|
||||||
const uprDeletetionExtraLen = 18
|
const uprDeletetionExtraLen = 18
|
||||||
const uprDeletetionWithDeletionTimeExtraLen = 21
|
const uprDeletetionWithDeletionTimeExtraLen = 21
|
||||||
const uprSnapshotExtraLen = 20
|
const uprSnapshotExtraLen = 20
|
||||||
|
const dcpSystemEventExtraLen = 13
|
||||||
const bufferAckThreshold = 0.2
|
const bufferAckThreshold = 0.2
|
||||||
const opaqueOpen = 0xBEAF0001
|
const opaqueOpen = 0xBEAF0001
|
||||||
const opaqueFailover = 0xDEADBEEF
|
const opaqueFailover = 0xDEADBEEF
|
||||||
|
@ -27,32 +28,6 @@ const uprDefaultNoopInterval = 120
|
||||||
// Counter on top of opaqueOpen that others can draw from for open and control msgs
|
// Counter on top of opaqueOpen that others can draw from for open and control msgs
|
||||||
var opaqueOpenCtrlWell uint32 = opaqueOpen
|
var opaqueOpenCtrlWell uint32 = opaqueOpen
|
||||||
|
|
||||||
// UprEvent memcached events for UPR streams.
|
|
||||||
type UprEvent struct {
|
|
||||||
Opcode gomemcached.CommandCode // Type of event
|
|
||||||
Status gomemcached.Status // Response status
|
|
||||||
VBucket uint16 // VBucket this event applies to
|
|
||||||
DataType uint8 // data type
|
|
||||||
Opaque uint16 // 16 MSB of opaque
|
|
||||||
VBuuid uint64 // This field is set by downstream
|
|
||||||
Flags uint32 // Item flags
|
|
||||||
Expiry uint32 // Item expiration time
|
|
||||||
Key, Value []byte // Item key/value
|
|
||||||
OldValue []byte // TODO: TBD: old document value
|
|
||||||
Cas uint64 // CAS value of the item
|
|
||||||
Seqno uint64 // sequence number of the mutation
|
|
||||||
RevSeqno uint64 // rev sequence number : deletions
|
|
||||||
LockTime uint32 // Lock time
|
|
||||||
MetadataSize uint16 // Metadata size
|
|
||||||
SnapstartSeq uint64 // start sequence number of this snapshot
|
|
||||||
SnapendSeq uint64 // End sequence number of the snapshot
|
|
||||||
SnapshotType uint32 // 0: disk 1: memory
|
|
||||||
FailoverLog *FailoverLog // Failover log containing vvuid and sequnce number
|
|
||||||
Error error // Error value in case of a failure
|
|
||||||
ExtMeta []byte
|
|
||||||
AckSize uint32 // The number of bytes that can be Acked to DCP
|
|
||||||
}
|
|
||||||
|
|
||||||
type PriorityType string
|
type PriorityType string
|
||||||
|
|
||||||
// high > medium > disabled > low
|
// high > medium > disabled > low
|
||||||
|
@ -63,13 +38,39 @@ const (
|
||||||
PriorityHigh PriorityType = "high"
|
PriorityHigh PriorityType = "high"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
type DcpStreamType int32
|
||||||
|
|
||||||
|
var UninitializedStream DcpStreamType = -1
|
||||||
|
|
||||||
|
const (
|
||||||
|
NonCollectionStream DcpStreamType = 0
|
||||||
|
CollectionsNonStreamId DcpStreamType = iota
|
||||||
|
CollectionsStreamId DcpStreamType = iota
|
||||||
|
)
|
||||||
|
|
||||||
|
func (t DcpStreamType) String() string {
|
||||||
|
switch t {
|
||||||
|
case UninitializedStream:
|
||||||
|
return "Un-Initialized Stream"
|
||||||
|
case NonCollectionStream:
|
||||||
|
return "Traditional Non-Collection Stream"
|
||||||
|
case CollectionsNonStreamId:
|
||||||
|
return "Collections Stream without StreamID"
|
||||||
|
case CollectionsStreamId:
|
||||||
|
return "Collection Stream with StreamID"
|
||||||
|
default:
|
||||||
|
return "Unknown Stream Type"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// UprStream is per stream data structure over an UPR Connection.
|
// UprStream is per stream data structure over an UPR Connection.
|
||||||
type UprStream struct {
|
type UprStream struct {
|
||||||
Vbucket uint16 // Vbucket id
|
Vbucket uint16 // Vbucket id
|
||||||
Vbuuid uint64 // vbucket uuid
|
Vbuuid uint64 // vbucket uuid
|
||||||
StartSeq uint64 // start sequence number
|
StartSeq uint64 // start sequence number
|
||||||
EndSeq uint64 // end sequence number
|
EndSeq uint64 // end sequence number
|
||||||
connected bool
|
connected bool
|
||||||
|
StreamType DcpStreamType
|
||||||
}
|
}
|
||||||
|
|
||||||
type FeedState int
|
type FeedState int
|
||||||
|
@ -113,6 +114,7 @@ type UprFeatures struct {
|
||||||
IncludeDeletionTime bool
|
IncludeDeletionTime bool
|
||||||
DcpPriority PriorityType
|
DcpPriority PriorityType
|
||||||
EnableExpiry bool
|
EnableExpiry bool
|
||||||
|
EnableStreamId bool
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -274,9 +276,15 @@ type UprFeed struct {
|
||||||
// if flag is true, upr feed will use ack from client to determine whether/when to send ack to DCP
|
// if flag is true, upr feed will use ack from client to determine whether/when to send ack to DCP
|
||||||
// if flag is false, upr feed will track how many bytes it has sent to client
|
// if flag is false, upr feed will track how many bytes it has sent to client
|
||||||
// and use that to determine whether/when to send ack to DCP
|
// and use that to determine whether/when to send ack to DCP
|
||||||
ackByClient bool
|
ackByClient bool
|
||||||
feedState FeedState
|
feedState FeedState
|
||||||
muFeedState sync.RWMutex
|
muFeedState sync.RWMutex
|
||||||
|
activatedFeatures UprFeatures
|
||||||
|
collectionEnabled bool // This is needed separately because parsing depends on this
|
||||||
|
// DCP StreamID allows multiple filtered collection streams to share a single DCP Stream
|
||||||
|
// It is not allowed once a regular/legacy stream was started originally
|
||||||
|
streamsType DcpStreamType
|
||||||
|
initStreamTypeOnce sync.Once
|
||||||
}
|
}
|
||||||
|
|
||||||
// Exported interface - to allow for mocking
|
// Exported interface - to allow for mocking
|
||||||
|
@ -296,6 +304,9 @@ type UprFeedIface interface {
|
||||||
UprRequestStream(vbno, opaqueMSB uint16, flags uint32, vuuid, startSequence, endSequence, snapStart, snapEnd uint64) error
|
UprRequestStream(vbno, opaqueMSB uint16, flags uint32, vuuid, startSequence, endSequence, snapStart, snapEnd uint64) error
|
||||||
// Set DCP priority on an existing DCP connection. The command is sent asynchronously without waiting for a response
|
// Set DCP priority on an existing DCP connection. The command is sent asynchronously without waiting for a response
|
||||||
SetPriorityAsync(p PriorityType) error
|
SetPriorityAsync(p PriorityType) error
|
||||||
|
|
||||||
|
// Various Collection-Type RequestStreams
|
||||||
|
UprRequestCollectionsStream(vbno, opaqueMSB uint16, flags uint32, vbuuid, startSeq, endSeq, snapStart, snapEnd uint64, filter *CollectionsFilter) error
|
||||||
}
|
}
|
||||||
|
|
||||||
type UprStats struct {
|
type UprStats struct {
|
||||||
|
@ -305,9 +316,6 @@ type UprStats struct {
|
||||||
TotalSnapShot uint64
|
TotalSnapShot uint64
|
||||||
}
|
}
|
||||||
|
|
||||||
// FailoverLog containing vvuid and sequnce number
|
|
||||||
type FailoverLog [][2]uint64
|
|
||||||
|
|
||||||
// error codes
|
// error codes
|
||||||
var ErrorInvalidLog = errors.New("couchbase.errorInvalidLog")
|
var ErrorInvalidLog = errors.New("couchbase.errorInvalidLog")
|
||||||
|
|
||||||
|
@ -320,76 +328,6 @@ func (flogp *FailoverLog) Latest() (vbuuid, seqno uint64, err error) {
|
||||||
return vbuuid, seqno, ErrorInvalidLog
|
return vbuuid, seqno, ErrorInvalidLog
|
||||||
}
|
}
|
||||||
|
|
||||||
func makeUprEvent(rq gomemcached.MCRequest, stream *UprStream, bytesReceivedFromDCP int) *UprEvent {
|
|
||||||
event := &UprEvent{
|
|
||||||
Opcode: rq.Opcode,
|
|
||||||
VBucket: stream.Vbucket,
|
|
||||||
VBuuid: stream.Vbuuid,
|
|
||||||
Key: rq.Key,
|
|
||||||
Value: rq.Body,
|
|
||||||
Cas: rq.Cas,
|
|
||||||
ExtMeta: rq.ExtMeta,
|
|
||||||
DataType: rq.DataType,
|
|
||||||
}
|
|
||||||
|
|
||||||
// set AckSize for events that need to be acked to DCP,
|
|
||||||
// i.e., events with CommandCodes that need to be buffered in DCP
|
|
||||||
if _, ok := gomemcached.BufferedCommandCodeMap[rq.Opcode]; ok {
|
|
||||||
event.AckSize = uint32(bytesReceivedFromDCP)
|
|
||||||
}
|
|
||||||
|
|
||||||
// 16 LSBits are used by client library to encode vbucket number.
|
|
||||||
// 16 MSBits are left for application to multiplex on opaque value.
|
|
||||||
event.Opaque = appOpaque(rq.Opaque)
|
|
||||||
|
|
||||||
if len(rq.Extras) >= uprMutationExtraLen &&
|
|
||||||
event.Opcode == gomemcached.UPR_MUTATION {
|
|
||||||
|
|
||||||
event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8])
|
|
||||||
event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16])
|
|
||||||
event.Flags = binary.BigEndian.Uint32(rq.Extras[16:20])
|
|
||||||
event.Expiry = binary.BigEndian.Uint32(rq.Extras[20:24])
|
|
||||||
event.LockTime = binary.BigEndian.Uint32(rq.Extras[24:28])
|
|
||||||
event.MetadataSize = binary.BigEndian.Uint16(rq.Extras[28:30])
|
|
||||||
|
|
||||||
} else if len(rq.Extras) >= uprDeletetionWithDeletionTimeExtraLen &&
|
|
||||||
event.Opcode == gomemcached.UPR_DELETION {
|
|
||||||
|
|
||||||
event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8])
|
|
||||||
event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16])
|
|
||||||
event.Expiry = binary.BigEndian.Uint32(rq.Extras[16:20])
|
|
||||||
|
|
||||||
} else if len(rq.Extras) >= uprDeletetionExtraLen &&
|
|
||||||
event.Opcode == gomemcached.UPR_DELETION ||
|
|
||||||
event.Opcode == gomemcached.UPR_EXPIRATION {
|
|
||||||
|
|
||||||
event.Seqno = binary.BigEndian.Uint64(rq.Extras[:8])
|
|
||||||
event.RevSeqno = binary.BigEndian.Uint64(rq.Extras[8:16])
|
|
||||||
event.MetadataSize = binary.BigEndian.Uint16(rq.Extras[16:18])
|
|
||||||
|
|
||||||
} else if len(rq.Extras) >= uprSnapshotExtraLen &&
|
|
||||||
event.Opcode == gomemcached.UPR_SNAPSHOT {
|
|
||||||
|
|
||||||
event.SnapstartSeq = binary.BigEndian.Uint64(rq.Extras[:8])
|
|
||||||
event.SnapendSeq = binary.BigEndian.Uint64(rq.Extras[8:16])
|
|
||||||
event.SnapshotType = binary.BigEndian.Uint32(rq.Extras[16:20])
|
|
||||||
}
|
|
||||||
|
|
||||||
return event
|
|
||||||
}
|
|
||||||
|
|
||||||
func (event *UprEvent) String() string {
|
|
||||||
name := gomemcached.CommandNames[event.Opcode]
|
|
||||||
if name == "" {
|
|
||||||
name = fmt.Sprintf("#%d", event.Opcode)
|
|
||||||
}
|
|
||||||
return name
|
|
||||||
}
|
|
||||||
|
|
||||||
func (event *UprEvent) IsSnappyDataType() bool {
|
|
||||||
return event.Opcode == gomemcached.UPR_MUTATION && (event.DataType&SnappyDataType > 0)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (feed *UprFeed) sendCommands(mc *Client) {
|
func (feed *UprFeed) sendCommands(mc *Client) {
|
||||||
transmitCh := feed.transmitCh
|
transmitCh := feed.transmitCh
|
||||||
transmitCl := feed.transmitCl
|
transmitCl := feed.transmitCl
|
||||||
|
@ -420,6 +358,10 @@ func (feed *UprFeed) activateStream(vbno, opaque uint16, stream *UprStream) erro
|
||||||
feed.muVbstreams.Lock()
|
feed.muVbstreams.Lock()
|
||||||
defer feed.muVbstreams.Unlock()
|
defer feed.muVbstreams.Unlock()
|
||||||
|
|
||||||
|
if feed.collectionEnabled {
|
||||||
|
stream.StreamType = feed.streamsType
|
||||||
|
}
|
||||||
|
|
||||||
// Set this stream as the officially connected stream for this vb
|
// Set this stream as the officially connected stream for this vb
|
||||||
stream.connected = true
|
stream.connected = true
|
||||||
feed.vbstreams[vbno] = stream
|
feed.vbstreams[vbno] = stream
|
||||||
|
@ -440,14 +382,15 @@ func (mc *Client) NewUprFeed() (*UprFeed, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (mc *Client) NewUprFeedWithConfig(ackByClient bool) (*UprFeed, error) {
|
func (mc *Client) NewUprFeedWithConfig(ackByClient bool) (*UprFeed, error) {
|
||||||
|
|
||||||
feed := &UprFeed{
|
feed := &UprFeed{
|
||||||
conn: mc,
|
conn: mc,
|
||||||
closer: make(chan bool, 1),
|
closer: make(chan bool, 1),
|
||||||
vbstreams: make(map[uint16]*UprStream),
|
vbstreams: make(map[uint16]*UprStream),
|
||||||
transmitCh: make(chan *gomemcached.MCRequest),
|
transmitCh: make(chan *gomemcached.MCRequest),
|
||||||
transmitCl: make(chan bool),
|
transmitCl: make(chan bool),
|
||||||
ackByClient: ackByClient,
|
ackByClient: ackByClient,
|
||||||
|
collectionEnabled: mc.CollectionEnabled(),
|
||||||
|
streamsType: UninitializedStream,
|
||||||
}
|
}
|
||||||
|
|
||||||
feed.negotiator.initialize()
|
feed.negotiator.initialize()
|
||||||
|
@ -642,7 +585,22 @@ func (feed *UprFeed) uprOpen(name string, sequence uint32, bufSize uint32, featu
|
||||||
activatedFeatures.EnableExpiry = true
|
activatedFeatures.EnableExpiry = true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if features.EnableStreamId {
|
||||||
|
rq := &gomemcached.MCRequest{
|
||||||
|
Opcode: gomemcached.UPR_CONTROL,
|
||||||
|
Key: []byte("enable_stream_id"),
|
||||||
|
Body: []byte("true"),
|
||||||
|
Opaque: getUprOpenCtrlOpaque(),
|
||||||
|
}
|
||||||
|
err = sendMcRequestSync(feed.conn, rq)
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
activatedFeatures.EnableStreamId = true
|
||||||
|
}
|
||||||
|
|
||||||
// everything is ok so far, set upr feed to open state
|
// everything is ok so far, set upr feed to open state
|
||||||
|
feed.activatedFeatures = activatedFeatures
|
||||||
feed.setOpen()
|
feed.setOpen()
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
@ -689,10 +647,60 @@ func (mc *Client) UprGetFailoverLog(
|
||||||
func (feed *UprFeed) UprRequestStream(vbno, opaqueMSB uint16, flags uint32,
|
func (feed *UprFeed) UprRequestStream(vbno, opaqueMSB uint16, flags uint32,
|
||||||
vuuid, startSequence, endSequence, snapStart, snapEnd uint64) error {
|
vuuid, startSequence, endSequence, snapStart, snapEnd uint64) error {
|
||||||
|
|
||||||
|
return feed.UprRequestCollectionsStream(vbno, opaqueMSB, flags, vuuid, startSequence, endSequence, snapStart, snapEnd, nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (feed *UprFeed) initStreamType(filter *CollectionsFilter) (err error) {
|
||||||
|
if filter != nil && filter.UseStreamId && !feed.activatedFeatures.EnableStreamId {
|
||||||
|
err = fmt.Errorf("Cannot use streamID based filter if the feed was not started with the streamID feature")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
streamInitFunc := func() {
|
||||||
|
if feed.streamsType != UninitializedStream {
|
||||||
|
// Shouldn't happen
|
||||||
|
err = fmt.Errorf("The current feed has already been started in %v mode", feed.streamsType.String())
|
||||||
|
} else {
|
||||||
|
if !feed.collectionEnabled {
|
||||||
|
feed.streamsType = NonCollectionStream
|
||||||
|
} else {
|
||||||
|
if filter != nil && filter.UseStreamId {
|
||||||
|
feed.streamsType = CollectionsStreamId
|
||||||
|
} else {
|
||||||
|
feed.streamsType = CollectionsNonStreamId
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
feed.initStreamTypeOnce.Do(streamInitFunc)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func (feed *UprFeed) UprRequestCollectionsStream(vbno, opaqueMSB uint16, flags uint32,
|
||||||
|
vbuuid, startSequence, endSequence, snapStart, snapEnd uint64, filter *CollectionsFilter) error {
|
||||||
|
|
||||||
|
err := feed.initStreamType(filter)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
var mcRequestBody []byte
|
||||||
|
if filter != nil {
|
||||||
|
err = filter.IsValid()
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
mcRequestBody, err = filter.ToStreamReqBody()
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
rq := &gomemcached.MCRequest{
|
rq := &gomemcached.MCRequest{
|
||||||
Opcode: gomemcached.UPR_STREAMREQ,
|
Opcode: gomemcached.UPR_STREAMREQ,
|
||||||
VBucket: vbno,
|
VBucket: vbno,
|
||||||
Opaque: composeOpaque(vbno, opaqueMSB),
|
Opaque: composeOpaque(vbno, opaqueMSB),
|
||||||
|
Body: mcRequestBody,
|
||||||
}
|
}
|
||||||
|
|
||||||
rq.Extras = make([]byte, 48) // #Extras
|
rq.Extras = make([]byte, 48) // #Extras
|
||||||
|
@ -700,15 +708,15 @@ func (feed *UprFeed) UprRequestStream(vbno, opaqueMSB uint16, flags uint32,
|
||||||
binary.BigEndian.PutUint32(rq.Extras[4:8], uint32(0))
|
binary.BigEndian.PutUint32(rq.Extras[4:8], uint32(0))
|
||||||
binary.BigEndian.PutUint64(rq.Extras[8:16], startSequence)
|
binary.BigEndian.PutUint64(rq.Extras[8:16], startSequence)
|
||||||
binary.BigEndian.PutUint64(rq.Extras[16:24], endSequence)
|
binary.BigEndian.PutUint64(rq.Extras[16:24], endSequence)
|
||||||
binary.BigEndian.PutUint64(rq.Extras[24:32], vuuid)
|
binary.BigEndian.PutUint64(rq.Extras[24:32], vbuuid)
|
||||||
binary.BigEndian.PutUint64(rq.Extras[32:40], snapStart)
|
binary.BigEndian.PutUint64(rq.Extras[32:40], snapStart)
|
||||||
binary.BigEndian.PutUint64(rq.Extras[40:48], snapEnd)
|
binary.BigEndian.PutUint64(rq.Extras[40:48], snapEnd)
|
||||||
|
|
||||||
feed.negotiator.registerRequest(vbno, opaqueMSB, vuuid, startSequence, endSequence)
|
feed.negotiator.registerRequest(vbno, opaqueMSB, vbuuid, startSequence, endSequence)
|
||||||
// Any client that has ever called this method, regardless of return code,
|
// Any client that has ever called this method, regardless of return code,
|
||||||
// should expect a potential UPR_CLOSESTREAM message due to this new map entry prior to Transmit.
|
// should expect a potential UPR_CLOSESTREAM message due to this new map entry prior to Transmit.
|
||||||
|
|
||||||
if err := feed.conn.Transmit(rq); err != nil {
|
if err = feed.conn.Transmit(rq); err != nil {
|
||||||
logging.Errorf("Error in StreamRequest %s", err.Error())
|
logging.Errorf("Error in StreamRequest %s", err.Error())
|
||||||
// If an error occurs during transmit, then the UPRFeed will keep the stream
|
// If an error occurs during transmit, then the UPRFeed will keep the stream
|
||||||
// in the vbstreams map. This is to prevent nil lookup from any previously
|
// in the vbstreams map. This is to prevent nil lookup from any previously
|
||||||
|
@ -973,6 +981,12 @@ loop:
|
||||||
if err := feed.conn.TransmitResponse(noop); err != nil {
|
if err := feed.conn.TransmitResponse(noop); err != nil {
|
||||||
logging.Warnf("failed to transmit command %s. Error %s", noop.Opcode.String(), err.Error())
|
logging.Warnf("failed to transmit command %s. Error %s", noop.Opcode.String(), err.Error())
|
||||||
}
|
}
|
||||||
|
case gomemcached.DCP_SYSTEM_EVENT:
|
||||||
|
if stream == nil {
|
||||||
|
logging.Infof("Stream not found for vb %d: %#v", vb, pkt)
|
||||||
|
break loop
|
||||||
|
}
|
||||||
|
event = makeUprEvent(pkt, stream, bytes)
|
||||||
default:
|
default:
|
||||||
logging.Infof("Recived an unknown response for vbucket %d", vb)
|
logging.Infof("Recived an unknown response for vbucket %d", vb)
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,381 @@
|
||||||
|
package gomemcached
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/binary"
|
||||||
|
"fmt"
|
||||||
|
)
|
||||||
|
|
||||||
|
type FrameObjType int
|
||||||
|
|
||||||
|
const (
|
||||||
|
FrameBarrier FrameObjType = iota
|
||||||
|
FrameDurability FrameObjType = iota
|
||||||
|
FrameDcpStreamId FrameObjType = iota
|
||||||
|
FrameOpenTracing FrameObjType = iota
|
||||||
|
)
|
||||||
|
|
||||||
|
type FrameInfo struct {
|
||||||
|
ObjId FrameObjType
|
||||||
|
ObjLen int
|
||||||
|
ObjData []byte
|
||||||
|
}
|
||||||
|
|
||||||
|
var ErrorInvalidOp error = fmt.Errorf("Specified method is not applicable")
|
||||||
|
var ErrorObjLenNotMatch error = fmt.Errorf("Object length does not match data")
|
||||||
|
|
||||||
|
func (f *FrameInfo) Validate() error {
|
||||||
|
switch f.ObjId {
|
||||||
|
case FrameBarrier:
|
||||||
|
if f.ObjLen != 0 {
|
||||||
|
return fmt.Errorf("Invalid FrameBarrier - length is %v\n", f.ObjLen)
|
||||||
|
} else if f.ObjLen != len(f.ObjData) {
|
||||||
|
return ErrorObjLenNotMatch
|
||||||
|
}
|
||||||
|
case FrameDurability:
|
||||||
|
if f.ObjLen != 1 && f.ObjLen != 3 {
|
||||||
|
return fmt.Errorf("Invalid FrameDurability - length is %v\n", f.ObjLen)
|
||||||
|
} else if f.ObjLen != len(f.ObjData) {
|
||||||
|
return ErrorObjLenNotMatch
|
||||||
|
}
|
||||||
|
case FrameDcpStreamId:
|
||||||
|
if f.ObjLen != 2 {
|
||||||
|
return fmt.Errorf("Invalid FrameDcpStreamId - length is %v\n", f.ObjLen)
|
||||||
|
} else if f.ObjLen != len(f.ObjData) {
|
||||||
|
return ErrorObjLenNotMatch
|
||||||
|
}
|
||||||
|
case FrameOpenTracing:
|
||||||
|
if f.ObjLen == 0 {
|
||||||
|
return fmt.Errorf("Invalid FrameOpenTracing - length must be > 0")
|
||||||
|
} else if f.ObjLen != len(f.ObjData) {
|
||||||
|
return ErrorObjLenNotMatch
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
return fmt.Errorf("Unknown FrameInfo type")
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (f *FrameInfo) GetStreamId() (uint16, error) {
|
||||||
|
if f.ObjId != FrameDcpStreamId {
|
||||||
|
return 0, ErrorInvalidOp
|
||||||
|
}
|
||||||
|
|
||||||
|
var output uint16
|
||||||
|
output = uint16(f.ObjData[0])
|
||||||
|
output = output << 8
|
||||||
|
output |= uint16(f.ObjData[1])
|
||||||
|
return output, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
type DurabilityLvl uint8
|
||||||
|
|
||||||
|
const (
|
||||||
|
DuraInvalid DurabilityLvl = iota // Not used (0x0)
|
||||||
|
DuraMajority DurabilityLvl = iota // (0x01)
|
||||||
|
DuraMajorityAndPersistOnMaster DurabilityLvl = iota // (0x02)
|
||||||
|
DuraPersistToMajority DurabilityLvl = iota // (0x03)
|
||||||
|
)
|
||||||
|
|
||||||
|
func (f *FrameInfo) GetDurabilityRequirements() (lvl DurabilityLvl, timeoutProvided bool, timeoutMs uint16, err error) {
|
||||||
|
if f.ObjId != FrameDurability {
|
||||||
|
err = ErrorInvalidOp
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if f.ObjLen != 1 && f.ObjLen != 3 {
|
||||||
|
err = ErrorObjLenNotMatch
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
lvl = DurabilityLvl(uint8(f.ObjData[0]))
|
||||||
|
|
||||||
|
if f.ObjLen == 3 {
|
||||||
|
timeoutProvided = true
|
||||||
|
timeoutMs = binary.BigEndian.Uint16(f.ObjData[1:2])
|
||||||
|
}
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func incrementMarker(bitsToBeIncremented, byteIncrementCnt *int, framingElen, curObjIdx int) (int, error) {
|
||||||
|
for *bitsToBeIncremented >= 8 {
|
||||||
|
*byteIncrementCnt++
|
||||||
|
*bitsToBeIncremented -= 8
|
||||||
|
}
|
||||||
|
marker := curObjIdx + *byteIncrementCnt
|
||||||
|
if marker > framingElen {
|
||||||
|
return -1, fmt.Errorf("Out of bounds")
|
||||||
|
}
|
||||||
|
return marker, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Right now, halfByteRemaining will always be false, because ObjID and Len haven't gotten that large yet
|
||||||
|
func (f *FrameInfo) Bytes() (output []byte, halfByteRemaining bool) {
|
||||||
|
// ObjIdentifier - 4 bits + ObjLength - 4 bits
|
||||||
|
var idAndLen uint8
|
||||||
|
idAndLen |= uint8(f.ObjId) << 4
|
||||||
|
idAndLen |= uint8(f.ObjLen)
|
||||||
|
output = append(output, byte(idAndLen))
|
||||||
|
|
||||||
|
// Rest is Data
|
||||||
|
output = append(output, f.ObjData...)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func parseFrameInfoObjects(buf []byte, framingElen int) (objs []FrameInfo, err error, halfByteRemaining bool) {
|
||||||
|
var curObjIdx int
|
||||||
|
var byteIncrementCnt int
|
||||||
|
var bitsToBeIncremented int
|
||||||
|
var marker int
|
||||||
|
|
||||||
|
// Parse frameInfo objects
|
||||||
|
for curObjIdx = 0; curObjIdx < framingElen; curObjIdx += byteIncrementCnt {
|
||||||
|
byteIncrementCnt = 0
|
||||||
|
var oneFrameObj FrameInfo
|
||||||
|
|
||||||
|
// First get the objId
|
||||||
|
// -------------------------
|
||||||
|
var objId int
|
||||||
|
var objHeader uint8 = buf[curObjIdx]
|
||||||
|
var objIdentifierRaw uint8
|
||||||
|
if bitsToBeIncremented == 0 {
|
||||||
|
// ObjHeader
|
||||||
|
// 0 1 2 3 4 5 6 7
|
||||||
|
// ^-----^
|
||||||
|
// ObjIdentifierRaw
|
||||||
|
objIdentifierRaw = (objHeader & 0xf0) >> 4
|
||||||
|
} else {
|
||||||
|
// ObjHeader
|
||||||
|
// 0 1 2 3 4 5 6 7
|
||||||
|
// ^-----^
|
||||||
|
// ObjIdentifierRaw
|
||||||
|
objIdentifierRaw = (objHeader & 0x0f)
|
||||||
|
}
|
||||||
|
bitsToBeIncremented += 4
|
||||||
|
|
||||||
|
marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx)
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Value is 0-14
|
||||||
|
objId = int(objIdentifierRaw & 0xe)
|
||||||
|
// If bit 15 is set, ID is 15 + value of next byte
|
||||||
|
if objIdentifierRaw&0x1 > 0 {
|
||||||
|
if bitsToBeIncremented > 0 {
|
||||||
|
// ObjHeader
|
||||||
|
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
||||||
|
// ^-----^ ^---------------^
|
||||||
|
// ObjId1 Extension
|
||||||
|
// ^ marker
|
||||||
|
buffer := uint16(buf[marker])
|
||||||
|
buffer = buffer << 8
|
||||||
|
buffer |= uint16(buf[marker+1])
|
||||||
|
var extension uint8 = uint8(buffer & 0xff0 >> 4)
|
||||||
|
objId += int(extension)
|
||||||
|
} else {
|
||||||
|
// ObjHeader
|
||||||
|
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
||||||
|
// ^-----^ ^-------------------^
|
||||||
|
// ObjId1 extension
|
||||||
|
// ^ marker
|
||||||
|
var extension uint8 = uint8(buf[marker])
|
||||||
|
objId += int(extension)
|
||||||
|
}
|
||||||
|
bitsToBeIncremented += 8
|
||||||
|
}
|
||||||
|
|
||||||
|
marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx)
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
oneFrameObj.ObjId = FrameObjType(objId)
|
||||||
|
|
||||||
|
// Then get the obj length
|
||||||
|
// -------------------------
|
||||||
|
var objLenRaw uint8
|
||||||
|
var objLen int
|
||||||
|
if bitsToBeIncremented > 0 {
|
||||||
|
// ObjHeader
|
||||||
|
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
||||||
|
// ^ ^---------^
|
||||||
|
// marker objLen
|
||||||
|
objLenRaw = uint8(buf[marker]) & 0x0f
|
||||||
|
} else {
|
||||||
|
// ObjHeader
|
||||||
|
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
|
||||||
|
// ^--------^
|
||||||
|
// objLen
|
||||||
|
// ^ marker
|
||||||
|
objLenRaw = uint8(buf[marker]) & 0xf0 >> 4
|
||||||
|
}
|
||||||
|
bitsToBeIncremented += 4
|
||||||
|
|
||||||
|
marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx)
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Length is 0-14
|
||||||
|
objLen = int(objLenRaw & 0xe)
|
||||||
|
// If bit 15 is set, lenghth is 15 + value of next byte
|
||||||
|
if objLenRaw&0x1 > 0 {
|
||||||
|
if bitsToBeIncremented == 0 {
|
||||||
|
// ObjHeader
|
||||||
|
// 12 13 14 15 16 17 18 19 20 21 22 23
|
||||||
|
// ^---------^ ^--------------------^
|
||||||
|
// objLen extension
|
||||||
|
// ^ marker
|
||||||
|
var extension uint8 = uint8(buf[marker])
|
||||||
|
objLen += int(extension)
|
||||||
|
} else {
|
||||||
|
// ObjHeader
|
||||||
|
// 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31
|
||||||
|
// ^--------^ ^---------------------^
|
||||||
|
// objLen extension
|
||||||
|
// ^ marker var buffer uint16
|
||||||
|
buffer := uint16(buf[marker])
|
||||||
|
buffer = buffer << 8
|
||||||
|
buffer |= uint16(buf[marker+1])
|
||||||
|
var extension uint8 = uint8(buffer & 0xff0 >> 4)
|
||||||
|
objLen += int(extension)
|
||||||
|
}
|
||||||
|
bitsToBeIncremented += 8
|
||||||
|
}
|
||||||
|
|
||||||
|
marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx)
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
oneFrameObj.ObjLen = objLen
|
||||||
|
|
||||||
|
// The rest is N-bytes of data based on the length
|
||||||
|
if bitsToBeIncremented == 0 {
|
||||||
|
// No weird alignment needed
|
||||||
|
oneFrameObj.ObjData = buf[marker : marker+objLen]
|
||||||
|
} else {
|
||||||
|
// 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31
|
||||||
|
// ^--------^ ^---------------------^ ^--------->
|
||||||
|
// objLen extension data
|
||||||
|
// ^ marker
|
||||||
|
oneFrameObj.ObjData = ShiftByteSliceLeft4Bits(buf[marker : marker+objLen+1])
|
||||||
|
}
|
||||||
|
err = oneFrameObj.Validate()
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
objs = append(objs, oneFrameObj)
|
||||||
|
|
||||||
|
bitsToBeIncremented += 8 * objLen
|
||||||
|
marker, err = incrementMarker(&bitsToBeIncremented, &byteIncrementCnt, framingElen, curObjIdx)
|
||||||
|
}
|
||||||
|
|
||||||
|
if bitsToBeIncremented > 0 {
|
||||||
|
halfByteRemaining = true
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func ShiftByteSliceLeft4Bits(slice []byte) (replacement []byte) {
|
||||||
|
var buffer uint16
|
||||||
|
var i int
|
||||||
|
sliceLen := len(slice)
|
||||||
|
|
||||||
|
if sliceLen < 2 {
|
||||||
|
// Let's not shift less than 16 bits
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
replacement = make([]byte, sliceLen, cap(slice))
|
||||||
|
|
||||||
|
for i = 0; i < sliceLen-1; i++ {
|
||||||
|
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
||||||
|
// ^-----^ ^---------------^ ^-----------
|
||||||
|
// garbage data byte 0 data byte 1
|
||||||
|
buffer = uint16(slice[i])
|
||||||
|
buffer = buffer << 8
|
||||||
|
buffer |= uint16(slice[i+1])
|
||||||
|
replacement[i] = uint8(buffer & 0xff0 >> 4)
|
||||||
|
}
|
||||||
|
|
||||||
|
if i < sliceLen {
|
||||||
|
lastByte := slice[sliceLen-1]
|
||||||
|
lastByte = lastByte << 4
|
||||||
|
replacement[i] = lastByte
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// The following is used to theoretically support frameInfo ObjID extensions
|
||||||
|
// for completeness, but they are not very efficient though
|
||||||
|
func ShiftByteSliceRight4Bits(slice []byte) (replacement []byte) {
|
||||||
|
var buffer uint16
|
||||||
|
var i int
|
||||||
|
var leftovers uint8 // 4 bits only
|
||||||
|
var replacementUnit uint16
|
||||||
|
var first bool = true
|
||||||
|
var firstLeftovers uint8
|
||||||
|
var lastLeftovers uint8
|
||||||
|
sliceLen := len(slice)
|
||||||
|
|
||||||
|
if sliceLen < 2 {
|
||||||
|
// Let's not shift less than 16 bits
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if slice[sliceLen-1]&0xf == 0 {
|
||||||
|
replacement = make([]byte, sliceLen, cap(slice))
|
||||||
|
} else {
|
||||||
|
replacement = make([]byte, sliceLen+1, cap(slice)+1)
|
||||||
|
}
|
||||||
|
|
||||||
|
for i = 0; i < sliceLen-1; i++ {
|
||||||
|
buffer = binary.BigEndian.Uint16(slice[i : i+2])
|
||||||
|
// (buffer)
|
||||||
|
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
|
||||||
|
// ^-------------^ ^-------------------^
|
||||||
|
// data byte 0 data byte 1
|
||||||
|
//
|
||||||
|
// into
|
||||||
|
//
|
||||||
|
// 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
|
||||||
|
// ^-----^ ^---------------^ ^--------------------^ ^----------^
|
||||||
|
// zeroes data byte 0 data byte 1 zeroes
|
||||||
|
|
||||||
|
if first {
|
||||||
|
// The leftover OR'ing will overwrite the first 4 bits of data byte 0. Save them
|
||||||
|
firstLeftovers = uint8(buffer & 0xf000 >> 12)
|
||||||
|
first = false
|
||||||
|
}
|
||||||
|
replacementUnit = 0
|
||||||
|
replacementUnit |= uint16(leftovers) << 12
|
||||||
|
replacementUnit |= (buffer & 0xff00) >> 4 // data byte 0
|
||||||
|
replacementUnit |= buffer & 0xff >> 4 // data byte 1 first 4 bits
|
||||||
|
lastLeftovers = uint8(buffer&0xf) << 4
|
||||||
|
|
||||||
|
replacement[i+1] = byte(replacementUnit)
|
||||||
|
|
||||||
|
leftovers = uint8((buffer & 0x000f) << 4)
|
||||||
|
}
|
||||||
|
|
||||||
|
replacement[0] = byte(uint8(replacement[0]) | firstLeftovers)
|
||||||
|
if lastLeftovers > 0 {
|
||||||
|
replacement[sliceLen] = byte(lastLeftovers)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func Merge2HalfByteSlices(src1, src2 []byte) (output []byte) {
|
||||||
|
src1Len := len(src1)
|
||||||
|
src2Len := len(src2)
|
||||||
|
output = make([]byte, src1Len+src2Len-1)
|
||||||
|
|
||||||
|
var mergeByte uint8 = src1[src1Len-1]
|
||||||
|
mergeByte |= uint8(src2[0])
|
||||||
|
|
||||||
|
copy(output, src1)
|
||||||
|
copy(output[src1Len:], src2[1:])
|
||||||
|
|
||||||
|
output[src1Len-1] = byte(mergeByte)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
|
@ -6,8 +6,10 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
REQ_MAGIC = 0x80
|
REQ_MAGIC = 0x80
|
||||||
RES_MAGIC = 0x81
|
RES_MAGIC = 0x81
|
||||||
|
FLEX_MAGIC = 0x08
|
||||||
|
FLEX_RES_MAGIC = 0x18
|
||||||
)
|
)
|
||||||
|
|
||||||
// CommandCode for memcached packets.
|
// CommandCode for memcached packets.
|
||||||
|
@ -99,6 +101,8 @@ const (
|
||||||
SUBDOC_GET = CommandCode(0xc5) // Get subdoc. Returns with xattrs
|
SUBDOC_GET = CommandCode(0xc5) // Get subdoc. Returns with xattrs
|
||||||
SUBDOC_MULTI_LOOKUP = CommandCode(0xd0) // Multi lookup. Doc xattrs and meta.
|
SUBDOC_MULTI_LOOKUP = CommandCode(0xd0) // Multi lookup. Doc xattrs and meta.
|
||||||
|
|
||||||
|
DCP_SYSTEM_EVENT = CommandCode(0x5f) // A system event has occurred
|
||||||
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// command codes that are counted toward DCP control buffer
|
// command codes that are counted toward DCP control buffer
|
||||||
|
|
|
@ -25,11 +25,17 @@ type MCRequest struct {
|
||||||
Extras, Key, Body, ExtMeta []byte
|
Extras, Key, Body, ExtMeta []byte
|
||||||
// Datatype identifier
|
// Datatype identifier
|
||||||
DataType uint8
|
DataType uint8
|
||||||
|
// len() calls are expensive - cache this in case for collection
|
||||||
|
Keylen int
|
||||||
|
// Flexible Framing Extras
|
||||||
|
FramingExtras []FrameInfo
|
||||||
|
// Stored length of incoming framing extras
|
||||||
|
FramingElen int
|
||||||
}
|
}
|
||||||
|
|
||||||
// Size gives the number of bytes this request requires.
|
// Size gives the number of bytes this request requires.
|
||||||
func (req *MCRequest) Size() int {
|
func (req *MCRequest) Size() int {
|
||||||
return HDR_LEN + len(req.Extras) + len(req.Key) + len(req.Body) + len(req.ExtMeta)
|
return HDR_LEN + len(req.Extras) + len(req.Key) + len(req.Body) + len(req.ExtMeta) + req.FramingElen
|
||||||
}
|
}
|
||||||
|
|
||||||
// A debugging string representation of this request
|
// A debugging string representation of this request
|
||||||
|
@ -38,7 +44,23 @@ func (req MCRequest) String() string {
|
||||||
req.Opcode, len(req.Body), req.Key)
|
req.Opcode, len(req.Body), req.Key)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (req *MCRequest) fillHeaderBytes(data []byte) int {
|
func (req *MCRequest) fillRegularHeaderBytes(data []byte) int {
|
||||||
|
// Byte/ 0 | 1 | 2 | 3 |
|
||||||
|
// / | | | |
|
||||||
|
// |0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|
|
||||||
|
// +---------------+---------------+---------------+---------------+
|
||||||
|
// 0| Magic | Opcode | Key length |
|
||||||
|
// +---------------+---------------+---------------+---------------+
|
||||||
|
// 4| Extras length | Data type | vbucket id |
|
||||||
|
// +---------------+---------------+---------------+---------------+
|
||||||
|
// 8| Total body length |
|
||||||
|
// +---------------+---------------+---------------+---------------+
|
||||||
|
// 12| Opaque |
|
||||||
|
// +---------------+---------------+---------------+---------------+
|
||||||
|
// 16| CAS |
|
||||||
|
// | |
|
||||||
|
// +---------------+---------------+---------------+---------------+
|
||||||
|
// Total 24 bytes
|
||||||
|
|
||||||
pos := 0
|
pos := 0
|
||||||
data[pos] = REQ_MAGIC
|
data[pos] = REQ_MAGIC
|
||||||
|
@ -84,16 +106,130 @@ func (req *MCRequest) fillHeaderBytes(data []byte) int {
|
||||||
copy(data[pos:pos+len(req.Key)], req.Key)
|
copy(data[pos:pos+len(req.Key)], req.Key)
|
||||||
pos += len(req.Key)
|
pos += len(req.Key)
|
||||||
}
|
}
|
||||||
|
|
||||||
return pos
|
return pos
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Returns pos and if trailing by half byte
|
||||||
|
func (req *MCRequest) fillFlexHeaderBytes(data []byte) (int, bool) {
|
||||||
|
|
||||||
|
// Byte/ 0 | 1 | 2 | 3 |
|
||||||
|
// / | | | |
|
||||||
|
// |0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|0 1 2 3 4 5 6 7|
|
||||||
|
// +---------------+---------------+---------------+---------------+
|
||||||
|
// 0| Magic (0x08) | Opcode | Framing extras| Key Length |
|
||||||
|
// +---------------+---------------+---------------+---------------+
|
||||||
|
// 4| Extras length | Data type | vbucket id |
|
||||||
|
// +---------------+---------------+---------------+---------------+
|
||||||
|
// 8| Total body length |
|
||||||
|
// +---------------+---------------+---------------+---------------+
|
||||||
|
// 12| Opaque |
|
||||||
|
// +---------------+---------------+---------------+---------------+
|
||||||
|
// 16| CAS |
|
||||||
|
// | |
|
||||||
|
// +---------------+---------------+---------------+---------------+
|
||||||
|
// Total 24 bytes
|
||||||
|
|
||||||
|
data[0] = FLEX_MAGIC
|
||||||
|
data[1] = byte(req.Opcode)
|
||||||
|
data[2] = byte(req.FramingElen)
|
||||||
|
data[3] = byte(req.Keylen)
|
||||||
|
elen := len(req.Extras)
|
||||||
|
data[4] = byte(elen)
|
||||||
|
if req.DataType != 0 {
|
||||||
|
data[5] = byte(req.DataType)
|
||||||
|
}
|
||||||
|
binary.BigEndian.PutUint16(data[6:8], req.VBucket)
|
||||||
|
binary.BigEndian.PutUint32(data[8:12],
|
||||||
|
uint32(len(req.Body)+req.Keylen+elen+len(req.ExtMeta)+req.FramingElen))
|
||||||
|
binary.BigEndian.PutUint32(data[12:16], req.Opaque)
|
||||||
|
if req.Cas != 0 {
|
||||||
|
binary.BigEndian.PutUint64(data[16:24], req.Cas)
|
||||||
|
}
|
||||||
|
pos := HDR_LEN
|
||||||
|
|
||||||
|
// Add framing infos
|
||||||
|
var framingExtras []byte
|
||||||
|
var outputBytes []byte
|
||||||
|
var mergeModeSrc []byte
|
||||||
|
var frameBytes int
|
||||||
|
var halfByteMode bool
|
||||||
|
var mergeMode bool
|
||||||
|
for _, frameInfo := range req.FramingExtras {
|
||||||
|
if !mergeMode {
|
||||||
|
outputBytes, halfByteMode = frameInfo.Bytes()
|
||||||
|
if !halfByteMode {
|
||||||
|
framingExtras = append(framingExtras, outputBytes...)
|
||||||
|
frameBytes += len(outputBytes)
|
||||||
|
} else {
|
||||||
|
mergeMode = true
|
||||||
|
mergeModeSrc = outputBytes
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
outputBytes, halfByteMode = frameInfo.Bytes()
|
||||||
|
outputBytes := ShiftByteSliceRight4Bits(outputBytes)
|
||||||
|
if halfByteMode {
|
||||||
|
// Previous halfbyte merge with this halfbyte will result in a complete byte
|
||||||
|
mergeMode = false
|
||||||
|
outputBytes = Merge2HalfByteSlices(mergeModeSrc, outputBytes)
|
||||||
|
framingExtras = append(framingExtras, outputBytes...)
|
||||||
|
frameBytes += len(outputBytes)
|
||||||
|
} else {
|
||||||
|
// Merge half byte with a non-half byte will result in a combined half-byte that will
|
||||||
|
// become the source for the next iteration
|
||||||
|
mergeModeSrc = Merge2HalfByteSlices(mergeModeSrc, outputBytes)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if mergeMode {
|
||||||
|
// Commit the temporary merge area into framingExtras
|
||||||
|
framingExtras = append(framingExtras, mergeModeSrc...)
|
||||||
|
frameBytes += len(mergeModeSrc)
|
||||||
|
}
|
||||||
|
|
||||||
|
copy(data[pos:pos+frameBytes], framingExtras)
|
||||||
|
|
||||||
|
pos += frameBytes
|
||||||
|
|
||||||
|
// Add Extras
|
||||||
|
if len(req.Extras) > 0 {
|
||||||
|
if mergeMode {
|
||||||
|
outputBytes = ShiftByteSliceRight4Bits(req.Extras)
|
||||||
|
data = Merge2HalfByteSlices(data, outputBytes)
|
||||||
|
} else {
|
||||||
|
copy(data[pos:pos+elen], req.Extras)
|
||||||
|
}
|
||||||
|
pos += elen
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add keys
|
||||||
|
if req.Keylen > 0 {
|
||||||
|
if mergeMode {
|
||||||
|
outputBytes = ShiftByteSliceRight4Bits(req.Key)
|
||||||
|
data = Merge2HalfByteSlices(data, outputBytes)
|
||||||
|
} else {
|
||||||
|
copy(data[pos:pos+req.Keylen], req.Key)
|
||||||
|
}
|
||||||
|
pos += req.Keylen
|
||||||
|
}
|
||||||
|
|
||||||
|
return pos, mergeMode
|
||||||
|
}
|
||||||
|
|
||||||
|
func (req *MCRequest) FillHeaderBytes(data []byte) (int, bool) {
|
||||||
|
if req.FramingElen == 0 {
|
||||||
|
return req.fillRegularHeaderBytes(data), false
|
||||||
|
} else {
|
||||||
|
return req.fillFlexHeaderBytes(data)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// HeaderBytes will return the wire representation of the request header
|
// HeaderBytes will return the wire representation of the request header
|
||||||
// (with the extras and key).
|
// (with the extras and key).
|
||||||
func (req *MCRequest) HeaderBytes() []byte {
|
func (req *MCRequest) HeaderBytes() []byte {
|
||||||
data := make([]byte, HDR_LEN+len(req.Extras)+len(req.Key))
|
data := make([]byte, HDR_LEN+len(req.Extras)+len(req.Key)+req.FramingElen)
|
||||||
|
|
||||||
req.fillHeaderBytes(data)
|
req.FillHeaderBytes(data)
|
||||||
|
|
||||||
return data
|
return data
|
||||||
}
|
}
|
||||||
|
@ -102,16 +238,27 @@ func (req *MCRequest) HeaderBytes() []byte {
|
||||||
func (req *MCRequest) Bytes() []byte {
|
func (req *MCRequest) Bytes() []byte {
|
||||||
data := make([]byte, req.Size())
|
data := make([]byte, req.Size())
|
||||||
|
|
||||||
pos := req.fillHeaderBytes(data)
|
pos, halfByteMode := req.FillHeaderBytes(data)
|
||||||
|
// TODO - the halfByteMode should be revisited for a more efficient
|
||||||
|
// way of doing things
|
||||||
|
|
||||||
if len(req.Body) > 0 {
|
if len(req.Body) > 0 {
|
||||||
copy(data[pos:pos+len(req.Body)], req.Body)
|
if halfByteMode {
|
||||||
|
shifted := ShiftByteSliceRight4Bits(req.Body)
|
||||||
|
data = Merge2HalfByteSlices(data, shifted)
|
||||||
|
} else {
|
||||||
|
copy(data[pos:pos+len(req.Body)], req.Body)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(req.ExtMeta) > 0 {
|
if len(req.ExtMeta) > 0 {
|
||||||
copy(data[pos+len(req.Body):pos+len(req.Body)+len(req.ExtMeta)], req.ExtMeta)
|
if halfByteMode {
|
||||||
|
shifted := ShiftByteSliceRight4Bits(req.ExtMeta)
|
||||||
|
data = Merge2HalfByteSlices(data, shifted)
|
||||||
|
} else {
|
||||||
|
copy(data[pos+len(req.Body):pos+len(req.Body)+len(req.ExtMeta)], req.ExtMeta)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return data
|
return data
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -130,40 +277,44 @@ func (req *MCRequest) Transmit(w io.Writer) (n int, err error) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Receive will fill this MCRequest with the data from a reader.
|
func (req *MCRequest) receiveHeaderCommon(hdrBytes []byte) (elen, totalBodyLen int) {
|
||||||
func (req *MCRequest) Receive(r io.Reader, hdrBytes []byte) (int, error) {
|
elen = int(hdrBytes[4])
|
||||||
if len(hdrBytes) < HDR_LEN {
|
|
||||||
hdrBytes = []byte{
|
|
||||||
0, 0, 0, 0, 0, 0, 0, 0,
|
|
||||||
0, 0, 0, 0, 0, 0, 0, 0,
|
|
||||||
0, 0, 0, 0, 0, 0, 0, 0}
|
|
||||||
}
|
|
||||||
n, err := io.ReadFull(r, hdrBytes)
|
|
||||||
if err != nil {
|
|
||||||
return n, err
|
|
||||||
}
|
|
||||||
|
|
||||||
if hdrBytes[0] != RES_MAGIC && hdrBytes[0] != REQ_MAGIC {
|
|
||||||
return n, fmt.Errorf("bad magic: 0x%02x", hdrBytes[0])
|
|
||||||
}
|
|
||||||
|
|
||||||
klen := int(binary.BigEndian.Uint16(hdrBytes[2:]))
|
|
||||||
elen := int(hdrBytes[4])
|
|
||||||
// Data type at 5
|
// Data type at 5
|
||||||
req.DataType = uint8(hdrBytes[5])
|
req.DataType = uint8(hdrBytes[5])
|
||||||
|
|
||||||
req.Opcode = CommandCode(hdrBytes[1])
|
req.Opcode = CommandCode(hdrBytes[1])
|
||||||
// Vbucket at 6:7
|
// Vbucket at 6:7
|
||||||
req.VBucket = binary.BigEndian.Uint16(hdrBytes[6:])
|
req.VBucket = binary.BigEndian.Uint16(hdrBytes[6:])
|
||||||
totalBodyLen := int(binary.BigEndian.Uint32(hdrBytes[8:]))
|
totalBodyLen = int(binary.BigEndian.Uint32(hdrBytes[8:]))
|
||||||
|
|
||||||
req.Opaque = binary.BigEndian.Uint32(hdrBytes[12:])
|
req.Opaque = binary.BigEndian.Uint32(hdrBytes[12:])
|
||||||
req.Cas = binary.BigEndian.Uint64(hdrBytes[16:])
|
req.Cas = binary.BigEndian.Uint64(hdrBytes[16:])
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func (req *MCRequest) receiveRegHeader(hdrBytes []byte) (elen, totalBodyLen int) {
|
||||||
|
elen, totalBodyLen = req.receiveHeaderCommon(hdrBytes)
|
||||||
|
req.Keylen = int(binary.BigEndian.Uint16(hdrBytes[2:]))
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func (req *MCRequest) receiveFlexibleFramingHeader(hdrBytes []byte) (elen, totalBodyLen, framingElen int) {
|
||||||
|
elen, totalBodyLen = req.receiveHeaderCommon(hdrBytes)
|
||||||
|
|
||||||
|
// For flexible framing header, key length is a single byte at byte index 3
|
||||||
|
req.Keylen = int(binary.BigEndian.Uint16(hdrBytes[2:]) & 0x0ff)
|
||||||
|
// Flexible framing lengh is a single byte at index 2
|
||||||
|
framingElen = int(binary.BigEndian.Uint16(hdrBytes[2:]) >> 8)
|
||||||
|
req.FramingElen = framingElen
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func (req *MCRequest) populateRegularBody(r io.Reader, totalBodyLen, elen int) (int, error) {
|
||||||
|
var m int
|
||||||
|
var err error
|
||||||
if totalBodyLen > 0 {
|
if totalBodyLen > 0 {
|
||||||
buf := make([]byte, totalBodyLen)
|
buf := make([]byte, totalBodyLen)
|
||||||
m, err := io.ReadFull(r, buf)
|
m, err = io.ReadFull(r, buf)
|
||||||
n += m
|
|
||||||
if err == nil {
|
if err == nil {
|
||||||
if req.Opcode >= TAP_MUTATION &&
|
if req.Opcode >= TAP_MUTATION &&
|
||||||
req.Opcode <= TAP_CHECKPOINT_END &&
|
req.Opcode <= TAP_CHECKPOINT_END &&
|
||||||
|
@ -175,7 +326,7 @@ func (req *MCRequest) Receive(r io.Reader, hdrBytes []byte) (int, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
req.Extras = buf[0:elen]
|
req.Extras = buf[0:elen]
|
||||||
req.Key = buf[elen : klen+elen]
|
req.Key = buf[elen : req.Keylen+elen]
|
||||||
|
|
||||||
// get the length of extended metadata
|
// get the length of extended metadata
|
||||||
extMetaLen := 0
|
extMetaLen := 0
|
||||||
|
@ -183,15 +334,149 @@ func (req *MCRequest) Receive(r io.Reader, hdrBytes []byte) (int, error) {
|
||||||
extMetaLen = int(binary.BigEndian.Uint16(req.Extras[28:30]))
|
extMetaLen = int(binary.BigEndian.Uint16(req.Extras[28:30]))
|
||||||
}
|
}
|
||||||
|
|
||||||
bodyLen := totalBodyLen - klen - elen - extMetaLen
|
bodyLen := totalBodyLen - req.Keylen - elen - extMetaLen
|
||||||
if bodyLen > MaxBodyLen {
|
if bodyLen > MaxBodyLen {
|
||||||
return n, fmt.Errorf("%d is too big (max %d)",
|
return m, fmt.Errorf("%d is too big (max %d)",
|
||||||
bodyLen, MaxBodyLen)
|
bodyLen, MaxBodyLen)
|
||||||
}
|
}
|
||||||
|
|
||||||
req.Body = buf[klen+elen : klen+elen+bodyLen]
|
req.Body = buf[req.Keylen+elen : req.Keylen+elen+bodyLen]
|
||||||
req.ExtMeta = buf[klen+elen+bodyLen:]
|
req.ExtMeta = buf[req.Keylen+elen+bodyLen:]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return n, err
|
return m, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (req *MCRequest) populateFlexBody(r io.Reader, totalBodyLen, elen, framingElen int) (int, error) {
|
||||||
|
var m int
|
||||||
|
var err error
|
||||||
|
if totalBodyLen > 0 {
|
||||||
|
buf := make([]byte, totalBodyLen)
|
||||||
|
m, err = io.ReadFull(r, buf)
|
||||||
|
if err != nil {
|
||||||
|
return m, err
|
||||||
|
}
|
||||||
|
err = req.populateFlexBodyInternal(buf, totalBodyLen, elen, framingElen)
|
||||||
|
}
|
||||||
|
return m, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (req *MCRequest) populateFlexBodyInternal(buf []byte, totalBodyLen, elen, framingElen int) error {
|
||||||
|
var halfByteOffset bool
|
||||||
|
var err error
|
||||||
|
if framingElen > 0 {
|
||||||
|
var objs []FrameInfo
|
||||||
|
objs, err, halfByteOffset = parseFrameInfoObjects(buf, framingElen)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
req.FramingExtras = objs
|
||||||
|
}
|
||||||
|
|
||||||
|
err = req.populateFlexBodyAfterFrames(buf, totalBodyLen, elen, framingElen, halfByteOffset)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (req *MCRequest) populateFlexBodyAfterFrames(buf []byte, totalBodyLen, elen, framingElen int, halfByteOffset bool) error {
|
||||||
|
var idxCursor int = framingElen
|
||||||
|
if req.Opcode >= TAP_MUTATION && req.Opcode <= TAP_CHECKPOINT_END && len(buf[idxCursor:]) > 1 {
|
||||||
|
// In these commands there is "engine private"
|
||||||
|
// data at the end of the extras. The first 2
|
||||||
|
// bytes of extra data give its length.
|
||||||
|
if !halfByteOffset {
|
||||||
|
elen += int(binary.BigEndian.Uint16(buf[idxCursor:]))
|
||||||
|
} else {
|
||||||
|
// 0 1 2 3 4 .... 19 20 21 22 ... 32
|
||||||
|
// ^-----^ ^-------^ ^------------^
|
||||||
|
// offset data do not care
|
||||||
|
var buffer uint32 = binary.BigEndian.Uint32(buf[idxCursor:])
|
||||||
|
buffer &= 0xffff000
|
||||||
|
elen += int(buffer >> 12)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Get the extras
|
||||||
|
if !halfByteOffset {
|
||||||
|
req.Extras = buf[idxCursor : idxCursor+elen]
|
||||||
|
} else {
|
||||||
|
preShift := buf[idxCursor : idxCursor+elen+1]
|
||||||
|
req.Extras = ShiftByteSliceLeft4Bits(preShift)
|
||||||
|
}
|
||||||
|
idxCursor += elen
|
||||||
|
|
||||||
|
// Get the Key
|
||||||
|
if !halfByteOffset {
|
||||||
|
req.Key = buf[idxCursor : idxCursor+req.Keylen]
|
||||||
|
} else {
|
||||||
|
preShift := buf[idxCursor : idxCursor+req.Keylen+1]
|
||||||
|
req.Key = ShiftByteSliceLeft4Bits(preShift)
|
||||||
|
}
|
||||||
|
idxCursor += req.Keylen
|
||||||
|
|
||||||
|
// get the length of extended metadata
|
||||||
|
extMetaLen := 0
|
||||||
|
if elen > 29 {
|
||||||
|
extMetaLen = int(binary.BigEndian.Uint16(req.Extras[28:30]))
|
||||||
|
}
|
||||||
|
idxCursor += extMetaLen
|
||||||
|
|
||||||
|
bodyLen := totalBodyLen - req.Keylen - elen - extMetaLen - framingElen
|
||||||
|
if bodyLen > MaxBodyLen {
|
||||||
|
return fmt.Errorf("%d is too big (max %d)",
|
||||||
|
bodyLen, MaxBodyLen)
|
||||||
|
}
|
||||||
|
|
||||||
|
if !halfByteOffset {
|
||||||
|
req.Body = buf[idxCursor : idxCursor+bodyLen]
|
||||||
|
idxCursor += bodyLen
|
||||||
|
} else {
|
||||||
|
preShift := buf[idxCursor : idxCursor+bodyLen+1]
|
||||||
|
req.Body = ShiftByteSliceLeft4Bits(preShift)
|
||||||
|
idxCursor += bodyLen
|
||||||
|
}
|
||||||
|
|
||||||
|
if extMetaLen > 0 {
|
||||||
|
if !halfByteOffset {
|
||||||
|
req.ExtMeta = buf[idxCursor:]
|
||||||
|
} else {
|
||||||
|
preShift := buf[idxCursor:]
|
||||||
|
req.ExtMeta = ShiftByteSliceLeft4Bits(preShift)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Receive will fill this MCRequest with the data from a reader.
|
||||||
|
func (req *MCRequest) Receive(r io.Reader, hdrBytes []byte) (int, error) {
|
||||||
|
if len(hdrBytes) < HDR_LEN {
|
||||||
|
hdrBytes = []byte{
|
||||||
|
0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
|
0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
|
0, 0, 0, 0, 0, 0, 0, 0}
|
||||||
|
}
|
||||||
|
n, err := io.ReadFull(r, hdrBytes)
|
||||||
|
if err != nil {
|
||||||
|
fmt.Printf("Err %v\n", err)
|
||||||
|
return n, err
|
||||||
|
}
|
||||||
|
|
||||||
|
switch hdrBytes[0] {
|
||||||
|
case RES_MAGIC:
|
||||||
|
fallthrough
|
||||||
|
case REQ_MAGIC:
|
||||||
|
elen, totalBodyLen := req.receiveRegHeader(hdrBytes)
|
||||||
|
bodyRead, err := req.populateRegularBody(r, totalBodyLen, elen)
|
||||||
|
return n + bodyRead, err
|
||||||
|
case FLEX_MAGIC:
|
||||||
|
elen, totalBodyLen, framingElen := req.receiveFlexibleFramingHeader(hdrBytes)
|
||||||
|
bodyRead, err := req.populateFlexBody(r, totalBodyLen, elen, framingElen)
|
||||||
|
return n + bodyRead, err
|
||||||
|
default:
|
||||||
|
return n, fmt.Errorf("bad magic: 0x%02x", hdrBytes[0])
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -153,6 +153,13 @@ func (res *MCResponse) Transmit(w io.Writer) (n int, err error) {
|
||||||
|
|
||||||
// Receive will fill this MCResponse with the data from this reader.
|
// Receive will fill this MCResponse with the data from this reader.
|
||||||
func (res *MCResponse) Receive(r io.Reader, hdrBytes []byte) (n int, err error) {
|
func (res *MCResponse) Receive(r io.Reader, hdrBytes []byte) (n int, err error) {
|
||||||
|
return res.ReceiveWithBuf(r, hdrBytes, nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ReceiveWithBuf takes an optional pre-allocated []byte buf which
|
||||||
|
// will be used if its capacity is large enough, otherwise a new
|
||||||
|
// []byte slice is allocated.
|
||||||
|
func (res *MCResponse) ReceiveWithBuf(r io.Reader, hdrBytes, buf []byte) (n int, err error) {
|
||||||
if len(hdrBytes) < HDR_LEN {
|
if len(hdrBytes) < HDR_LEN {
|
||||||
hdrBytes = []byte{
|
hdrBytes = []byte{
|
||||||
0, 0, 0, 0, 0, 0, 0, 0,
|
0, 0, 0, 0, 0, 0, 0, 0,
|
||||||
|
@ -187,7 +194,13 @@ func (res *MCResponse) Receive(r io.Reader, hdrBytes []byte) (n int, err error)
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
buf := make([]byte, klen+elen+bodyLen)
|
bufNeed := klen + elen + bodyLen
|
||||||
|
if buf != nil && cap(buf) >= bufNeed {
|
||||||
|
buf = buf[0:bufNeed]
|
||||||
|
} else {
|
||||||
|
buf = make([]byte, bufNeed)
|
||||||
|
}
|
||||||
|
|
||||||
m, err := io.ReadFull(r, buf)
|
m, err := io.ReadFull(r, buf)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
res.Extras = buf[0:elen]
|
res.Extras = buf[0:elen]
|
||||||
|
|
|
@ -1,47 +1,202 @@
|
||||||
COUCHBASE INC. COMMUNITY EDITION LICENSE AGREEMENT
|
Apache License
|
||||||
|
Version 2.0, January 2004
|
||||||
|
http://www.apache.org/licenses/
|
||||||
|
|
||||||
IMPORTANT-READ CAREFULLY: BY CLICKING THE "I ACCEPT" BOX OR INSTALLING,
|
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
||||||
DOWNLOADING OR OTHERWISE USING THIS SOFTWARE AND ANY ASSOCIATED
|
|
||||||
DOCUMENTATION, YOU, ON BEHALF OF YOURSELF OR AS AN AUTHORIZED
|
|
||||||
REPRESENTATIVE ON BEHALF OF AN ENTITY ("LICENSEE") AGREE TO ALL THE
|
|
||||||
TERMS OF THIS COMMUNITY EDITION LICENSE AGREEMENT (THE "AGREEMENT")
|
|
||||||
REGARDING YOUR USE OF THE SOFTWARE. YOU REPRESENT AND WARRANT THAT YOU
|
|
||||||
HAVE FULL LEGAL AUTHORITY TO BIND THE LICENSEE TO THIS AGREEMENT. IF YOU
|
|
||||||
DO NOT AGREE WITH ALL OF THESE TERMS, DO NOT SELECT THE "I ACCEPT" BOX
|
|
||||||
AND DO NOT INSTALL, DOWNLOAD OR OTHERWISE USE THE SOFTWARE. THE
|
|
||||||
EFFECTIVE DATE OF THIS AGREEMENT IS THE DATE ON WHICH YOU CLICK "I
|
|
||||||
ACCEPT" OR OTHERWISE INSTALL, DOWNLOAD OR USE THE SOFTWARE.
|
|
||||||
|
|
||||||
1. License Grant. Couchbase Inc. hereby grants Licensee, free of charge,
|
1. Definitions.
|
||||||
the non-exclusive right to use, copy, merge, publish, distribute,
|
|
||||||
sublicense, and/or sell copies of the Software, and to permit persons to
|
|
||||||
whom the Software is furnished to do so, subject to Licensee including
|
|
||||||
the following copyright notice in all copies or substantial portions of
|
|
||||||
the Software:
|
|
||||||
|
|
||||||
Couchbase (r) http://www.Couchbase.com Copyright 2016 Couchbase, Inc.
|
"License" shall mean the terms and conditions for use, reproduction,
|
||||||
|
and distribution as defined by Sections 1 through 9 of this document.
|
||||||
|
|
||||||
As used in this Agreement, "Software" means the object code version of
|
"Licensor" shall mean the copyright owner or entity authorized by
|
||||||
the applicable elastic data management server software provided by
|
the copyright owner that is granting the License.
|
||||||
Couchbase Inc.
|
|
||||||
|
|
||||||
2. Restrictions. Licensee will not reverse engineer, disassemble, or
|
"Legal Entity" shall mean the union of the acting entity and all
|
||||||
decompile the Software (except to the extent such restrictions are
|
other entities that control, are controlled by, or are under common
|
||||||
prohibited by law).
|
control with that entity. For the purposes of this definition,
|
||||||
|
"control" means (i) the power, direct or indirect, to cause the
|
||||||
|
direction or management of such entity, whether by contract or
|
||||||
|
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
||||||
|
outstanding shares, or (iii) beneficial ownership of such entity.
|
||||||
|
|
||||||
3. Support. Couchbase, Inc. will provide Licensee with access to, and
|
"You" (or "Your") shall mean an individual or Legal Entity
|
||||||
use of, the Couchbase, Inc. support forum available at the following
|
exercising permissions granted by this License.
|
||||||
URL: http://www.couchbase.org/forums/. Couchbase, Inc. may, at its
|
|
||||||
discretion, modify, suspend or terminate support at any time upon notice
|
"Source" form shall mean the preferred form for making modifications,
|
||||||
to Licensee.
|
including but not limited to software source code, documentation
|
||||||
|
source, and configuration files.
|
||||||
|
|
||||||
|
"Object" form shall mean any form resulting from mechanical
|
||||||
|
transformation or translation of a Source form, including but
|
||||||
|
not limited to compiled object code, generated documentation,
|
||||||
|
and conversions to other media types.
|
||||||
|
|
||||||
|
"Work" shall mean the work of authorship, whether in Source or
|
||||||
|
Object form, made available under the License, as indicated by a
|
||||||
|
copyright notice that is included in or attached to the work
|
||||||
|
(an example is provided in the Appendix below).
|
||||||
|
|
||||||
|
"Derivative Works" shall mean any work, whether in Source or Object
|
||||||
|
form, that is based on (or derived from) the Work and for which the
|
||||||
|
editorial revisions, annotations, elaborations, or other modifications
|
||||||
|
represent, as a whole, an original work of authorship. For the purposes
|
||||||
|
of this License, Derivative Works shall not include works that remain
|
||||||
|
separable from, or merely link (or bind by name) to the interfaces of,
|
||||||
|
the Work and Derivative Works thereof.
|
||||||
|
|
||||||
|
"Contribution" shall mean any work of authorship, including
|
||||||
|
the original version of the Work and any modifications or additions
|
||||||
|
to that Work or Derivative Works thereof, that is intentionally
|
||||||
|
submitted to Licensor for inclusion in the Work by the copyright owner
|
||||||
|
or by an individual or Legal Entity authorized to submit on behalf of
|
||||||
|
the copyright owner. For the purposes of this definition, "submitted"
|
||||||
|
means any form of electronic, verbal, or written communication sent
|
||||||
|
to the Licensor or its representatives, including but not limited to
|
||||||
|
communication on electronic mailing lists, source code control systems,
|
||||||
|
and issue tracking systems that are managed by, or on behalf of, the
|
||||||
|
Licensor for the purpose of discussing and improving the Work, but
|
||||||
|
excluding communication that is conspicuously marked or otherwise
|
||||||
|
designated in writing by the copyright owner as "Not a Contribution."
|
||||||
|
|
||||||
|
"Contributor" shall mean Licensor and any individual or Legal Entity
|
||||||
|
on behalf of whom a Contribution has been received by Licensor and
|
||||||
|
subsequently incorporated within the Work.
|
||||||
|
|
||||||
|
2. Grant of Copyright License. Subject to the terms and conditions of
|
||||||
|
this License, each Contributor hereby grants to You a perpetual,
|
||||||
|
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||||
|
copyright license to reproduce, prepare Derivative Works of,
|
||||||
|
publicly display, publicly perform, sublicense, and distribute the
|
||||||
|
Work and such Derivative Works in Source or Object form.
|
||||||
|
|
||||||
|
3. Grant of Patent License. Subject to the terms and conditions of
|
||||||
|
this License, each Contributor hereby grants to You a perpetual,
|
||||||
|
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||||
|
(except as stated in this section) patent license to make, have made,
|
||||||
|
use, offer to sell, sell, import, and otherwise transfer the Work,
|
||||||
|
where such license applies only to those patent claims licensable
|
||||||
|
by such Contributor that are necessarily infringed by their
|
||||||
|
Contribution(s) alone or by combination of their Contribution(s)
|
||||||
|
with the Work to which such Contribution(s) was submitted. If You
|
||||||
|
institute patent litigation against any entity (including a
|
||||||
|
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
||||||
|
or a Contribution incorporated within the Work constitutes direct
|
||||||
|
or contributory patent infringement, then any patent licenses
|
||||||
|
granted to You under this License for that Work shall terminate
|
||||||
|
as of the date such litigation is filed.
|
||||||
|
|
||||||
|
4. Redistribution. You may reproduce and distribute copies of the
|
||||||
|
Work or Derivative Works thereof in any medium, with or without
|
||||||
|
modifications, and in Source or Object form, provided that You
|
||||||
|
meet the following conditions:
|
||||||
|
|
||||||
|
(a) You must give any other recipients of the Work or
|
||||||
|
Derivative Works a copy of this License; and
|
||||||
|
|
||||||
|
(b) You must cause any modified files to carry prominent notices
|
||||||
|
stating that You changed the files; and
|
||||||
|
|
||||||
|
(c) You must retain, in the Source form of any Derivative Works
|
||||||
|
that You distribute, all copyright, patent, trademark, and
|
||||||
|
attribution notices from the Source form of the Work,
|
||||||
|
excluding those notices that do not pertain to any part of
|
||||||
|
the Derivative Works; and
|
||||||
|
|
||||||
|
(d) If the Work includes a "NOTICE" text file as part of its
|
||||||
|
distribution, then any Derivative Works that You distribute must
|
||||||
|
include a readable copy of the attribution notices contained
|
||||||
|
within such NOTICE file, excluding those notices that do not
|
||||||
|
pertain to any part of the Derivative Works, in at least one
|
||||||
|
of the following places: within a NOTICE text file distributed
|
||||||
|
as part of the Derivative Works; within the Source form or
|
||||||
|
documentation, if provided along with the Derivative Works; or,
|
||||||
|
within a display generated by the Derivative Works, if and
|
||||||
|
wherever such third-party notices normally appear. The contents
|
||||||
|
of the NOTICE file are for informational purposes only and
|
||||||
|
do not modify the License. You may add Your own attribution
|
||||||
|
notices within Derivative Works that You distribute, alongside
|
||||||
|
or as an addendum to the NOTICE text from the Work, provided
|
||||||
|
that such additional attribution notices cannot be construed
|
||||||
|
as modifying the License.
|
||||||
|
|
||||||
|
You may add Your own copyright statement to Your modifications and
|
||||||
|
may provide additional or different license terms and conditions
|
||||||
|
for use, reproduction, or distribution of Your modifications, or
|
||||||
|
for any such Derivative Works as a whole, provided Your use,
|
||||||
|
reproduction, and distribution of the Work otherwise complies with
|
||||||
|
the conditions stated in this License.
|
||||||
|
|
||||||
|
5. Submission of Contributions. Unless You explicitly state otherwise,
|
||||||
|
any Contribution intentionally submitted for inclusion in the Work
|
||||||
|
by You to the Licensor shall be under the terms and conditions of
|
||||||
|
this License, without any additional terms or conditions.
|
||||||
|
Notwithstanding the above, nothing herein shall supersede or modify
|
||||||
|
the terms of any separate license agreement you may have executed
|
||||||
|
with Licensor regarding such Contributions.
|
||||||
|
|
||||||
|
6. Trademarks. This License does not grant permission to use the trade
|
||||||
|
names, trademarks, service marks, or product names of the Licensor,
|
||||||
|
except as required for reasonable and customary use in describing the
|
||||||
|
origin of the Work and reproducing the content of the NOTICE file.
|
||||||
|
|
||||||
|
7. Disclaimer of Warranty. Unless required by applicable law or
|
||||||
|
agreed to in writing, Licensor provides the Work (and each
|
||||||
|
Contributor provides its Contributions) on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||||
|
implied, including, without limitation, any warranties or conditions
|
||||||
|
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
||||||
|
PARTICULAR PURPOSE. You are solely responsible for determining the
|
||||||
|
appropriateness of using or redistributing the Work and assume any
|
||||||
|
risks associated with Your exercise of permissions under this License.
|
||||||
|
|
||||||
|
8. Limitation of Liability. In no event and under no legal theory,
|
||||||
|
whether in tort (including negligence), contract, or otherwise,
|
||||||
|
unless required by applicable law (such as deliberate and grossly
|
||||||
|
negligent acts) or agreed to in writing, shall any Contributor be
|
||||||
|
liable to You for damages, including any direct, indirect, special,
|
||||||
|
incidental, or consequential damages of any character arising as a
|
||||||
|
result of this License or out of the use or inability to use the
|
||||||
|
Work (including but not limited to damages for loss of goodwill,
|
||||||
|
work stoppage, computer failure or malfunction, or any and all
|
||||||
|
other commercial damages or losses), even if such Contributor
|
||||||
|
has been advised of the possibility of such damages.
|
||||||
|
|
||||||
|
9. Accepting Warranty or Additional Liability. While redistributing
|
||||||
|
the Work or Derivative Works thereof, You may choose to offer,
|
||||||
|
and charge a fee for, acceptance of support, warranty, indemnity,
|
||||||
|
or other liability obligations and/or rights consistent with this
|
||||||
|
License. However, in accepting such obligations, You may act only
|
||||||
|
on Your own behalf and on Your sole responsibility, not on behalf
|
||||||
|
of any other Contributor, and only if You agree to indemnify,
|
||||||
|
defend, and hold each Contributor harmless for any liability
|
||||||
|
incurred by, or claims asserted against, such Contributor by reason
|
||||||
|
of your accepting any such warranty or additional liability.
|
||||||
|
|
||||||
|
END OF TERMS AND CONDITIONS
|
||||||
|
|
||||||
|
APPENDIX: How to apply the Apache License to your work.
|
||||||
|
|
||||||
|
To apply the Apache License to your work, attach the following
|
||||||
|
boilerplate notice, with the fields enclosed by brackets "{}"
|
||||||
|
replaced with your own identifying information. (Don't include
|
||||||
|
the brackets!) The text should be enclosed in the appropriate
|
||||||
|
comment syntax for the file format. We also recommend that a
|
||||||
|
file or class name and description of purpose be included on the
|
||||||
|
same "printed page" as the copyright notice for easier
|
||||||
|
identification within third-party archives.
|
||||||
|
|
||||||
|
Copyright {yyyy} {name of copyright owner}
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
|
||||||
4. Warranty Disclaimer and Limitation of Liability. THE SOFTWARE IS
|
|
||||||
PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED,
|
|
||||||
INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS
|
|
||||||
FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
||||||
COUCHBASE INC. OR THE AUTHORS OR COPYRIGHT HOLDERS IN THE SOFTWARE BE
|
|
||||||
LIABLE FOR ANY CLAIM, DAMAGES (IINCLUDING, WITHOUT LIMITATION, DIRECT,
|
|
||||||
INDIRECT OR CONSEQUENTIAL DAMAGES) OR OTHER LIABILITY, WHETHER IN AN
|
|
||||||
ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
||||||
CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
||||||
SOFTWARE.
|
|
||||||
|
|
|
@ -99,10 +99,10 @@ github.com/boombuler/barcode/qr
|
||||||
github.com/boombuler/barcode/utils
|
github.com/boombuler/barcode/utils
|
||||||
# github.com/bradfitz/gomemcache v0.0.0-20190329173943-551aad21a668
|
# github.com/bradfitz/gomemcache v0.0.0-20190329173943-551aad21a668
|
||||||
github.com/bradfitz/gomemcache/memcache
|
github.com/bradfitz/gomemcache/memcache
|
||||||
# github.com/couchbase/gomemcached v0.0.0-20190515232915-c4b4ca0eb21d
|
# github.com/couchbase/gomemcached v0.0.0-20191004160342-7b5da2ec40b2
|
||||||
github.com/couchbase/gomemcached
|
github.com/couchbase/gomemcached
|
||||||
github.com/couchbase/gomemcached/client
|
github.com/couchbase/gomemcached/client
|
||||||
# github.com/couchbase/goutils v0.0.0-20190315194238-f9d42b11473b
|
# github.com/couchbase/goutils v0.0.0-20191018232750-b49639060d85
|
||||||
github.com/couchbase/goutils/logging
|
github.com/couchbase/goutils/logging
|
||||||
github.com/couchbase/goutils/scramsha
|
github.com/couchbase/goutils/scramsha
|
||||||
# github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd
|
# github.com/couchbase/vellum v0.0.0-20190829182332-ef2e028c01fd
|
||||||
|
|
Loading…
Reference in New Issue