23
`github.com/bytedance/sonic/internal/caching`
24
`github.com/bytedance/sonic/internal/native/types`
25
`github.com/bytedance/sonic/internal/rt`
31
_MaxStackBytes = _MaxStack * _PtrBytes
32
_MaxDigitNums = types.MaxDigitNums
36
_PtrBytes = _PTR_SIZE / 8
37
_FsmOffset = (_MaxStack + 1) * _PtrBytes
38
_DbufOffset = _FsmOffset + int64(unsafe.Sizeof(types.StateMachine{})) + types.MAX_RECURSE * _PtrBytes
39
_StackSize = unsafe.Sizeof(_Stack{})
43
stackPool = sync.Pool{}
44
valueCache = []unsafe.Pointer(nil)
45
fieldCache = []*caching.FieldMap(nil)
46
fieldCacheMux = sync.Mutex{}
47
programCache = caching.CreateProgramCache()
52
sb [_MaxStack]unsafe.Pointer
54
vp [types.MAX_RECURSE]unsafe.Pointer
55
dp [_MaxDigitNums]byte
68
var _KeepAlive struct {
80
frame_decoder [_FP_offs]byte
81
frame_generic [_VD_offs]byte
85
argPtrs = []bool{true, false, false, true, true, false, true, false, true}
90
argPtrs_generic = []bool{true}
91
localPtrs_generic = []bool{}
94
func newStack() *_Stack {
95
if ret := stackPool.Get(); ret == nil {
102
func resetStack(p *_Stack) {
103
memclrNoHeapPointers(unsafe.Pointer(p), _StackSize)
106
func freeStack(p *_Stack) {
111
func freezeValue(v unsafe.Pointer) uintptr {
112
valueCache = append(valueCache, v)
116
func freezeFields(v *caching.FieldMap) int64 {
118
fieldCache = append(fieldCache, v)
119
fieldCacheMux.Unlock()
120
return referenceFields(v)
123
func referenceFields(v *caching.FieldMap) int64 {
124
return int64(uintptr(unsafe.Pointer(v)))
127
func makeDecoder(vt *rt.GoType, _ ...interface{}) (interface{}, error) {
128
if pp, err := newCompiler().compile(vt.Pack()); err != nil {
131
return newAssembler(pp).Load(), nil
135
func findOrCompile(vt *rt.GoType) (_Decoder, error) {
136
if val := programCache.Get(vt); val != nil {
137
return val.(_Decoder), nil
138
} else if ret, err := programCache.Compute(vt, makeDecoder); err == nil {
139
return ret.(_Decoder), nil