1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
|
// Copyright 2021 The gVisor Authors.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//go:build linux
// +build linux
package sharedmem
import (
"golang.org/x/sys/unix"
"gvisor.dev/gvisor/pkg/cleanup"
"gvisor.dev/gvisor/pkg/eventfd"
"gvisor.dev/gvisor/pkg/tcpip/buffer"
"gvisor.dev/gvisor/pkg/tcpip/link/sharedmem/pipe"
"gvisor.dev/gvisor/pkg/tcpip/link/sharedmem/queue"
)
// serverTx represents the server end of the sharedmem queue and is used to send
// packets to the peer in the buffers posted by the peer in the fillPipe.
type serverTx struct {
// fillPipe represents the receive end of the pipe that carries the RxBuffers
// posted by the peer.
fillPipe pipe.Rx
// completionPipe represents the transmit end of the pipe that carries the
// descriptors for filled RxBuffers.
completionPipe pipe.Tx
// data represents the buffer area where the packet payload is held.
data []byte
// eventFD is used to notify the peer when fill requests are fulfilled.
eventFD eventfd.Eventfd
// sharedData the memory region to use to enable/disable notifications.
sharedData []byte
}
// init initializes all tstate needed by the serverTx queue based on the
// information provided.
//
// The caller always retains ownership of all file descriptors passed in. The
// queue implementation will duplicate any that it may need in the future.
func (s *serverTx) init(c *QueueConfig) error {
// Map in all buffers.
fillPipeMem, err := getBuffer(c.TxPipeFD)
if err != nil {
return err
}
cu := cleanup.Make(func() { unix.Munmap(fillPipeMem) })
defer cu.Clean()
completionPipeMem, err := getBuffer(c.RxPipeFD)
if err != nil {
return err
}
cu.Add(func() { unix.Munmap(completionPipeMem) })
data, err := getBuffer(c.DataFD)
if err != nil {
return err
}
cu.Add(func() { unix.Munmap(data) })
sharedData, err := getBuffer(c.SharedDataFD)
if err != nil {
return err
}
cu.Add(func() { unix.Munmap(sharedData) })
// Duplicate the eventFD so that caller can close it but we can still
// use it.
efd, err := c.EventFD.Dup()
if err != nil {
return err
}
cu.Add(func() { efd.Close() })
cu.Release()
s.fillPipe.Init(fillPipeMem)
s.completionPipe.Init(completionPipeMem)
s.data = data
s.eventFD = efd
s.sharedData = sharedData
return nil
}
func (s *serverTx) cleanup() {
unix.Munmap(s.fillPipe.Bytes())
unix.Munmap(s.completionPipe.Bytes())
unix.Munmap(s.data)
unix.Munmap(s.sharedData)
s.eventFD.Close()
}
// fillPacket copies the data in the provided views into buffers pulled from the
// fillPipe and returns a slice of RxBuffers that contain the copied data as
// well as the total number of bytes copied.
//
// To avoid allocations the filledBuffers are appended to the buffers slice
// which will be grown as required.
func (s *serverTx) fillPacket(views []buffer.View, buffers []queue.RxBuffer) (filledBuffers []queue.RxBuffer, totalCopied uint32) {
filledBuffers = buffers[:0]
// fillBuffer copies as much of the views as possible into the provided buffer
// and returns any left over views (if any).
fillBuffer := func(buffer *queue.RxBuffer, views []buffer.View) (left []buffer.View) {
if len(views) == 0 {
return nil
}
availBytes := buffer.Size
copied := uint64(0)
for availBytes > 0 && len(views) > 0 {
n := copy(s.data[buffer.Offset+copied:][:uint64(buffer.Size)-copied], views[0])
views[0].TrimFront(n)
if !views[0].IsEmpty() {
break
}
views = views[1:]
copied += uint64(n)
availBytes -= uint32(n)
}
buffer.Size = uint32(copied)
return views
}
for len(views) > 0 {
var b []byte
// Spin till we get a free buffer reposted by the peer.
for {
if b = s.fillPipe.Pull(); b != nil {
break
}
}
rxBuffer := queue.DecodeRxBufferHeader(b)
// Copy the packet into the posted buffer.
views = fillBuffer(&rxBuffer, views)
totalCopied += rxBuffer.Size
filledBuffers = append(filledBuffers, rxBuffer)
}
return filledBuffers, totalCopied
}
func (s *serverTx) transmit(views []buffer.View) bool {
buffers := make([]queue.RxBuffer, 8)
buffers, totalCopied := s.fillPacket(views, buffers)
b := s.completionPipe.Push(queue.RxCompletionSize(len(buffers)))
if b == nil {
return false
}
queue.EncodeRxCompletion(b, totalCopied, 0 /* reserved */)
for i := 0; i < len(buffers); i++ {
queue.EncodeRxCompletionBuffer(b, i, buffers[i])
}
s.completionPipe.Flush()
s.fillPipe.Flush()
return true
}
func (s *serverTx) notify() {
s.eventFD.Notify()
}
|