mirror of
https://github.com/netbirdio/netbird.git
synced 2024-12-16 03:41:26 +01:00
720 lines
20 KiB
Go
720 lines
20 KiB
Go
package peer
|
|
|
|
import (
|
|
"context"
|
|
"math/rand"
|
|
"net"
|
|
"os"
|
|
"runtime"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/cenkalti/backoff/v4"
|
|
"github.com/pion/ice/v3"
|
|
log "github.com/sirupsen/logrus"
|
|
"golang.zx2c4.com/wireguard/wgctrl/wgtypes"
|
|
|
|
"github.com/netbirdio/netbird/client/internal/stdnet"
|
|
"github.com/netbirdio/netbird/client/internal/wgproxy"
|
|
"github.com/netbirdio/netbird/iface"
|
|
relayClient "github.com/netbirdio/netbird/relay/client"
|
|
"github.com/netbirdio/netbird/route"
|
|
nbnet "github.com/netbirdio/netbird/util/net"
|
|
)
|
|
|
|
type ConnPriority int
|
|
|
|
const (
|
|
defaultWgKeepAlive = 25 * time.Second
|
|
|
|
connPriorityRelay ConnPriority = 1
|
|
connPriorityICETurn ConnPriority = 1
|
|
connPriorityICEP2P ConnPriority = 2
|
|
)
|
|
|
|
type WgConfig struct {
|
|
WgListenPort int
|
|
RemoteKey string
|
|
WgInterface iface.IWGIface
|
|
AllowedIps string
|
|
PreSharedKey *wgtypes.Key
|
|
}
|
|
|
|
// ConnConfig is a peer Connection configuration
|
|
type ConnConfig struct {
|
|
// Key is a public key of a remote peer
|
|
Key string
|
|
// LocalKey is a public key of a local peer
|
|
LocalKey string
|
|
|
|
Timeout time.Duration
|
|
|
|
WgConfig WgConfig
|
|
|
|
LocalWgPort int
|
|
|
|
// RosenpassPubKey is this peer's Rosenpass public key
|
|
RosenpassPubKey []byte
|
|
// RosenpassPubKey is this peer's RosenpassAddr server address (IP:port)
|
|
RosenpassAddr string
|
|
|
|
// ICEConfig ICE protocol configuration
|
|
ICEConfig ICEConfig
|
|
}
|
|
|
|
type WorkerCallbacks struct {
|
|
OnRelayReadyCallback func(info RelayConnInfo)
|
|
OnRelayStatusChanged func(ConnStatus)
|
|
|
|
OnICEConnReadyCallback func(ConnPriority, ICEConnInfo)
|
|
OnICEStatusChanged func(ConnStatus)
|
|
}
|
|
|
|
type Conn struct {
|
|
log *log.Entry
|
|
mu sync.Mutex
|
|
ctx context.Context
|
|
ctxCancel context.CancelFunc
|
|
config ConnConfig
|
|
statusRecorder *Status
|
|
wgProxyFactory *wgproxy.Factory
|
|
wgProxyICE wgproxy.Proxy
|
|
wgProxyRelay wgproxy.Proxy
|
|
signaler *Signaler
|
|
relayManager *relayClient.Manager
|
|
allowedIPsIP string
|
|
handshaker *Handshaker
|
|
|
|
onConnected func(remoteWireGuardKey string, remoteRosenpassPubKey []byte, wireGuardIP string, remoteRosenpassAddr string)
|
|
onDisconnected func(remotePeer string, wgIP string)
|
|
|
|
statusRelay ConnStatus
|
|
statusICE ConnStatus
|
|
currentConnType ConnPriority
|
|
opened bool // this flag is used to prevent close in case of not opened connection
|
|
|
|
workerICE *WorkerICE
|
|
workerRelay *WorkerRelay
|
|
|
|
connID nbnet.ConnectionID
|
|
beforeAddPeerHooks []nbnet.AddHookFunc
|
|
afterRemovePeerHooks []nbnet.RemoveHookFunc
|
|
|
|
endpointRelay *net.UDPAddr
|
|
|
|
// for reconnection operations
|
|
iCEDisconnected chan bool
|
|
relayDisconnected chan bool
|
|
}
|
|
|
|
// NewConn creates a new not opened Conn to the remote peer.
|
|
// To establish a connection run Conn.Open
|
|
func NewConn(engineCtx context.Context, config ConnConfig, statusRecorder *Status, wgProxyFactory *wgproxy.Factory, signaler *Signaler, iFaceDiscover stdnet.ExternalIFaceDiscover, relayManager *relayClient.Manager) (*Conn, error) {
|
|
_, allowedIPsIP, err := net.ParseCIDR(config.WgConfig.AllowedIps)
|
|
if err != nil {
|
|
log.Errorf("failed to parse allowedIPS: %v", err)
|
|
return nil, err
|
|
}
|
|
|
|
ctx, ctxCancel := context.WithCancel(engineCtx)
|
|
connLog := log.WithField("peer", config.Key)
|
|
|
|
var conn = &Conn{
|
|
log: connLog,
|
|
ctx: ctx,
|
|
ctxCancel: ctxCancel,
|
|
config: config,
|
|
statusRecorder: statusRecorder,
|
|
wgProxyFactory: wgProxyFactory,
|
|
signaler: signaler,
|
|
relayManager: relayManager,
|
|
allowedIPsIP: allowedIPsIP.String(),
|
|
statusRelay: StatusDisconnected,
|
|
statusICE: StatusDisconnected,
|
|
iCEDisconnected: make(chan bool, 1),
|
|
relayDisconnected: make(chan bool, 1),
|
|
}
|
|
|
|
rFns := WorkerRelayCallbacks{
|
|
OnConnReady: conn.relayConnectionIsReady,
|
|
OnDisconnected: conn.onWorkerRelayStateDisconnected,
|
|
}
|
|
|
|
wFns := WorkerICECallbacks{
|
|
OnConnReady: conn.iCEConnectionIsReady,
|
|
OnStatusChanged: conn.onWorkerICEStateDisconnected,
|
|
}
|
|
|
|
conn.handshaker = NewHandshaker(ctx, connLog, config, signaler)
|
|
conn.workerRelay = NewWorkerRelay(ctx, connLog, config, relayManager, rFns)
|
|
|
|
relayIsSupportedLocally := conn.workerRelay.RelayIsSupportedLocally()
|
|
conn.workerICE, err = NewWorkerICE(ctx, connLog, config, config.ICEConfig, signaler, iFaceDiscover, statusRecorder, relayIsSupportedLocally, wFns)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
conn.handshaker.AddOnNewOfferListener(conn.workerRelay.OnNewOffer)
|
|
if os.Getenv("NB_FORCE_RELAY") != "true" {
|
|
conn.handshaker.AddOnNewOfferListener(conn.workerICE.OnNewOffer)
|
|
}
|
|
|
|
go conn.handshaker.Listen()
|
|
|
|
return conn, nil
|
|
}
|
|
|
|
// Open opens connection to the remote peer
|
|
// It will try to establish a connection using ICE and in parallel with relay. The higher priority connection type will
|
|
// be used.
|
|
func (conn *Conn) Open() {
|
|
conn.log.Debugf("open connection to peer")
|
|
conn.mu.Lock()
|
|
defer conn.mu.Unlock()
|
|
conn.opened = true
|
|
|
|
peerState := State{
|
|
PubKey: conn.config.Key,
|
|
IP: strings.Split(conn.config.WgConfig.AllowedIps, "/")[0],
|
|
ConnStatusUpdate: time.Now(),
|
|
ConnStatus: StatusDisconnected,
|
|
Mux: new(sync.RWMutex),
|
|
}
|
|
err := conn.statusRecorder.UpdatePeerState(peerState)
|
|
if err != nil {
|
|
conn.log.Warnf("error while updating the state err: %v", err)
|
|
}
|
|
|
|
conn.waitInitialRandomSleepTime()
|
|
|
|
err = conn.doHandshake()
|
|
if err != nil {
|
|
conn.log.Errorf("failed to send offer: %v", err)
|
|
}
|
|
|
|
if conn.workerRelay.IsController() {
|
|
go conn.reconnectLoopWithRetry()
|
|
} else {
|
|
go conn.reconnectLoopForOnDisconnectedEvent()
|
|
}
|
|
}
|
|
|
|
// Close closes this peer Conn issuing a close event to the Conn closeCh
|
|
func (conn *Conn) Close() {
|
|
conn.mu.Lock()
|
|
defer conn.mu.Unlock()
|
|
|
|
conn.ctxCancel()
|
|
|
|
if !conn.opened {
|
|
log.Infof("IGNORE close connection to peer")
|
|
return
|
|
}
|
|
|
|
if conn.wgProxyRelay != nil {
|
|
err := conn.wgProxyRelay.CloseConn()
|
|
if err != nil {
|
|
conn.log.Errorf("failed to close wg proxy for relay: %v", err)
|
|
}
|
|
conn.wgProxyRelay = nil
|
|
}
|
|
|
|
if conn.wgProxyICE != nil {
|
|
err := conn.wgProxyICE.CloseConn()
|
|
if err != nil {
|
|
conn.log.Errorf("failed to close wg proxy for ice: %v", err)
|
|
}
|
|
conn.wgProxyICE = nil
|
|
}
|
|
|
|
err := conn.config.WgConfig.WgInterface.RemovePeer(conn.config.WgConfig.RemoteKey)
|
|
if err != nil {
|
|
conn.log.Errorf("failed to remove wg endpoint: %v", err)
|
|
}
|
|
|
|
if conn.connID != "" {
|
|
for _, hook := range conn.afterRemovePeerHooks {
|
|
if err := hook(conn.connID); err != nil {
|
|
conn.log.Errorf("After remove peer hook failed: %v", err)
|
|
}
|
|
}
|
|
conn.connID = ""
|
|
}
|
|
|
|
if conn.evalStatus() == StatusConnected && conn.onDisconnected != nil {
|
|
conn.onDisconnected(conn.config.WgConfig.RemoteKey, conn.config.WgConfig.AllowedIps)
|
|
}
|
|
|
|
conn.statusRelay = StatusDisconnected
|
|
conn.statusICE = StatusDisconnected
|
|
|
|
peerState := State{
|
|
PubKey: conn.config.Key,
|
|
ConnStatus: StatusDisconnected,
|
|
ConnStatusUpdate: time.Now(),
|
|
Mux: new(sync.RWMutex),
|
|
}
|
|
err = conn.statusRecorder.UpdatePeerState(peerState)
|
|
if err != nil {
|
|
// pretty common error because by that time Engine can already remove the peer and status won't be available.
|
|
// todo rethink status updates
|
|
conn.log.Debugf("error while updating peer's state, err: %v", err)
|
|
}
|
|
if err := conn.statusRecorder.UpdateWireGuardPeerState(conn.config.Key, iface.WGStats{}); err != nil {
|
|
conn.log.Debugf("failed to reset wireguard stats for peer: %s", err)
|
|
}
|
|
}
|
|
|
|
// OnRemoteAnswer handles an offer from the remote peer and returns true if the message was accepted, false otherwise
|
|
// doesn't block, discards the message if connection wasn't ready
|
|
func (conn *Conn) OnRemoteAnswer(answer OfferAnswer) bool {
|
|
conn.log.Debugf("OnRemoteAnswer, status ICE: %s, status relay: %s", conn.statusICE, conn.statusRelay)
|
|
return conn.handshaker.OnRemoteAnswer(answer)
|
|
}
|
|
|
|
// OnRemoteCandidate Handles ICE connection Candidate provided by the remote peer.
|
|
func (conn *Conn) OnRemoteCandidate(candidate ice.Candidate, haRoutes route.HAMap) {
|
|
conn.workerICE.OnRemoteCandidate(candidate, haRoutes)
|
|
}
|
|
|
|
func (conn *Conn) AddBeforeAddPeerHook(hook nbnet.AddHookFunc) {
|
|
conn.beforeAddPeerHooks = append(conn.beforeAddPeerHooks, hook)
|
|
}
|
|
func (conn *Conn) AddAfterRemovePeerHook(hook nbnet.RemoveHookFunc) {
|
|
conn.afterRemovePeerHooks = append(conn.afterRemovePeerHooks, hook)
|
|
}
|
|
|
|
// SetOnConnected sets a handler function to be triggered by Conn when a new connection to a remote peer established
|
|
func (conn *Conn) SetOnConnected(handler func(remoteWireGuardKey string, remoteRosenpassPubKey []byte, wireGuardIP string, remoteRosenpassAddr string)) {
|
|
conn.onConnected = handler
|
|
}
|
|
|
|
// SetOnDisconnected sets a handler function to be triggered by Conn when a connection to a remote disconnected
|
|
func (conn *Conn) SetOnDisconnected(handler func(remotePeer string, wgIP string)) {
|
|
conn.onDisconnected = handler
|
|
}
|
|
|
|
func (conn *Conn) OnRemoteOffer(offer OfferAnswer) bool {
|
|
conn.log.Debugf("OnRemoteOffer, on status ICE: %s, status Relay: %s", conn.statusICE, conn.statusRelay)
|
|
return conn.handshaker.OnRemoteOffer(offer)
|
|
}
|
|
|
|
// WgConfig returns the WireGuard config
|
|
func (conn *Conn) WgConfig() WgConfig {
|
|
return conn.config.WgConfig
|
|
}
|
|
|
|
// Status returns current status of the Conn
|
|
func (conn *Conn) Status() ConnStatus {
|
|
conn.mu.Lock()
|
|
defer conn.mu.Unlock()
|
|
return conn.evalStatus()
|
|
}
|
|
|
|
func (conn *Conn) GetKey() string {
|
|
return conn.config.Key
|
|
}
|
|
|
|
func (conn *Conn) reconnectLoopWithRetry() {
|
|
// give chance to the peer to establish the initial connection
|
|
select {
|
|
case <-conn.ctx.Done():
|
|
case <-time.After(3 * time.Second):
|
|
}
|
|
|
|
bo := backoff.WithContext(&backoff.ExponentialBackOff{
|
|
InitialInterval: 800 * time.Millisecond,
|
|
RandomizationFactor: 0,
|
|
Multiplier: 1.99,
|
|
MaxInterval: conn.config.Timeout * time.Second,
|
|
MaxElapsedTime: 0,
|
|
Stop: backoff.Stop,
|
|
Clock: backoff.SystemClock,
|
|
}, conn.ctx)
|
|
|
|
ticker := backoff.NewTicker(bo)
|
|
defer ticker.Stop()
|
|
|
|
<-ticker.C // consume the initial tick what is happening right after the ticker has been created
|
|
|
|
no := time.Now()
|
|
for {
|
|
select {
|
|
case t := <-ticker.C:
|
|
if t.IsZero() {
|
|
// in case if the ticker has been canceled by context then avoid the temporary loop
|
|
return
|
|
}
|
|
// checks if there is peer connection is established via relay or ice and that it has a wireguard handshake and skip offer
|
|
// todo check wg handshake
|
|
conn.log.Tracef("ticker timedout, relay state: %s, ice state: %s, elapsed time: %s", conn.statusRelay, conn.statusICE, time.Since(no))
|
|
no = time.Now()
|
|
|
|
if conn.statusRelay == StatusConnected && conn.statusICE == StatusConnected {
|
|
continue
|
|
}
|
|
|
|
log.Debugf("ticker timed out, retry to do handshake")
|
|
err := conn.doHandshake()
|
|
if err != nil {
|
|
conn.log.Errorf("failed to do handshake: %v", err)
|
|
}
|
|
case changed := <-conn.relayDisconnected:
|
|
if !changed {
|
|
continue
|
|
}
|
|
conn.log.Debugf("Relay state changed, reset reconnect timer")
|
|
bo.Reset()
|
|
case changed := <-conn.iCEDisconnected:
|
|
if !changed {
|
|
continue
|
|
}
|
|
conn.log.Debugf("ICE state changed, reset reconnect timer")
|
|
bo.Reset()
|
|
case <-conn.ctx.Done():
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// reconnectLoopForOnDisconnectedEvent is used when the peer is not a controller and it should reconnect to the peer
|
|
// when the connection is lost. It will try to establish a connection only once time if before the connection was established
|
|
// It track separately the ice and relay connection status. Just because a lover priority connection reestablished it does not
|
|
// mean that to switch to it. We always force to use the higher priority connection.
|
|
func (conn *Conn) reconnectLoopForOnDisconnectedEvent() {
|
|
for {
|
|
select {
|
|
case changed := <-conn.relayDisconnected:
|
|
if !changed {
|
|
continue
|
|
}
|
|
conn.log.Debugf("Relay state changed, try to send new offer")
|
|
case changed := <-conn.iCEDisconnected:
|
|
if !changed {
|
|
continue
|
|
}
|
|
conn.log.Debugf("ICE state changed, try to send new offer")
|
|
case <-conn.ctx.Done():
|
|
return
|
|
}
|
|
|
|
err := conn.doHandshake()
|
|
if err != nil {
|
|
conn.log.Errorf("failed to do handshake: %v", err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// configureConnection starts proxying traffic from/to local Wireguard and sets connection status to StatusConnected
|
|
func (conn *Conn) iCEConnectionIsReady(priority ConnPriority, iceConnInfo ICEConnInfo) {
|
|
conn.mu.Lock()
|
|
defer conn.mu.Unlock()
|
|
|
|
if conn.ctx.Err() != nil {
|
|
return
|
|
}
|
|
|
|
conn.log.Debugf("ICE connection is ready")
|
|
|
|
conn.statusICE = StatusConnected
|
|
|
|
if conn.currentConnType > priority {
|
|
return
|
|
}
|
|
|
|
if conn.currentConnType != 0 {
|
|
conn.log.Infof("update connection to ICE type")
|
|
} else {
|
|
conn.log.Infof("set ICE to active connection")
|
|
}
|
|
|
|
var (
|
|
endpoint net.Addr
|
|
wgProxy wgproxy.Proxy
|
|
)
|
|
if iceConnInfo.RelayedOnLocal {
|
|
conn.log.Debugf("setup ice turn connection")
|
|
wgProxy = conn.wgProxyFactory.GetProxy(conn.ctx)
|
|
ep, err := wgProxy.AddTurnConn(iceConnInfo.RemoteConn)
|
|
if err != nil {
|
|
conn.log.Errorf("failed to add turn net.Conn to local proxy: %v", err)
|
|
err = wgProxy.CloseConn()
|
|
if err != nil {
|
|
conn.log.Warnf("failed to close turn proxy connection: %v", err)
|
|
}
|
|
return
|
|
}
|
|
endpoint = ep
|
|
} else {
|
|
endpoint = iceConnInfo.RemoteConn.RemoteAddr()
|
|
}
|
|
|
|
endpointUdpAddr, _ := net.ResolveUDPAddr(endpoint.Network(), endpoint.String())
|
|
conn.log.Debugf("Conn resolved IP is %s for endopint %s", endpoint, endpointUdpAddr.IP)
|
|
|
|
conn.connID = nbnet.GenerateConnID()
|
|
for _, hook := range conn.beforeAddPeerHooks {
|
|
if err := hook(conn.connID, endpointUdpAddr.IP); err != nil {
|
|
conn.log.Errorf("Before add peer hook failed: %v", err)
|
|
}
|
|
}
|
|
|
|
err := conn.config.WgConfig.WgInterface.UpdatePeer(conn.config.WgConfig.RemoteKey, conn.config.WgConfig.AllowedIps, defaultWgKeepAlive, endpointUdpAddr, conn.config.WgConfig.PreSharedKey)
|
|
if err != nil {
|
|
if wgProxy != nil {
|
|
if err := wgProxy.CloseConn(); err != nil {
|
|
conn.log.Warnf("Failed to close turn connection: %v", err)
|
|
}
|
|
}
|
|
conn.log.Warnf("Failed to update wg peer configuration: %v", err)
|
|
return
|
|
}
|
|
|
|
if conn.wgProxyICE != nil {
|
|
if err := conn.wgProxyICE.CloseConn(); err != nil {
|
|
conn.log.Warnf("failed to close depracated wg proxy conn: %v", err)
|
|
}
|
|
}
|
|
conn.wgProxyICE = wgProxy
|
|
|
|
conn.currentConnType = priority
|
|
|
|
peerState := State{
|
|
LocalIceCandidateType: iceConnInfo.LocalIceCandidateType,
|
|
RemoteIceCandidateType: iceConnInfo.RemoteIceCandidateType,
|
|
LocalIceCandidateEndpoint: iceConnInfo.LocalIceCandidateEndpoint,
|
|
RemoteIceCandidateEndpoint: iceConnInfo.RemoteIceCandidateEndpoint,
|
|
Direct: iceConnInfo.Direct,
|
|
Relayed: iceConnInfo.Relayed,
|
|
}
|
|
|
|
conn.updateStatus(peerState, iceConnInfo.RosenpassPubKey, iceConnInfo.RosenpassAddr)
|
|
}
|
|
|
|
// todo review to make sense to handle connection and disconnected status also?
|
|
func (conn *Conn) onWorkerICEStateDisconnected(newState ConnStatus) {
|
|
conn.mu.Lock()
|
|
defer conn.mu.Unlock()
|
|
|
|
conn.log.Tracef("ICE connection state changed to %s", newState)
|
|
defer func() {
|
|
|
|
changed := conn.statusICE != newState && newState != StatusConnecting
|
|
conn.statusICE = newState
|
|
|
|
select {
|
|
case conn.iCEDisconnected <- changed:
|
|
default:
|
|
}
|
|
}()
|
|
|
|
// switch back to relay connection
|
|
if conn.endpointRelay != nil {
|
|
conn.log.Debugf("ICE disconnected, set Relay to active connection")
|
|
err := conn.configureWGEndpoint(conn.endpointRelay)
|
|
if err != nil {
|
|
conn.log.Errorf("failed to switch to relay conn: %v", err)
|
|
}
|
|
// todo update status to relay related things
|
|
return
|
|
}
|
|
|
|
if conn.statusRelay == StatusConnected {
|
|
return
|
|
}
|
|
|
|
if conn.evalStatus() == newState {
|
|
return
|
|
}
|
|
|
|
if newState > conn.statusICE {
|
|
peerState := State{
|
|
PubKey: conn.config.Key,
|
|
ConnStatus: newState,
|
|
ConnStatusUpdate: time.Now(),
|
|
Mux: new(sync.RWMutex),
|
|
}
|
|
_ = conn.statusRecorder.UpdatePeerState(peerState)
|
|
}
|
|
}
|
|
|
|
func (conn *Conn) relayConnectionIsReady(rci RelayConnInfo) {
|
|
conn.mu.Lock()
|
|
defer conn.mu.Unlock()
|
|
|
|
if conn.ctx.Err() != nil {
|
|
return
|
|
}
|
|
|
|
conn.log.Debugf("Relay connection is ready to use")
|
|
conn.statusRelay = StatusConnected
|
|
|
|
wgProxy := conn.wgProxyFactory.GetProxy(conn.ctx)
|
|
endpoint, err := wgProxy.AddTurnConn(rci.relayedConn)
|
|
if err != nil {
|
|
conn.log.Errorf("failed to add relayed net.Conn to local proxy: %v", err)
|
|
return
|
|
}
|
|
|
|
endpointUdpAddr, _ := net.ResolveUDPAddr(endpoint.Network(), endpoint.String())
|
|
conn.endpointRelay = endpointUdpAddr
|
|
conn.log.Debugf("conn resolved IP for %s: %s", endpoint, endpointUdpAddr.IP)
|
|
|
|
if conn.currentConnType > connPriorityRelay {
|
|
if conn.statusICE == StatusConnected {
|
|
log.Debugf("do not switch to relay because current priority is: %v", conn.currentConnType)
|
|
return
|
|
}
|
|
}
|
|
|
|
conn.connID = nbnet.GenerateConnID()
|
|
for _, hook := range conn.beforeAddPeerHooks {
|
|
if err := hook(conn.connID, endpointUdpAddr.IP); err != nil {
|
|
conn.log.Errorf("Before add peer hook failed: %v", err)
|
|
}
|
|
}
|
|
|
|
err = conn.configureWGEndpoint(endpointUdpAddr)
|
|
if err != nil {
|
|
if err := wgProxy.CloseConn(); err != nil {
|
|
conn.log.Warnf("Failed to close relay connection: %v", err)
|
|
}
|
|
conn.log.Errorf("Failed to update wg peer configuration: %v", err)
|
|
return
|
|
}
|
|
|
|
if conn.wgProxyRelay != nil {
|
|
if err := conn.wgProxyRelay.CloseConn(); err != nil {
|
|
conn.log.Warnf("failed to close depracated wg proxy conn: %v", err)
|
|
}
|
|
}
|
|
conn.wgProxyRelay = wgProxy
|
|
conn.currentConnType = connPriorityRelay
|
|
|
|
peerState := State{
|
|
Direct: false,
|
|
Relayed: true,
|
|
}
|
|
|
|
conn.log.Infof("start to communicate with peer via relay")
|
|
conn.updateStatus(peerState, rci.rosenpassPubKey, rci.rosenpassAddr)
|
|
}
|
|
|
|
func (conn *Conn) onWorkerRelayStateDisconnected() {
|
|
conn.mu.Lock()
|
|
defer conn.mu.Unlock()
|
|
defer func() {
|
|
changed := conn.statusRelay != StatusDisconnected
|
|
conn.statusRelay = StatusDisconnected
|
|
|
|
select {
|
|
case conn.relayDisconnected <- changed:
|
|
default:
|
|
}
|
|
}()
|
|
|
|
if conn.wgProxyRelay != nil {
|
|
conn.endpointRelay = nil
|
|
_ = conn.wgProxyRelay.CloseConn()
|
|
conn.wgProxyRelay = nil
|
|
}
|
|
|
|
if conn.statusICE == StatusConnected {
|
|
return
|
|
}
|
|
|
|
if conn.evalStatus() == StatusDisconnected {
|
|
return
|
|
}
|
|
|
|
if StatusDisconnected > conn.statusRelay {
|
|
peerState := State{
|
|
PubKey: conn.config.Key,
|
|
ConnStatus: StatusDisconnected,
|
|
ConnStatusUpdate: time.Now(),
|
|
Mux: new(sync.RWMutex),
|
|
}
|
|
_ = conn.statusRecorder.UpdatePeerState(peerState)
|
|
}
|
|
}
|
|
|
|
func (conn *Conn) configureWGEndpoint(addr *net.UDPAddr) error {
|
|
return conn.config.WgConfig.WgInterface.UpdatePeer(
|
|
conn.config.WgConfig.RemoteKey,
|
|
conn.config.WgConfig.AllowedIps,
|
|
defaultWgKeepAlive,
|
|
addr,
|
|
conn.config.WgConfig.PreSharedKey,
|
|
)
|
|
}
|
|
func (conn *Conn) updateStatus(peerState State, remoteRosenpassPubKey []byte, remoteRosenpassAddr string) {
|
|
peerState.PubKey = conn.config.Key
|
|
peerState.ConnStatus = StatusConnected
|
|
peerState.ConnStatusUpdate = time.Now()
|
|
peerState.RosenpassEnabled = isRosenpassEnabled(remoteRosenpassPubKey)
|
|
peerState.Mux = new(sync.RWMutex)
|
|
|
|
err := conn.statusRecorder.UpdatePeerState(peerState)
|
|
if err != nil {
|
|
conn.log.Warnf("unable to save peer's state, got error: %v", err)
|
|
}
|
|
|
|
if runtime.GOOS == "ios" {
|
|
runtime.GC()
|
|
}
|
|
|
|
if conn.onConnected != nil {
|
|
conn.onConnected(conn.config.Key, remoteRosenpassPubKey, conn.allowedIPsIP, remoteRosenpassAddr)
|
|
}
|
|
}
|
|
|
|
func (conn *Conn) doHandshake() error {
|
|
if !conn.signaler.Ready() {
|
|
return ErrSignalIsNotReady
|
|
}
|
|
|
|
var (
|
|
ha HandshakeArgs
|
|
err error
|
|
)
|
|
ha.IceUFrag, ha.IcePwd = conn.workerICE.GetLocalUserCredentials()
|
|
addr, err := conn.workerRelay.RelayInstanceAddress()
|
|
if err == nil {
|
|
ha.RelayAddr = addr
|
|
}
|
|
|
|
conn.log.Tracef("do handshake with args: %v", ha)
|
|
return conn.handshaker.SendOffer(ha)
|
|
}
|
|
|
|
func (conn *Conn) waitInitialRandomSleepTime() {
|
|
minWait := 100
|
|
maxWait := 800
|
|
duration := time.Duration(rand.Intn(maxWait-minWait)+minWait) * time.Millisecond
|
|
|
|
timeout := time.NewTimer(duration)
|
|
defer timeout.Stop()
|
|
|
|
select {
|
|
case <-conn.ctx.Done():
|
|
case <-timeout.C:
|
|
}
|
|
}
|
|
|
|
func (conn *Conn) evalStatus() ConnStatus {
|
|
if conn.statusRelay == StatusConnected || conn.statusICE == StatusConnected {
|
|
return StatusConnected
|
|
}
|
|
|
|
if conn.statusRelay == StatusConnecting || conn.statusICE == StatusConnecting {
|
|
return StatusConnecting
|
|
}
|
|
|
|
return StatusDisconnected
|
|
}
|
|
|
|
func isRosenpassEnabled(remoteRosenpassPubKey []byte) bool {
|
|
return remoteRosenpassPubKey != nil
|
|
}
|