You cannot select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
lal/pkg/rtsp/base_in_session.go

337 lines
9.5 KiB
Go

// Copyright 2020, Chef. All rights reserved.
// https://github.com/q191201771/lal
//
// Use of this source code is governed by a MIT-style license
// that can be found in the License file.
//
// Author: Chef (191201771@qq.com)
package rtsp
import (
"net"
"strings"
"sync"
"sync/atomic"
"github.com/q191201771/lal/pkg/base"
"github.com/q191201771/lal/pkg/rtprtcp"
"github.com/q191201771/lal/pkg/sdp"
"github.com/q191201771/naza/pkg/connection"
"github.com/q191201771/naza/pkg/nazalog"
"github.com/q191201771/naza/pkg/nazanet"
)
type BaseCommandSession interface {
Write(channel int, b []byte) error
Dispose() error
}
// 聚合PubSession和PullSession也即流数据是输入类型的session
// BaseInSession会向上层回调两种格式的数据
// 1. 原始的rtp packet
// 2. rtp合并后的av packet
type BaseInSessionObserver interface {
OnRTPPacket(pkt rtprtcp.RTPPacket)
// @param asc: AAC AudioSpecificConfig注意如果不存在音频则为nil
// @param vps: 视频为H264时为nil视频为H265时不为nil
OnAVConfig(asc, vps, sps, pps []byte)
// @param pkt: pkt结构体中字段含义见rtprtcp.OnAVPacket
OnAVPacket(pkt base.AVPacket)
}
type BaseInSession struct {
UniqueKey string // 使用上层Session的值
stat base.StatPub
cmdSession BaseCommandSession
observer BaseInSessionObserver
audioRTPConn *nazanet.UDPConnection
videoRTPConn *nazanet.UDPConnection
audioRTCPConn *nazanet.UDPConnection
videoRTCPConn *nazanet.UDPConnection
audioRTPChannel int
audioRTCPChannel int
videoRTPChannel int
videoRTCPChannel int
currConnStat connection.Stat
prevConnStat connection.Stat
staleStat *connection.Stat
m sync.Mutex
rawSDP []byte // const after set
sdpLogicCtx sdp.LogicContext // const after set
avPacketQueue *AVPacketQueue
audioUnpacker *rtprtcp.RTPUnpacker
videoUnpacker *rtprtcp.RTPUnpacker
audioRRProducer *rtprtcp.RRProducer
videoRRProducer *rtprtcp.RRProducer
audioSsrc uint32
videoSsrc uint32
}
func (s *BaseInSession) InitWithSDP(rawSDP []byte, sdpLogicCtx sdp.LogicContext) {
s.m.Lock()
s.rawSDP = rawSDP
s.sdpLogicCtx = sdpLogicCtx
s.m.Unlock()
s.audioUnpacker = rtprtcp.NewRTPUnpacker(s.sdpLogicCtx.AudioPayloadType, s.sdpLogicCtx.AudioClockRate, unpackerItemMaxSize, s.onAVPacketUnpacked)
s.videoUnpacker = rtprtcp.NewRTPUnpacker(s.sdpLogicCtx.VideoPayloadType, s.sdpLogicCtx.VideoClockRate, unpackerItemMaxSize, s.onAVPacketUnpacked)
s.audioRRProducer = rtprtcp.NewRRProducer(s.sdpLogicCtx.AudioClockRate)
s.videoRRProducer = rtprtcp.NewRRProducer(s.sdpLogicCtx.VideoClockRate)
if s.sdpLogicCtx.AudioPayloadType != 0 && s.sdpLogicCtx.VideoPayloadType != 0 {
s.avPacketQueue = NewAVPacketQueue(s.onAVPacket)
}
if s.observer != nil {
s.observer.OnAVConfig(s.sdpLogicCtx.ASC, s.sdpLogicCtx.VPS, s.sdpLogicCtx.SPS, s.sdpLogicCtx.PPS)
}
}
func (s *BaseInSession) SetObserver(observer BaseInSessionObserver) {
s.observer = observer
if s.sdpLogicCtx.ASC != nil && s.sdpLogicCtx.SPS != nil {
s.observer.OnAVConfig(s.sdpLogicCtx.ASC, s.sdpLogicCtx.VPS, s.sdpLogicCtx.SPS, s.sdpLogicCtx.PPS)
}
}
func (s *BaseInSession) SetupWithConn(uri string, rtpConn, rtcpConn *nazanet.UDPConnection) error {
if strings.HasSuffix(uri, s.sdpLogicCtx.AudioAControl) {
s.audioRTPConn = rtpConn
s.audioRTCPConn = rtcpConn
} else if strings.HasSuffix(uri, s.sdpLogicCtx.VideoAControl) {
s.videoRTPConn = rtpConn
s.videoRTCPConn = rtcpConn
} else {
return ErrRTSP
}
go rtpConn.RunLoop(s.onReadRTPPacket)
go rtcpConn.RunLoop(s.onReadRTCPPacket)
return nil
}
func (s *BaseInSession) SetupWithChannel(uri string, rtpChannel, rtcpChannel int, remoteAddr string) error {
s.stat.RemoteAddr = remoteAddr
if strings.HasSuffix(uri, s.sdpLogicCtx.AudioAControl) {
s.audioRTPChannel = rtpChannel
s.audioRTCPChannel = rtcpChannel
return nil
}
if strings.HasSuffix(uri, s.sdpLogicCtx.VideoAControl) {
s.videoRTPChannel = rtpChannel
s.videoRTCPChannel = rtcpChannel
return nil
}
return ErrRTSP
}
func (s *BaseInSession) Dispose() {
nazalog.Infof("[%s] lifecycle dispose rtsp BaseInSession.", s.UniqueKey)
if s.audioRTPConn != nil {
_ = s.audioRTPConn.Dispose()
}
if s.audioRTCPConn != nil {
_ = s.audioRTCPConn.Dispose()
}
if s.videoRTPConn != nil {
_ = s.videoRTPConn.Dispose()
}
if s.videoRTCPConn != nil {
_ = s.videoRTCPConn.Dispose()
}
_ = s.cmdSession.Dispose()
}
func (s *BaseInSession) GetSDP() ([]byte, sdp.LogicContext) {
s.m.Lock()
defer s.m.Unlock()
return s.rawSDP, s.sdpLogicCtx
}
func (s *BaseInSession) HandleInterleavedPacket(b []byte, channel int) {
switch channel {
case s.audioRTPChannel:
fallthrough
case s.videoRTPChannel:
_ = s.handleRTPPacket(b)
case s.audioRTCPChannel:
fallthrough
case s.videoRTCPChannel:
// TODO chef: 这个地方有bug处理RTCP包则推流会失败有可能是我的RTCP RR包打的有问题
//_ = p.handleRTCPPacket(b, nil)
default:
nazalog.Errorf("[%s] read interleaved packet but channel invalid. channel=%d", s.UniqueKey, channel)
}
}
func (s *BaseInSession) GetStat() base.StatPub {
s.stat.ReadBytesSum = atomic.LoadUint64(&s.currConnStat.ReadBytesSum)
s.stat.WroteBytesSum = atomic.LoadUint64(&s.currConnStat.WroteBytesSum)
return s.stat
}
func (s *BaseInSession) UpdateStat(interval uint32) {
readBytesSum := atomic.LoadUint64(&s.currConnStat.ReadBytesSum)
wroteBytesSum := atomic.LoadUint64(&s.currConnStat.WroteBytesSum)
diff := readBytesSum - s.prevConnStat.ReadBytesSum
s.stat.Bitrate = int(diff * 8 / 1024 / uint64(interval))
s.prevConnStat.ReadBytesSum = readBytesSum
s.prevConnStat.WroteBytesSum = wroteBytesSum
}
func (s *BaseInSession) IsAlive(interval uint32) (ret bool) {
readBytesSum := atomic.LoadUint64(&s.currConnStat.ReadBytesSum)
wroteBytesSum := atomic.LoadUint64(&s.currConnStat.WroteBytesSum)
if s.staleStat == nil {
s.staleStat = new(connection.Stat)
s.staleStat.ReadBytesSum = readBytesSum
s.staleStat.WroteBytesSum = wroteBytesSum
return true
}
ret = !(readBytesSum-s.staleStat.ReadBytesSum == 0)
s.staleStat.ReadBytesSum = readBytesSum
s.staleStat.WroteBytesSum = wroteBytesSum
return ret
}
// callback by RTPUnpacker
func (s *BaseInSession) onAVPacketUnpacked(pkt base.AVPacket) {
if s.avPacketQueue != nil {
s.avPacketQueue.Feed(pkt)
} else {
s.observer.OnAVPacket(pkt)
}
}
// callback by avpacket queue
func (s *BaseInSession) onAVPacket(pkt base.AVPacket) {
s.observer.OnAVPacket(pkt)
}
// callback by UDPConnection
func (s *BaseInSession) onReadRTPPacket(b []byte, rAddr *net.UDPAddr, err error) bool {
if err != nil {
nazalog.Errorf("[%s] read udp packet failed. err=%+v", s.UniqueKey, err)
return true
}
_ = s.handleRTPPacket(b)
if s.stat.RemoteAddr == "" {
s.stat.RemoteAddr = rAddr.String()
}
return true
}
// callback by UDPConnection
func (s *BaseInSession) onReadRTCPPacket(b []byte, rAddr *net.UDPAddr, err error) bool {
if err != nil {
nazalog.Errorf("[%s] read udp packet failed. err=%+v", s.UniqueKey, err)
return true
}
_ = s.handleRTCPPacket(b, rAddr)
return true
}
// @param rAddr 对端地址往对端发送数据时使用注意如果nil则表示是interleaved模式我们直接往TCP连接发数据
func (s *BaseInSession) handleRTCPPacket(b []byte, rAddr *net.UDPAddr) error {
switch b[1] {
case rtprtcp.RTCPPacketTypeSR:
sr := rtprtcp.ParseSR(b)
//nazalog.Debugf("%+v", sr)
var rrBuf []byte
switch sr.SenderSSRC {
case s.audioSsrc:
rrBuf = s.audioRRProducer.Produce(sr.GetMiddleNTP())
if rrBuf != nil {
if rAddr != nil {
_ = s.audioRTCPConn.Write2Addr(rrBuf, rAddr)
} else {
_ = s.cmdSession.Write(s.audioRTCPChannel, rrBuf)
}
atomic.AddUint64(&s.currConnStat.WroteBytesSum, uint64(len(b)))
}
case s.videoSsrc:
rrBuf = s.videoRRProducer.Produce(sr.GetMiddleNTP())
if rrBuf != nil {
if rAddr != nil {
_ = s.videoRTCPConn.Write2Addr(rrBuf, rAddr)
} else {
_ = s.cmdSession.Write(s.videoRTCPChannel, rrBuf)
}
atomic.AddUint64(&s.currConnStat.WroteBytesSum, uint64(len(b)))
}
default:
// ffmpeg推流时会在发送第一个RTP包之前就发送一个SR所以关闭这个警告日志
//nazalog.Warnf("[%s] read rtcp sr but senderSSRC invalid. senderSSRC=%d, audio=%d, video=%d",
// p.UniqueKey, sr.SenderSSRC, p.audioSsrc, p.videoSsrc)
return ErrRTSP
}
default:
nazalog.Errorf("[%s] read rtcp packet but type invalid. type=%d", s.UniqueKey, b[1])
return ErrRTSP
}
return nil
}
func (s *BaseInSession) handleRTPPacket(b []byte) error {
if len(b) < rtprtcp.RTPFixedHeaderLength {
nazalog.Errorf("[%s] read udp packet length invalid. len=%d", s.UniqueKey, len(b))
return ErrRTSP
}
packetType := b[1] & 0x7F
if packetType != base.RTPPacketTypeAVCOrHEVC && packetType != base.RTPPacketTypeAAC {
return ErrRTSP
}
h, err := rtprtcp.ParseRTPPacket(b)
if err != nil {
nazalog.Errorf("[%s] read invalid rtp packet. err=%+v", s.UniqueKey, err)
return err
}
atomic.AddUint64(&s.currConnStat.ReadBytesSum, uint64(len(b)))
//nazalog.Debugf("%+v", h)
var pkt rtprtcp.RTPPacket
pkt.Header = h
pkt.Raw = b
switch packetType {
case base.RTPPacketTypeAVCOrHEVC:
s.videoSsrc = h.Ssrc
s.observer.OnRTPPacket(pkt)
s.videoUnpacker.Feed(pkt)
s.videoRRProducer.FeedRTPPacket(h.Seq)
case base.RTPPacketTypeAAC:
s.audioSsrc = h.Ssrc
s.observer.OnRTPPacket(pkt)
s.audioUnpacker.Feed(pkt)
s.audioRRProducer.FeedRTPPacket(h.Seq)
default:
// 因为前面已经判断过type了所以永远不会走到这
}
return nil
}