summaryrefslogtreecommitdiffstats
path: root/dom/media/webrtc/tests/mochitests/test_peerConnection_simulcastAnswer_oldSetParameters.html
blob: bc0b9f71cc1cafa70f27f12d35a8b64dfb15325b (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
<!DOCTYPE HTML>
<html>
<head>
  <script type="application/javascript" src="pc.js"></script>
  <script type="application/javascript" src="/tests/dom/canvas/test/captureStream_common.js"></script>
  <script type="application/javascript" src="helpers_from_wpt/sdp.js"></script>
  <script type="application/javascript" src="simulcast.js"></script>
  <script type="application/javascript" src="stats.js"></script>
</head>
<body>
<pre id="test">
<script type="application/javascript">
  createHTML({
    bug: "1231507",
    title: "Basic video-only peer connection with Simulcast answer",
    visible: true
  });

  runNetworkTest(async () => {
    await pushPrefs(
        ['media.peerconnection.simulcast', true],
        // 180Kbps was determined empirically, set well-higher than
        // the 80Kbps+overhead needed for the two simulcast streams.
        // 100Kbps was apparently too low.
        ['media.peerconnection.video.min_bitrate_estimate', 180*1000]);


    const offerer = new RTCPeerConnection();
    const answerer = new RTCPeerConnection();

    const add = (pc, can, failed) => can && pc.addIceCandidate(can).catch(failed);
    offerer.onicecandidate = e => add(answerer, e.candidate, generateErrorCallback());
    answerer.onicecandidate = e => add(offerer, e.candidate, generateErrorCallback());

    const metadataToBeLoaded = [];
    offerer.ontrack = (e) => {
      metadataToBeLoaded.push(getPlaybackWithLoadedMetadata(e.track));
    };

    // Two recv transceivers, one for each simulcast stream
    offerer.addTransceiver('video', { direction: 'recvonly' });
    offerer.addTransceiver('video', { direction: 'recvonly' });

    // One send transceiver, that will be used to send both simulcast streams
    const emitter = new VideoFrameEmitter();
    const videoStream = emitter.stream();
    answerer.addTrack(videoStream.getVideoTracks()[0], videoStream);
    emitter.start();

    const offer = await offerer.createOffer();

    const mungedOffer = midToRid(offer);
    info(`Transformed recv offer to simulcast: ${offer.sdp} to ${mungedOffer}`);

    await answerer.setRemoteDescription({type: "offer", sdp: mungedOffer});
    await offerer.setLocalDescription(offer);

    const rids = offerer.getTransceivers().map(t => t.mid);
    is(rids.length, 2, 'Should have 2 mids in offer');
    ok(rids[0] != '', 'First mid should be non-empty');
    ok(rids[1] != '', 'Second mid should be non-empty');
    info(`rids: ${JSON.stringify(rids)}`);

    const sender = answerer.getSenders()[0];
    await sender.setParameters({
        encodings: [
            { rid: rids[0], maxBitrate: 40000 },
            { rid: rids[1], maxBitrate: 40000, scaleResolutionDownBy: 2 }
          ]
    });

    const answer = await answerer.createAnswer();

    const mungedAnswer = ridToMid(answer);
    info(`Transformed send simulcast answer to multiple m-sections: ${answer.sdp} to ${mungedAnswer}`);
    await offerer.setRemoteDescription({type: "answer", sdp: mungedAnswer});
    await answerer.setLocalDescription(answer);

    is(metadataToBeLoaded.length, 2, 'Offerer should have gotten 2 ontrack events');
    info('Waiting for 2 loadedmetadata events');
    const videoElems = await Promise.all(metadataToBeLoaded);

    const statsReady =
      Promise.all([waitForSyncedRtcp(offerer), waitForSyncedRtcp(answerer)]);

    const helper = new VideoStreamHelper();
    info('Waiting for first video element to start playing');
    await helper.checkVideoPlaying(videoElems[0]);
    info('Waiting for second video element to start playing');
    await helper.checkVideoPlaying(videoElems[1]);

    is(videoElems[0].videoWidth, 50,
       "sink is same width as source, modulo our cropping algorithm");
    is(videoElems[0].videoHeight, 50,
       "sink is same height as source, modulo our cropping algorithm");
    is(videoElems[1].videoWidth, 25,
       "sink is 1/2 width of source, modulo our cropping algorithm");
    is(videoElems[1].videoHeight, 25,
       "sink is 1/2 height of source, modulo our cropping algorithm");

    await statsReady;
    const senderStats = await sender.getStats();
    checkSenderStats(senderStats, 2);
    checkExpectedFields(senderStats);
    pedanticChecks(senderStats);

    emitter.stop();
    videoStream.getVideoTracks()[0].stop();
    offerer.close();
    answerer.close();
  });
</script>
</pre>
</body>
</html>