Release notes
This page provides the release notes for Video Calling.
Video SDK
v6.2.2
v6.2.2 was released on August 1, 2023.
New features
-
Wildcard token
This release introduces wildcard tokens. Agora supports setting the channel name used for generating a token as a wildcard character. The token generated can be used to join any channel if you use the same user id. In scenarios involving multiple channels, such as switching between different channels, using a wildcard token can avoid repeated application of tokens every time users joining a new channel, which reduces the pressure on your token server. See Secure authentication with tokens.
All 4.x SDKs support using wildcard tokens. -
Preloading channels
This release adds
preloadChannel
andpreloadChannelWithUserAccount
methods, which allows a user whose role is set as audience to preload channels before joining one. Calling the method can help shortening the time of joining a channel, thus reducing the time it takes for audience members to hear and see the host.When preloading more than one channels, Agora recommends that you use a wildcard token for preloading to avoid repeated application of tokens every time you joining a new channel, thus saving the time for switching between channels. See Secure authentication with tokens.
-
Customized background color of video canvas
In this release, the
backgroundColor
member has been added toVideoCanvas
, which allows you to customize the background color of the video canvas when setting the properties of local or remote video display. -
Publishing video streams from different sources (Windows, macOS)
This release adds the following members in
ChannelMediaOptions
to allow you publish video streams captured from the third and fourth camera or screen:publishThirdCameraTrack
: Publishing the video stream captured from the third camera.publishFourthCameraTrack
: Publishing the video stream captured from the fourth camera.publishThirdScreenTrack
: Publishing the video stream captured from the third screen.publishFourthScreenTrack
: Publishing the video stream captured from the fourth screen.
For oneRtcConnection
, Agora supports publishing multiple audio streams and one video stream at the same time.
Improvements
-
Improved camera capture effect (Android, iOS)
This release has improved camera capture effect in the following aspects:
-
Support for camera exposure adjustment
This release adds
isCameraExposureSupported
to query whether the device supports exposure adjustment andsetCameraExposureFactor
to set the exposure ratio of the camera. -
Optimization of default camera selection (iOS)
Since this release, the default camera selection behavior of the SDK is aligned with that of the iOS system camera. If the device has multiple rear cameras, better shooting perspectives, zooming capabilities, low-light performance, and depth sensing can be achieved during video capture, thereby improving the quality of video capture.
-
-
Virtual Background Algorithm Upgrade
This version has upgraded the portrait segmentation algorithm of the virtual background, which comprehensively improves the accuracy of portrait segmentation, the smoothness of the portrait edge with the virtual background, and the fit of the edge when the person moves. In addition, it optimizes the precision of the person's edge in scenarios such as meetings, offices, homes, and under backlight or weak light conditions.
-
Channel media relay
The number of target channels for media relay has been increased to 6. When calling
startOrUpdateChannelMediaRelay
andstartOrUpdateChannelMediaRelayEx
, you can specify up to 6 target channels. -
Enhancement in video codec query capability
To improve the video codec query capability, this release adds the
codecLevels
member inCodecCapInfo
. After successfully callingqueryCodecCapability
, you can obtain the hardware and software decoding capability levels of the device for H.264 and H.265 video formats throughcodecLevels
.
This release includes the following additional improvements:
- To improve the switching experience between multiple audio routes, this release adds the
setRouteInCommunicationMode
method. This method can switch the audio route from a Bluetooth headphone to the earpiece, wired headphone or speaker in communication volume mode (MODE_IN_COMMUNICATION
). (Android) - The SDK automacially adjusts the frame rate of the sending end based on the screen sharing scenario. Especially in document sharing scenarios, this feature avoids exceeding the expected video bitrate on the sending end to improve transmission efficiency and reduce network burden.
- To help users understand the reasons for more types of remote video state changes, the
remoteVideoStateReasonCodecNotSupport
enumeration has been added to theonRemoteVideoStateChanged
callback, indicating that the local video decoder does not support decoding the received remote video stream.
Issues fixed
This release fixed the following issues:
- Occasionally, noise occurred when the local user listened to their own and remote audio after joining the channel. (macOS)
- Slow channel reconnection after the connection was interrupted due to network reasons.
- In screen sharing scenarios, the delay of seeing the shared screen was occasionally higher than expected on some devices.
- In custom video capturing scenarios,
setBeautyEffectOptions
,setLowlightEnhanceOptions
,setVideoDenoiserOptions
, andsetColorEnhanceOptions
could not load extensions automatically. - In multi-device audio recording scenarios, after repeatedly plugging and unplugging or enabling/disabling the audio recording device, no sound could be heard occasionally when calling the
startRecordingDeviceTest
to start an audio capturing device test. (Windows)
API changes
Added
setCameraExposureFactor
(Android, iOS)isCameraExposureSupported
(Android, iOS)preloadChannel
preloadChannelWithUserAccount
updatePreloadChannelToken
setRouteInCommunicationMode
(Android)- The following members in
ChannelMediaOptions
:publishThirdCameraTrack
(Windows, macOS)publishFourthCameraTrack
(Windows, macOS)publishThirdScreenTrack
(Windows, macOS)publishFourthScreenTrack
(Windows, macOS)
CodecCapLevels
VideoCodecCapabilityLevel
backgroundColor
inVideoCanvas
codecLevels
inCodecCapInfo
remoteVideoStateReasonCodecNotSupport
inRemoteVideoStateReason
v6.2.1
This version was released on June 21, 2023.
Improvements
This version improves the network transmission strategy, enhancing the smoothness of audio and video interactions.
Issues fixed
This version fixed the following issues:
- Inability to join channels caused by SDK's incompatibility with some older versions of AccessToken.
- After the sending end called
setAINSMode
to activate AI noise reduction, occasional echo was observed by the receiving end. - Brief noise occurred while playing media files using the media player.
- After enabling the screen sharing feature on the sending end, there was high delay until the receiving end could see the shared screen. (macOS)
- Occasional crash after calling the
destroyMediaPlayer
method. (iOS) - When the sending end mixed and published two streams of video captured by two cameras locally, the video from the second camera was occasionally missing on the receiving end. (Windows)
- In screen sharing scenarios, some Android devices experienced choppy video on the receiving end. (Android)
v6.2.0
v6.2.0 was released on May 29, 2023.
Compatibility changes
If you use the features mentioned in this section, ensure that you modify the implementation of the relevant features after upgrading the SDK.
1. Video capture (Windows, iOS)
This release optimizes the APIs for camera and screen capture function. As of v6.2.0, ensure you use the alternative methods listed in the table below and specify the video source by setting the sourceType
parameter:
Deleted methods | Alternative methods | |
---|---|
startPrimaryCameraCapture (Windows)startSecondaryCameraCapture (Windows, iOS) | startCameraCapture |
stopPrimaryCameraCapture (Windows)stopSecondaryCameraCapture (Windows, iOS) | stopCameraCapture |
startPrimaryScreenCapture (Windows)startSecondaryScreenCapture (Windows) | startScreenCaptureBySourceType (Windows) |
stopPrimaryScreenCapture (Windows)stopSecondaryScreenCapture (Windows) | stopScreenCaptureBySourceType (Windows) |
2. Video data acquisition
-
The
onCaptureVideoFrame
andonPreEncodeVideoFrame
callbacks are added with a new parameter calledsourceType
, which is used to indicate the specific video source type. -
The following callbacks are deleted. Get the video source type through the
sourceType
parameter in theonPreEncodeVideoFrame
andonCaptureVideoFrame
callbacks (Windows):onSecondaryPreEncodeCameraVideoFrame
onScreenCaptureVideoFrame
onPreEncodeScreenVideoFrame
onSecondaryPreEncodeScreenVideoFrame
3. Channel media options
publishCustomAudioTrackEnableAec
inChannelMediaOptions
is deleted. UsepublishCustomAudioTrack
instead.publishTrancodedVideoTrack
inChannelMediaOptions
is renamed topublishTranscodedVideoTrack
.publishCustomAudioSourceId
inChannelMediaOptions
is renamed topublishCustomAudioTrackId
.
4. Local video mixing (Windows)
- The
VideoInputStreams
inLocalTranscoderConfiguration
is changed tovideoInputStreams
. - The
MediaSourceType
inTranscodingVideoStream
is changed toVideoSourceType
.
5. Video encoder algorithm
As of this release, the SDK optimizes the video encoder algorithm and upgrades the default video encoding resolution from 640 × 360 to 960 × 540 to accommodate improvements in device performance and network bandwidth, providing users with a full-link HD experience in various audio and video interaction scenarios.
Call the setVideoEncoderConfiguration
method to set the expected video encoding resolution in the video encoding parameters configuration.
The increase in the default resolution affects the aggregate resolution and thus the billing rate. See Pricing.
6. Miscellaneous
onApiCallExecuted
is deleted. Agora recommends getting the results of the API implementation through relevant channels and media callbacks.- The
IAudioFrameObserver
class is renamed toIAudioPcmFrameSink
, thus the prototypes of the following methods are updated accordingly:onFrame
registerAudioFrameObserver
andunregisterAudioFrameObserver
inMediaPlayer
startChannelMediaRelay
,updateChannelMediaRelay
,startChannelMediaRelayEx
, andupdateChannelMediaRelayEx
are deprecated. UsestartOrUpdateChannelMediaRelay
andstartOrUpdateChannelMediaRelayEx
instead.
New features
1. AI Noise Suppression
This release introduces public APIs for the AI Noise Suppression function. Once enabled, the SDK automatically detects and reduces background noises. Whether in bustling public venues or real-time competitive arenas that demand lightning-fast responsiveness, this function guarantees optimal audio clarity, providing users with an elevated audio experience. You can enable this function through the newly-introduced setAINSMode
method and set the noise suppression mode to balance, aggressive, or low latency according to your scenarios.
2. Enhanced Virtual Background
To increase the fun of real-time video calls and protect user privacy, this version has enhanced the Virtual Background feature. You can now set custom backgrounds of various types by calling the enableVirtualBackground
method, including:
- Process the background as alpha information without replacement, only separating the portrait and the background. This can be combined with the local video mixing feature to achieve a portrait-in-picture effect.
- Replace the background with various formats of local videos.
See Virtual Background documentation.
3. Video scenario settings
This release introduces setVideoScenario
for setting the video application scene. The SDK will automatically enable the best practice strategy based on different scenes, adjusting key performance indicators to optimize video quality and improve user experience. Whether it is a formal business meeting or a casual online gathering, this feature ensures that the video quality meets the requirements.
Currently, this feature provides targeted optimizations for real-time video conferencing scenarios, including:
- Automatically activate multiple anti-weak-network technologies to enhance the capability and performance of low-quality video streams in meeting scenarios where high bitrate is required, ensuring smoothness when multiple streams are subscribed to by the receiving end.
- Monitor the number of subscribers for the high-quality and low-quality video streams in real time, dynamically adjusting the configuration of the high-quality stream and dynamically enabling or disabling the low-quality stream, to save uplink bandwidth and consumption.
4. Local video mixing (macOS, iOS, Android)
This release adds the local video mixing feature. You can use the startLocalVideoTranscoder
method to mix and render multiple video streams locally, such as camera-captured video, screen sharing streams, video files, images, etc. This allows you to achieve custom layouts and effects, making it easy to create personalized video display effects to meet various scenario requirements, such as remote meetings, live streaming, online education, while also supporting features like portrait-in-picture effect.
Additionally, the SDK provides the updateLocalTranscoderConfiguration
method and the onLocalVideoTranscoderError
callback. After enabling local video mixing, you can use the updateLocalTranscoderConfiguration
method to update the video mixing configuration. Where an error occurs in starting the local video mixing or updating the configuration, you can get the reason for the failure through the onLocalVideoTranscoderError
callback.
5. Local video mixing (Windows)
This release adds the onLocalVideoTranscoderError
callback. When there is an error in starting or updating the local video mixing, the SDK triggers this callback to report the reason for the failure.
6. Cross-device synchronization
In real-time collaborative singing scenarios, network issues can cause inconsistencies in the downlinks of different client devices. To address this, this release introduces getNtpWallTimeInMs
for obtaining the current Network Time Protocol (NTP) time. By using this method to synchronize lyrics and music across multiple client devices, users can achieve synchronized singing and lyrics progression, resulting in a better collaborative experience.
7. Instant frame rendering
This release adds the enableInstantMediaRendering
method to enable instant rendering mode for audio and video frames, which can speed up the first video or audio frame rendering after the user joins the channel.
8. Video rendering tracing
This release adds the startMediaRenderingTracing
and startMediaRenderingTracingEx
methods. The SDK starts tracing the rendering status of the video frames in the channel from the moment this method is called and reports information about the event through the onVideoRenderingTracingResult
callback.
Agora recommends that you use this method in conjunction with the UI settings, such as buttons and sliders, in your app. For example, call this method when the user clicks Join Channel button and then get the indicators in the video frame rendering process through the onVideoRenderingTracingResult
callback. This enables developers to optimize the indicators and improve the user experience.
Improvements
1. Voice changer
This release introduces the setLocalVoiceFormant
method that allows you to adjust the formant ratio to change the timbre of the voice. This method can be used together with the setLocalVoicePitch
method to adjust the pitch and timbre of voice at the same time, enabling a wider range of voice transformation effects.
2. Enhanced screen share (Android, iOS)
This release adds the queryScreenCaptureCapability
method, which is used to query the screen capture capabilities of the current device. To ensure optimal screen sharing performance, particularly in enabling high frame rates like 60 fps, Agora recommends you to query the device's maximum supported frame rate using this method beforehand.
This release also adds the setScreenCaptureScenario
method, which is used to set the scenario type for screen sharing. The SDK automatically adjusts the smoothness and clarity of the shared screen based on the scenario type you set.
3. Improved compatibility with audio file types (Android)
As of this release, you can use the following methods to open files with a URI starting with content://
:
startAudioMixing
playEffect
openWithMediaSource
4. Enhanced rendering compatibility (Windows)
This release enhances the rendering compatibility of the SDK. Issues like black screens caused by rendering failures on certain devices are fixed.
5. Audio and video synchronization
For custom video and audio capture scenarios, this release introduces getCurrentMonotonicTimeInMs
for obtaining the current Monotonic Time. By passing this value into the timestamps of audio and video frames, developers can accurately control the timing of their audio and video streams, ensuring proper synchronization.
6. Multi-camera capture and multi-screen capture
This release introduces startScreenCaptureBySourceType
(PC only) and startCameraCapture
. By calling these methods multiple times and specifying the sourceType
parameter, developers can start capturing video streams from multiple cameras and screens for local video mixing or multi-channel publishing. This is particularly useful for scenarios such as remote medical care and online education, where multiple cameras and displays need to be connected.
7. Channel media relay
This release introduces startOrUpdateChannelMediaRelay
and startOrUpdateChannelMediaRelayEx
, allowing for a simpler and smoother way to start and update media relay across channels. With these methods, developers can easily start the media relay across channels and update the target channels for media relay with a single method. Additionally, the internal interaction frequency has been optimized, effectively reducing latency in function calls.
8. Custom audio tracks
To better meet the needs of custom audio capture scenarios, this release adds createCustomAudioTrack
and destroyCustomAudioTrack
for creating and destroying custom audio tracks. Two types of audio tracks are also provided for users to choose from, further improving the flexibility of capturing external audio source:
- Mixable audio track: Supports mixing multiple external audio sources and publishing them to the same channel, suitable for multi-channel audio capture scenarios.
- Direct audio track: Only supports publishing one external audio source to a single channel, suitable for low-latency audio capture scenarios.
9. Video frame observer
As of this release, the SDK optimizes the onRenderVideoFrame
callback, and the meaning of the return value is different depending on the video processing mode:
- When the video processing mode is
processModeReadOnly
, the return value is reserved for future use. - When the video processing mode is
processModeReadWrite
, the SDK receives the video frame when the return value istrue
. The video frame is discarded when the return value isfalse
.
Issues fixed
This release fixed the following issues:
- Occasional crashes occurred on Android devices when users joined or left a channel (Android).
- When the host frequently switched the user role between broadcaster and audience in a short period of time, the audience members could not hear the audio of the host.
- Occasional failure when enabling in-ear monitoring (Android).
- Occasional loss of the
onFirstRemoteVideoFrame
callback during channel media relay (iOS). - The receiver actively subscribed to the high-quality stream but unexpectedly received a low-quality stream (iOS).
- The receiver was receiving the low-quality stream originally and automatically switched to the high-quality stream after a few seconds (macOS).
- Occasional echo (Android).
- Occasional screen jitter during screen sharing (macOS).
- Abnormal client status caused by an exception in the
onRemoteAudioStateChanged
callback (Android, iOS). - When using Agora Media Player to play RTSP video streams, the video images sometimes appeared pixelated (Windows).
- Playing audio files with a sample rate of 48 kHz failed.
- Crashes occurred after users set the video resolution as 3840 × 2160 and started CDN streaming on Xiaomi Redmi 9A devices (Android).
- If the rendering view of the player was set as a
UIViewController
's view, the video was zoomed from the bottom-left corner to the middle of the screen when entering full-screen mode (macOS). - Adding an alpha channel to an image in the PNG or GIF format failed when the local client mixed video streams (Windows).
- In real-time chorus scenarios, remote users heard noises and echoes when an OPPO R11 device joined the channel in the loudspeaker mode (Android).
- When the playback of the local music finished, the
onAudioMixingFinished
callback was not properly triggered (Android). - After joining the channel, remotes users saw a watermark even though the watermark was deleted (Windows).
- If a watermark was added after starting screen sharing, the watermark was not displayed on the screen (Windows).
- When joining a channel and accessing an external camera, calling
setDevice
to specify the video capture device as the external camera did not take effect (macOS, Windows). - When using a video frame observer, the first video frame was occasionally missed on the receiver's end (Android).
- When sharing screens in scenarios involving multiple channels, remote users occasionally saw black screens (Android).
- When trying to outline the shared window and put it on top, the shared window did not stay on top of other windows (Windows).
- Switching to the rear camera with the virtual background enabled occasionally caused the background to be inverted (Android).
- When there were multiple video streams in a channel, calling some video enhancement APIs occasionally failed.
- At the moment when a user left a channel, a request for leaving was not sent to the server and the leaving behavior was incorrectly determined by the server as timed out.
API changes
Added
startCameraCapture
stopCameraCapture
startScreenCaptureBySourceType
(Windows, macOS)stopScreenCaptureBySourceType
(Windows, macOS)startOrUpdateChannelMediaRelay
startOrUpdateChannelMediaRelayEx
getNtpWallTimeInMs
setVideoScenario
getCurrentMonotonicTimeInMs
onLocalVideoTranscoderError
startLocalVideoTranscoder
(macOS, iOS, Android)updateLocalTranscoderConfiguration
(macOS, iOS, Android)queryScreenCaptureCapability
(iOS, Android)setScreenCaptureScenario
(iOS, Android)setAINSMode
createAudioCustomTrack
destroyAudioCustomTrack
AudioTrackConfig
AudioAinsMode
AudioTrackType
VideoApplicationScenarioType
ScreenCaptureFramerateCapability
- The
domainLimit
andautoRegisterAgoraExtensions
members inRtcEngineContext
- The
sourceType
parameter inonCaptureVideoFrame
andonPreEncodeVideoFrame
callbacks - The
backgroundNone
andbackgroundVideo
enumerators inbackgroundSourceType
Deprecated
startChannelMediaRelay
startChannelMediaRelayEx
updateChannelMediaRelay
updateChannelMediaRelayEx
onChannelMediaRelayEvent
ChannelMediaRelayEvent
enableInstantMediaRendering
startMediaRenderingTracing
startMediaRenderingTracingEx
onVideoRenderingTracingResult
MediaTraceEvent
VideoRenderingTracingInfo
Deleted
startPrimaryScreenCapture
(Windows)startSecondaryScreenCapture
(Windows)stopPrimaryScreenCapture
(Windows)stopSecondaryScreenCapture
(Window)startPrimaryCameraCapture
(Windows)startSecondaryCameraCapture
(Windows, iOS)stopPrimaryCameraCapture
(Windows)stopSecondaryCameraCapture
(Windows, iOS)onSecondaryPreEncodeCameraVideoFrame
(Windows)onScreenCaptureVideoFrame
(Windows)onPreEncodeScreenVideoFrame
(Windows)onSecondaryPreEncodeScreenVideoFrame
(Windows)onApiCallExecuted
publishCustomAudioTrackEnableAec
inChannelMediaOptions
inChannelMediaOptions
enableRemoteSuperResolution
superResolutionType
inRemoteVideoStats
v6.1.0
v6.1.0 was released on December 20, 2022.
Compatibility changes
This release deletes the sourceType
parameter in enableDualStreamMode
, because the SDK supports enabling dual-stream mode for various video sources captured by custom capture or SDK; you no longer need to specify the video source type.
New features
1. In-ear monitoring
This release adds support for in-ear monitoring. You can call enableInEarMonitoring
to enable the in-ear monitoring function.
After successfully enabling the in-ear monitoring function, you can call registerAudioFrameObserver
to register the audio observer, and the SDK triggers the onEarMonitoringAudioFrame
callback to report the audio frame data. You can use your own audio effect processing module to preprocess the audio frame data of the in-ear monitoring to implement custom audio effects. Agora recommends that you call the setEarMonitoringAudioFrameParameters
method to set the audio data format of in-ear monitoring. The SDK calculates the sampling interval based on the parameters in this method and triggers the onEarMonitoringAudioFrame
callback based on the sampling interval.
To adjust the in-ear monitoring volume, you can call setInEarMonitoringVolume
.
2. Audio capture device test (Android)
This release adds support for testing local audio capture devices before joining a channel. You can call startRecordingDeviceTest
to start the audio capture device test. After the test is complete, call the stopPlaybackDeviceTest
method to stop the audio capture device test.
3. Local network connection types
To make it easier for users to know the connection type of the local network at any stage, this release adds the getNetworkType
method. You can use this method to get the type of network connection in use. The available values are UNKNOWN
, DISCONNECTED
, LAN
, WIFI
, 2G
, 3G
, 4G
, and 5G
. When the local network connection type changes, the SDK triggers the onNetworkTypeChanged
callback to report the current network connection type.
4. Audio stream filter
This release introduces filtering audio streams based on volume. Once this function is enabled, the Agora server ranks all audio streams by volume and transports the three audio streams with the highest volumes to the receivers by default. The number of audio streams to be transported can be adjusted; contact support@agora.io to adjust this number according to your scenarios.
Agora also supports publishers in choosing whether the audio streams being published are to be filtered based on volume. Streams that are not filtered bypass this filter mechanism and are transported directly to the receivers. In scenarios with a number of publishers, enabling this function helps reduce the bandwidth and device system pressure for the receivers.
5. Dual-stream mode
This release optimizes the dual-stream mode. You can call enableDualStreamMode
and enableDualStreamModeEx
before and after joining a channel.
The implementation of subscribing to a low-quality video stream is expanded. The SDK enables the low-quality video stream auto mode on the sender by default (the SDK does not send low-quality video streams). Follow these steps to enable sending low-quality video streams:
- The host at the receiving end calls
setRemoteVideoStreamType
orsetRemoteDefaultVideoStreamType
to initiate a low-quality video stream request. - After receiving the application, the sender automatically switches to sending low-quality video stream mode.
If you want to modify this default behavior, you can call setDualStreamMode
and set the mode
parameter to disableSimulcastStream
(never send low-quality video streams) or enableSimulcastStream
(always send low-quality video streams).
6. Loopback device (Windows)
The SDK uses the playback device as the loopback device by default. As of 6.1.0, you can specify a loopback device separately and publish the captured audio to the remote end.
setLoopbackDevice
: Specifies the loopback device. If you do not want the current playback device to be the loopback device, you can call this method to specify another device as the loopback device.getLoopbackDevice
: Gets the current loopback device.followSystemLoopbackDevice
: Whether the loopback device follows the default playback device of the system.
7. Spatial audio effect
This release adds the following features applicable to spatial audio effect scenarios, which can effectively enhance the user's sense-of-presence experience in virtual interactive scenarios.
- Sound insulation area: You can set a sound insulation area and sound attenuation parameter by calling
setZones
. When the sound source (which can be a user or the media player) and the listener belong to the inside and outside of the sound insulation area, the listener experiences an attenuation effect similar to that of the sound in the real environment when it encounters a building partition. You can also set the sound attenuation parameter for the media player and the user by callingsetPlayerAttenuation
andsetRemoteAudioAttenuation
respectively, and specify whether to use that setting to force an override of the sound attenuation parameter insetZones
. - Doppler sound: You can enable Doppler sound by setting the
enableDoppler
parameter inSpatialAudioParams
. The receiver experiences noticeable tonal changes in the event of a high-speed relative displacement between the source and receiver (such as in a racing game scenario). - Headphone equalizer: You can use a preset headphone equalization effect by calling the
setHeadphoneEQPreset
method to improve the audio experience for users with headphones.
8. Multiple cameras for video capture (iOS)
This release supports multi-camera video capture. You can call enableMultiCamera
to enable multi-camera capture mode, call startSecondaryCameraCapture
to start capturing video from the second camera, and then publish the captured video to the second channel.
To stop using multi-camera capture, you need to call stopSecondaryCameraCapture
to stop the second camera capture, then call enableMultiCamera
and set enabled
to false
.
9. Headphone equalization effect
This release adds the setHeadphoneEQParameters
method, which is used to adjust the low- and high-frequency parameters of the headphone EQ. This is mainly useful in spatial audio scenarios. If you cannot achieve the expected headphone EQ effect after calling setHeadphoneEQPreset
, you can call setHeadphoneEQParameters
to adjust the EQ.
10. Encoded video frame observer
This release adds the setRemoteVideoSubscriptionOptions
and setRemoteVideoSubscriptionOptionsEx
methods. When you call the registerVideoEncodedFrameObserver
method to register a video frame observer for the encoded video frames, the SDK subscribes to the encoded video frames by default. If you want to change the subscription options, you can call these new methods to set them.
For more information about registering video observers and subscription options, see the API reference.
11. MPUDP (MultiPath UDP) (Beta)
As of this release, the SDK supports MPUDP protocol, which enables you to connect and use multiple paths to maximize the use of channel resources based on the UDP protocol. You can use different physical NICs on both mobile and desktop and aggregate them to effectively combat network jitter and improve transmission quality.
12. Register extensions (Windows)
This release adds the registerExtension
method for registering extensions. When using a third-party extension, you need to call the extension-related APIs in the following order:
loadExtensionProvider
-> registerExtension
-> setExtensionProviderProperty
-> enableExtension
13. Device management (Windows, macOS)
This release adds a series of callbacks to help you better understand the status of your audio and video devices:
onVideoDeviceStateChanged
: Occurs when the status of the video device changes.onAudioDeviceStateChanged
: Occurs when the status of the audio device changes.onAudioDeviceVolumeChanged
: Occurs when the volume of an audio device or app changes.
14. Camera capture options
This release adds the followEncodeDimensionRatio
member in CameraCapturerConfiguration
, which enables you to set whether to follow the video aspect ratio already set in setVideoEncoderConfiguration
when capturing video with the camera.
15. Multi-channel management
This release adds a series of multi-channel related methods that you can call to manage audio and video streams in multi-channel scenarios.
- The
muteLocalAudioStreamEx
andmuteLocalVideoStreamEx
methods are used to cancel or resume publishing a local audio or video stream, respectively. - The
muteAllRemoteAudioStreamsEx
andmuteAllRemoteVideoStreamsEx
are used to cancel or resume the subscription of all remote users to audio or video streams, respectively. - The
startRtmpStreamWithoutTranscodingEx
,startRtmpStreamWithTranscodingEx
,updateRtmpTranscodingEx
, andstopRtmpStreamEx
methods are used to implement Media Push in multi-channel scenarios. - The
startChannelMediaRelayEx
,updateChannelMediaRelayEx
,pauseAllChannelMediaRelayEx
,resumeAllChannelMediaRelayEx
, andstopChannelMediaRelayEx
methods are used to relay media streams across channels in multi-channel scenarios. - The
options
parameter in theleaveChannelEx
method, which is used to choose whether to stop recording with the microphone when leaving a channel in a multi-channel scenario.
16. Video encoding preferences
In general scenarios, the default video encoding configuration meets most requirements. For certain specific scenarios, this release adds the advanceOptions
member in VideoEncoderConfiguration
for advanced settings of video encoding properties:
CompressionPreference
: The compression preferences for video encoding, which is used to select low-latency or high-quality video preferences.EncodingPreference
: The video encoder preference, which is used to select adaptive preference, software encoder preference, or hardware encoder video preferences.
17. Client role switching
In order to enable users to know whether the switched user role is low-latency or ultra-low-latency, this release adds the newRoleOptions
parameter to the onClientRoleChanged
callback. The value of this parameter is as follows:
audienceLatencyLevelLowLatency (1)
: Low latency.audienceLatencyLevelUltraLowLatency (2)
: Ultra-low latency.
18. Brand-new AI Noise Suppression
The SDK supports a new version of Noise Suppression (versus the basic AI Noise Suppression in agora_rtc_engine: ^5.x). Compared to the previous Noise Suppression, the new AI Noise Suppression has better vocal fidelity, cleaner noise suppression, and add the Dereverberation capability. To enable this feature, contact support@agora.io.
Improvements
1. Video information change callback
This release optimizes the trigger logic of onVideoSizeChanged
, which can also be triggered and report the local video size change when startPreview
is called separately.
2. First video frame rendering (Windows)
This release speeds up the first video frame rendering time to improve the video experience.
3. Screen sharing
In addition to the usability enhancements detailed in the fixed issued section, this release includes a number of functional improvements to screen sharing, as follows:
Windows
- New
minimizeWindow
member inScreenCaptureSourceInfo
to indicate whether the target window is minimized. - New
enableHighLight
,highLightColor
, andhighLightWidth
members inScreenCaptureParameters
so that you can place a border around the target window or screen when screen sharing. - Compatibility with a greater number of mainstream apps, including WPS Office, Microsoft Office PowerPoint, Visual Studio Code, Adobe Photoshop, Windows Media Player, and Scratch.
- Compatibility with additional devices and operating systems, including: Window 8 systems, devices without discrete graphics cards, and dual graphics devices.
- Support for Ultra HD video (4K, 60 fps) on devices that meet the requirements. Agora recommends a device with an Intel Core i7-9750H CPU @ 2.60 GHz or better.
macOS
- Compatibility with additional devices and scenarios, including: dual graphics devices and screen sharing using external screens.
- Support for Ultra HD video (4K, 60 fps) on devices that meet the requirements. Agora recommends a 2021 16" M1 Macbook Pro or better.
4. Bluetooth permissions (Android)
To simplify integration, as of this release, you can use the SDK to enable Android users to use Bluetooth normally without adding the BLUETOOTH_CONNECT
permission.
5. Relaying media streams across channels
This release optimizes the updateChannelMediaRelay
method as follows:
- Before v6.1.0: If the target channel update fails due to internal reasons in the server, the SDK returns the error code
relayEventPacketUpdateDestChannelRefused (8)
, and you need to call theupdateChannelMediaRelay
method again. - v6.1.0 and later: If the target channel update fails due to internal server reasons, the SDK retries the update until the target channel update is successful.
6. Reconstructed AIAEC algorithm
This release reconstructs the AEC algorithm based on the AI method. Compared with the traditional AEC algorithm, the new algorithm can preserve the complete, clear, and smooth near-end vocals under poor echo-to-signal conditions, significantly improving the system's echo cancellation and dual-talk performance. This gives users a more comfortable call and live-broadcast experience. AIAEC is suitable for conference calls, chats, karaoke, and other scenarios.
7. Virtual background
This release optimizes the virtual background algorithm. Improvements include the following:
- The boundaries of virtual backgrounds are handled in a more nuanced way and image matting is is now extremely thin.
- The stability of the virtual background is improved whether the portrait is still or moving, effectively eliminating the problem of background flickering and exceeding the range of the picture.
- More application scenarios are now supported, and a user obtains a good virtual background effect day or night, indoors or out.
- A larger variety of postures are now recognized, when half the body is motionless, the body is shaking, the hands are swinging, or there is fine finger movement. This helps to achieve a good virtual background effect in conjunction with many different gestures.
Other improvements
This release includes the following additional improvements:
- Reduces the latency when pushing external audio sources.
- Improves the performance of echo cancellation when using the
audioScenarioMeeting
scenario. - Improves the smoothness of SDK video rendering.
- Reduces the CPU usage and power consumption of the local device when the host calls the
muteLocalVideoStream
method. (Windows, macOS) - Enhances the ability to identify different network protocol stacks and improves the SDK's access capabilities in multiple-operator network scenarios.
Issues fixed
This release fixes the following issues:
All
- When calling
setVideoEncoderConfigurationEx
in the channel to increase the resolution of the video, it occasionally failed. - When using the Agora media player to play videos, after you played and paused the video, and then called the
seek
method to specify a new position for playback, the video image occasionally remained unchanged; if you called theresume
method to resume playback, the video was sometimes played at a speed faster than the original one. - When entering a live streaming room that has been played for a long time as an audience, the time for the first frame to be rendered was shortened.
- The call
getExtensionProperty
failed and returned an empty string. - Audience members heard buzzing noises when the host switched between speakers and earphones during live streaming.
Android
- In online meeting scenarios, the local user and the remote user occasionally could not hear each other after the local user was interrupted by a call.
- After calling
setCloudProxy
to set the cloud proxy, callingjoinChannelEx
to join multiple channels failed.
iOS
- Calling
startAudioMixing
to play music files in the ipod-library://item path failed. - Different timestamps for audio and video data were obtained simultaneously and separately via
onRecordAudioFrame
andonCaptureVideoFrame
callbacks.
Windows
- When
stopPreview
was called to disable the local video preview, the virtual background that had been set up was occasionally invalidated. - Crashes occasionally occurred when exiting a channel and joining it multiple times with virtual background enabled and set to blur effect.
- If the local client used a camera with a resolution set to 1920 × 1080 as the video capture source, the resolution of the remote video was occasionally inconsistent with the local client.
- When capturing video through the camera, if the video aspect ratio set in
CameraCapturerConfiguration
was inconsistent with that set insetVideoEncoderConfiguration
, the aspect ratio of the local video preview was not rendered according to the latter setting. - In screen sharing scenarios, when the user minimized and then restored the shared window, the remote video occasionally switched to the low-quality stream.
- When the host started screen sharing during live streaming, the audience members sometimes heard echoes.
- In screen sharing scenarios, the system volume of the local user occasionally decreased.
- In screen sharing scenarios, a black screen appeared when sharing a screen between a landscape monitor and a portrait monitor.
- In screen sharing scenarios with a window excluded, the application crashed when the specified shared area exceeded the screen resolution.
- The application failed to exclude a window using the
startScreenCaptureByDisplayId
method for screen sharing. - In screen sharing scenarios, the screen seen by the remote user occasionally crashed, lagged, or displayed a black screen.
- The uplink network quality reported by the
onNetworkQuality
callback was inaccurate for the user who was sharing a screen. - In screen sharing scenarios, when the user shared the screen by window, the mouse in the shared screen was not in its actual position.
- When switching from a non-screen sharing scenario to a screen sharing one, the application occasionally crashed if the user did not switch the resolution accordingly.
macOS
- When capturing video through the camera, if the video aspect ratio set in
CameraCapturerConfiguration
was inconsistent with that set insetVideoEncoderConfiguration
, the aspect ratio of the local video preview was not rendered according to the latter setting. - In screen sharing scenarios on Mac devices, when the user minimized or closed a shared application window, another window of the same application was automatically shared.
- In screen sharing scenarios, the system volume of the local user occasionally decreased.
- In screen sharing scenarios, the shared window of a split screen was not highlighted correctly.
- In screen sharing scenarios, the screen seen by the remote user occasionally crashed, lagged, or displayed a black screen.
- The uplink network quality reported by the
onNetworkQuality
callback was inaccurate for the user who was sharing a screen. - After starting and stopping the audio capture device test, there was no sound when the audio playback device was subsequently started.
- The
onVideoPublishStateChanged
callback reported an inaccurate video source type.
API changes
Added
getNativeHandle
getPlaybackDefaultDevice
getRecordingDefaultDevice
enableDualStreamModeEx
setDualStreamMode
setDualStreamModeEx
SimulcastStreamMode
getNetworkType
setLoopbackDevice
(Windows)getLoopbackDevice
(Windows)followSystemLoopbackDevice
(Windows)setZones
setPlayerAttenuation
setRemoteAudioAttenuation
setHeadphoneEQPreset
setHeadphoneEQParameters
HeadphoneEqualizerPreset
enableMultiCamera
(iOS)setRemoteVideoSubscriptionOptions
setRemoteVideoSubscriptionOptionsEx
VideoSubscriptionOptions
AdvanceOptions
EncodingPreference
CompressionPreference
adjustUserPlaybackSignalVolumeEx
onRhythmPlayerStateChanged
(Android, iOS)RhythmPlayerStateType
RhythmPlayerErrorType
enableAudioVolumeIndicationEx
onVideoDeviceStateChanged
(Windows, macOS)onAudioDeviceStateChanged
(Windows, macOS)onAudioDeviceVolumeChanged
(Windows, macOS)registerExtension
(Windows)muteLocalAudioStreamEx
muteLocalVideoStreamEx
muteAllRemoteAudioStreamsEx
muteAllRemoteVideoStreamsEx
startRtmpStreamWithoutTranscodingEx
startRtmpStreamWithTranscodingEx
updateRtmpTranscodingEx
stopRtmpStreamEx
startChannelMediaRelayEx
updateChannelMediaRelayEx
pauseAllChannelMediaRelayEx
resumeAllChannelMediaRelayEx
stopChannelMediaRelayEx
Modified
- Adds
isAudioFilterable
inChannelMediaOptions
- Adds
enableDoppler
inSpatialAudioParams
- Adds
minimizeWindow
inScreenCaptureSourceInfo
- Adds
followEncodeDimensionRatio
inCameraCapturerConfiguration
- Adds
options
inleaveChannelEx
- Adds
advanceOptions
inVideoEncoderConfiguration
- Adds
newRoleOptions
inonClientRoleChanged
- Adds
setupMode
,mediaPlayerId
, andcropArea
inVideoCanvas
- Adds
hwEncoderAccelerating
inLocalVideoStats
- Removes
sourceType
inenableDualStreamMode
enableInEarMonitoring
: Supports Windows and macOSsetEarMonitoringAudioFrameParameters
: Supports Windows and macOSsetInEarMonitoringVolume
: Supports Windows and macOSonEarMonitoringAudioFrame
: Supports Windows and macOSScreenCaptureParameters
: Supports Windows
Deprecated
onApiCallExecuted
: Use the callbacks triggered by specific methods instead.relayEventPacketUpdateDestChannelRefused
(8) inChannelMediaRelayEvent
v6.0.0
v6.0.0 was released on September 29, 2022. This release is compatible with Video Calling 4.x.
Compatibility changes
This release changed SDK package name from agora_rtc_ng
to agora_rtc_engine
, and optimized the implementation of some features, resulting in incompatibility with v5.x. The following are the main features with compatibility changes:
- Multiple channel
- Media stream publishing control
- Custom video capture and renderer (Media IO)
- Error codes and warning codes
After upgrading the SDK, you need to update the code in your app according to the business scenarios. See Migration guide for details.
New features
1. Multiple media tracks
This release supports one RtcEngine
instance to collect multiple audio and video sources at the same time and publish them to the remote users by setting RtcEngineEx
and ChannelMediaOptions.
- After calling
joinChannel
to join the first channel, calljoinChannelEx
multiple times to join multiple channels, and publish the specified stream to different channels through different user ID (localUid
) andChannelMediaOptions
settings. - You can simultaneously publish multiple sets of video streams captured by multiple cameras or screen sharing by setting
publishSecondaryCameraTrack
andpublishSecondaryScreenTrack
inChannelMediaOptions
. (Windows)
This release adds createCustomVideoTrack
method to implement video custom capture. You can refer to the following steps to publish multiple custom captured video in the channel:
- Create a custom video track: Call the
createCustomVideoTrack
method to create a video track, and get the video track ID. - Set the custom video track to be published in the channel: In each channel's
ChannelMediaOptions
, set thecustomVideoTrackId
parameter to the ID of the video track you want to publish, and setpublishCustomVideoTrack
totrue
. - Pushing an external video source: Call
pushVideoFrame
, and specifyvideoTrackId
as the ID of the custom video track in step 2 in order to publish the corresponding custom video source in multiple channels.
You can also experience the following features with the multi-channel capability:
- Publish multiple sets of audio and video streams to the remote users through different user IDs (
uid
). - Mix multiple audio streams and publish to the remote users through a user ID (
uid
). - Combine multiple video streams and publish them to the remote users through a user ID (
uid
).
2. Ultra HD resolution (Beta) (Windows, macOS)
In order to improve the interactive video experience, the SDK optimizes the whole process of video capture, encoding, decoding and rendering, and now supports 4K resolution. The improved FEC (Forward Error Correction) algorithm enables adaptive switches according to the frame rate and number of video frame packets, which further reduces the video stuttering rate in 4K scenes.
Additionally, you can set the encoding resolution to 4K (3840 × 2160) and the frame rate to 60 fps when calling setVideoEncoderConfiguration
. The SDK supports automatic fallback to the appropriate resolution and frame rate if your device does not support 4K.
This feature has certain requirements with regards to device performance and network bandwidth, and the supported upstream and downstream frame rates vary on different platforms. To enable this feature, contact support@agora.io.
The increase in the default resolution affects the aggregate resolution and thus the billing rate. See Pricing.
3. Full HD and Ultra HD resolution (Beta) (Android, iOS)
In order to improve the interactive video experience, the SDK optimizes the whole process of video capturing, encoding, decoding and rendering. Starting from this version, it supports Full HD (FHD) and Ultra HD (UHD) video resolutions. You can set the dimensions
parameter to 1920 × 1080 or higher when calling the setVideoEncoderConfiguration
method. If your device does not support high resolutions, the SDK will automatically fall back to an appropriate resolution.
The UHD resolution (4K, 60 fps) is currently in beta and requires certain device performance and network bandwidth. If you want to enable this feature, contact technical support.
High resolution typically means higher performance consumption. To avoid a decrease in experience due to insufficient device performance, Agora recommends that you enable FHD and UHD video resolutions on devices with better performance.
The increase in the default resolution affects the aggregate resolution and thus the billing rate. See Pricing.
4. Agora media player
To make it easier for users to integrate the Agora SDK and reduce the SDK's package size, this release introduces the Agora media player. After calling the createMediaPlayer
method to create a media player object, you can then call the methods in the MediaPlayer
class to experience a series of functions, such as playing local and online media files, preloading a media file, changing the CDN route for playing according to your network conditions, or sharing the audio and video streams being played with remote users.
- Plays local and online media files.
- Preloads media files.
- Changes the CDN route for playing media files according your network conditions.
- Shares the audio and video streams being played with remote users.
- Caches data when playing media files.
To make audios sound clearer and stay true to the original sound of audio files, this release adds the ultraHighQualityVoice
enumeration. In scenarios that involve human voices, such as chatting or singing, you can call setVoiceBeautifierPreset
and use this enumeration to experience ultra-high audio quality.
5. Spatial audio
NOTE: This feature is in experimental status. To enable this feature, contact support@agora.io. Contact Technical Support if needed.
You can set the spatial audio for the remote user as following:
-
Local Cartesian Coordinate System Calculation Solution: This solution uses the
LocalSpatialAudioEngine
class to implement spatial audio by calculating the spatial coordinates of the remote user. You need to callupdateSelfPosition
andupdateRemotePosition
to update the spatial coordinates of the local and remote users respectively, so that the local user can hear the spatial audio of the remote user. -
Local Cartesian Coordinate System Calculation Solution: This solution uses the
LocalSpatialAudioEngine
class to implement spatial audio. You need to callupdateSelfPosition
andupdatePlayerPositionInfo
to update the spatial coordinates of the local user and media player, respectively, so that the local user can hear the spatial audio effect of media player.
6. Real-time chorus
This release gives real-time chorus the following abilities:
- Two or more choruses are supported.
- Each singer is independent of each other. If one singer fails or quits the chorus, the other singers can continue to sing.
- Very low latency experience. Each singer can hear each other in real time, and the audience can also hear each singer in real time.
This release adds the audioScenarioChorus
enumeration in AudioScenarioType
. With this enumeration, users can experience ultra-low latency in real-time chorus when the network conditions are good.
7. Enhanced channel management
To meet the channel management requirements of various business scenarios, this release adds the following functions to the ChannelMediaOptions
structure:
- Sets or switches the publishing of multiple audio and video sources.
- Sets or switches channel profile and user role.
- Sets or switches the stream type of the subscribed video.
- Controls audio publishing delay.
Set ChannelMediaOptions
when calling joinChannel
or joinChannelEx
to specify the publishing and subscription behavior of a media stream, for example, whether to publish video streams captured by cameras or screen sharing, and whether to subscribe to the audio and video streams of remote users. After joining the channel, call updateChannelMediaOptions
to update the settings in ChannelMediaOptions
at any time, for example, to switch the published audio and video sources.
8. Screen sharing
This release optimizes the screen sharing function. You can enable this function through the following ways.
On Windows and macOS:
- Call the
startScreenCaptureByDisplayId
method before joining a channel, and then calljoinChannel
[2/2] to join a channel and setpublishScreenTrack
orpublishSecondaryScreenTrack
astrue
. - Call the
startScreenCaptureByDisplayId
method after joining a channel, and then callupdateChannelMediaOptions
to setpublishScreenTrack
orpublishSecondaryScreenTrack
astrue
.
On Android and iOS:
- Call the
startScreenCapture
method before joining a channel, and then calljoinChannel
[2/2] to join a channel and setpublishScreenCaptureVideo
astrue
. - Call the
startScreenCapture
method after joining a channel, and then callupdateChannelMediaOptions
to setpublishScreenCaptureVideo
as true.
9. Subscription allowlists and blocklists
This release introduces subscription allowlists and blocklists for remote audio and video streams. You can add the user ID that you want to subscribe to in your allowlist, or in your blocklist if you do not want to subscribe to. You can experience this feature through the following APIs, and in scenarios that involve multiple channels, you can call the following methods in the RtcEngineEx
interface.
setSubscribeAudioBlacklist
: Set the audio subscription blocklist.setSubscribeAudioWhitelist
: Set the audio subscription allowlist.setSubscribeVideoBlacklist
: Set the video subscription blocklist.setSubscribeVideoWhitelist
: Set the video subscription allowlist.
If a user is added in a blocklist and allowlist at the same time, only the blocklist takes effect.
10. Set audio scenarios
To make it easier to change audio scenarios, this release adds the setAudioScenario
method. For example, if you want to change the audio scenario from AudioScenarioDefault
to AudioScenarioGameStreaming
when you are in a channel, you can call this method.
11. Replace video feeds with images
This release supports replacing video feeds with images when publishing video streams. You can call the enableVideoImageSource
method to enable this function and choose your own images through the options
parameter. If you disable this function, the remote users see the video feeds that you publish.
12. Local video transcoder
This release introduces local video transcoder with which you can locally merge multiple video streams into one. Common scenarios are as follows:
When streaming or using media push, you can merge the frames of multiple anchors into one locally.
To do this, you should merge multi-channel video streams collected locally (such as video captured by camera, screen sharing stream, video files and pictures) into one video stream. Then, release the merged stream in the channel.
You can call the startLocalVideoTranscoder
method to enable the local video transcoder and call the stopLocalVideoTranscoder
method to stop the local video transcoder; After the local video transcoder is enabled, you can call updateLocalTranscoderConfiguration
to update the configuration of the local video transcoder.
13. Video device management
A video capture device may support a variety of video formats, each of which supports a different combination of video frame width, video frame height, and frame rate.
This release introduces the numberOfCapabilities
and getCapability
method to obtain the number of video formats supported by video capture devices and the details of video frames under specified video formats. When you call the startPrimaryCameraCapture
or startSecondaryCameraCapture
method to capture video using camera, you can obtain the video with the specified video format.
The SDK will automatically select the video format for video capture equipment according to your settings in VideoEncoderConfiguration
. In general, you don't need to use this set of new methods.
Improvements
1. Fast channel switching
This release can achieve the same switching speed as switchChannel
in agora_rtc_engine: ^5.x through the leaveChannel
and joinChannel
methods so that you don't need to take the time to call the switchChannel
method.
2. Push external video frames
This releases supports pushing video frames in I422 format. You can call the pushVideoFrame
method to push such video frames to the SDK.
3. Voice pitch of the local user
This release adds voicePitch
in AudioVolumeInfo
of onAudioVolumeIndication
. You can use voicePitch
to get the local user's voice pitch and perform business functions such as rating for singing.
4. Video preview
This release improves the implementation logic of startPreview
. You can call the startPreview
method to enable video preview at any time.
5. Video types of subscription
You can call the setRemoteDefaultVideoStreamType
method to choose the video stream type when subscribing streams.
Notifications
2022.10
- After you enable Notifications, your server receives the events that you subscribe to in the form of HTTPS requests.
- To improve communication security between the Notifications and your server, Agora SD-RTN™ uses signatures for identity verification.
- As of this release, you can use Notifications in conjunction with this product.
AI Noise Suppression
Agora charges additionally for this extension. See Pricing.
v1.1.0
Improvement
This release improves the calculation performance of the AI-powered noise suppression algorithm.
New features
This release adds the following APIs and parameters:
- APIs:
checkCompatibility
: Checks whether the AI Noise Suppression extension is supported on the current browser.setMode
: Sets the noise suppression mode as AI noise suppression or stationary noise suppression.setLevel
: Sets the AI noise suppression level.
- Parameters:
elapsedTime
inonoverload
: Reports the time in ms that the extension needs to process one audio frame.
For API details, see AI Noise Suppression.
Compatibility changes
This release brings the following changes:
- AI Noise Suppression supports Agora Video SDK for Web v4.15.0 or later.
- The extension has Wasm dependencies only. Because JS dependencies are removed, you need to publish the Wasm files located in the
node_modules/agora-extension-ai-denoiser/external
directory again. If you have enabled the Content Security Policy (CSP), you need to modify the CSP configuration. See AI Noise Suppression for details. - The audio data is dumped in PCM format instead of WAV format.
- To adjust the intensity of noise suppression, best practice is to call
setLevel
.
v1.0.0
First release.
Virtual Background
You may be charged for the usage of this extension. Contact support@agora.io for details.
v1.1.3
Fixed issues
This release fixes the occasional issue of jagged background images on Chrome for Android.
v1.1.2
New features
You can now specify the fit
property when calling setOptions
. This sets how the background is resized to fit the container. For API details, see Virtual background.
Compatibility changes
Virtual Background supports Agora Video SDK for Web v4.15.0 or later.
v1.1.1
New features
You can now call checkCompatibility
and test if AI Noise Suppression extension is supported on the current browser. For API details, see Virtual background.
Fixed issues
A black bar is no longer displayed to the left of the virtual background.
v1.1.0
New features
You can create multiple VirtualBackgroundProcessor
instances to process multiple video streams.
v1.0.0
First release.