There is no simple / easy answer for your questions at this time. While most of your requirements can be met by ACS (e.g., interactivity, chat, 1:1, real time video (i.e., 200- 300 ms range for latency) etc.) Other capabilities can be met by Azure Media Services (AMS) - (e.g., Premium video quality, scale to very large audiences, etc.). ACS uses the WebRTC protocol for video delivery which prioritizes conversational video quality vs broadcast video quality. AMS leverages adaptive bitrate (ABR) protocols (HLS, DASH, etc.) which focuses on broadcast video quality and typically has had glass-to-glass latencies of 15 or more seconds. Recent improvements in HLS and DASH protocols (LL-HLS and LL-DASH) allow for much lower latencies in the 2-6 second range. These improvements will be available later this year for AMS.
Overall, the streaming industry is working towards an inflection point in the future where real time video becomes indistinguishable from broadcast video. Until this is achieved, solutions that need real time / interactive capabilities as well as large scale and broadcast video quality capabilities are typically taking a hybrid approach of using a combination of WebRTC and ABR solutions. Cost is also a driving factor to a hybrid approach as delivery costs for WebRTC products are typically an order of magnitude larger vs ABR products. Teams is an example of this hybrid approach where WebRTC is used up to a specific audience size and then ABR path is taken (leveraging AMS) for very large audiences.
With regard to AMS library for Android & iOS one would just use the default player frameworks for each of these. For building a native app for iOS one would just use AVPlayer framework. To build a native app for Android one would use ExoPlayer framework. One can then just point an HLS URL to either of these for playback.