Throttling Concurrent Outgoing HTTP Requests in .NET Core
In my last post, we implemented the rate limiting for an API endpoint on the server side. The server side rate limiting is useful for API providers to ensure system performance and/or to realize business values. On the client side, the API consumers then should throttle the rate of concurrent HTTP requests in order to comply with the rate limits of the endpoints and moderate the usage of client side resources. This post will go over how to make concurrent outgoing HTTP requests on the client…