1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
|
/**
* Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
* SPDX-License-Identifier: Apache-2.0.
*/
#include <aws/mqtt/private/v5/rate_limiters.h>
#include <aws/common/clock.h>
static int s_rate_limit_time_fn(const struct aws_rate_limiter_token_bucket_options *options, uint64_t *current_time) {
if (options->clock_fn != NULL) {
return (*options->clock_fn)(current_time);
}
return aws_high_res_clock_get_ticks(current_time);
}
int aws_rate_limiter_token_bucket_init(
struct aws_rate_limiter_token_bucket *limiter,
const struct aws_rate_limiter_token_bucket_options *options) {
AWS_ZERO_STRUCT(*limiter);
if (options->tokens_per_second == 0 || options->maximum_token_count == 0) {
return aws_raise_error(AWS_ERROR_INVALID_ARGUMENT);
}
limiter->config = *options;
aws_rate_limiter_token_bucket_reset(limiter);
return AWS_OP_SUCCESS;
}
void aws_rate_limiter_token_bucket_reset(struct aws_rate_limiter_token_bucket *limiter) {
limiter->current_token_count =
aws_min_u64(limiter->config.initial_token_count, limiter->config.maximum_token_count);
limiter->fractional_nanos = 0;
limiter->fractional_nano_tokens = 0;
uint64_t now = 0;
AWS_FATAL_ASSERT(s_rate_limit_time_fn(&limiter->config, &now) == AWS_OP_SUCCESS);
limiter->last_service_time = now;
}
static void s_regenerate_tokens(struct aws_rate_limiter_token_bucket *limiter) {
uint64_t now = 0;
AWS_FATAL_ASSERT(s_rate_limit_time_fn(&limiter->config, &now) == AWS_OP_SUCCESS);
if (now <= limiter->last_service_time) {
return;
}
uint64_t nanos_elapsed = now - limiter->last_service_time;
/*
* We break the regeneration calculation into two distinct steps:
* (1) Perform regeneration based on whole seconds elapsed (nice and easy just multiply times the regen rate)
* (2) Perform regeneration based on the remaining fraction of a second elapsed
*
* We do this to minimize the chances of multiplication saturation before the divide necessary to normalize to
* nanos.
*
* In particular, by doing this, we won't see saturation unless a regeneration rate in the multi-billions is used
* or elapsed_seconds is in the billions. This is similar reasoning to what we do in aws_timestamp_convert_u64.
*
* Additionally, we use a (sub-second) fractional counter/accumulator (fractional_nanos, fractional_nano_tokens)
* in order to prevent error accumulation due to integer division rounding.
*/
/* break elapsed time into seconds and remainder nanos */
uint64_t remainder_nanos = 0;
uint64_t elapsed_seconds =
aws_timestamp_convert(nanos_elapsed, AWS_TIMESTAMP_NANOS, AWS_TIMESTAMP_SECS, &remainder_nanos);
/* apply seconds-based regeneration */
uint64_t tokens_regenerated = aws_mul_u64_saturating(elapsed_seconds, limiter->config.tokens_per_second);
/* apply fractional remainder regeneration */
limiter->fractional_nanos += remainder_nanos;
/* fractional overflow check */
if (limiter->fractional_nanos < AWS_TIMESTAMP_NANOS) {
/*
* no overflow, just do the division to figure out how many tokens are represented by the updated
* fractional nanos
*/
uint64_t new_fractional_tokens =
aws_mul_u64_saturating(limiter->fractional_nanos, limiter->config.tokens_per_second) / AWS_TIMESTAMP_NANOS;
/*
* update token count by how much fractional tokens changed
*/
tokens_regenerated += new_fractional_tokens - limiter->fractional_nano_tokens;
limiter->fractional_nano_tokens = new_fractional_tokens;
} else {
/*
* overflow. In this case, update token count by the remaining tokens left to regenerate to make the
* original fractional nano amount equal to one second. This is the key part (a pseudo-reset) that lets us
* avoid error accumulation due to integer division rounding over time.
*/
tokens_regenerated += limiter->config.tokens_per_second - limiter->fractional_nano_tokens;
/*
* subtract off a second from the fractional part. Guaranteed to be less than a second afterwards.
*/
limiter->fractional_nanos -= AWS_TIMESTAMP_NANOS;
/*
* Calculate the new fractional nano token amount, and add them in.
*/
limiter->fractional_nano_tokens =
aws_mul_u64_saturating(limiter->fractional_nanos, limiter->config.tokens_per_second) / AWS_TIMESTAMP_NANOS;
tokens_regenerated += limiter->fractional_nano_tokens;
}
limiter->current_token_count = aws_add_u64_saturating(tokens_regenerated, limiter->current_token_count);
if (limiter->current_token_count > limiter->config.maximum_token_count) {
limiter->current_token_count = limiter->config.maximum_token_count;
}
limiter->last_service_time = now;
}
bool aws_rate_limiter_token_bucket_can_take_tokens(
struct aws_rate_limiter_token_bucket *limiter,
uint64_t token_count) {
s_regenerate_tokens(limiter);
return limiter->current_token_count >= token_count;
}
int aws_rate_limiter_token_bucket_take_tokens(struct aws_rate_limiter_token_bucket *limiter, uint64_t token_count) {
s_regenerate_tokens(limiter);
if (limiter->current_token_count < token_count) {
/* TODO: correct error once seated in aws-c-common */
return aws_raise_error(AWS_ERROR_INVALID_STATE);
}
limiter->current_token_count -= token_count;
return AWS_OP_SUCCESS;
}
uint64_t aws_rate_limiter_token_bucket_compute_wait_for_tokens(
struct aws_rate_limiter_token_bucket *limiter,
uint64_t token_count) {
s_regenerate_tokens(limiter);
if (limiter->current_token_count >= token_count) {
return 0;
}
uint64_t token_rate = limiter->config.tokens_per_second;
AWS_FATAL_ASSERT(limiter->fractional_nanos < AWS_TIMESTAMP_NANOS);
AWS_FATAL_ASSERT(limiter->fractional_nano_tokens <= token_rate);
uint64_t expected_wait = 0;
uint64_t deficit = token_count - limiter->current_token_count;
uint64_t remaining_fractional_tokens = token_rate - limiter->fractional_nano_tokens;
if (deficit < remaining_fractional_tokens) {
/*
* case 1:
* The token deficit is less than what will be regenerated by waiting for the fractional nanos accumulator
* to reach one second's worth of time.
*
* In this case, base the calculation off of just a wait from fractional nanos.
*/
uint64_t target_fractional_tokens = aws_add_u64_saturating(deficit, limiter->fractional_nano_tokens);
uint64_t remainder_wait_unnormalized = aws_mul_u64_saturating(target_fractional_tokens, AWS_TIMESTAMP_NANOS);
expected_wait = remainder_wait_unnormalized / token_rate - limiter->fractional_nanos;
/* If the fractional wait is itself, fractional, then add one more nano second to push us over the edge */
if (remainder_wait_unnormalized % token_rate) {
++expected_wait;
}
} else {
/*
* case 2:
* The token deficit requires regeneration for a time interval at least as large as what is needed
* to overflow the fractional nanos accumulator.
*/
/* First account for making the fractional nano accumulator exactly one second */
expected_wait = AWS_TIMESTAMP_NANOS - limiter->fractional_nanos;
deficit -= remaining_fractional_tokens;
/*
* Now, for the remaining tokens, split into tokens from whole seconds worth of regeneration as well
* as a remainder requiring a fractional regeneration
*/
uint64_t expected_wait_seconds = deficit / token_rate;
uint64_t deficit_remainder = deficit % token_rate;
/*
* Account for seconds worth of waiting
*/
expected_wait += aws_mul_u64_saturating(expected_wait_seconds, AWS_TIMESTAMP_NANOS);
/*
* And finally, calculate the fractional wait to give us the last few tokens
*/
uint64_t remainder_wait_unnormalized = aws_mul_u64_saturating(deficit_remainder, AWS_TIMESTAMP_NANOS);
expected_wait += remainder_wait_unnormalized / token_rate;
/* If the fractional wait is itself, fractional, then add one more nano second to push us over the edge */
if (remainder_wait_unnormalized % token_rate) {
++expected_wait;
}
}
return expected_wait;
}
|