1// SPDX-License-Identifier: GPL-2.0 2/* 3 * Copyright (C) 2016 CNEX Labs 4 * Initial release: Javier Gonzalez <javier@cnexlabs.com> 5 * Matias Bjorling <matias@cnexlabs.com> 6 * 7 * This program is free software; you can redistribute it and/or 8 * modify it under the terms of the GNU General Public License version 9 * 2 as published by the Free Software Foundation. 10 * 11 * This program is distributed in the hope that it will be useful, but 12 * WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 14 * General Public License for more details. 15 * 16 * pblk-rl.c - pblk's rate limiter for user I/O 17 * 18 */ 19 20#include "pblk.h" 21 22static void pblk_rl_kick_u_timer(struct pblk_rl *rl) 23{ 24 mod_timer(&rl->u_timer, jiffies + msecs_to_jiffies(5000)); 25} 26 27int pblk_rl_is_limit(struct pblk_rl *rl) 28{ 29 int rb_space; 30 31 rb_space = atomic_read(&rl->rb_space); 32 33 return (rb_space == 0); 34} 35 36int pblk_rl_user_may_insert(struct pblk_rl *rl, int nr_entries) 37{ 38 int rb_user_cnt = atomic_read(&rl->rb_user_cnt); 39 int rb_space = atomic_read(&rl->rb_space); 40 41 if (unlikely(rb_space >= 0) && (rb_space - nr_entries < 0)) 42 return NVM_IO_ERR; 43 44 if (rb_user_cnt >= rl->rb_user_max) 45 return NVM_IO_REQUEUE; 46 47 return NVM_IO_OK; 48} 49 50void pblk_rl_inserted(struct pblk_rl *rl, int nr_entries) 51{ 52 int rb_space = atomic_read(&rl->rb_space); 53 54 if (unlikely(rb_space >= 0)) 55 atomic_sub(nr_entries, &rl->rb_space); 56} 57 58int pblk_rl_gc_may_insert(struct pblk_rl *rl, int nr_entries) 59{ 60 int rb_gc_cnt = atomic_read(&rl->rb_gc_cnt); 61 int rb_user_active; 62 63 /* If there is no user I/O let GC take over space on the write buffer */ 64 rb_user_active = READ_ONCE(rl->rb_user_active); 65 return (!(rb_gc_cnt >= rl->rb_gc_max && rb_user_active)); 66} 67 68void pblk_rl_user_in(struct pblk_rl *rl, int nr_entries) 69{ 70 atomic_add(nr_entries, &rl->rb_user_cnt); 71 72 /* Release user I/O state. Protect from GC */ 73 smp_store_release(&rl->rb_user_active, 1); 74 pblk_rl_kick_u_timer(rl); 75} 76 77void pblk_rl_werr_line_in(struct pblk_rl *rl) 78{ 79 atomic_inc(&rl->werr_lines); 80} 81 82void pblk_rl_werr_line_out(struct pblk_rl *rl) 83{ 84 atomic_dec(&rl->werr_lines); 85} 86 87void pblk_rl_gc_in(struct pblk_rl *rl, int nr_entries) 88{ 89 atomic_add(nr_entries, &rl->rb_gc_cnt); 90} 91 92void pblk_rl_out(struct pblk_rl *rl, int nr_user, int nr_gc) 93{ 94 atomic_sub(nr_user, &rl->rb_user_cnt); 95 atomic_sub(nr_gc, &rl->rb_gc_cnt); 96} 97 98unsigned long pblk_rl_nr_free_blks(struct pblk_rl *rl) 99{ 100 return atomic_read(&rl->free_blocks); 101} 102 103unsigned long pblk_rl_nr_user_free_blks(struct pblk_rl *rl) 104{ 105 return atomic_read(&rl->free_user_blocks); 106} 107 108static void __pblk_rl_update_rates(struct pblk_rl *rl, 109 unsigned long free_blocks) 110{ 111 struct pblk *pblk = container_of(rl, struct pblk, rl); 112 int max = rl->rb_budget; 113 int werr_gc_needed = atomic_read(&rl->werr_lines); 114 115 if (free_blocks >= rl->high) { 116 if (werr_gc_needed) { 117 /* Allocate a small budget for recovering 118 * lines with write errors 119 */ 120 rl->rb_gc_max = 1 << rl->rb_windows_pw; 121 rl->rb_user_max = max - rl->rb_gc_max; 122 rl->rb_state = PBLK_RL_WERR; 123 } else { 124 rl->rb_user_max = max; 125 rl->rb_gc_max = 0; 126 rl->rb_state = PBLK_RL_OFF; 127 } 128 } else if (free_blocks < rl->high) { 129 int shift = rl->high_pw - rl->rb_windows_pw; 130 int user_windows = free_blocks >> shift; 131 int user_max = user_windows << ilog2(NVM_MAX_VLBA); 132 133 rl->rb_user_max = user_max; 134 rl->rb_gc_max = max - user_max; 135 136 if (free_blocks <= rl->rsv_blocks) { 137 rl->rb_user_max = 0; 138 rl->rb_gc_max = max; 139 } 140 141 /* In the worst case, we will need to GC lines in the low list 142 * (high valid sector count). If there are lines to GC on high 143 * or mid lists, these will be prioritized 144 */ 145 rl->rb_state = PBLK_RL_LOW; 146 } 147 148 if (rl->rb_state != PBLK_RL_OFF) 149 pblk_gc_should_start(pblk); 150 else 151 pblk_gc_should_stop(pblk); 152} 153 154void pblk_rl_update_rates(struct pblk_rl *rl) 155{ 156 __pblk_rl_update_rates(rl, pblk_rl_nr_user_free_blks(rl)); 157} 158 159void pblk_rl_free_lines_inc(struct pblk_rl *rl, struct pblk_line *line) 160{ 161 int blk_in_line = atomic_read(&line->blk_in_line); 162 int free_blocks; 163 164 atomic_add(blk_in_line, &rl->free_blocks); 165 free_blocks = atomic_add_return(blk_in_line, &rl->free_user_blocks); 166 167 __pblk_rl_update_rates(rl, free_blocks); 168} 169 170void pblk_rl_free_lines_dec(struct pblk_rl *rl, struct pblk_line *line, 171 bool used) 172{ 173 int blk_in_line = atomic_read(&line->blk_in_line); 174 int free_blocks; 175 176 atomic_sub(blk_in_line, &rl->free_blocks); 177 178 if (used) 179 free_blocks = atomic_sub_return(blk_in_line, 180 &rl->free_user_blocks); 181 else 182 free_blocks = atomic_read(&rl->free_user_blocks); 183 184 __pblk_rl_update_rates(rl, free_blocks); 185} 186 187int pblk_rl_high_thrs(struct pblk_rl *rl) 188{ 189 return rl->high; 190} 191 192int pblk_rl_max_io(struct pblk_rl *rl) 193{ 194 return rl->rb_max_io; 195} 196 197static void pblk_rl_u_timer(struct timer_list *t) 198{ 199 struct pblk_rl *rl = from_timer(rl, t, u_timer); 200 201 /* Release user I/O state. Protect from GC */ 202 smp_store_release(&rl->rb_user_active, 0); 203} 204 205void pblk_rl_free(struct pblk_rl *rl) 206{ 207 del_timer(&rl->u_timer); 208} 209 210void pblk_rl_init(struct pblk_rl *rl, int budget, int threshold) 211{ 212 struct pblk *pblk = container_of(rl, struct pblk, rl); 213 struct nvm_tgt_dev *dev = pblk->dev; 214 struct nvm_geo *geo = &dev->geo; 215 struct pblk_line_mgmt *l_mg = &pblk->l_mg; 216 struct pblk_line_meta *lm = &pblk->lm; 217 int sec_meta, blk_meta; 218 unsigned int rb_windows; 219 220 /* Consider sectors used for metadata */ 221 sec_meta = (lm->smeta_sec + lm->emeta_sec[0]) * l_mg->nr_free_lines; 222 blk_meta = DIV_ROUND_UP(sec_meta, geo->clba); 223 224 rl->high = pblk->op_blks - blk_meta - lm->blk_per_line; 225 rl->high_pw = get_count_order(rl->high); 226 227 rl->rsv_blocks = pblk_get_min_chks(pblk); 228 229 /* This will always be a power-of-2 */ 230 rb_windows = budget / NVM_MAX_VLBA; 231 rl->rb_windows_pw = get_count_order(rb_windows); 232 233 /* To start with, all buffer is available to user I/O writers */ 234 rl->rb_budget = budget; 235 rl->rb_user_max = budget; 236 rl->rb_gc_max = 0; 237 rl->rb_state = PBLK_RL_HIGH; 238 239 /* Maximize I/O size and ansure that back threshold is respected */ 240 if (threshold) 241 rl->rb_max_io = budget - pblk->min_write_pgs_data - threshold; 242 else 243 rl->rb_max_io = budget - pblk->min_write_pgs_data - 1; 244 245 atomic_set(&rl->rb_user_cnt, 0); 246 atomic_set(&rl->rb_gc_cnt, 0); 247 atomic_set(&rl->rb_space, -1); 248 atomic_set(&rl->werr_lines, 0); 249 250 timer_setup(&rl->u_timer, pblk_rl_u_timer, 0); 251 252 rl->rb_user_active = 0; 253 rl->rb_gc_active = 0; 254} 255