Lines Matching refs:map
3 // Register map access API
36 static inline bool regmap_should_log(struct regmap *map)
38 return (map->dev && strcmp(dev_name(map->dev), LOG_DEVICE) == 0);
41 static inline bool regmap_should_log(struct regmap *map) { return false; }
45 static int _regmap_update_bits(struct regmap *map, unsigned int reg,
74 bool regmap_check_range_table(struct regmap *map, unsigned int reg,
90 bool regmap_writeable(struct regmap *map, unsigned int reg)
92 if (map->max_register && reg > map->max_register)
95 if (map->writeable_reg)
96 return map->writeable_reg(map->dev, reg);
98 if (map->wr_table)
99 return regmap_check_range_table(map, reg, map->wr_table);
104 bool regmap_cached(struct regmap *map, unsigned int reg)
109 if (map->cache_type == REGCACHE_NONE)
112 if (!map->cache_ops)
115 if (map->max_register && reg > map->max_register)
118 map->lock(map->lock_arg);
119 ret = regcache_read(map, reg, &val);
120 map->unlock(map->lock_arg);
127 bool regmap_readable(struct regmap *map, unsigned int reg)
129 if (!map->reg_read)
132 if (map->max_register && reg > map->max_register)
135 if (map->format.format_write)
138 if (map->readable_reg)
139 return map->readable_reg(map->dev, reg);
141 if (map->rd_table)
142 return regmap_check_range_table(map, reg, map->rd_table);
147 bool regmap_volatile(struct regmap *map, unsigned int reg)
149 if (!map->format.format_write && !regmap_readable(map, reg))
152 if (map->volatile_reg)
153 return map->volatile_reg(map->dev, reg);
155 if (map->volatile_table)
156 return regmap_check_range_table(map, reg, map->volatile_table);
158 if (map->cache_ops)
164 bool regmap_precious(struct regmap *map, unsigned int reg)
166 if (!regmap_readable(map, reg))
169 if (map->precious_reg)
170 return map->precious_reg(map->dev, reg);
172 if (map->precious_table)
173 return regmap_check_range_table(map, reg, map->precious_table);
178 bool regmap_writeable_noinc(struct regmap *map, unsigned int reg)
180 if (map->writeable_noinc_reg)
181 return map->writeable_noinc_reg(map->dev, reg);
183 if (map->wr_noinc_table)
184 return regmap_check_range_table(map, reg, map->wr_noinc_table);
189 bool regmap_readable_noinc(struct regmap *map, unsigned int reg)
191 if (map->readable_noinc_reg)
192 return map->readable_noinc_reg(map->dev, reg);
194 if (map->rd_noinc_table)
195 return regmap_check_range_table(map, reg, map->rd_noinc_table);
200 static bool regmap_volatile_range(struct regmap *map, unsigned int reg,
206 if (!regmap_volatile(map, reg + regmap_get_offset(map, i)))
212 static void regmap_format_12_20_write(struct regmap *map,
215 u8 *out = map->work_buf;
224 static void regmap_format_2_6_write(struct regmap *map,
227 u8 *out = map->work_buf;
232 static void regmap_format_4_12_write(struct regmap *map,
235 __be16 *out = map->work_buf;
239 static void regmap_format_7_9_write(struct regmap *map,
242 __be16 *out = map->work_buf;
246 static void regmap_format_10_14_write(struct regmap *map,
249 u8 *out = map->work_buf;
451 struct regmap *map = __map;
453 hwspin_lock_timeout(map->hwlock, UINT_MAX);
458 struct regmap *map = __map;
460 hwspin_lock_timeout_irq(map->hwlock, UINT_MAX);
465 struct regmap *map = __map;
467 hwspin_lock_timeout_irqsave(map->hwlock, UINT_MAX,
468 &map->spinlock_flags);
473 struct regmap *map = __map;
475 hwspin_unlock(map->hwlock);
480 struct regmap *map = __map;
482 hwspin_unlock_irq(map->hwlock);
487 struct regmap *map = __map;
489 hwspin_unlock_irqrestore(map->hwlock, &map->spinlock_flags);
499 struct regmap *map = __map;
500 mutex_lock(&map->mutex);
505 struct regmap *map = __map;
506 mutex_unlock(&map->mutex);
510 __acquires(&map->spinlock)
512 struct regmap *map = __map;
515 spin_lock_irqsave(&map->spinlock, flags);
516 map->spinlock_flags = flags;
520 __releases(&map->spinlock)
522 struct regmap *map = __map;
523 spin_unlock_irqrestore(&map->spinlock, map->spinlock_flags);
535 static bool _regmap_range_add(struct regmap *map,
538 struct rb_root *root = &map->range_tree;
560 static struct regmap_range_node *_regmap_range_lookup(struct regmap *map,
563 struct rb_node *node = map->range_tree.rb_node;
580 static void regmap_range_exit(struct regmap *map)
585 next = rb_first(&map->range_tree);
589 rb_erase(&range_node->node, &map->range_tree);
593 kfree(map->selector_work_buf);
596 static int regmap_set_name(struct regmap *map, const struct regmap_config *config)
604 kfree_const(map->name);
605 map->name = name;
611 int regmap_attach_dev(struct device *dev, struct regmap *map,
617 map->dev = dev;
619 ret = regmap_set_name(map, config);
623 regmap_debugfs_exit(map);
624 regmap_debugfs_init(map);
629 regmap_debugfs_exit(map);
632 *m = map;
709 struct regmap *map;
717 map = kzalloc(sizeof(*map), GFP_KERNEL);
718 if (map == NULL) {
723 ret = regmap_set_name(map, config);
730 map->lock = map->unlock = regmap_lock_unlock_none;
731 map->can_sleep = config->can_sleep;
732 regmap_debugfs_disable(map);
734 map->lock = config->lock;
735 map->unlock = config->unlock;
736 map->lock_arg = config->lock_arg;
737 map->can_sleep = config->can_sleep;
739 map->hwlock = hwspin_lock_request_specific(config->hwlock_id);
740 if (!map->hwlock) {
747 map->lock = regmap_lock_hwlock_irqsave;
748 map->unlock = regmap_unlock_hwlock_irqrestore;
751 map->lock = regmap_lock_hwlock_irq;
752 map->unlock = regmap_unlock_hwlock_irq;
755 map->lock = regmap_lock_hwlock;
756 map->unlock = regmap_unlock_hwlock;
760 map->lock_arg = map;
764 spin_lock_init(&map->spinlock);
765 map->lock = regmap_lock_spinlock;
766 map->unlock = regmap_unlock_spinlock;
767 lockdep_set_class_and_name(&map->spinlock,
770 mutex_init(&map->mutex);
771 map->lock = regmap_lock_mutex;
772 map->unlock = regmap_unlock_mutex;
773 map->can_sleep = true;
774 lockdep_set_class_and_name(&map->mutex,
777 map->lock_arg = map;
785 map->alloc_flags = GFP_ATOMIC;
787 map->alloc_flags = GFP_KERNEL;
789 map->format.reg_bytes = DIV_ROUND_UP(config->reg_bits, 8);
790 map->format.pad_bytes = config->pad_bits / 8;
791 map->format.val_bytes = DIV_ROUND_UP(config->val_bits, 8);
792 map->format.buf_size = DIV_ROUND_UP(config->reg_bits +
794 map->reg_shift = config->pad_bits % 8;
796 map->reg_stride = config->reg_stride;
798 map->reg_stride = 1;
799 if (is_power_of_2(map->reg_stride))
800 map->reg_stride_order = ilog2(map->reg_stride);
802 map->reg_stride_order = -1;
803 map->use_single_read = config->use_single_read || !bus || !bus->read;
804 map->use_single_write = config->use_single_write || !bus || !bus->write;
805 map->can_multi_write = config->can_multi_write && bus && bus->write;
807 map->max_raw_read = bus->max_raw_read;
808 map->max_raw_write = bus->max_raw_write;
810 map->dev = dev;
811 map->bus = bus;
812 map->bus_context = bus_context;
813 map->max_register = config->max_register;
814 map->wr_table = config->wr_table;
815 map->rd_table = config->rd_table;
816 map->volatile_table = config->volatile_table;
817 map->precious_table = config->precious_table;
818 map->wr_noinc_table = config->wr_noinc_table;
819 map->rd_noinc_table = config->rd_noinc_table;
820 map->writeable_reg = config->writeable_reg;
821 map->readable_reg = config->readable_reg;
822 map->volatile_reg = config->volatile_reg;
823 map->precious_reg = config->precious_reg;
824 map->writeable_noinc_reg = config->writeable_noinc_reg;
825 map->readable_noinc_reg = config->readable_noinc_reg;
826 map->cache_type = config->cache_type;
828 spin_lock_init(&map->async_lock);
829 INIT_LIST_HEAD(&map->async_list);
830 INIT_LIST_HEAD(&map->async_free);
831 init_waitqueue_head(&map->async_waitq);
836 map->read_flag_mask = config->read_flag_mask;
837 map->write_flag_mask = config->write_flag_mask;
839 map->read_flag_mask = bus->read_flag_mask;
843 map->reg_read = config->reg_read;
844 map->reg_write = config->reg_write;
846 map->defer_caching = false;
849 map->reg_read = _regmap_bus_reg_read;
850 map->reg_write = _regmap_bus_reg_write;
851 map->reg_update_bits = bus->reg_update_bits;
853 map->defer_caching = false;
856 map->reg_read = _regmap_bus_read;
857 map->reg_update_bits = bus->reg_update_bits;
863 switch (config->reg_bits + map->reg_shift) {
867 map->format.format_write = regmap_format_2_6_write;
877 map->format.format_write = regmap_format_4_12_write;
887 map->format.format_write = regmap_format_7_9_write;
897 map->format.format_write = regmap_format_10_14_write;
907 map->format.format_write = regmap_format_12_20_write;
915 map->format.format_reg = regmap_format_8;
921 map->format.format_reg = regmap_format_16_be;
924 map->format.format_reg = regmap_format_16_le;
927 map->format.format_reg = regmap_format_16_native;
937 map->format.format_reg = regmap_format_24;
943 map->format.format_reg = regmap_format_32_be;
946 map->format.format_reg = regmap_format_32_le;
949 map->format.format_reg = regmap_format_32_native;
960 map->format.format_reg = regmap_format_64_be;
963 map->format.format_reg = regmap_format_64_le;
966 map->format.format_reg = regmap_format_64_native;
979 map->format.parse_inplace = regmap_parse_inplace_noop;
983 map->format.format_val = regmap_format_8;
984 map->format.parse_val = regmap_parse_8;
985 map->format.parse_inplace = regmap_parse_inplace_noop;
990 map->format.format_val = regmap_format_16_be;
991 map->format.parse_val = regmap_parse_16_be;
992 map->format.parse_inplace = regmap_parse_16_be_inplace;
995 map->format.format_val = regmap_format_16_le;
996 map->format.parse_val = regmap_parse_16_le;
997 map->format.parse_inplace = regmap_parse_16_le_inplace;
1000 map->format.format_val = regmap_format_16_native;
1001 map->format.parse_val = regmap_parse_16_native;
1010 map->format.format_val = regmap_format_24;
1011 map->format.parse_val = regmap_parse_24;
1016 map->format.format_val = regmap_format_32_be;
1017 map->format.parse_val = regmap_parse_32_be;
1018 map->format.parse_inplace = regmap_parse_32_be_inplace;
1021 map->format.format_val = regmap_format_32_le;
1022 map->format.parse_val = regmap_parse_32_le;
1023 map->format.parse_inplace = regmap_parse_32_le_inplace;
1026 map->format.format_val = regmap_format_32_native;
1027 map->format.parse_val = regmap_parse_32_native;
1037 map->format.format_val = regmap_format_64_be;
1038 map->format.parse_val = regmap_parse_64_be;
1039 map->format.parse_inplace = regmap_parse_64_be_inplace;
1042 map->format.format_val = regmap_format_64_le;
1043 map->format.parse_val = regmap_parse_64_le;
1044 map->format.parse_inplace = regmap_parse_64_le_inplace;
1047 map->format.format_val = regmap_format_64_native;
1048 map->format.parse_val = regmap_parse_64_native;
1057 if (map->format.format_write) {
1061 map->use_single_write = true;
1064 if (!map->format.format_write &&
1065 !(map->format.format_reg && map->format.format_val))
1068 map->work_buf = kzalloc(map->format.buf_size, GFP_KERNEL);
1069 if (map->work_buf == NULL) {
1074 if (map->format.format_write) {
1075 map->defer_caching = false;
1076 map->reg_write = _regmap_bus_formatted_write;
1077 } else if (map->format.format_val) {
1078 map->defer_caching = true;
1079 map->reg_write = _regmap_bus_raw_write;
1084 map->range_tree = RB_ROOT;
1091 dev_err(map->dev, "Invalid range %d: %d < %d\n", i,
1096 if (range_cfg->range_max > map->max_register) {
1097 dev_err(map->dev, "Invalid range %d: %d > %d\n", i,
1098 range_cfg->range_max, map->max_register);
1102 if (range_cfg->selector_reg > map->max_register) {
1103 dev_err(map->dev,
1104 "Invalid range %d: selector out of map\n", i);
1109 dev_err(map->dev, "Invalid range %d: window_len 0\n",
1128 dev_err(map->dev,
1136 dev_err(map->dev,
1149 new->map = map;
1159 if (!_regmap_range_add(map, new)) {
1160 dev_err(map->dev, "Failed to add range %d\n", i);
1165 if (map->selector_work_buf == NULL) {
1166 map->selector_work_buf =
1167 kzalloc(map->format.buf_size, GFP_KERNEL);
1168 if (map->selector_work_buf == NULL) {
1175 ret = regcache_init(map, config);
1180 ret = regmap_attach_dev(dev, map, config);
1184 regmap_debugfs_init(map);
1187 return map;
1190 regcache_exit(map);
1192 regmap_range_exit(map);
1193 kfree(map->work_buf);
1195 if (map->hwlock)
1196 hwspin_lock_free(map->hwlock);
1198 kfree_const(map->name);
1200 kfree(map);
1431 * @map: Register map to operate on.
1434 * Discard any existing register cache for the map and initialize a
1442 int regmap_reinit_cache(struct regmap *map, const struct regmap_config *config)
1446 regcache_exit(map);
1447 regmap_debugfs_exit(map);
1449 map->max_register = config->max_register;
1450 map->writeable_reg = config->writeable_reg;
1451 map->readable_reg = config->readable_reg;
1452 map->volatile_reg = config->volatile_reg;
1453 map->precious_reg = config->precious_reg;
1454 map->writeable_noinc_reg = config->writeable_noinc_reg;
1455 map->readable_noinc_reg = config->readable_noinc_reg;
1456 map->cache_type = config->cache_type;
1458 ret = regmap_set_name(map, config);
1462 regmap_debugfs_init(map);
1464 map->cache_bypass = false;
1465 map->cache_only = false;
1467 return regcache_init(map, config);
1472 * regmap_exit() - Free a previously allocated register map
1474 * @map: Register map to operate on.
1476 void regmap_exit(struct regmap *map)
1480 regcache_exit(map);
1481 regmap_debugfs_exit(map);
1482 regmap_range_exit(map);
1483 if (map->bus && map->bus->free_context)
1484 map->bus->free_context(map->bus_context);
1485 kfree(map->work_buf);
1486 while (!list_empty(&map->async_free)) {
1487 async = list_first_entry_or_null(&map->async_free,
1494 if (map->hwlock)
1495 hwspin_lock_free(map->hwlock);
1496 if (map->lock == regmap_lock_mutex)
1497 mutex_destroy(&map->mutex);
1498 kfree_const(map->name);
1499 kfree(map->patch);
1500 kfree(map);
1522 * @dev: Device to retrieve the map for
1523 * @name: Optional name for the register map, usually NULL.
1545 * @map: Register map to operate on.
1549 struct device *regmap_get_device(struct regmap *map)
1551 return map->dev;
1555 static int _regmap_select_page(struct regmap *map, unsigned int *reg,
1584 orig_work_buf = map->work_buf;
1585 map->work_buf = map->selector_work_buf;
1587 ret = _regmap_update_bits(map, range->selector_reg,
1592 map->work_buf = orig_work_buf;
1603 static void regmap_set_work_buf_flag_mask(struct regmap *map, int max_bytes,
1609 if (!mask || !map->work_buf)
1612 buf = map->work_buf;
1618 static int _regmap_raw_write_impl(struct regmap *map, unsigned int reg,
1623 void *work_val = map->work_buf + map->format.reg_bytes +
1624 map->format.pad_bytes;
1630 WARN_ON(!map->bus);
1635 if (!regmap_writeable_noinc(map, reg)) {
1636 for (i = 0; i < val_len / map->format.val_bytes; i++) {
1638 reg + regmap_get_offset(map, i);
1639 if (!regmap_writeable(map, element) ||
1640 regmap_writeable_noinc(map, element))
1645 if (!map->cache_bypass && map->format.parse_val) {
1647 int val_bytes = map->format.val_bytes;
1652 ival = map->format.parse_val(val + i);
1653 offset = noinc ? 0 : regmap_get_offset(map, i / val_bytes);
1654 ret = regcache_write(map, reg + offset, ival);
1656 dev_err(map->dev,
1662 if (map->cache_only) {
1663 map->cache_dirty = true;
1668 range = _regmap_range_lookup(map, reg);
1670 int val_num = val_len / map->format.val_bytes;
1676 dev_dbg(map->dev, "Writing window %d/%zu\n",
1677 win_residue, val_len / map->format.val_bytes);
1678 ret = _regmap_raw_write_impl(map, reg, val,
1680 map->format.val_bytes, noinc);
1686 val += win_residue * map->format.val_bytes;
1687 val_len -= win_residue * map->format.val_bytes;
1694 ret = _regmap_select_page(map, ®, range, noinc ? 1 : val_num);
1699 map->format.format_reg(map->work_buf, reg, map->reg_shift);
1700 regmap_set_work_buf_flag_mask(map, map->format.reg_bytes,
1701 map->write_flag_mask);
1708 if (val != work_val && val_len == map->format.val_bytes) {
1709 memcpy(work_val, val, map->format.val_bytes);
1713 if (map->async && map->bus->async_write) {
1716 trace_regmap_async_write_start(map, reg, val_len);
1718 spin_lock_irqsave(&map->async_lock, flags);
1719 async = list_first_entry_or_null(&map->async_free,
1724 spin_unlock_irqrestore(&map->async_lock, flags);
1727 async = map->bus->async_alloc();
1731 async->work_buf = kzalloc(map->format.buf_size,
1739 async->map = map;
1742 memcpy(async->work_buf, map->work_buf, map->format.pad_bytes +
1743 map->format.reg_bytes + map->format.val_bytes);
1745 spin_lock_irqsave(&map->async_lock, flags);
1746 list_add_tail(&async->list, &map->async_list);
1747 spin_unlock_irqrestore(&map->async_lock, flags);
1750 ret = map->bus->async_write(map->bus_context,
1752 map->format.reg_bytes +
1753 map->format.pad_bytes,
1756 ret = map->bus->async_write(map->bus_context,
1758 map->format.reg_bytes +
1759 map->format.pad_bytes +
1763 dev_err(map->dev, "Failed to schedule write: %d\n",
1766 spin_lock_irqsave(&map->async_lock, flags);
1767 list_move(&async->list, &map->async_free);
1768 spin_unlock_irqrestore(&map->async_lock, flags);
1774 trace_regmap_hw_write_start(map, reg, val_len / map->format.val_bytes);
1781 ret = map->bus->write(map->bus_context, map->work_buf,
1782 map->format.reg_bytes +
1783 map->format.pad_bytes +
1785 else if (map->bus->gather_write)
1786 ret = map->bus->gather_write(map->bus_context, map->work_buf,
1787 map->format.reg_bytes +
1788 map->format.pad_bytes,
1795 len = map->format.reg_bytes + map->format.pad_bytes + val_len;
1800 memcpy(buf, map->work_buf, map->format.reg_bytes);
1801 memcpy(buf + map->format.reg_bytes + map->format.pad_bytes,
1803 ret = map->bus->write(map->bus_context, buf, len);
1806 } else if (ret != 0 && !map->cache_bypass && map->format.parse_val) {
1808 * thus call map->cache_ops->drop() directly
1810 if (map->cache_ops && map->cache_ops->drop)
1811 map->cache_ops->drop(map, reg, reg + 1);
1814 trace_regmap_hw_write_done(map, reg, val_len / map->format.val_bytes);
1822 * @map: Map to check.
1824 bool regmap_can_raw_write(struct regmap *map)
1826 return map->bus && map->bus->write && map->format.format_val &&
1827 map->format.format_reg;
1834 * @map: Map to check.
1836 size_t regmap_get_raw_read_max(struct regmap *map)
1838 return map->max_raw_read;
1845 * @map: Map to check.
1847 size_t regmap_get_raw_write_max(struct regmap *map)
1849 return map->max_raw_write;
1858 struct regmap *map = context;
1860 WARN_ON(!map->bus || !map->format.format_write);
1862 range = _regmap_range_lookup(map, reg);
1864 ret = _regmap_select_page(map, ®, range, 1);
1869 map->format.format_write(map, reg, val);
1871 trace_regmap_hw_write_start(map, reg, 1);
1873 ret = map->bus->write(map->bus_context, map->work_buf,
1874 map->format.buf_size);
1876 trace_regmap_hw_write_done(map, reg, 1);
1884 struct regmap *map = context;
1886 return map->bus->reg_write(map->bus_context, reg, val);
1892 struct regmap *map = context;
1894 WARN_ON(!map->bus || !map->format.format_val);
1896 map->format.format_val(map->work_buf + map->format.reg_bytes
1897 + map->format.pad_bytes, val, 0);
1898 return _regmap_raw_write_impl(map, reg,
1899 map->work_buf +
1900 map->format.reg_bytes +
1901 map->format.pad_bytes,
1902 map->format.val_bytes,
1906 static inline void *_regmap_map_get_context(struct regmap *map)
1908 return (map->bus) ? map : map->bus_context;
1911 int _regmap_write(struct regmap *map, unsigned int reg,
1915 void *context = _regmap_map_get_context(map);
1917 if (!regmap_writeable(map, reg))
1920 if (!map->cache_bypass && !map->defer_caching) {
1921 ret = regcache_write(map, reg, val);
1924 if (map->cache_only) {
1925 map->cache_dirty = true;
1930 if (regmap_should_log(map))
1931 dev_info(map->dev, "%x <= %x\n", reg, val);
1933 trace_regmap_reg_write(map, reg, val);
1935 return map->reg_write(context, reg, val);
1941 * @map: Register map to write to
1948 int regmap_write(struct regmap *map, unsigned int reg, unsigned int val)
1952 if (!IS_ALIGNED(reg, map->reg_stride))
1955 map->lock(map->lock_arg);
1957 ret = _regmap_write(map, reg, val);
1959 map->unlock(map->lock_arg);
1968 * @map: Register map to write to
1975 int regmap_write_async(struct regmap *map, unsigned int reg, unsigned int val)
1979 if (!IS_ALIGNED(reg, map->reg_stride))
1982 map->lock(map->lock_arg);
1984 map->async = true;
1986 ret = _regmap_write(map, reg, val);
1988 map->async = false;
1990 map->unlock(map->lock_arg);
1996 int _regmap_raw_write(struct regmap *map, unsigned int reg,
1999 size_t val_bytes = map->format.val_bytes;
2008 if (map->use_single_write)
2010 else if (map->max_raw_write && val_len > map->max_raw_write)
2011 chunk_regs = map->max_raw_write / val_bytes;
2018 ret = _regmap_raw_write_impl(map, reg, val, chunk_bytes, noinc);
2022 reg += regmap_get_offset(map, chunk_regs);
2029 ret = _regmap_raw_write_impl(map, reg, val, val_len, noinc);
2037 * @map: Register map to write to
2050 int regmap_raw_write(struct regmap *map, unsigned int reg,
2055 if (!regmap_can_raw_write(map))
2057 if (val_len % map->format.val_bytes)
2060 map->lock(map->lock_arg);
2062 ret = _regmap_raw_write(map, reg, val, val_len, false);
2064 map->unlock(map->lock_arg);
2074 * @map: Register map to write to
2091 int regmap_noinc_write(struct regmap *map, unsigned int reg,
2097 if (!map->bus)
2099 if (!map->bus->write)
2101 if (val_len % map->format.val_bytes)
2103 if (!IS_ALIGNED(reg, map->reg_stride))
2108 map->lock(map->lock_arg);
2110 if (!regmap_volatile(map, reg) || !regmap_writeable_noinc(map, reg)) {
2116 if (map->max_raw_write && map->max_raw_write < val_len)
2117 write_len = map->max_raw_write;
2120 ret = _regmap_raw_write(map, reg, val, write_len, true);
2128 map->unlock(map->lock_arg);
2196 * @map: Register map to write to
2207 int regmap_bulk_write(struct regmap *map, unsigned int reg, const void *val,
2211 size_t val_bytes = map->format.val_bytes;
2213 if (!IS_ALIGNED(reg, map->reg_stride))
2220 if (!map->bus || !map->format.parse_inplace) {
2221 map->lock(map->lock_arg);
2245 ret = _regmap_write(map,
2246 reg + regmap_get_offset(map, i),
2252 map->unlock(map->lock_arg);
2256 wval = kmemdup(val, val_count * val_bytes, map->alloc_flags);
2261 map->format.parse_inplace(wval + i);
2263 ret = regmap_raw_write(map, reg, wval, val_bytes * val_count);
2278 static int _regmap_raw_multi_reg_write(struct regmap *map,
2286 size_t val_bytes = map->format.val_bytes;
2287 size_t reg_bytes = map->format.reg_bytes;
2288 size_t pad_bytes = map->format.pad_bytes;
2306 trace_regmap_hw_write_start(map, reg, 1);
2307 map->format.format_reg(u8, reg, map->reg_shift);
2309 map->format.format_val(u8, val, 0);
2313 *u8 |= map->write_flag_mask;
2315 ret = map->bus->write(map->bus_context, buf, len);
2321 trace_regmap_hw_write_done(map, reg, 1);
2326 static unsigned int _regmap_register_page(struct regmap *map,
2335 static int _regmap_range_multi_paged_reg_write(struct regmap *map,
2355 range = _regmap_range_lookup(map, reg);
2357 unsigned int win_page = _regmap_register_page(map, reg,
2384 ret = _regmap_raw_multi_reg_write(map, base, n);
2389 if (map->can_sleep)
2399 ret = _regmap_select_page(map,
2412 return _regmap_raw_multi_reg_write(map, base, n);
2416 static int _regmap_multi_reg_write(struct regmap *map,
2423 if (!map->can_multi_write) {
2425 ret = _regmap_write(map, regs[i].reg, regs[i].def);
2430 if (map->can_sleep)
2439 if (!map->format.parse_inplace)
2442 if (map->writeable_reg)
2445 if (!map->writeable_reg(map->dev, reg))
2447 if (!IS_ALIGNED(reg, map->reg_stride))
2451 if (!map->cache_bypass) {
2455 ret = regcache_write(map, reg, val);
2457 dev_err(map->dev,
2463 if (map->cache_only) {
2464 map->cache_dirty = true;
2469 WARN_ON(!map->bus);
2478 range = _regmap_range_lookup(map, reg);
2485 ret = _regmap_range_multi_paged_reg_write(map, base,
2492 return _regmap_raw_multi_reg_write(map, regs, num_regs);
2498 * @map: Register map to write to
2514 int regmap_multi_reg_write(struct regmap *map, const struct reg_sequence *regs,
2519 map->lock(map->lock_arg);
2521 ret = _regmap_multi_reg_write(map, regs, num_regs);
2523 map->unlock(map->lock_arg);
2533 * @map: Register map to write to
2547 int regmap_multi_reg_write_bypassed(struct regmap *map,
2554 map->lock(map->lock_arg);
2556 bypass = map->cache_bypass;
2557 map->cache_bypass = true;
2559 ret = _regmap_multi_reg_write(map, regs, num_regs);
2561 map->cache_bypass = bypass;
2563 map->unlock(map->lock_arg);
2573 * @map: Register map to write to
2591 int regmap_raw_write_async(struct regmap *map, unsigned int reg,
2596 if (val_len % map->format.val_bytes)
2598 if (!IS_ALIGNED(reg, map->reg_stride))
2601 map->lock(map->lock_arg);
2603 map->async = true;
2605 ret = _regmap_raw_write(map, reg, val, val_len, false);
2607 map->async = false;
2609 map->unlock(map->lock_arg);
2615 static int _regmap_raw_read(struct regmap *map, unsigned int reg, void *val,
2621 WARN_ON(!map->bus);
2623 if (!map->bus || !map->bus->read)
2626 range = _regmap_range_lookup(map, reg);
2628 ret = _regmap_select_page(map, ®, range,
2629 noinc ? 1 : val_len / map->format.val_bytes);
2634 map->format.format_reg(map->work_buf, reg, map->reg_shift);
2635 regmap_set_work_buf_flag_mask(map, map->format.reg_bytes,
2636 map->read_flag_mask);
2637 trace_regmap_hw_read_start(map, reg, val_len / map->format.val_bytes);
2639 ret = map->bus->read(map->bus_context, map->work_buf,
2640 map->format.reg_bytes + map->format.pad_bytes,
2643 trace_regmap_hw_read_done(map, reg, val_len / map->format.val_bytes);
2651 struct regmap *map = context;
2653 return map->bus->reg_read(map->bus_context, reg, val);
2660 struct regmap *map = context;
2661 void *work_val = map->work_buf + map->format.reg_bytes +
2662 map->format.pad_bytes;
2664 if (!map->format.parse_val)
2667 ret = _regmap_raw_read(map, reg, work_val, map->format.val_bytes, false);
2669 *val = map->format.parse_val(work_val);
2674 static int _regmap_read(struct regmap *map, unsigned int reg,
2678 void *context = _regmap_map_get_context(map);
2680 if (!map->cache_bypass) {
2681 ret = regcache_read(map, reg, val);
2686 if (map->cache_only)
2689 if (!regmap_readable(map, reg))
2692 ret = map->reg_read(context, reg, val);
2694 if (regmap_should_log(map))
2695 dev_info(map->dev, "%x => %x\n", reg, *val);
2697 trace_regmap_reg_read(map, reg, *val);
2699 if (!map->cache_bypass)
2700 regcache_write(map, reg, *val);
2709 * @map: Register map to read from
2716 int regmap_read(struct regmap *map, unsigned int reg, unsigned int *val)
2720 if (!IS_ALIGNED(reg, map->reg_stride))
2723 map->lock(map->lock_arg);
2725 ret = _regmap_read(map, reg, val);
2727 map->unlock(map->lock_arg);
2736 * @map: Register map to read from
2744 int regmap_raw_read(struct regmap *map, unsigned int reg, void *val,
2747 size_t val_bytes = map->format.val_bytes;
2752 if (!map->bus)
2754 if (val_len % map->format.val_bytes)
2756 if (!IS_ALIGNED(reg, map->reg_stride))
2761 map->lock(map->lock_arg);
2763 if (regmap_volatile_range(map, reg, val_count) || map->cache_bypass ||
2764 map->cache_type == REGCACHE_NONE) {
2768 if (!map->bus->read) {
2773 if (map->use_single_read)
2775 else if (map->max_raw_read && val_len > map->max_raw_read)
2776 chunk_regs = map->max_raw_read / val_bytes;
2783 ret = _regmap_raw_read(map, reg, val, chunk_bytes, false);
2787 reg += regmap_get_offset(map, chunk_regs);
2794 ret = _regmap_raw_read(map, reg, val, val_len, false);
2803 ret = _regmap_read(map, reg + regmap_get_offset(map, i),
2808 map->format.format_val(val + (i * val_bytes), v, 0);
2813 map->unlock(map->lock_arg);
2823 * @map: Register map to read from
2840 int regmap_noinc_read(struct regmap *map, unsigned int reg,
2846 if (!map->bus)
2848 if (!map->bus->read)
2850 if (val_len % map->format.val_bytes)
2852 if (!IS_ALIGNED(reg, map->reg_stride))
2857 map->lock(map->lock_arg);
2859 if (!regmap_volatile(map, reg) || !regmap_readable_noinc(map, reg)) {
2865 if (map->max_raw_read && map->max_raw_read < val_len)
2866 read_len = map->max_raw_read;
2869 ret = _regmap_raw_read(map, reg, val, read_len, true);
2877 map->unlock(map->lock_arg);
2943 * @map: Register map to read from
2951 int regmap_bulk_read(struct regmap *map, unsigned int reg, void *val,
2955 size_t val_bytes = map->format.val_bytes;
2956 bool vol = regmap_volatile_range(map, reg, val_count);
2958 if (!IS_ALIGNED(reg, map->reg_stride))
2963 if (map->bus && map->format.parse_inplace && (vol || map->cache_type == REGCACHE_NONE)) {
2964 ret = regmap_raw_read(map, reg, val, val_bytes * val_count);
2969 map->format.parse_inplace(val + i);
2978 map->lock(map->lock_arg);
2983 ret = _regmap_read(map, reg + regmap_get_offset(map, i),
2988 switch (map->format.val_bytes) {
3010 map->unlock(map->lock_arg);
3017 static int _regmap_update_bits(struct regmap *map, unsigned int reg,
3027 if (regmap_volatile(map, reg) && map->reg_update_bits) {
3028 ret = map->reg_update_bits(map->bus_context, reg, mask, val);
3032 ret = _regmap_read(map, reg, &orig);
3040 ret = _regmap_write(map, reg, tmp);
3052 * @map: Register map to update
3060 * Perform a read/modify/write cycle on a register map with change, async, force
3071 int regmap_update_bits_base(struct regmap *map, unsigned int reg,
3077 map->lock(map->lock_arg);
3079 map->async = async;
3081 ret = _regmap_update_bits(map, reg, mask, val, change, force);
3083 map->async = false;
3085 map->unlock(map->lock_arg);
3094 * @map: Register map to operate on
3102 int regmap_test_bits(struct regmap *map, unsigned int reg, unsigned int bits)
3106 ret = regmap_read(map, reg, &val);
3116 struct regmap *map = async->map;
3119 trace_regmap_async_io_complete(map);
3121 spin_lock(&map->async_lock);
3122 list_move(&async->list, &map->async_free);
3123 wake = list_empty(&map->async_list);
3126 map->async_ret = ret;
3128 spin_unlock(&map->async_lock);
3131 wake_up(&map->async_waitq);
3135 static int regmap_async_is_done(struct regmap *map)
3140 spin_lock_irqsave(&map->async_lock, flags);
3141 ret = list_empty(&map->async_list);
3142 spin_unlock_irqrestore(&map->async_lock, flags);
3150 * @map: Map to operate on.
3155 int regmap_async_complete(struct regmap *map)
3161 if (!map->bus || !map->bus->async_write)
3164 trace_regmap_async_complete_start(map);
3166 wait_event(map->async_waitq, regmap_async_is_done(map));
3168 spin_lock_irqsave(&map->async_lock, flags);
3169 ret = map->async_ret;
3170 map->async_ret = 0;
3171 spin_unlock_irqrestore(&map->async_lock, flags);
3173 trace_regmap_async_complete_done(map);
3183 * @map: Register map to apply updates to.
3196 int regmap_register_patch(struct regmap *map, const struct reg_sequence *regs,
3207 p = krealloc(map->patch,
3208 sizeof(struct reg_sequence) * (map->patch_regs + num_regs),
3211 memcpy(p + map->patch_regs, regs, num_regs * sizeof(*regs));
3212 map->patch = p;
3213 map->patch_regs += num_regs;
3218 map->lock(map->lock_arg);
3220 bypass = map->cache_bypass;
3222 map->cache_bypass = true;
3223 map->async = true;
3225 ret = _regmap_multi_reg_write(map, regs, num_regs);
3227 map->async = false;
3228 map->cache_bypass = bypass;
3230 map->unlock(map->lock_arg);
3232 regmap_async_complete(map);
3241 * @map: Register map to operate on.
3246 int regmap_get_val_bytes(struct regmap *map)
3248 if (map->format.format_write)
3251 return map->format.val_bytes;
3258 * @map: Register map to operate on.
3263 int regmap_get_max_register(struct regmap *map)
3265 return map->max_register ? map->max_register : -EINVAL;
3272 * @map: Register map to operate on.
3277 int regmap_get_reg_stride(struct regmap *map)
3279 return map->reg_stride;
3283 int regmap_parse_val(struct regmap *map, const void *buf,
3286 if (!map->format.parse_val)
3289 *val = map->format.parse_val(buf);