treewide: mark stuff as __ro_after_init

__read_mostly predates __ro_after_init. Many variables which are marked
__read_mostly should have been __ro_after_init from day 1.

Also, mark some stuff as "const" and "__init" while I'm at it.

[akpm@linux-foundation.org: revert sysctl_nr_open_min, sysctl_nr_open_max changes due to arm warning]
[akpm@linux-foundation.org: coding-style cleanups]
Link: https://lkml.kernel.org/r/4f6bb9c0-abba-4ee4-a7aa-89265e886817@p183
Signed-off-by: Alexey Dobriyan <adobriyan@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
Alexey Dobriyan 2023-10-11 19:55:00 +03:00 committed by Andrew Morton
parent 1b13a70305
commit 68279f9c9f
26 changed files with 64 additions and 63 deletions

View file

@ -292,7 +292,7 @@ EXPORT_SYMBOL(thaw_bdev);
*/ */
static __cacheline_aligned_in_smp DEFINE_MUTEX(bdev_lock); static __cacheline_aligned_in_smp DEFINE_MUTEX(bdev_lock);
static struct kmem_cache * bdev_cachep __read_mostly; static struct kmem_cache *bdev_cachep __ro_after_init;
static struct inode *bdev_alloc_inode(struct super_block *sb) static struct inode *bdev_alloc_inode(struct super_block *sb)
{ {
@ -361,13 +361,13 @@ static struct file_system_type bd_type = {
.kill_sb = kill_anon_super, .kill_sb = kill_anon_super,
}; };
struct super_block *blockdev_superblock __read_mostly; struct super_block *blockdev_superblock __ro_after_init;
EXPORT_SYMBOL_GPL(blockdev_superblock); EXPORT_SYMBOL_GPL(blockdev_superblock);
void __init bdev_cache_init(void) void __init bdev_cache_init(void)
{ {
int err; int err;
static struct vfsmount *bd_mnt; static struct vfsmount *bd_mnt __ro_after_init;
bdev_cachep = kmem_cache_create("bdev_cache", sizeof(struct bdev_inode), bdev_cachep = kmem_cache_create("bdev_cache", sizeof(struct bdev_inode),
0, (SLAB_HWCACHE_ALIGN|SLAB_RECLAIM_ACCOUNT| 0, (SLAB_HWCACHE_ALIGN|SLAB_RECLAIM_ACCOUNT|

View file

@ -24,8 +24,8 @@
#include <linux/uaccess.h> #include <linux/uaccess.h>
static struct vfsmount *anon_inode_mnt __read_mostly; static struct vfsmount *anon_inode_mnt __ro_after_init;
static struct inode *anon_inode_inode; static struct inode *anon_inode_inode __ro_after_init;
/* /*
* anon_inodefs_dname() is called from d_path(). * anon_inodefs_dname() is called from d_path().

View file

@ -2988,13 +2988,13 @@ EXPORT_SYMBOL(try_to_free_buffers);
/* /*
* Buffer-head allocation * Buffer-head allocation
*/ */
static struct kmem_cache *bh_cachep __read_mostly; static struct kmem_cache *bh_cachep __ro_after_init;
/* /*
* Once the number of bh's in the machine exceeds this level, we start * Once the number of bh's in the machine exceeds this level, we start
* stripping them in writeback. * stripping them in writeback.
*/ */
static unsigned long max_buffer_heads; static unsigned long max_buffer_heads __ro_after_init;
int buffer_heads_over_limit; int buffer_heads_over_limit;

View file

@ -25,7 +25,7 @@
#include "internal.h" #include "internal.h"
static struct kobj_map *cdev_map; static struct kobj_map *cdev_map __ro_after_init;
static DEFINE_MUTEX(chrdevs_lock); static DEFINE_MUTEX(chrdevs_lock);

View file

@ -78,7 +78,7 @@ __cacheline_aligned_in_smp DEFINE_SEQLOCK(rename_lock);
EXPORT_SYMBOL(rename_lock); EXPORT_SYMBOL(rename_lock);
static struct kmem_cache *dentry_cache __read_mostly; static struct kmem_cache *dentry_cache __ro_after_init;
const struct qstr empty_name = QSTR_INIT("", 0); const struct qstr empty_name = QSTR_INIT("", 0);
EXPORT_SYMBOL(empty_name); EXPORT_SYMBOL(empty_name);
@ -96,9 +96,9 @@ EXPORT_SYMBOL(dotdot_name);
* information, yet avoid using a prime hash-size or similar. * information, yet avoid using a prime hash-size or similar.
*/ */
static unsigned int d_hash_shift __read_mostly; static unsigned int d_hash_shift __ro_after_init;
static struct hlist_bl_head *dentry_hashtable __read_mostly; static struct hlist_bl_head *dentry_hashtable __ro_after_init;
static inline struct hlist_bl_head *d_hash(unsigned int hash) static inline struct hlist_bl_head *d_hash(unsigned int hash)
{ {
@ -3324,7 +3324,7 @@ static void __init dcache_init(void)
} }
/* SLAB cache for __getname() consumers */ /* SLAB cache for __getname() consumers */
struct kmem_cache *names_cachep __read_mostly; struct kmem_cache *names_cachep __ro_after_init;
EXPORT_SYMBOL(names_cachep); EXPORT_SYMBOL(names_cachep);
void __init vfs_caches_init_early(void) void __init vfs_caches_init_early(void)

View file

@ -151,7 +151,7 @@ struct dio {
}; };
} ____cacheline_aligned_in_smp; } ____cacheline_aligned_in_smp;
static struct kmem_cache *dio_cache __read_mostly; static struct kmem_cache *dio_cache __ro_after_init;
/* /*
* How many pages are in the queue? * How many pages are in the queue?

View file

@ -256,10 +256,10 @@ static u64 loop_check_gen = 0;
static struct eventpoll *inserting_into; static struct eventpoll *inserting_into;
/* Slab cache used to allocate "struct epitem" */ /* Slab cache used to allocate "struct epitem" */
static struct kmem_cache *epi_cache __read_mostly; static struct kmem_cache *epi_cache __ro_after_init;
/* Slab cache used to allocate "struct eppoll_entry" */ /* Slab cache used to allocate "struct eppoll_entry" */
static struct kmem_cache *pwq_cache __read_mostly; static struct kmem_cache *pwq_cache __ro_after_init;
/* /*
* List of files with newly added links, where we may need to limit the number * List of files with newly added links, where we may need to limit the number
@ -271,7 +271,7 @@ struct epitems_head {
}; };
static struct epitems_head *tfile_check_list = EP_UNACTIVE_PTR; static struct epitems_head *tfile_check_list = EP_UNACTIVE_PTR;
static struct kmem_cache *ephead_cache __read_mostly; static struct kmem_cache *ephead_cache __ro_after_init;
static inline void free_ephead(struct epitems_head *head) static inline void free_ephead(struct epitems_head *head)
{ {

View file

@ -844,7 +844,7 @@ int send_sigurg(struct fown_struct *fown)
} }
static DEFINE_SPINLOCK(fasync_lock); static DEFINE_SPINLOCK(fasync_lock);
static struct kmem_cache *fasync_cache __read_mostly; static struct kmem_cache *fasync_cache __ro_after_init;
static void fasync_free_rcu(struct rcu_head *head) static void fasync_free_rcu(struct rcu_head *head)
{ {

View file

@ -40,7 +40,7 @@ static struct files_stat_struct files_stat = {
}; };
/* SLAB cache for file structures */ /* SLAB cache for file structures */
static struct kmem_cache *filp_cachep __read_mostly; static struct kmem_cache *filp_cachep __ro_after_init;
static struct percpu_counter nr_files __cacheline_aligned_in_smp; static struct percpu_counter nr_files __cacheline_aligned_in_smp;

View file

@ -54,9 +54,9 @@
* inode_hash_lock * inode_hash_lock
*/ */
static unsigned int i_hash_mask __read_mostly; static unsigned int i_hash_mask __ro_after_init;
static unsigned int i_hash_shift __read_mostly; static unsigned int i_hash_shift __ro_after_init;
static struct hlist_head *inode_hashtable __read_mostly; static struct hlist_head *inode_hashtable __ro_after_init;
static __cacheline_aligned_in_smp DEFINE_SPINLOCK(inode_hash_lock); static __cacheline_aligned_in_smp DEFINE_SPINLOCK(inode_hash_lock);
/* /*
@ -70,7 +70,7 @@ EXPORT_SYMBOL(empty_aops);
static DEFINE_PER_CPU(unsigned long, nr_inodes); static DEFINE_PER_CPU(unsigned long, nr_inodes);
static DEFINE_PER_CPU(unsigned long, nr_unused); static DEFINE_PER_CPU(unsigned long, nr_unused);
static struct kmem_cache *inode_cachep __read_mostly; static struct kmem_cache *inode_cachep __ro_after_init;
static long get_nr_inodes(void) static long get_nr_inodes(void)
{ {

View file

@ -21,8 +21,9 @@
#include "kernfs-internal.h" #include "kernfs-internal.h"
struct kmem_cache *kernfs_node_cache, *kernfs_iattrs_cache; struct kmem_cache *kernfs_node_cache __ro_after_init;
struct kernfs_global_locks *kernfs_locks; struct kmem_cache *kernfs_iattrs_cache __ro_after_init;
struct kernfs_global_locks *kernfs_locks __ro_after_init;
static int kernfs_sop_show_options(struct seq_file *sf, struct dentry *dentry) static int kernfs_sop_show_options(struct seq_file *sf, struct dentry *dentry)
{ {

View file

@ -167,8 +167,8 @@ static DEFINE_HASHTABLE(blocked_hash, BLOCKED_HASH_BITS);
*/ */
static DEFINE_SPINLOCK(blocked_lock_lock); static DEFINE_SPINLOCK(blocked_lock_lock);
static struct kmem_cache *flctx_cache __read_mostly; static struct kmem_cache *flctx_cache __ro_after_init;
static struct kmem_cache *filelock_cache __read_mostly; static struct kmem_cache *filelock_cache __ro_after_init;
static struct file_lock_context * static struct file_lock_context *
locks_get_lock_context(struct inode *inode, int type) locks_get_lock_context(struct inode *inode, int type)

View file

@ -39,10 +39,10 @@
/* Maximum number of mounts in a mount namespace */ /* Maximum number of mounts in a mount namespace */
static unsigned int sysctl_mount_max __read_mostly = 100000; static unsigned int sysctl_mount_max __read_mostly = 100000;
static unsigned int m_hash_mask __read_mostly; static unsigned int m_hash_mask __ro_after_init;
static unsigned int m_hash_shift __read_mostly; static unsigned int m_hash_shift __ro_after_init;
static unsigned int mp_hash_mask __read_mostly; static unsigned int mp_hash_mask __ro_after_init;
static unsigned int mp_hash_shift __read_mostly; static unsigned int mp_hash_shift __ro_after_init;
static __initdata unsigned long mhash_entries; static __initdata unsigned long mhash_entries;
static int __init set_mhash_entries(char *str) static int __init set_mhash_entries(char *str)
@ -68,9 +68,9 @@ static u64 event;
static DEFINE_IDA(mnt_id_ida); static DEFINE_IDA(mnt_id_ida);
static DEFINE_IDA(mnt_group_ida); static DEFINE_IDA(mnt_group_ida);
static struct hlist_head *mount_hashtable __read_mostly; static struct hlist_head *mount_hashtable __ro_after_init;
static struct hlist_head *mountpoint_hashtable __read_mostly; static struct hlist_head *mountpoint_hashtable __ro_after_init;
static struct kmem_cache *mnt_cache __read_mostly; static struct kmem_cache *mnt_cache __ro_after_init;
static DECLARE_RWSEM(namespace_sem); static DECLARE_RWSEM(namespace_sem);
static HLIST_HEAD(unmounted); /* protected by namespace_sem */ static HLIST_HEAD(unmounted); /* protected by namespace_sem */
static LIST_HEAD(ex_mountpoints); /* protected by namespace_sem */ static LIST_HEAD(ex_mountpoints); /* protected by namespace_sem */
@ -86,7 +86,7 @@ struct mount_kattr {
}; };
/* /sys/fs */ /* /sys/fs */
struct kobject *fs_kobj; struct kobject *fs_kobj __ro_after_init;
EXPORT_SYMBOL_GPL(fs_kobj); EXPORT_SYMBOL_GPL(fs_kobj);
/* /*

View file

@ -39,9 +39,9 @@ static void __init dnotify_sysctl_init(void)
#define dnotify_sysctl_init() do { } while (0) #define dnotify_sysctl_init() do { } while (0)
#endif #endif
static struct kmem_cache *dnotify_struct_cache __read_mostly; static struct kmem_cache *dnotify_struct_cache __ro_after_init;
static struct kmem_cache *dnotify_mark_cache __read_mostly; static struct kmem_cache *dnotify_mark_cache __ro_after_init;
static struct fsnotify_group *dnotify_group __read_mostly; static struct fsnotify_group *dnotify_group __ro_after_init;
/* /*
* dnotify will attach one of these to each inode (i_fsnotify_marks) which * dnotify will attach one of these to each inode (i_fsnotify_marks) which

View file

@ -112,10 +112,10 @@ static void __init fanotify_sysctls_init(void)
extern const struct fsnotify_ops fanotify_fsnotify_ops; extern const struct fsnotify_ops fanotify_fsnotify_ops;
struct kmem_cache *fanotify_mark_cache __read_mostly; struct kmem_cache *fanotify_mark_cache __ro_after_init;
struct kmem_cache *fanotify_fid_event_cachep __read_mostly; struct kmem_cache *fanotify_fid_event_cachep __ro_after_init;
struct kmem_cache *fanotify_path_event_cachep __read_mostly; struct kmem_cache *fanotify_path_event_cachep __ro_after_init;
struct kmem_cache *fanotify_perm_event_cachep __read_mostly; struct kmem_cache *fanotify_perm_event_cachep __ro_after_init;
#define FANOTIFY_EVENT_ALIGN 4 #define FANOTIFY_EVENT_ALIGN 4
#define FANOTIFY_FID_INFO_HDR_LEN \ #define FANOTIFY_FID_INFO_HDR_LEN \

View file

@ -49,7 +49,7 @@
/* configurable via /proc/sys/fs/inotify/ */ /* configurable via /proc/sys/fs/inotify/ */
static int inotify_max_queued_events __read_mostly; static int inotify_max_queued_events __read_mostly;
struct kmem_cache *inotify_inode_mark_cachep __read_mostly; struct kmem_cache *inotify_inode_mark_cachep __ro_after_init;
#ifdef CONFIG_SYSCTL #ifdef CONFIG_SYSCTL

View file

@ -854,7 +854,7 @@ void free_pipe_info(struct pipe_inode_info *pipe)
kfree(pipe); kfree(pipe);
} }
static struct vfsmount *pipe_mnt __read_mostly; static struct vfsmount *pipe_mnt __ro_after_init;
/* /*
* pipefs_dname() is called from d_path(). * pipefs_dname() is called from d_path().

View file

@ -49,7 +49,7 @@ static struct ctl_table vm_userfaultfd_table[] = {
}; };
#endif #endif
static struct kmem_cache *userfaultfd_ctx_cachep __read_mostly; static struct kmem_cache *userfaultfd_ctx_cachep __ro_after_init;
/* /*
* Start with fault_pending_wqh and fault_wqh so they're more likely * Start with fault_pending_wqh and fault_wqh so they're more likely

View file

@ -87,8 +87,8 @@ static struct task_struct *prune_thread;
* that makes a difference. Some. * that makes a difference. Some.
*/ */
static struct fsnotify_group *audit_tree_group; static struct fsnotify_group *audit_tree_group __ro_after_init;
static struct kmem_cache *audit_tree_mark_cachep __read_mostly; static struct kmem_cache *audit_tree_mark_cachep __ro_after_init;
static struct audit_tree *alloc_tree(const char *s) static struct audit_tree *alloc_tree(const char *s)
{ {

View file

@ -9903,7 +9903,7 @@ struct task_group root_task_group;
LIST_HEAD(task_groups); LIST_HEAD(task_groups);
/* Cacheline aligned slab cache for task_group */ /* Cacheline aligned slab cache for task_group */
static struct kmem_cache *task_group_cache __read_mostly; static struct kmem_cache *task_group_cache __ro_after_init;
#endif #endif
void __init sched_init(void) void __init sched_init(void)

View file

@ -22,7 +22,7 @@
#include <linux/bsearch.h> #include <linux/bsearch.h>
#include <linux/sort.h> #include <linux/sort.h>
static struct kmem_cache *user_ns_cachep __read_mostly; static struct kmem_cache *user_ns_cachep __ro_after_init;
static DEFINE_MUTEX(userns_state_mutex); static DEFINE_MUTEX(userns_state_mutex);
static bool new_idmap_permitted(const struct file *file, static bool new_idmap_permitted(const struct file *file,

View file

@ -418,21 +418,21 @@ static struct workqueue_attrs *ordered_wq_attrs[NR_STD_WORKER_POOLS];
* process context while holding a pool lock. Bounce to a dedicated kthread * process context while holding a pool lock. Bounce to a dedicated kthread
* worker to avoid A-A deadlocks. * worker to avoid A-A deadlocks.
*/ */
static struct kthread_worker *pwq_release_worker; static struct kthread_worker *pwq_release_worker __ro_after_init;
struct workqueue_struct *system_wq __read_mostly; struct workqueue_struct *system_wq __ro_after_init;
EXPORT_SYMBOL(system_wq); EXPORT_SYMBOL(system_wq);
struct workqueue_struct *system_highpri_wq __read_mostly; struct workqueue_struct *system_highpri_wq __ro_after_init;
EXPORT_SYMBOL_GPL(system_highpri_wq); EXPORT_SYMBOL_GPL(system_highpri_wq);
struct workqueue_struct *system_long_wq __read_mostly; struct workqueue_struct *system_long_wq __ro_after_init;
EXPORT_SYMBOL_GPL(system_long_wq); EXPORT_SYMBOL_GPL(system_long_wq);
struct workqueue_struct *system_unbound_wq __read_mostly; struct workqueue_struct *system_unbound_wq __ro_after_init;
EXPORT_SYMBOL_GPL(system_unbound_wq); EXPORT_SYMBOL_GPL(system_unbound_wq);
struct workqueue_struct *system_freezable_wq __read_mostly; struct workqueue_struct *system_freezable_wq __ro_after_init;
EXPORT_SYMBOL_GPL(system_freezable_wq); EXPORT_SYMBOL_GPL(system_freezable_wq);
struct workqueue_struct *system_power_efficient_wq __read_mostly; struct workqueue_struct *system_power_efficient_wq __ro_after_init;
EXPORT_SYMBOL_GPL(system_power_efficient_wq); EXPORT_SYMBOL_GPL(system_power_efficient_wq);
struct workqueue_struct *system_freezable_power_efficient_wq __read_mostly; struct workqueue_struct *system_freezable_power_efficient_wq __ro_after_init;
EXPORT_SYMBOL_GPL(system_freezable_power_efficient_wq); EXPORT_SYMBOL_GPL(system_freezable_power_efficient_wq);
static int worker_thread(void *__worker); static int worker_thread(void *__worker);

View file

@ -89,7 +89,7 @@ static int debug_objects_pool_size __read_mostly
static int debug_objects_pool_min_level __read_mostly static int debug_objects_pool_min_level __read_mostly
= ODEBUG_POOL_MIN_LEVEL; = ODEBUG_POOL_MIN_LEVEL;
static const struct debug_obj_descr *descr_test __read_mostly; static const struct debug_obj_descr *descr_test __read_mostly;
static struct kmem_cache *obj_cache __read_mostly; static struct kmem_cache *obj_cache __ro_after_init;
/* /*
* Track numbers of kmem_cache_alloc()/free() calls done. * Track numbers of kmem_cache_alloc()/free() calls done.

View file

@ -91,7 +91,7 @@ static unsigned int khugepaged_max_ptes_shared __read_mostly;
#define MM_SLOTS_HASH_BITS 10 #define MM_SLOTS_HASH_BITS 10
static DEFINE_READ_MOSTLY_HASHTABLE(mm_slots_hash, MM_SLOTS_HASH_BITS); static DEFINE_READ_MOSTLY_HASHTABLE(mm_slots_hash, MM_SLOTS_HASH_BITS);
static struct kmem_cache *mm_slot_cache __read_mostly; static struct kmem_cache *mm_slot_cache __ro_after_init;
struct collapse_control { struct collapse_control {
bool is_khugepaged; bool is_khugepaged;

View file

@ -42,7 +42,7 @@
#include <linux/iversion.h> #include <linux/iversion.h>
#include "swap.h" #include "swap.h"
static struct vfsmount *shm_mnt; static struct vfsmount *shm_mnt __ro_after_init;
#ifdef CONFIG_SHMEM #ifdef CONFIG_SHMEM
/* /*
@ -4394,7 +4394,7 @@ static const struct fs_context_operations shmem_fs_context_ops = {
#endif #endif
}; };
static struct kmem_cache *shmem_inode_cachep; static struct kmem_cache *shmem_inode_cachep __ro_after_init;
static struct inode *shmem_alloc_inode(struct super_block *sb) static struct inode *shmem_alloc_inode(struct super_block *sb)
{ {
@ -4426,14 +4426,14 @@ static void shmem_init_inode(void *foo)
inode_init_once(&info->vfs_inode); inode_init_once(&info->vfs_inode);
} }
static void shmem_init_inodecache(void) static void __init shmem_init_inodecache(void)
{ {
shmem_inode_cachep = kmem_cache_create("shmem_inode_cache", shmem_inode_cachep = kmem_cache_create("shmem_inode_cache",
sizeof(struct shmem_inode_info), sizeof(struct shmem_inode_info),
0, SLAB_PANIC|SLAB_ACCOUNT, shmem_init_inode); 0, SLAB_PANIC|SLAB_ACCOUNT, shmem_init_inode);
} }
static void shmem_destroy_inodecache(void) static void __init shmem_destroy_inodecache(void)
{ {
kmem_cache_destroy(shmem_inode_cachep); kmem_cache_destroy(shmem_inode_cachep);
} }

View file

@ -23,7 +23,7 @@
static struct rb_root integrity_iint_tree = RB_ROOT; static struct rb_root integrity_iint_tree = RB_ROOT;
static DEFINE_RWLOCK(integrity_iint_lock); static DEFINE_RWLOCK(integrity_iint_lock);
static struct kmem_cache *iint_cache __read_mostly; static struct kmem_cache *iint_cache __ro_after_init;
struct dentry *integrity_dir; struct dentry *integrity_dir;