root/include/linux/livepatch.h

/* [<][>][^][v][top][bottom][index][help] */

INCLUDED FROM


DEFINITIONS

This source file includes following definitions.
  1. klp_patch_pending
  2. klp_have_reliable_stack
  3. klp_module_coming
  4. klp_module_going
  5. klp_patch_pending
  6. klp_update_patch_state
  7. klp_copy_process

   1 /* SPDX-License-Identifier: GPL-2.0-or-later */
   2 /*
   3  * livepatch.h - Kernel Live Patching Core
   4  *
   5  * Copyright (C) 2014 Seth Jennings <sjenning@redhat.com>
   6  * Copyright (C) 2014 SUSE
   7  */
   8 
   9 #ifndef _LINUX_LIVEPATCH_H_
  10 #define _LINUX_LIVEPATCH_H_
  11 
  12 #include <linux/module.h>
  13 #include <linux/ftrace.h>
  14 #include <linux/completion.h>
  15 #include <linux/list.h>
  16 
  17 #if IS_ENABLED(CONFIG_LIVEPATCH)
  18 
  19 #include <asm/livepatch.h>
  20 
  21 /* task patch states */
  22 #define KLP_UNDEFINED   -1
  23 #define KLP_UNPATCHED    0
  24 #define KLP_PATCHED      1
  25 
  26 /**
  27  * struct klp_func - function structure for live patching
  28  * @old_name:   name of the function to be patched
  29  * @new_func:   pointer to the patched function code
  30  * @old_sympos: a hint indicating which symbol position the old function
  31  *              can be found (optional)
  32  * @old_func:   pointer to the function being patched
  33  * @kobj:       kobject for sysfs resources
  34  * @node:       list node for klp_object func_list
  35  * @stack_node: list node for klp_ops func_stack list
  36  * @old_size:   size of the old function
  37  * @new_size:   size of the new function
  38  * @nop:        temporary patch to use the original code again; dyn. allocated
  39  * @patched:    the func has been added to the klp_ops list
  40  * @transition: the func is currently being applied or reverted
  41  *
  42  * The patched and transition variables define the func's patching state.  When
  43  * patching, a func is always in one of the following states:
  44  *
  45  *   patched=0 transition=0: unpatched
  46  *   patched=0 transition=1: unpatched, temporary starting state
  47  *   patched=1 transition=1: patched, may be visible to some tasks
  48  *   patched=1 transition=0: patched, visible to all tasks
  49  *
  50  * And when unpatching, it goes in the reverse order:
  51  *
  52  *   patched=1 transition=0: patched, visible to all tasks
  53  *   patched=1 transition=1: patched, may be visible to some tasks
  54  *   patched=0 transition=1: unpatched, temporary ending state
  55  *   patched=0 transition=0: unpatched
  56  */
  57 struct klp_func {
  58         /* external */
  59         const char *old_name;
  60         void *new_func;
  61         /*
  62          * The old_sympos field is optional and can be used to resolve
  63          * duplicate symbol names in livepatch objects. If this field is zero,
  64          * it is expected the symbol is unique, otherwise patching fails. If
  65          * this value is greater than zero then that occurrence of the symbol
  66          * in kallsyms for the given object is used.
  67          */
  68         unsigned long old_sympos;
  69 
  70         /* internal */
  71         void *old_func;
  72         struct kobject kobj;
  73         struct list_head node;
  74         struct list_head stack_node;
  75         unsigned long old_size, new_size;
  76         bool nop;
  77         bool patched;
  78         bool transition;
  79 };
  80 
  81 struct klp_object;
  82 
  83 /**
  84  * struct klp_callbacks - pre/post live-(un)patch callback structure
  85  * @pre_patch:          executed before code patching
  86  * @post_patch:         executed after code patching
  87  * @pre_unpatch:        executed before code unpatching
  88  * @post_unpatch:       executed after code unpatching
  89  * @post_unpatch_enabled:       flag indicating if post-unpatch callback
  90  *                              should run
  91  *
  92  * All callbacks are optional.  Only the pre-patch callback, if provided,
  93  * will be unconditionally executed.  If the parent klp_object fails to
  94  * patch for any reason, including a non-zero error status returned from
  95  * the pre-patch callback, no further callbacks will be executed.
  96  */
  97 struct klp_callbacks {
  98         int (*pre_patch)(struct klp_object *obj);
  99         void (*post_patch)(struct klp_object *obj);
 100         void (*pre_unpatch)(struct klp_object *obj);
 101         void (*post_unpatch)(struct klp_object *obj);
 102         bool post_unpatch_enabled;
 103 };
 104 
 105 /**
 106  * struct klp_object - kernel object structure for live patching
 107  * @name:       module name (or NULL for vmlinux)
 108  * @funcs:      function entries for functions to be patched in the object
 109  * @callbacks:  functions to be executed pre/post (un)patching
 110  * @kobj:       kobject for sysfs resources
 111  * @func_list:  dynamic list of the function entries
 112  * @node:       list node for klp_patch obj_list
 113  * @mod:        kernel module associated with the patched object
 114  *              (NULL for vmlinux)
 115  * @dynamic:    temporary object for nop functions; dynamically allocated
 116  * @patched:    the object's funcs have been added to the klp_ops list
 117  */
 118 struct klp_object {
 119         /* external */
 120         const char *name;
 121         struct klp_func *funcs;
 122         struct klp_callbacks callbacks;
 123 
 124         /* internal */
 125         struct kobject kobj;
 126         struct list_head func_list;
 127         struct list_head node;
 128         struct module *mod;
 129         bool dynamic;
 130         bool patched;
 131 };
 132 
 133 /**
 134  * struct klp_patch - patch structure for live patching
 135  * @mod:        reference to the live patch module
 136  * @objs:       object entries for kernel objects to be patched
 137  * @replace:    replace all actively used patches
 138  * @list:       list node for global list of actively used patches
 139  * @kobj:       kobject for sysfs resources
 140  * @obj_list:   dynamic list of the object entries
 141  * @enabled:    the patch is enabled (but operation may be incomplete)
 142  * @forced:     was involved in a forced transition
 143  * @free_work:  patch cleanup from workqueue-context
 144  * @finish:     for waiting till it is safe to remove the patch module
 145  */
 146 struct klp_patch {
 147         /* external */
 148         struct module *mod;
 149         struct klp_object *objs;
 150         bool replace;
 151 
 152         /* internal */
 153         struct list_head list;
 154         struct kobject kobj;
 155         struct list_head obj_list;
 156         bool enabled;
 157         bool forced;
 158         struct work_struct free_work;
 159         struct completion finish;
 160 };
 161 
 162 #define klp_for_each_object_static(patch, obj) \
 163         for (obj = patch->objs; obj->funcs || obj->name; obj++)
 164 
 165 #define klp_for_each_object_safe(patch, obj, tmp_obj)           \
 166         list_for_each_entry_safe(obj, tmp_obj, &patch->obj_list, node)
 167 
 168 #define klp_for_each_object(patch, obj) \
 169         list_for_each_entry(obj, &patch->obj_list, node)
 170 
 171 #define klp_for_each_func_static(obj, func) \
 172         for (func = obj->funcs; \
 173              func->old_name || func->new_func || func->old_sympos; \
 174              func++)
 175 
 176 #define klp_for_each_func_safe(obj, func, tmp_func)                     \
 177         list_for_each_entry_safe(func, tmp_func, &obj->func_list, node)
 178 
 179 #define klp_for_each_func(obj, func)    \
 180         list_for_each_entry(func, &obj->func_list, node)
 181 
 182 int klp_enable_patch(struct klp_patch *);
 183 
 184 void arch_klp_init_object_loaded(struct klp_patch *patch,
 185                                  struct klp_object *obj);
 186 
 187 /* Called from the module loader during module coming/going states */
 188 int klp_module_coming(struct module *mod);
 189 void klp_module_going(struct module *mod);
 190 
 191 void klp_copy_process(struct task_struct *child);
 192 void klp_update_patch_state(struct task_struct *task);
 193 
 194 static inline bool klp_patch_pending(struct task_struct *task)
 195 {
 196         return test_tsk_thread_flag(task, TIF_PATCH_PENDING);
 197 }
 198 
 199 static inline bool klp_have_reliable_stack(void)
 200 {
 201         return IS_ENABLED(CONFIG_STACKTRACE) &&
 202                IS_ENABLED(CONFIG_HAVE_RELIABLE_STACKTRACE);
 203 }
 204 
 205 typedef int (*klp_shadow_ctor_t)(void *obj,
 206                                  void *shadow_data,
 207                                  void *ctor_data);
 208 typedef void (*klp_shadow_dtor_t)(void *obj, void *shadow_data);
 209 
 210 void *klp_shadow_get(void *obj, unsigned long id);
 211 void *klp_shadow_alloc(void *obj, unsigned long id,
 212                        size_t size, gfp_t gfp_flags,
 213                        klp_shadow_ctor_t ctor, void *ctor_data);
 214 void *klp_shadow_get_or_alloc(void *obj, unsigned long id,
 215                               size_t size, gfp_t gfp_flags,
 216                               klp_shadow_ctor_t ctor, void *ctor_data);
 217 void klp_shadow_free(void *obj, unsigned long id, klp_shadow_dtor_t dtor);
 218 void klp_shadow_free_all(unsigned long id, klp_shadow_dtor_t dtor);
 219 
 220 #else /* !CONFIG_LIVEPATCH */
 221 
 222 static inline int klp_module_coming(struct module *mod) { return 0; }
 223 static inline void klp_module_going(struct module *mod) {}
 224 static inline bool klp_patch_pending(struct task_struct *task) { return false; }
 225 static inline void klp_update_patch_state(struct task_struct *task) {}
 226 static inline void klp_copy_process(struct task_struct *child) {}
 227 
 228 #endif /* CONFIG_LIVEPATCH */
 229 
 230 #endif /* _LINUX_LIVEPATCH_H_ */

/* [<][>][^][v][top][bottom][index][help] */