lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20221229183027.6f2c75ba0b2924a2bfd0499d@kernel.org>
Date:   Thu, 29 Dec 2022 18:30:27 +0900
From:   Masami Hiramatsu (Google) <mhiramat@...nel.org>
To:     Song Chen <chensong_2000@....cn>
Cc:     rostedt@...dmis.org, arnd@...db.de, linux-kernel@...r.kernel.org,
        linux-trace-kernel@...r.kernel.org, linux-arch@...r.kernel.org
Subject: Re: [PATCH v4 2/2] kernel/trace: Provide default impelentations
 defined in trace_probe_tmpl.h

On Wed, 28 Dec 2022 15:08:25 +0800
Song Chen <chensong_2000@....cn> wrote:

> There are 7 function definitions in trace_probe_tmpl.h, they are:
> 
> 1, process_fetch_insn

Other functions except for this are OK for me.
But I think this function should have different instance for kprobe
and eprobe, because the passed 'void *rec' is completely different.
For the kprobe's process_fetch_insn() will take a 'pt_regs *', on the
other hand, the eprobe's process_fetch_insn() will take a
'ftrace_event_field *'.
I also, don't want to expose get_event_field() function because
it is only for the eprobes.

What about making a new common function and call it in default case
from both process_fetch_insn()? Your goal is reducing the redundant
code, then it is enough to share only the common code. :)

For example, in trace_probe_kernel.h,

process_common_fetch_insn(struct fetch_insn *code, void *dest, void *base, unsigned long *val)
{
	switch (code->op) {
	case FETCH_OP_IMM:
		*val = code->immediate;
		break;
	case FETCH_OP_COMM:
		*val = (unsigned long)current->comm;
		break;
	case FETCH_OP_DATA:
		*val = (unsigned long)code->data;
		break;
	default:
		return -EILSEQ;
	}
	return 0;
}

For kprobe events:

process_fetch_insn(struct fetch_insn *code, void *rec, void *dest, void *base)
{
	while (code->op == FETCH_NOP_SYMBOL)
		code++;
	switch (code->op) {
	case FETCH_OP_REG:
		val = regs_get_register(regs, code->param);
		...
	default:
		ret = process_common_fetch_insn(code, dest, base, &val);
		if (ret < 0)
			return ret;
	}
	...
}

For eprobe events:

process_fetch_insn(struct fetch_insn *code, void *rec, void *dest, void *base)
{
	while (code->op == FETCH_NOP_SYMBOL)
		code++;
	if (code->op == FETCH_OP_TP_ARG) {
		val = get_event_field(code, rec);
	} else {
		ret = process_common_fetch_insn(code, dest, base, &val);
		if (ret < 0)
			return ret;
	}
	...
}

You also don't see the error related to CONFIG_HAVE_REGS_AND_STACK_ACCESS_API. :)

Thank you,

> 2, fetch_store_strlen
> 3, fetch_store_string
> 4, fetch_store_strlen_user
> 5, fetch_store_string_user
> 6, probe_mem_read
> 7, probe_mem_read_user
> 
> Every C file which includes trace_probe_tmpl.h has to implement them,
> otherwise it gets warnings and errors. However, some of them are identical,
> like kprobe and eprobe, as a result, there is a lot redundant code in those
> 2 files.
> 
> This patch would like to provide default behaviors for those functions
> which kprobe and eprobe can share by just including trace_probe_kernel.h
> with trace_probe_tmpl.h together.
> 
> It removes redundant code, increases readability, and more importantly,
> makes it easier to introduce a new feature based on trace probe
> (it's possible).
> 
> Signed-off-by: Song Chen <chensong_2000@....cn>
> Reported-by: kernel test robot <lkp@...el.com>
> 
> ---
> v2:
> 1, reorganize patchset
> 
> v3:
> 1, mark nokprobe_inline for get_event_field
> 2, remove warnings reported from kernel test robot
> 3, fix errors reported from kernel test robot
> 
> v4:
> 1, reset changes in v3(2) and v3(3), they are not reasonable fix.
> 2, I intended to introduce a new header file to exclude
> trace_events_synth.c out of the patch. However, after looking
> further, i found the errors in [1] are introduced by this patch,
> but the warning in [1] are not. The errors can be fixed by adding
> "#ifdef CONFIG_HAVE_REGS_AND_STACK_ACCESS_API".
> 
> [1]:https://lore.kernel.org/lkml/202211301946.pkLE4PDp-lkp@intel.com/
> ---
>  kernel/trace/trace_eprobe.c       | 144 ------------------------------
>  kernel/trace/trace_events_synth.c |   7 +-
>  kernel/trace/trace_kprobe.c       | 102 ---------------------
>  kernel/trace/trace_probe_kernel.h | 143 +++++++++++++++++++++++++++--
>  4 files changed, 141 insertions(+), 255 deletions(-)
> 
> diff --git a/kernel/trace/trace_eprobe.c b/kernel/trace/trace_eprobe.c
> index bdb26eee7a0c..60ced4a7a25d 100644
> --- a/kernel/trace/trace_eprobe.c
> +++ b/kernel/trace/trace_eprobe.c
> @@ -319,64 +319,6 @@ print_eprobe_event(struct trace_iterator *iter, int flags,
>  	return trace_handle_return(s);
>  }
>  
> -static unsigned long get_event_field(struct fetch_insn *code, void *rec)
> -{
> -	struct ftrace_event_field *field = code->data;
> -	unsigned long val;
> -	void *addr;
> -
> -	addr = rec + field->offset;
> -
> -	if (is_string_field(field)) {
> -		switch (field->filter_type) {
> -		case FILTER_DYN_STRING:
> -			val = (unsigned long)(rec + (*(unsigned int *)addr & 0xffff));
> -			break;
> -		case FILTER_RDYN_STRING:
> -			val = (unsigned long)(addr + (*(unsigned int *)addr & 0xffff));
> -			break;
> -		case FILTER_STATIC_STRING:
> -			val = (unsigned long)addr;
> -			break;
> -		case FILTER_PTR_STRING:
> -			val = (unsigned long)(*(char *)addr);
> -			break;
> -		default:
> -			WARN_ON_ONCE(1);
> -			return 0;
> -		}
> -		return val;
> -	}
> -
> -	switch (field->size) {
> -	case 1:
> -		if (field->is_signed)
> -			val = *(char *)addr;
> -		else
> -			val = *(unsigned char *)addr;
> -		break;
> -	case 2:
> -		if (field->is_signed)
> -			val = *(short *)addr;
> -		else
> -			val = *(unsigned short *)addr;
> -		break;
> -	case 4:
> -		if (field->is_signed)
> -			val = *(int *)addr;
> -		else
> -			val = *(unsigned int *)addr;
> -		break;
> -	default:
> -		if (field->is_signed)
> -			val = *(long *)addr;
> -		else
> -			val = *(unsigned long *)addr;
> -		break;
> -	}
> -	return val;
> -}
> -
>  static int get_eprobe_size(struct trace_probe *tp, void *rec)
>  {
>  	struct fetch_insn *code;
> @@ -419,92 +361,6 @@ static int get_eprobe_size(struct trace_probe *tp, void *rec)
>  	return ret;
>  }
>  
> -/* Kprobe specific fetch functions */
> -
> -/* Note that we don't verify it, since the code does not come from user space */
> -static int
> -process_fetch_insn(struct fetch_insn *code, void *rec, void *dest,
> -		   void *base)
> -{
> -	unsigned long val;
> -
> - retry:
> -	switch (code->op) {
> -	case FETCH_OP_TP_ARG:
> -		val = get_event_field(code, rec);
> -		break;
> -	case FETCH_OP_IMM:
> -		val = code->immediate;
> -		break;
> -	case FETCH_OP_COMM:
> -		val = (unsigned long)current->comm;
> -		break;
> -	case FETCH_OP_DATA:
> -		val = (unsigned long)code->data;
> -		break;
> -	case FETCH_NOP_SYMBOL:	/* Ignore a place holder */
> -		code++;
> -		goto retry;
> -	default:
> -		return -EILSEQ;
> -	}
> -	code++;
> -	return process_fetch_insn_bottom(code, val, dest, base);
> -}
> -NOKPROBE_SYMBOL(process_fetch_insn)
> -
> -/* Return the length of string -- including null terminal byte */
> -static nokprobe_inline int
> -fetch_store_strlen_user(unsigned long addr)
> -{
> -	return kern_fetch_store_strlen_user(addr);
> -}
> -
> -/* Return the length of string -- including null terminal byte */
> -static nokprobe_inline int
> -fetch_store_strlen(unsigned long addr)
> -{
> -	return kern_fetch_store_strlen(addr);
> -}
> -
> -/*
> - * Fetch a null-terminated string from user. Caller MUST set *(u32 *)buf
> - * with max length and relative data location.
> - */
> -static nokprobe_inline int
> -fetch_store_string_user(unsigned long addr, void *dest, void *base)
> -{
> -	return kern_fetch_store_string_user(addr, dest, base);
> -}
> -
> -/*
> - * Fetch a null-terminated string. Caller MUST set *(u32 *)buf with max
> - * length and relative data location.
> - */
> -static nokprobe_inline int
> -fetch_store_string(unsigned long addr, void *dest, void *base)
> -{
> -	return kern_fetch_store_string(addr, dest, base);
> -}
> -
> -static nokprobe_inline int
> -probe_mem_read_user(void *dest, void *src, size_t size)
> -{
> -	const void __user *uaddr =  (__force const void __user *)src;
> -
> -	return copy_from_user_nofault(dest, uaddr, size);
> -}
> -
> -static nokprobe_inline int
> -probe_mem_read(void *dest, void *src, size_t size)
> -{
> -#ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE
> -	if ((unsigned long)src < TASK_SIZE)
> -		return probe_mem_read_user(dest, src, size);
> -#endif
> -	return copy_from_kernel_nofault(dest, src, size);
> -}
> -
>  /* eprobe handler */
>  static inline void
>  __eprobe_trace_func(struct eprobe_data *edata, void *rec)
> diff --git a/kernel/trace/trace_events_synth.c b/kernel/trace/trace_events_synth.c
> index e310052dc83c..7460f18ba973 100644
> --- a/kernel/trace/trace_events_synth.c
> +++ b/kernel/trace/trace_events_synth.c
> @@ -18,6 +18,7 @@
>  #include <linux/trace_events.h>
>  #include <trace/events/mmflags.h>
>  #include "trace_probe.h"
> +#include "trace_probe_tmpl.h"
>  #include "trace_probe_kernel.h"
>  
>  #include "trace_synth.h"
> @@ -420,12 +421,12 @@ static unsigned int trace_string(struct synth_trace_event *entry,
>  		data_offset += event->n_u64 * sizeof(u64);
>  		data_offset += data_size;
>  
> -		len = kern_fetch_store_strlen((unsigned long)str_val);
> +		len = fetch_store_strlen((unsigned long)str_val);
>  
>  		data_offset |= len << 16;
>  		*(u32 *)&entry->fields[*n_u64] = data_offset;
>  
> -		ret = kern_fetch_store_string((unsigned long)str_val, &entry->fields[*n_u64], entry);
> +		ret = fetch_store_string((unsigned long)str_val, &entry->fields[*n_u64], entry);
>  
>  		(*n_u64)++;
>  	} else {
> @@ -473,7 +474,7 @@ static notrace void trace_event_raw_event_synth(void *__data,
>  		val_idx = var_ref_idx[field_pos];
>  		str_val = (char *)(long)var_ref_vals[val_idx];
>  
> -		len = kern_fetch_store_strlen((unsigned long)str_val);
> +		len = fetch_store_strlen((unsigned long)str_val);
>  
>  		fields_size += len;
>  	}
> diff --git a/kernel/trace/trace_kprobe.c b/kernel/trace/trace_kprobe.c
> index a4ffa864dbb7..c2e0b741ae82 100644
> --- a/kernel/trace/trace_kprobe.c
> +++ b/kernel/trace/trace_kprobe.c
> @@ -1218,108 +1218,6 @@ static const struct file_operations kprobe_profile_ops = {
>  	.release        = seq_release,
>  };
>  
> -/* Kprobe specific fetch functions */
> -
> -/* Return the length of string -- including null terminal byte */
> -static nokprobe_inline int
> -fetch_store_strlen_user(unsigned long addr)
> -{
> -	return kern_fetch_store_strlen_user(addr);
> -}
> -
> -/* Return the length of string -- including null terminal byte */
> -static nokprobe_inline int
> -fetch_store_strlen(unsigned long addr)
> -{
> -	return kern_fetch_store_strlen(addr);
> -}
> -
> -/*
> - * Fetch a null-terminated string from user. Caller MUST set *(u32 *)buf
> - * with max length and relative data location.
> - */
> -static nokprobe_inline int
> -fetch_store_string_user(unsigned long addr, void *dest, void *base)
> -{
> -	return kern_fetch_store_string_user(addr, dest, base);
> -}
> -
> -/*
> - * Fetch a null-terminated string. Caller MUST set *(u32 *)buf with max
> - * length and relative data location.
> - */
> -static nokprobe_inline int
> -fetch_store_string(unsigned long addr, void *dest, void *base)
> -{
> -	return kern_fetch_store_string(addr, dest, base);
> -}
> -
> -static nokprobe_inline int
> -probe_mem_read_user(void *dest, void *src, size_t size)
> -{
> -	const void __user *uaddr =  (__force const void __user *)src;
> -
> -	return copy_from_user_nofault(dest, uaddr, size);
> -}
> -
> -static nokprobe_inline int
> -probe_mem_read(void *dest, void *src, size_t size)
> -{
> -#ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE
> -	if ((unsigned long)src < TASK_SIZE)
> -		return probe_mem_read_user(dest, src, size);
> -#endif
> -	return copy_from_kernel_nofault(dest, src, size);
> -}
> -
> -/* Note that we don't verify it, since the code does not come from user space */
> -static int
> -process_fetch_insn(struct fetch_insn *code, void *rec, void *dest,
> -		   void *base)
> -{
> -	struct pt_regs *regs = rec;
> -	unsigned long val;
> -
> -retry:
> -	/* 1st stage: get value from context */
> -	switch (code->op) {
> -	case FETCH_OP_REG:
> -		val = regs_get_register(regs, code->param);
> -		break;
> -	case FETCH_OP_STACK:
> -		val = regs_get_kernel_stack_nth(regs, code->param);
> -		break;
> -	case FETCH_OP_STACKP:
> -		val = kernel_stack_pointer(regs);
> -		break;
> -	case FETCH_OP_RETVAL:
> -		val = regs_return_value(regs);
> -		break;
> -	case FETCH_OP_IMM:
> -		val = code->immediate;
> -		break;
> -	case FETCH_OP_COMM:
> -		val = (unsigned long)current->comm;
> -		break;
> -	case FETCH_OP_DATA:
> -		val = (unsigned long)code->data;
> -		break;
> -#ifdef CONFIG_HAVE_FUNCTION_ARG_ACCESS_API
> -	case FETCH_OP_ARG:
> -		val = regs_get_kernel_argument(regs, code->param);
> -		break;
> -#endif
> -	case FETCH_NOP_SYMBOL:	/* Ignore a place holder */
> -		code++;
> -		goto retry;
> -	default:
> -		return -EILSEQ;
> -	}
> -	code++;
> -
> -	return process_fetch_insn_bottom(code, val, dest, base);
> -}
> -NOKPROBE_SYMBOL(process_fetch_insn)
>  
>  /* Kprobe handler */
>  static nokprobe_inline void
> diff --git a/kernel/trace/trace_probe_kernel.h b/kernel/trace/trace_probe_kernel.h
> index 77dbd9ff9782..39f44513ec4e 100644
> --- a/kernel/trace/trace_probe_kernel.h
> +++ b/kernel/trace/trace_probe_kernel.h
> @@ -12,7 +12,7 @@
>   */
>  /* Return the length of string -- including null terminal byte */
>  static nokprobe_inline int
> -kern_fetch_store_strlen_user(unsigned long addr)
> +fetch_store_strlen_user(unsigned long addr)
>  {
>  	const void __user *uaddr =  (__force const void __user *)addr;
>  	int ret;
> @@ -29,14 +29,14 @@ kern_fetch_store_strlen_user(unsigned long addr)
>  
>  /* Return the length of string -- including null terminal byte */
>  static nokprobe_inline int
> -kern_fetch_store_strlen(unsigned long addr)
> +fetch_store_strlen(unsigned long addr)
>  {
>  	int ret, len = 0;
>  	u8 c;
>  
>  #ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE
>  	if (addr < TASK_SIZE)
> -		return kern_fetch_store_strlen_user(addr);
> +		return fetch_store_strlen_user(addr);
>  #endif
>  
>  	do {
> @@ -63,7 +63,7 @@ static nokprobe_inline void set_data_loc(int ret, void *dest, void *__dest, void
>   * with max length and relative data location.
>   */
>  static nokprobe_inline int
> -kern_fetch_store_string_user(unsigned long addr, void *dest, void *base)
> +fetch_store_string_user(unsigned long addr, void *dest, void *base)
>  {
>  	const void __user *uaddr =  (__force const void __user *)addr;
>  	int maxlen = get_loc_len(*(u32 *)dest);
> @@ -86,7 +86,7 @@ kern_fetch_store_string_user(unsigned long addr, void *dest, void *base)
>   * length and relative data location.
>   */
>  static nokprobe_inline int
> -kern_fetch_store_string(unsigned long addr, void *dest, void *base)
> +fetch_store_string(unsigned long addr, void *dest, void *base)
>  {
>  	int maxlen = get_loc_len(*(u32 *)dest);
>  	void *__dest;
> @@ -94,7 +94,7 @@ kern_fetch_store_string(unsigned long addr, void *dest, void *base)
>  
>  #ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE
>  	if ((unsigned long)addr < TASK_SIZE)
> -		return kern_fetch_store_string_user(addr, dest, base);
> +		return fetch_store_string_user(addr, dest, base);
>  #endif
>  
>  	if (unlikely(!maxlen))
> @@ -112,4 +112,135 @@ kern_fetch_store_string(unsigned long addr, void *dest, void *base)
>  	return ret;
>  }
>  
> +static nokprobe_inline int
> +probe_mem_read_user(void *dest, void *src, size_t size)
> +{
> +	const void __user *uaddr =  (__force const void __user *)src;
> +
> +	return copy_from_user_nofault(dest, uaddr, size);
> +}
> +
> +static nokprobe_inline int
> +probe_mem_read(void *dest, void *src, size_t size)
> +{
> +#ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE
> +	if ((unsigned long)src < TASK_SIZE)
> +		return probe_mem_read_user(dest, src, size);
> +#endif
> +	return copy_from_kernel_nofault(dest, src, size);
> +}
> +
> +static nokprobe_inline unsigned long
> +get_event_field(struct fetch_insn *code, void *rec)
> +{
> +	struct ftrace_event_field *field = code->data;
> +	unsigned long val;
> +	void *addr;
> +
> +	addr = rec + field->offset;
> +
> +	if (is_string_field(field)) {
> +		switch (field->filter_type) {
> +		case FILTER_DYN_STRING:
> +			val = (unsigned long)(rec + (*(unsigned int *)addr & 0xffff));
> +			break;
> +		case FILTER_RDYN_STRING:
> +			val = (unsigned long)(addr + (*(unsigned int *)addr & 0xffff));
> +			break;
> +		case FILTER_STATIC_STRING:
> +			val = (unsigned long)addr;
> +			break;
> +		case FILTER_PTR_STRING:
> +			val = (unsigned long)(*(char *)addr);
> +			break;
> +		default:
> +			WARN_ON_ONCE(1);
> +			return 0;
> +		}
> +		return val;
> +	}
> +
> +	switch (field->size) {
> +	case 1:
> +		if (field->is_signed)
> +			val = *(char *)addr;
> +		else
> +			val = *(unsigned char *)addr;
> +		break;
> +	case 2:
> +		if (field->is_signed)
> +			val = *(short *)addr;
> +		else
> +			val = *(unsigned short *)addr;
> +		break;
> +	case 4:
> +		if (field->is_signed)
> +			val = *(int *)addr;
> +		else
> +			val = *(unsigned int *)addr;
> +		break;
> +	default:
> +		if (field->is_signed)
> +			val = *(long *)addr;
> +		else
> +			val = *(unsigned long *)addr;
> +		break;
> +	}
> +	return val;
> +}
> +
> +/* Note that we don't verify it, since the code does not come from user space */
> +static int
> +process_fetch_insn(struct fetch_insn *code, void *rec, void *dest,
> +		   void *base)
> +{
> +	struct pt_regs *regs = rec;
> +	unsigned long val;
> +
> +retry:
> +	/* 1st stage: get value from context */
> +	switch (code->op) {
> +#ifdef CONFIG_HAVE_REGS_AND_STACK_ACCESS_API
> +	case FETCH_OP_REG:
> +		val = regs_get_register(regs, code->param);
> +		break;
> +	case FETCH_OP_STACK:
> +		val = regs_get_kernel_stack_nth(regs, code->param);
> +		break;
> +	case FETCH_OP_STACKP:
> +		val = kernel_stack_pointer(regs);
> +		break;
> +	case FETCH_OP_RETVAL:
> +		val = regs_return_value(regs);
> +		break;
> +#endif
> +	case FETCH_OP_IMM:
> +		val = code->immediate;
> +		break;
> +	case FETCH_OP_COMM:
> +		val = (unsigned long)current->comm;
> +		break;
> +	case FETCH_OP_DATA:
> +		val = (unsigned long)code->data;
> +		break;
> +#ifdef CONFIG_HAVE_FUNCTION_ARG_ACCESS_API
> +	case FETCH_OP_ARG:
> +		val = regs_get_kernel_argument(regs, code->param);
> +		break;
> +#endif
> +	case FETCH_NOP_SYMBOL:	/* Ignore a place holder */
> +		code++;
> +		goto retry;
> +	case FETCH_OP_TP_ARG:
> +		val = get_event_field(code, rec);
> +		break;
> +	default:
> +		return -EILSEQ;
> +	}
> +	code++;
> +
> +	return process_fetch_insn_bottom(code, val, dest, base);
> +}
> +NOKPROBE_SYMBOL(process_fetch_insn)
> +
>  #endif /* __TRACE_PROBE_KERNEL_H_ */
> -- 
> 2.25.1
> 


-- 
Masami Hiramatsu (Google) <mhiramat@...nel.org>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ