lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Tue, 28 May 2019 10:22:30 -0700
From:   Wengang <wen.gang.wang@...cle.com>
To:     Gang He <ghe@...e.com>, mark@...heh.com, jlbec@...lplan.org,
        jiangqi903@...il.com
Cc:     linux-kernel@...r.kernel.org, ocfs2-devel@....oracle.com
Subject: Re: [Ocfs2-devel] [PATCH 1/2] ocfs2: add last unlock times in
 locking_state

Hi Gang,

This idea sounds cool!
Some comments in lines:

On 05/23/2019 03:40 AM, Gang He wrote:
> ocfs2 file system uses locking_state file under debugfs to dump
> each ocfs2 file system's dlm lock resources, but the dlm lock
> resources in memory are becoming more and more after the files
> were touched by the user. it will become a bit difficult to analyze
> these dlm lock resource records in locking_state file by the upper
> scripts, though some files are not active for now, which were
> accessed long time ago.
> Then, I'd like to add last pr/ex unlock times in locking_state file
> for each dlm lock resource record, the the upper scripts can use
> last unlock time to filter inactive dlm lock resource record.
>
> Signed-off-by: Gang He <ghe@...e.com>
> Reviewed-by: Joseph Qi <joseph.qi@...ux.alibaba.com>
> ---
>   fs/ocfs2/dlmglue.c | 21 +++++++++++++++++----
>   fs/ocfs2/ocfs2.h   |  1 +
>   2 files changed, 18 insertions(+), 4 deletions(-)
>
> diff --git a/fs/ocfs2/dlmglue.c b/fs/ocfs2/dlmglue.c
> index af405586c5b1..dccf4136f8c1 100644
> --- a/fs/ocfs2/dlmglue.c
> +++ b/fs/ocfs2/dlmglue.c
> @@ -448,7 +448,7 @@ static void ocfs2_update_lock_stats(struct ocfs2_lock_res *res, int level,
>   				    struct ocfs2_mask_waiter *mw, int ret)
>   {
>   	u32 usec;
> -	ktime_t kt;
> +	ktime_t last, kt;
>   	struct ocfs2_lock_stats *stats;
>   
>   	if (level == LKM_PRMODE)
> @@ -458,7 +458,8 @@ static void ocfs2_update_lock_stats(struct ocfs2_lock_res *res, int level,
>   	else
>   		return;
>   
> -	kt = ktime_sub(ktime_get(), mw->mw_lock_start);
> +	last = ktime_get();
Will ktime_get_real() be better than ktime_get() here?
Per description,
ktime_get:
Useful for reliable timestamps and measuring short time intervals 
accurately. Starts at system boot time but stops during suspend.
ktime_get_real:
Returns the time in relative to the UNIX epoch starting in 1970 using 
the Coordinated Universal Time (UTC), same as gettimeofday() user space.

Since ktime_get() returnis time since boot time, this value is 
meaningless when compared to those from a different node in cluster, right?

And we need a "__kernel_long_t" to rather than a "u32"?


> +	kt = ktime_sub(last, mw->mw_lock_start);
>   	usec = ktime_to_us(kt);
>   
>   	stats->ls_gets++;
> @@ -474,6 +475,8 @@ static void ocfs2_update_lock_stats(struct ocfs2_lock_res *res, int level,
>   
>   	if (ret)
>   		stats->ls_fail++;
> +
> +	stats->ls_last = ktime_to_timespec(last).tv_sec;
>   }
>   
Though maybe ocfs2_update_lock_stats() is designed to be called for each 
successful lock request,
seems current code calls it even when it returns with -EAGAIN which 
breaks the design.  That's not introduced by your change, well, it may 
lead to wrong stats...

thanks,
wengang

>   static inline void ocfs2_track_lock_refresh(struct ocfs2_lock_res *lockres)
> @@ -3093,8 +3096,10 @@ static void *ocfs2_dlm_seq_next(struct seq_file *m, void *v, loff_t *pos)
>    *	- Lock stats printed
>    * New in version 3
>    *	- Max time in lock stats is in usecs (instead of nsecs)
> + * New in version 4
> + *	- Add last pr/ex unlock times in secs
>    */
> -#define OCFS2_DLM_DEBUG_STR_VERSION 3
> +#define OCFS2_DLM_DEBUG_STR_VERSION 4
>   static int ocfs2_dlm_seq_show(struct seq_file *m, void *v)
>   {
>   	int i;
> @@ -3145,6 +3150,8 @@ static int ocfs2_dlm_seq_show(struct seq_file *m, void *v)
>   # define lock_max_prmode(_l)		((_l)->l_lock_prmode.ls_max)
>   # define lock_max_exmode(_l)		((_l)->l_lock_exmode.ls_max)
>   # define lock_refresh(_l)		((_l)->l_lock_refresh)
> +# define lock_last_prmode(_l)		((_l)->l_lock_prmode.ls_last)
> +# define lock_last_exmode(_l)		((_l)->l_lock_exmode.ls_last)
>   #else
>   # define lock_num_prmode(_l)		(0)
>   # define lock_num_exmode(_l)		(0)
> @@ -3155,6 +3162,8 @@ static int ocfs2_dlm_seq_show(struct seq_file *m, void *v)
>   # define lock_max_prmode(_l)		(0)
>   # define lock_max_exmode(_l)		(0)
>   # define lock_refresh(_l)		(0)
> +# define lock_last_prmode(_l)		(0)
> +# define lock_last_exmode(_l)		(0)
>   #endif
>   	/* The following seq_print was added in version 2 of this output */
>   	seq_printf(m, "%u\t"
> @@ -3165,6 +3174,8 @@ static int ocfs2_dlm_seq_show(struct seq_file *m, void *v)
>   		   "%llu\t"
>   		   "%u\t"
>   		   "%u\t"
> +		   "%u\t"
> +		   "%u\t"
>   		   "%u\t",
>   		   lock_num_prmode(lockres),
>   		   lock_num_exmode(lockres),
> @@ -3174,7 +3185,9 @@ static int ocfs2_dlm_seq_show(struct seq_file *m, void *v)
>   		   lock_total_exmode(lockres),
>   		   lock_max_prmode(lockres),
>   		   lock_max_exmode(lockres),
> -		   lock_refresh(lockres));
> +		   lock_refresh(lockres),
> +		   lock_last_prmode(lockres),
> +		   lock_last_exmode(lockres));
>   
>   	/* End the line */
>   	seq_printf(m, "\n");
> diff --git a/fs/ocfs2/ocfs2.h b/fs/ocfs2/ocfs2.h
> index 1f029fbe8b8d..8efa022684f4 100644
> --- a/fs/ocfs2/ocfs2.h
> +++ b/fs/ocfs2/ocfs2.h
> @@ -164,6 +164,7 @@ struct ocfs2_lock_stats {
>   
>   	/* Storing max wait in usecs saves 24 bytes per inode */
>   	u32		ls_max;		/* Max wait in USEC */
> +	u32		ls_last;	/* Last unlock time in SEC */
>   };
>   #endif
>   

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ