lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <202205190911.n5iX1ftB-lkp@intel.com>
Date:   Thu, 19 May 2022 09:55:17 +0800
From:   kernel test robot <lkp@...el.com>
To:     Mel Gorman <mgorman@...e.de>
Cc:     kbuild-all@...ts.01.org, linux-kernel@...r.kernel.org,
        Peter Zijlstra <peterz@...radead.org>
Subject: [peterz-queue:sched/core 6/8] kernel/sched/fair.c:9199:30: error:
 implicit declaration of function 'adjust_numa_imbalance'

tree:   https://git.kernel.org/pub/scm/linux/kernel/git/peterz/queue.git sched/core
head:   45ff65aa1bfd4826331c9c4daafdca21ef8f79f8
commit: c81394419b54c2df2644a34892a6d6434fd922b3 [6/8] sched/numa: Apply imbalance limitations consistently
config: arc-allyesconfig (https://download.01.org/0day-ci/archive/20220519/202205190911.n5iX1ftB-lkp@intel.com/config)
compiler: arceb-elf-gcc (GCC) 11.3.0
reproduce (this is a W=1 build):
        wget https://raw.githubusercontent.com/intel/lkp-tests/master/sbin/make.cross -O ~/bin/make.cross
        chmod +x ~/bin/make.cross
        # https://git.kernel.org/pub/scm/linux/kernel/git/peterz/queue.git/commit/?id=c81394419b54c2df2644a34892a6d6434fd922b3
        git remote add peterz-queue https://git.kernel.org/pub/scm/linux/kernel/git/peterz/queue.git
        git fetch --no-tags peterz-queue sched/core
        git checkout c81394419b54c2df2644a34892a6d6434fd922b3
        # save the config file
        mkdir build_dir && cp config build_dir/.config
        COMPILER_INSTALL_PATH=$HOME/0day COMPILER=gcc-11.3.0 make.cross W=1 O=build_dir ARCH=arc SHELL=/bin/bash

If you fix the issue, kindly add following tag as appropriate
Reported-by: kernel test robot <lkp@...el.com>

All errors (new ones prefixed by >>):

   kernel/sched/fair.c: In function 'find_idlest_group':
>> kernel/sched/fair.c:9199:30: error: implicit declaration of function 'adjust_numa_imbalance' [-Werror=implicit-function-declaration]
    9199 |                         if (!adjust_numa_imbalance(imbalance,
         |                              ^~~~~~~~~~~~~~~~~~~~~
   cc1: some warnings being treated as errors


vim +/adjust_numa_imbalance +9199 kernel/sched/fair.c

  9060	
  9061	/*
  9062	 * find_idlest_group() finds and returns the least busy CPU group within the
  9063	 * domain.
  9064	 *
  9065	 * Assumes p is allowed on at least one CPU in sd.
  9066	 */
  9067	static struct sched_group *
  9068	find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu)
  9069	{
  9070		struct sched_group *idlest = NULL, *local = NULL, *group = sd->groups;
  9071		struct sg_lb_stats local_sgs, tmp_sgs;
  9072		struct sg_lb_stats *sgs;
  9073		unsigned long imbalance;
  9074		struct sg_lb_stats idlest_sgs = {
  9075				.avg_load = UINT_MAX,
  9076				.group_type = group_overloaded,
  9077		};
  9078	
  9079		do {
  9080			int local_group;
  9081	
  9082			/* Skip over this group if it has no CPUs allowed */
  9083			if (!cpumask_intersects(sched_group_span(group),
  9084						p->cpus_ptr))
  9085				continue;
  9086	
  9087			/* Skip over this group if no cookie matched */
  9088			if (!sched_group_cookie_match(cpu_rq(this_cpu), p, group))
  9089				continue;
  9090	
  9091			local_group = cpumask_test_cpu(this_cpu,
  9092						       sched_group_span(group));
  9093	
  9094			if (local_group) {
  9095				sgs = &local_sgs;
  9096				local = group;
  9097			} else {
  9098				sgs = &tmp_sgs;
  9099			}
  9100	
  9101			update_sg_wakeup_stats(sd, group, sgs, p);
  9102	
  9103			if (!local_group && update_pick_idlest(idlest, &idlest_sgs, group, sgs)) {
  9104				idlest = group;
  9105				idlest_sgs = *sgs;
  9106			}
  9107	
  9108		} while (group = group->next, group != sd->groups);
  9109	
  9110	
  9111		/* There is no idlest group to push tasks to */
  9112		if (!idlest)
  9113			return NULL;
  9114	
  9115		/* The local group has been skipped because of CPU affinity */
  9116		if (!local)
  9117			return idlest;
  9118	
  9119		/*
  9120		 * If the local group is idler than the selected idlest group
  9121		 * don't try and push the task.
  9122		 */
  9123		if (local_sgs.group_type < idlest_sgs.group_type)
  9124			return NULL;
  9125	
  9126		/*
  9127		 * If the local group is busier than the selected idlest group
  9128		 * try and push the task.
  9129		 */
  9130		if (local_sgs.group_type > idlest_sgs.group_type)
  9131			return idlest;
  9132	
  9133		switch (local_sgs.group_type) {
  9134		case group_overloaded:
  9135		case group_fully_busy:
  9136	
  9137			/* Calculate allowed imbalance based on load */
  9138			imbalance = scale_load_down(NICE_0_LOAD) *
  9139					(sd->imbalance_pct-100) / 100;
  9140	
  9141			/*
  9142			 * When comparing groups across NUMA domains, it's possible for
  9143			 * the local domain to be very lightly loaded relative to the
  9144			 * remote domains but "imbalance" skews the comparison making
  9145			 * remote CPUs look much more favourable. When considering
  9146			 * cross-domain, add imbalance to the load on the remote node
  9147			 * and consider staying local.
  9148			 */
  9149	
  9150			if ((sd->flags & SD_NUMA) &&
  9151			    ((idlest_sgs.avg_load + imbalance) >= local_sgs.avg_load))
  9152				return NULL;
  9153	
  9154			/*
  9155			 * If the local group is less loaded than the selected
  9156			 * idlest group don't try and push any tasks.
  9157			 */
  9158			if (idlest_sgs.avg_load >= (local_sgs.avg_load + imbalance))
  9159				return NULL;
  9160	
  9161			if (100 * local_sgs.avg_load <= sd->imbalance_pct * idlest_sgs.avg_load)
  9162				return NULL;
  9163			break;
  9164	
  9165		case group_imbalanced:
  9166		case group_asym_packing:
  9167			/* Those type are not used in the slow wakeup path */
  9168			return NULL;
  9169	
  9170		case group_misfit_task:
  9171			/* Select group with the highest max capacity */
  9172			if (local->sgc->max_capacity >= idlest->sgc->max_capacity)
  9173				return NULL;
  9174			break;
  9175	
  9176		case group_has_spare:
  9177			if (sd->flags & SD_NUMA) {
  9178	#ifdef CONFIG_NUMA_BALANCING
  9179				int idlest_cpu;
  9180				/*
  9181				 * If there is spare capacity at NUMA, try to select
  9182				 * the preferred node
  9183				 */
  9184				if (cpu_to_node(this_cpu) == p->numa_preferred_nid)
  9185					return NULL;
  9186	
  9187				idlest_cpu = cpumask_first(sched_group_span(idlest));
  9188				if (cpu_to_node(idlest_cpu) == p->numa_preferred_nid)
  9189					return idlest;
  9190	#endif
  9191				/*
  9192				 * Otherwise, keep the task close to the wakeup source
  9193				 * and improve locality if the number of running tasks
  9194				 * would remain below threshold where an imbalance is
  9195				 * allowed. If there is a real need of migration,
  9196				 * periodic load balance will take care of it.
  9197				 */
  9198				imbalance = abs(local_sgs.idle_cpus - idlest_sgs.idle_cpus);
> 9199				if (!adjust_numa_imbalance(imbalance,
  9200							   local_sgs.sum_nr_running + 1,
  9201							   sd->imb_numa_nr)) {
  9202					return NULL;
  9203				}
  9204			}
  9205	
  9206			/*
  9207			 * Select group with highest number of idle CPUs. We could also
  9208			 * compare the utilization which is more stable but it can end
  9209			 * up that the group has less spare capacity but finally more
  9210			 * idle CPUs which means more opportunity to run task.
  9211			 */
  9212			if (local_sgs.idle_cpus >= idlest_sgs.idle_cpus)
  9213				return NULL;
  9214			break;
  9215		}
  9216	
  9217		return idlest;
  9218	}
  9219	

-- 
0-DAY CI Kernel Test Service
https://01.org/lkp

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ