[<prev] [next>] [day] [month] [year] [list]
Message-ID: <202203291432.SQmLkNIv-lkp@intel.com>
Date: Tue, 29 Mar 2022 14:11:47 +0800
From: kernel test robot <lkp@...el.com>
To: Ohad Sharabi <osharabi@...ana.ai>
Cc: kbuild-all@...ts.01.org, linux-kernel@...r.kernel.org,
Oded Gabbay <ogabbay@...nel.org>
Subject: drivers/misc/habanalabs/common/memory.c:153:7: warning: cast from
pointer to integer of different size
tree: https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
head: 1930a6e739c4b4a654a69164dbe39e554d228915
commit: e8458e20e0a3c426ed5ed3ce590c05718c8b8e8e habanalabs: make sure device mem alloc is page aligned
date: 4 weeks ago
config: i386-allyesconfig (https://download.01.org/0day-ci/archive/20220329/202203291432.SQmLkNIv-lkp@intel.com/config)
compiler: gcc-9 (Ubuntu 9.4.0-1ubuntu1~20.04.1) 9.4.0
reproduce (this is a W=1 build):
# https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/?id=e8458e20e0a3c426ed5ed3ce590c05718c8b8e8e
git remote add linus https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
git fetch --no-tags linus master
git checkout e8458e20e0a3c426ed5ed3ce590c05718c8b8e8e
# save the config file to linux build tree
mkdir build_dir
make W=1 O=build_dir ARCH=i386 SHELL=/bin/bash drivers/misc/
If you fix the issue, kindly add following tag as appropriate
Reported-by: kernel test robot <lkp@...el.com>
All warnings (new ones prefixed by >>):
drivers/misc/habanalabs/common/memory.c: In function 'alloc_device_memory':
>> drivers/misc/habanalabs/common/memory.c:153:7: warning: cast from pointer to integer of different size [-Wpointer-to-int-cast]
153 | (u64) gen_pool_dma_alloc_align(vm->dram_pg_pool,
| ^
vim +153 drivers/misc/habanalabs/common/memory.c
51
52 /*
53 * The va ranges in context object contain a list with the available chunks of
54 * device virtual memory.
55 * There is one range for host allocations and one for DRAM allocations.
56 *
57 * On initialization each range contains one chunk of all of its available
58 * virtual range which is a half of the total device virtual range.
59 *
60 * On each mapping of physical pages, a suitable virtual range chunk (with a
61 * minimum size) is selected from the list. If the chunk size equals the
62 * requested size, the chunk is returned. Otherwise, the chunk is split into
63 * two chunks - one to return as result and a remainder to stay in the list.
64 *
65 * On each Unmapping of a virtual address, the relevant virtual chunk is
66 * returned to the list. The chunk is added to the list and if its edges match
67 * the edges of the adjacent chunks (means a contiguous chunk can be created),
68 * the chunks are merged.
69 *
70 * On finish, the list is checked to have only one chunk of all the relevant
71 * virtual range (which is a half of the device total virtual range).
72 * If not (means not all mappings were unmapped), a warning is printed.
73 */
74
75 /*
76 * alloc_device_memory() - allocate device memory.
77 * @ctx: pointer to the context structure.
78 * @args: host parameters containing the requested size.
79 * @ret_handle: result handle.
80 *
81 * This function does the following:
82 * - Allocate the requested size rounded up to 'dram_page_size' pages.
83 * - Return unique handle for later map/unmap/free.
84 */
85 static int alloc_device_memory(struct hl_ctx *ctx, struct hl_mem_in *args,
86 u32 *ret_handle)
87 {
88 struct hl_device *hdev = ctx->hdev;
89 struct hl_vm *vm = &hdev->vm;
90 struct hl_vm_phys_pg_pack *phys_pg_pack;
91 u64 paddr = 0, total_size, num_pgs, i;
92 u32 num_curr_pgs, page_size;
93 bool contiguous;
94 int handle, rc;
95
96 num_curr_pgs = 0;
97
98 rc = set_alloc_page_size(hdev, args, &page_size);
99 if (rc)
100 return rc;
101
102 num_pgs = DIV_ROUND_UP_ULL(args->alloc.mem_size, page_size);
103 total_size = num_pgs * page_size;
104
105 if (!total_size) {
106 dev_err(hdev->dev, "Cannot allocate 0 bytes\n");
107 return -EINVAL;
108 }
109
110 contiguous = args->flags & HL_MEM_CONTIGUOUS;
111
112 if (contiguous) {
113 if (is_power_of_2(page_size))
114 paddr = (u64) (uintptr_t) gen_pool_dma_alloc_align(vm->dram_pg_pool,
115 total_size, NULL, page_size);
116 else
117 paddr = (u64) (uintptr_t) gen_pool_alloc(vm->dram_pg_pool, total_size);
118 if (!paddr) {
119 dev_err(hdev->dev,
120 "failed to allocate %llu contiguous pages with total size of %llu\n",
121 num_pgs, total_size);
122 return -ENOMEM;
123 }
124 }
125
126 phys_pg_pack = kzalloc(sizeof(*phys_pg_pack), GFP_KERNEL);
127 if (!phys_pg_pack) {
128 rc = -ENOMEM;
129 goto pages_pack_err;
130 }
131
132 phys_pg_pack->vm_type = VM_TYPE_PHYS_PACK;
133 phys_pg_pack->asid = ctx->asid;
134 phys_pg_pack->npages = num_pgs;
135 phys_pg_pack->page_size = page_size;
136 phys_pg_pack->total_size = total_size;
137 phys_pg_pack->flags = args->flags;
138 phys_pg_pack->contiguous = contiguous;
139
140 phys_pg_pack->pages = kvmalloc_array(num_pgs, sizeof(u64), GFP_KERNEL);
141 if (ZERO_OR_NULL_PTR(phys_pg_pack->pages)) {
142 rc = -ENOMEM;
143 goto pages_arr_err;
144 }
145
146 if (phys_pg_pack->contiguous) {
147 for (i = 0 ; i < num_pgs ; i++)
148 phys_pg_pack->pages[i] = paddr + i * page_size;
149 } else {
150 for (i = 0 ; i < num_pgs ; i++) {
151 if (is_power_of_2(page_size))
152 phys_pg_pack->pages[i] =
> 153 (u64) gen_pool_dma_alloc_align(vm->dram_pg_pool,
154 page_size, NULL,
155 page_size);
156 else
157 phys_pg_pack->pages[i] = (u64) gen_pool_alloc(vm->dram_pg_pool,
158 page_size);
159 if (!phys_pg_pack->pages[i]) {
160 dev_err(hdev->dev,
161 "Failed to allocate device memory (out of memory)\n");
162 rc = -ENOMEM;
163 goto page_err;
164 }
165
166 num_curr_pgs++;
167 }
168 }
169
170 spin_lock(&vm->idr_lock);
171 handle = idr_alloc(&vm->phys_pg_pack_handles, phys_pg_pack, 1, 0,
172 GFP_ATOMIC);
173 spin_unlock(&vm->idr_lock);
174
175 if (handle < 0) {
176 dev_err(hdev->dev, "Failed to get handle for page\n");
177 rc = -EFAULT;
178 goto idr_err;
179 }
180
181 for (i = 0 ; i < num_pgs ; i++)
182 kref_get(&vm->dram_pg_pool_refcount);
183
184 phys_pg_pack->handle = handle;
185
186 atomic64_add(phys_pg_pack->total_size, &ctx->dram_phys_mem);
187 atomic64_add(phys_pg_pack->total_size, &hdev->dram_used_mem);
188
189 *ret_handle = handle;
190
191 return 0;
192
193 idr_err:
194 page_err:
195 if (!phys_pg_pack->contiguous)
196 for (i = 0 ; i < num_curr_pgs ; i++)
197 gen_pool_free(vm->dram_pg_pool, phys_pg_pack->pages[i],
198 page_size);
199
200 kvfree(phys_pg_pack->pages);
201 pages_arr_err:
202 kfree(phys_pg_pack);
203 pages_pack_err:
204 if (contiguous)
205 gen_pool_free(vm->dram_pg_pool, paddr, total_size);
206
207 return rc;
208 }
209
--
0-DAY CI Kernel Test Service
https://01.org/lkp
Powered by blists - more mailing lists