mirror of
git://nv-tegra.nvidia.com/linux-nvgpu.git
synced 2025-12-22 17:36:20 +03:00
- Earlier we whitelisted wrongly reported static analysis violations by tool, raised coverity tool bugs for these cases. - These bugs are fixed with new version of tool, so no need fo whitelisting. JIRA NVGPU-7119 Change-Id: Ib2341db0d46fa7fac4c0cc9a6c1bdc8704377ef1 Signed-off-by: srajum <srajum@nvidia.com> Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2604365 (cherry picked from commit dc2d8ddaa409aefe0e04e0bacb3a8a977f6dbd64) Reviewed-on: https://git-master.nvidia.com/r/c/linux-nvgpu/+/2677523 Reviewed-by: svc-mobile-coverity <svc-mobile-coverity@nvidia.com> Reviewed-by: svc-mobile-misra <svc-mobile-misra@nvidia.com> Reviewed-by: svc-mobile-cert <svc-mobile-cert@nvidia.com> Reviewed-by: Rajesh Devaraj <rdevaraj@nvidia.com> Reviewed-by: Vaibhav Kachore <vkachore@nvidia.com> GVS: Gerrit_Virtual_Submit
256 lines
6.6 KiB
C
256 lines
6.6 KiB
C
/*
|
|
* Copyright (c) 2018-2022, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
|
* DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <stdatomic.h>
|
|
|
|
#include <nvgpu/posix/utils.h>
|
|
#include <nvgpu/posix/bitops.h>
|
|
#include <nvgpu/posix/atomic.h>
|
|
|
|
static inline unsigned long get_mask(unsigned int bit)
|
|
{
|
|
unsigned long lbit = bit;
|
|
lbit %= BITS_PER_LONG;
|
|
return (1UL << lbit);
|
|
}
|
|
|
|
static inline unsigned int get_index(unsigned int bit)
|
|
{
|
|
unsigned long bpl = BITS_PER_LONG;
|
|
return (bit / nvgpu_safe_cast_u64_to_u32(bpl));
|
|
}
|
|
|
|
unsigned long nvgpu_posix_ffs(unsigned long word)
|
|
{
|
|
int ret = 0;
|
|
const int maxvalue = 64;
|
|
|
|
if ((word & (unsigned long) LONG_MAX) != 0UL) {
|
|
ret = __builtin_ffsl(
|
|
nvgpu_safe_cast_u64_to_s64(
|
|
(word & (unsigned long) LONG_MAX)));
|
|
} else {
|
|
if (word > (unsigned long) LONG_MAX) {
|
|
ret = maxvalue;
|
|
}
|
|
}
|
|
|
|
return nvgpu_safe_cast_s32_to_u64(ret);
|
|
}
|
|
|
|
unsigned long nvgpu_posix_fls(unsigned long word)
|
|
{
|
|
unsigned long ret;
|
|
|
|
if (word == 0UL) {
|
|
/* __builtin_clzl() below is undefined for 0, so we have
|
|
* to handle that as a special case.
|
|
*/
|
|
ret = 0UL;
|
|
} else {
|
|
ret = (sizeof(unsigned long) * 8UL) -
|
|
(nvgpu_safe_cast_s32_to_u64(__builtin_clzl(word)));
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static unsigned long nvgpu_posix_find_next_bit(const unsigned long *address,
|
|
unsigned long n,
|
|
unsigned long start,
|
|
bool invert)
|
|
{
|
|
unsigned long idx, idx_max;
|
|
unsigned long w;
|
|
unsigned long start_mask;
|
|
const unsigned long *base_addr = (const unsigned long *)&address[0];
|
|
|
|
/*
|
|
* We make a mask we can XOR into the word so that we can invert the
|
|
* word without requiring a branch. I.e instead of doing:
|
|
*
|
|
* w = invert ? ~addr[idx] : addr[idx]
|
|
*
|
|
* We can do:
|
|
*
|
|
* w = addr[idx] ^= invert_mask
|
|
*
|
|
* This saves us a branch every iteration through the loop. Now we can
|
|
* always just look for 1s.
|
|
*/
|
|
unsigned long invert_mask = invert ? ~0UL : 0UL;
|
|
|
|
if (start >= n) {
|
|
return n;
|
|
}
|
|
|
|
start_mask = ~0UL << (start & (BITS_PER_LONG - 1UL));
|
|
|
|
idx = start / BITS_PER_LONG;
|
|
w = (base_addr[idx] ^ invert_mask) & start_mask;
|
|
|
|
idx_max = (n - 1UL) / BITS_PER_LONG;
|
|
|
|
/*
|
|
* Find the first non-zero word taking into account start and
|
|
* invert.
|
|
*/
|
|
while (w == 0UL) {
|
|
idx = nvgpu_safe_add_u64(idx, 1UL);
|
|
if (idx > idx_max) {
|
|
return n;
|
|
}
|
|
|
|
w = base_addr[idx] ^ invert_mask;
|
|
}
|
|
|
|
return min(n, (nvgpu_safe_add_u64(((nvgpu_ffs(w)) - 1UL),
|
|
(nvgpu_safe_mult_u64(idx, BITS_PER_LONG)))));
|
|
}
|
|
|
|
unsigned long find_first_bit(const unsigned long *address, unsigned long size)
|
|
{
|
|
return nvgpu_posix_find_next_bit(address, size, 0, false);
|
|
}
|
|
|
|
unsigned long find_first_zero_bit(const unsigned long *address, unsigned long size)
|
|
{
|
|
return nvgpu_posix_find_next_bit(address, size, 0, true);
|
|
}
|
|
|
|
unsigned long find_next_bit(const unsigned long *address, unsigned long size,
|
|
unsigned long offset)
|
|
{
|
|
return nvgpu_posix_find_next_bit(address, size, offset, false);
|
|
}
|
|
|
|
static unsigned long find_next_zero_bit(const unsigned long *address,
|
|
unsigned long size,
|
|
unsigned long offset)
|
|
{
|
|
return nvgpu_posix_find_next_bit(address, size, offset, true);
|
|
}
|
|
|
|
void nvgpu_bitmap_set(unsigned long *map, unsigned int start, unsigned int len)
|
|
{
|
|
unsigned int end = start + len;
|
|
|
|
/*
|
|
* Super slow naive implementation. But speed isn't what matters here.
|
|
*/
|
|
while (start < end) {
|
|
nvgpu_set_bit(start++, map);
|
|
}
|
|
}
|
|
|
|
void nvgpu_bitmap_clear(unsigned long *map,
|
|
unsigned int start, unsigned int len)
|
|
{
|
|
unsigned int end = start + len;
|
|
|
|
while (start < end) {
|
|
nvgpu_clear_bit(start++, map);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* This is essentially a find-first-fit allocator: this searches a bitmap for
|
|
* the first space that is large enough to satisfy the requested size of bits.
|
|
* That means that this is not a vary smart allocator. But it is fast relative
|
|
* to an allocator that goes looking for an optimal location.
|
|
*/
|
|
unsigned long bitmap_find_next_zero_area(unsigned long *map,
|
|
unsigned long size,
|
|
unsigned long start,
|
|
unsigned int bit,
|
|
unsigned long align_mask)
|
|
{
|
|
unsigned long offs;
|
|
|
|
while ((nvgpu_safe_add_u64(start, (unsigned long)bit)) <= size) {
|
|
start = find_next_zero_bit(map, size, start);
|
|
|
|
start = ALIGN_MASK(start, align_mask);
|
|
|
|
/*
|
|
* Not enough space left to satisfy the requested area.
|
|
*/
|
|
if ((nvgpu_safe_add_u64(start, (unsigned long)bit)) > size) {
|
|
return size;
|
|
}
|
|
|
|
offs = find_next_bit(map, size, start);
|
|
|
|
if ((offs - start) >= bit) {
|
|
return start;
|
|
}
|
|
|
|
start = offs + 1UL;
|
|
}
|
|
|
|
return size;
|
|
}
|
|
|
|
bool nvgpu_test_bit(unsigned int bit, const volatile unsigned long *address)
|
|
{
|
|
return (1UL & (address[get_index(bit)] >>
|
|
(bit & (BITS_PER_LONG-1UL)))) != 0UL;
|
|
}
|
|
|
|
bool nvgpu_test_and_set_bit(unsigned int bit, volatile unsigned long *address)
|
|
{
|
|
unsigned long mask = get_mask(bit);
|
|
volatile unsigned _Atomic long *p =
|
|
(volatile unsigned _Atomic long *)address + get_index(bit);
|
|
|
|
return (atomic_fetch_or(p, mask) & mask) != 0ULL;
|
|
}
|
|
|
|
bool nvgpu_test_and_clear_bit(unsigned int bit, volatile unsigned long *address)
|
|
{
|
|
unsigned long mask = get_mask(bit);
|
|
volatile unsigned _Atomic long *p =
|
|
(volatile unsigned _Atomic long *)address + get_index(bit);
|
|
|
|
return (atomic_fetch_and(p, ~mask) & mask) != 0ULL;
|
|
}
|
|
|
|
void nvgpu_set_bit(unsigned int bit, volatile unsigned long *address)
|
|
{
|
|
unsigned long mask = get_mask(bit);
|
|
volatile unsigned _Atomic long *p =
|
|
(unsigned volatile _Atomic long *)address + get_index(bit);
|
|
|
|
(void)atomic_fetch_or(p, mask);
|
|
}
|
|
|
|
void nvgpu_clear_bit(unsigned int bit, volatile unsigned long *address)
|
|
{
|
|
unsigned long mask = get_mask(bit);
|
|
volatile unsigned _Atomic long *p =
|
|
(volatile unsigned _Atomic long *)address + get_index(bit);
|
|
|
|
(void)atomic_fetch_and(p, ~mask);
|
|
}
|