/*************************************************************************************************** * Copyright (c) 2017 - 2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved. * SPDX-License-Identifier: BSD-3-Clause * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, this * list of conditions and the following disclaimer. * * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * * 3. Neither the name of the copyright holder nor the names of its * contributors may be used to endorse or promote products derived from * this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. * **************************************************************************************************/ /* \file \brief */ #include "cutlass/profiler/device_context.h" namespace cutlass { namespace profiler { ///////////////////////////////////////////////////////////////////////////////////////////////// /// Allocates memory of a given type, capacity (elements), and name DeviceAllocation *DeviceContext::allocate_block( Options const &options, std::string const &name, library::NumericTypeID type, size_t capacity, size_t device_index) { int device = options.device.device_id(device_index); device_memory_.emplace_back(type, capacity, device); DeviceAllocation *allocation = &device_memory_.back(); allocations_[name] = allocation; return allocation; } /// Allocates memory of a given type, capacity (elements), and name DeviceAllocation *DeviceContext::allocate_tensor( Options const &options, std::string const &name, library::NumericTypeID type, library::LayoutTypeID layout_id, std::vector const &extent, std::vector const &stride, int batch_count, size_t device_index) { int device = options.device.device_id(device_index); device_memory_.emplace_back(type, layout_id, extent, stride, batch_count, device); DeviceAllocation *allocation = &device_memory_.back(); allocations_[name] = allocation; return allocation; } static void initialize_allocation_with_data_distribution( Options const &options, int seed_shift, DeviceAllocation *allocation, Distribution &data_distribution) { if (options.initialization.provider == library::Provider::kReferenceDevice) { if (data_distribution.kind == Distribution::Sequential) { allocation->initialize_sequential_device( data_distribution); } else { allocation->initialize_random_device( options.initialization.seed + seed_shift, data_distribution); } } else if (options.initialization.provider == library::Provider::kReferenceHost) { if (data_distribution.kind == Distribution::Sequential) { allocation->initialize_sequential_host( data_distribution); } else { allocation->initialize_random_host( options.initialization.seed + seed_shift, data_distribution); } } } /// Allocates memory of a given type, capacity (elements), and name DeviceAllocation *DeviceContext::allocate_and_initialize_tensor( Options const &options, std::string const &name, library::NumericTypeID type, library::LayoutTypeID layout_id, std::vector const &extent, std::vector const &stride, int batch_count, int seed_shift, size_t device_index) { DeviceAllocation *allocation = allocate_tensor(options, name, type, layout_id, extent, stride, batch_count, device_index); if (options.initialization.enabled) { Distribution data_distribution = options.initialization.data_distribution; // check if data distribution is allowed to change if(!options.initialization.fix_data_distribution) { // change data distribution based on bit width switch(type) { case library::NumericTypeID::kFE4M3: data_distribution.set_uniform(-1, 1, 0); break; case library::NumericTypeID::kFE5M2: data_distribution.set_uniform(-1, 1, 0); break; case library::NumericTypeID::kFE2M3: data_distribution.set_uniform(-2, 2, 0); break; case library::NumericTypeID::kFE3M2: data_distribution.set_uniform(-2, 2, 0); break; case library::NumericTypeID::kFE2M1: data_distribution.set_uniform(-2, 2, 0); break; case library::NumericTypeID::kFUE8M0: data_distribution.set_uniform(1, 4, 0); break; case library::NumericTypeID::kFUE4M3: data_distribution.set_uniform(1, 4, 0); break; case library::NumericTypeID::kF16: data_distribution.set_uniform(-3, 3, 0); break; case library::NumericTypeID::kB1: data_distribution.set_uniform(0, 1, 0); break; case library::NumericTypeID::kS2: data_distribution.set_uniform(-1, 1, 0); break; case library::NumericTypeID::kS4: data_distribution.set_uniform(-2, 2, 0); break; case library::NumericTypeID::kU2: data_distribution.set_uniform(0, 2, 0); break; case library::NumericTypeID::kU4: data_distribution.set_uniform(0, 2, 0); break; case library::NumericTypeID::kS8: data_distribution.set_uniform(-3, 3, 0); break; case library::NumericTypeID::kU8: data_distribution.set_uniform(0, 4, 0); break; default: break; } } // Override pnz for the A/B/C tensors if overridden for Gaussian distributions if (data_distribution.kind == Distribution::Gaussian) { double mean = data_distribution.gaussian.mean; double stddev = data_distribution.gaussian.stddev; int scale = data_distribution.int_scale; if (name == "A" && data_distribution.gaussian.pnzA != 1.0) { data_distribution.set_gaussian(mean, stddev, scale, data_distribution.gaussian.pnzA); } else if (name == "B" && data_distribution.gaussian.pnzB != 1.0) { data_distribution.set_gaussian(mean, stddev, scale, data_distribution.gaussian.pnzB); } else if (name == "C" && data_distribution.gaussian.pnzC != 1.0) { data_distribution.set_gaussian(mean, stddev, scale, data_distribution.gaussian.pnzC); } } initialize_allocation_with_data_distribution( options, seed_shift, allocation, data_distribution ); } return allocation; } /// Allocates memory for sparse meta data DeviceAllocation *DeviceContext::allocate_and_initialize_sparsemeta_tensor( Options const &options, std::string const &name, library::NumericTypeID type, library::LayoutTypeID layout_id, library::NumericTypeID type_a, std::vector const &extent, std::vector const &stride, int batch_count, int seed_shift, size_t device_index) { DeviceAllocation *allocation = allocate_tensor(options, name, type, layout_id, extent, stride, batch_count, device_index); if (options.initialization.enabled) { // TF32 has 4bit meta data. The rest has 2bit. int MetaSizeInBits = (cutlass::library::sizeof_bits(type_a) == 32) ? 4 : 2; if (options.initialization.provider == library::Provider::kReferenceDevice) { allocation->initialize_random_sparsemeta_device( options.initialization.seed + seed_shift, MetaSizeInBits); } else if (options.initialization.provider == library::Provider::kReferenceHost) { allocation->initialize_random_sparsemeta_host( options.initialization.seed + seed_shift, MetaSizeInBits); } } return allocation; } /// Clears named allocations (but does not necessarily free memory) void DeviceContext::clear() { allocations_.clear(); } /// Frees all device memory allocations void DeviceContext::free() { allocations_.clear(); device_memory_.clear(); } /// Gets the allocation by name DeviceAllocation &DeviceContext::at(std::string const &name) { return *allocations_.at(name); } size_t DeviceContext::size() const { return allocations_.size(); } DeviceContext::AllocationMap::iterator DeviceContext::begin() { return allocations_.begin(); } DeviceContext::AllocationMap::iterator DeviceContext::end() { return allocations_.end(); } ///////////////////////////////////////////////////////////////////////////////////////////////// } // namespace profiler } // namespace cutlass