VidChain-exercise
/
VTimeLLM
/flash-attention
/csrc
/cutlass
/tools
/profiler
/src
/device_context.cu
| /*************************************************************************************************** | |
| * Copyright (c) 2017 - 2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved. | |
| * SPDX-License-Identifier: BSD-3-Clause | |
| * | |
| * Redistribution and use in source and binary forms, with or without | |
| * modification, are permitted provided that the following conditions are met: | |
| * | |
| * 1. Redistributions of source code must retain the above copyright notice, this | |
| * list of conditions and the following disclaimer. | |
| * | |
| * 2. Redistributions in binary form must reproduce the above copyright notice, | |
| * this list of conditions and the following disclaimer in the documentation | |
| * and/or other materials provided with the distribution. | |
| * | |
| * 3. Neither the name of the copyright holder nor the names of its | |
| * contributors may be used to endorse or promote products derived from | |
| * this software without specific prior written permission. | |
| * | |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | |
| * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |
| * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE | |
| * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE | |
| * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL | |
| * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR | |
| * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER | |
| * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, | |
| * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE | |
| * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |
| * | |
| **************************************************************************************************/ | |
| /* \file | |
| \brief | |
| */ | |
| #include "cutlass/profiler/device_context.h" | |
| namespace cutlass { | |
| namespace profiler { | |
| ///////////////////////////////////////////////////////////////////////////////////////////////// | |
| /// Allocates memory of a given type, capacity (elements), and name | |
| DeviceAllocation *DeviceContext::allocate_block( | |
| Options const &options, | |
| std::string const &name, | |
| library::NumericTypeID type, | |
| size_t capacity, | |
| size_t device_index) { | |
| int device = options.device.device_id(device_index); | |
| device_memory_.emplace_back(type, capacity, device); | |
| DeviceAllocation *allocation = &device_memory_.back(); | |
| allocations_[name] = allocation; | |
| return allocation; | |
| } | |
| /// Allocates memory of a given type, capacity (elements), and name | |
| DeviceAllocation *DeviceContext::allocate_tensor( | |
| Options const &options, | |
| std::string const &name, | |
| library::NumericTypeID type, | |
| library::LayoutTypeID layout_id, | |
| std::vector<int> const &extent, | |
| std::vector<int64_t> const &stride, | |
| int batch_count, | |
| size_t device_index) { | |
| int device = options.device.device_id(device_index); | |
| device_memory_.emplace_back(type, layout_id, extent, stride, batch_count, | |
| device); | |
| DeviceAllocation *allocation = &device_memory_.back(); | |
| allocations_[name] = allocation; | |
| return allocation; | |
| } | |
| static void initialize_allocation_with_data_distribution( | |
| Options const &options, | |
| int seed_shift, | |
| DeviceAllocation *allocation, | |
| Distribution &data_distribution) { | |
| if (options.initialization.provider == library::Provider::kReferenceDevice) { | |
| if (data_distribution.kind == Distribution::Sequential) { | |
| allocation->initialize_sequential_device( | |
| data_distribution); | |
| } | |
| else { | |
| allocation->initialize_random_device( | |
| options.initialization.seed + seed_shift, | |
| data_distribution); | |
| } | |
| } | |
| else if (options.initialization.provider == library::Provider::kReferenceHost) { | |
| if (data_distribution.kind == Distribution::Sequential) { | |
| allocation->initialize_sequential_host( | |
| data_distribution); | |
| } | |
| else { | |
| allocation->initialize_random_host( | |
| options.initialization.seed + seed_shift, | |
| data_distribution); | |
| } | |
| } | |
| } | |
| /// Allocates memory of a given type, capacity (elements), and name | |
| DeviceAllocation *DeviceContext::allocate_and_initialize_tensor( | |
| Options const &options, | |
| std::string const &name, | |
| library::NumericTypeID type, | |
| library::LayoutTypeID layout_id, | |
| std::vector<int> const &extent, | |
| std::vector<int64_t> const &stride, | |
| int batch_count, | |
| int seed_shift, | |
| size_t device_index) { | |
| DeviceAllocation *allocation = | |
| allocate_tensor(options, name, type, layout_id, extent, stride, | |
| batch_count, device_index); | |
| if (options.initialization.enabled) { | |
| Distribution data_distribution = options.initialization.data_distribution; | |
| // check if data distribution is allowed to change | |
| if(!options.initialization.fix_data_distribution) { | |
| // change data distribution based on bit width | |
| switch(type) { | |
| case library::NumericTypeID::kFE4M3: | |
| data_distribution.set_uniform(-1, 1, 0); | |
| break; | |
| case library::NumericTypeID::kFE5M2: | |
| data_distribution.set_uniform(-1, 1, 0); | |
| break; | |
| case library::NumericTypeID::kFE2M3: | |
| data_distribution.set_uniform(-2, 2, 0); | |
| break; | |
| case library::NumericTypeID::kFE3M2: | |
| data_distribution.set_uniform(-2, 2, 0); | |
| break; | |
| case library::NumericTypeID::kFE2M1: | |
| data_distribution.set_uniform(-2, 2, 0); | |
| break; | |
| case library::NumericTypeID::kFUE8M0: | |
| data_distribution.set_uniform(1, 4, 0); | |
| break; | |
| case library::NumericTypeID::kFUE4M3: | |
| data_distribution.set_uniform(1, 4, 0); | |
| break; | |
| case library::NumericTypeID::kF16: | |
| data_distribution.set_uniform(-3, 3, 0); | |
| break; | |
| case library::NumericTypeID::kB1: | |
| data_distribution.set_uniform(0, 1, 0); | |
| break; | |
| case library::NumericTypeID::kS2: | |
| data_distribution.set_uniform(-1, 1, 0); | |
| break; | |
| case library::NumericTypeID::kS4: | |
| data_distribution.set_uniform(-2, 2, 0); | |
| break; | |
| case library::NumericTypeID::kU2: | |
| data_distribution.set_uniform(0, 2, 0); | |
| break; | |
| case library::NumericTypeID::kU4: | |
| data_distribution.set_uniform(0, 2, 0); | |
| break; | |
| case library::NumericTypeID::kS8: | |
| data_distribution.set_uniform(-3, 3, 0); | |
| break; | |
| case library::NumericTypeID::kU8: | |
| data_distribution.set_uniform(0, 4, 0); | |
| break; | |
| default: break; | |
| } | |
| } | |
| // Override pnz for the A/B/C tensors if overridden for Gaussian distributions | |
| if (data_distribution.kind == Distribution::Gaussian) { | |
| double mean = data_distribution.gaussian.mean; | |
| double stddev = data_distribution.gaussian.stddev; | |
| int scale = data_distribution.int_scale; | |
| if (name == "A" && data_distribution.gaussian.pnzA != 1.0) { | |
| data_distribution.set_gaussian(mean, stddev, scale, data_distribution.gaussian.pnzA); | |
| } | |
| else if (name == "B" && data_distribution.gaussian.pnzB != 1.0) { | |
| data_distribution.set_gaussian(mean, stddev, scale, data_distribution.gaussian.pnzB); | |
| } | |
| else if (name == "C" && data_distribution.gaussian.pnzC != 1.0) { | |
| data_distribution.set_gaussian(mean, stddev, scale, data_distribution.gaussian.pnzC); | |
| } | |
| } | |
| initialize_allocation_with_data_distribution( | |
| options, seed_shift, allocation, data_distribution | |
| ); | |
| } | |
| return allocation; | |
| } | |
| /// Allocates memory for sparse meta data | |
| DeviceAllocation *DeviceContext::allocate_and_initialize_sparsemeta_tensor( | |
| Options const &options, | |
| std::string const &name, | |
| library::NumericTypeID type, | |
| library::LayoutTypeID layout_id, | |
| library::NumericTypeID type_a, | |
| std::vector<int> const &extent, | |
| std::vector<int64_t> const &stride, | |
| int batch_count, | |
| int seed_shift, | |
| size_t device_index) { | |
| DeviceAllocation *allocation = | |
| allocate_tensor(options, name, type, layout_id, extent, stride, | |
| batch_count, device_index); | |
| if (options.initialization.enabled) { | |
| // TF32 has 4bit meta data. The rest has 2bit. | |
| int MetaSizeInBits = (cutlass::library::sizeof_bits(type_a) == 32) ? 4 : 2; | |
| if (options.initialization.provider == library::Provider::kReferenceDevice) { | |
| allocation->initialize_random_sparsemeta_device( | |
| options.initialization.seed + seed_shift, | |
| MetaSizeInBits); | |
| } | |
| else if (options.initialization.provider == library::Provider::kReferenceHost) { | |
| allocation->initialize_random_sparsemeta_host( | |
| options.initialization.seed + seed_shift, | |
| MetaSizeInBits); | |
| } | |
| } | |
| return allocation; | |
| } | |
| /// Clears named allocations (but does not necessarily free memory) | |
| void DeviceContext::clear() { | |
| allocations_.clear(); | |
| } | |
| /// Frees all device memory allocations | |
| void DeviceContext::free() { | |
| allocations_.clear(); | |
| device_memory_.clear(); | |
| } | |
| /// Gets the allocation by name | |
| DeviceAllocation &DeviceContext::at(std::string const &name) { | |
| return *allocations_.at(name); | |
| } | |
| size_t DeviceContext::size() const { | |
| return allocations_.size(); | |
| } | |
| DeviceContext::AllocationMap::iterator DeviceContext::begin() { | |
| return allocations_.begin(); | |
| } | |
| DeviceContext::AllocationMap::iterator DeviceContext::end() { | |
| return allocations_.end(); | |
| } | |
| ///////////////////////////////////////////////////////////////////////////////////////////////// | |
| } // namespace profiler | |
| } // namespace cutlass | |