File size: 3,515 Bytes
5f923cd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
// Copyright 2025 The ODML Authors.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//      http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#ifndef THIRD_PARTY_ODML_LITERT_LM_RUNTIME_COMPONENTS_LORA_H_
#define THIRD_PARTY_ODML_LITERT_LM_RUNTIME_COMPONENTS_LORA_H_

#include <memory>
#include <string>
#include <utility>

#include "absl/container/flat_hash_map.h"  // from @com_google_absl
#include "absl/status/status.h"  // from @com_google_absl
#include "absl/status/statusor.h"  // from @com_google_absl
#include "absl/strings/string_view.h"  // from @com_google_absl
#include "litert/cc/litert_compiled_model.h"  // from @litert
#include "litert/cc/litert_model.h"  // from @litert
#include "litert/cc/litert_tensor_buffer.h"  // from @litert
#include "runtime/util/lora_data.h"

namespace litert::lm {

// The LoRA interface for LiteRT-LM.
// It handles the LoRA weight loading, filling Lora weights into LiteRT objects
// (e.g. TensorBuffer), and weight rearranging.
// Note that LoRA backend resources are held in litert::TensorBuffer, which is
// essentially a shared_ptr to the real data, so the LoRA class is not the sore
// owner of the underlying resources. But we should still treat LoRA as the main
// owner of the lora data, and destroy it to free resources when necessary.
class LoRA {
 public:
  // Creates and initializes a LoRA object.
  //
  // @param lora_data The LoraData object containing the LoRA weights.
  // @param compiled_model The CompiledModel object containing model and
  // environment information. It is used for creating backend resources for
  // model buffers.
  // @return A unique_ptr to the LoRA instance, or an error status.
  static absl::StatusOr<std::unique_ptr<LoRA>> Create(
      std::unique_ptr<LoraData> lora_data,
      const litert::CompiledModel& compiled_model);

  virtual ~LoRA() = default;

  // Returns a duplicated TensorBuffer for the given LoRA tensor name.
  // TensorBuffer is a shared_ptr to the real data, so users are responsible
  // for destroying the TensorBuffer received after use to properly decrease
  // reference count to the underlying data.
  absl::StatusOr<litert::TensorBuffer> GetLoRABuffer(
      const std::string& name) const;

  // Returns a map of all the LoRA tensor names to their duplicated
  // TensorBuffers.
  // See GetLoRABuffer() for more details about resource ownership.
  absl::StatusOr<absl::flat_hash_map<absl::string_view, litert::TensorBuffer>>
  GetLoRABuffers() const;

 private:
  LoRA(std::unique_ptr<LoraData> lora_data,
       const litert::CompiledModel& compiled_model)
      : lora_data_(std::move(lora_data)), compiled_model_(compiled_model) {}

  // Initializes the LoRA object by creating TensorBuffers for all LoRA inputs
  // and copying the data from LoraData.
  absl::Status Init();

  std::unique_ptr<LoraData> lora_data_;
  const litert::CompiledModel& compiled_model_;
  absl::flat_hash_map<std::string, litert::TensorBuffer> lora_buffers_;
};

}  // namespace litert::lm

#endif  // THIRD_PARTY_ODML_LITERT_LM_RUNTIME_COMPONENTS_LORA_H_