optimum/habana/transformers/models/decilm/configuration_decilm.py (11 lines of code) (raw):
"""
Adapted from the following sources:
https://huggingface.co/Deci/DeciLM-7B/blob/main/configuration_decilm.py
"""
from transformers.models.llama.configuration_llama import LlamaConfig
class DeciLMConfig(LlamaConfig):
r"""
Args:
num_key_value_heads_per_layer (`List[int]`):
The number of key-value heads per layer.
"""
model_type = "deci"
def __init__(
self,
num_key_value_heads_per_layer: list = None,
**kwargs,
):
self.num_key_value_heads_per_layer = num_key_value_heads_per_layer
super().__init__(**kwargs)