experimental/llm_truss/llama2-7b-sample/config.yaml (53 lines of code) (raw):
apply_library_patches: true
base_image: null
build:
arguments: {}
model_server: TrussServer
bundled_packages_dir: packages
data_dir: data
description: null
environment_variables: {}
examples_filename: examples.yaml
external_data: null
external_package_dirs: []
input_type: Any
live_reload: false
model_cache: []
model_class_filename: model.py
model_class_name: Model
model_framework: custom
model_metadata: {}
model_module_dir: model
model_name: Llama2-7B
model_type: custom
python_version: py39
requirements:
- torch
- peft
- sentencepiece
- accelerate
- bitsandbytes
- einops
- scipy
- protobuf
- git+https://github.com/huggingface/transformers.git
resources:
accelerator: null
cpu: '3'
memory: 14Gi
use_gpu: true
runtime:
predict_concurrency: 1
secrets: {}
spec_version: '2.0'
system_packages: []
train:
resources:
accelerator: null
cpu: '1'
memory: 2Gi
use_gpu: false
training_class_filename: train.py
training_class_name: Train
training_module_dir: train
variables: {}