Skip to content

Commit

Permalink
update
Browse files Browse the repository at this point in the history
  • Loading branch information
zhulin1 committed Sep 9, 2024
1 parent fca3174 commit 8483e3f
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 10 deletions.
7 changes: 1 addition & 6 deletions autotest/benchmark/test_apiserver_performance.py
Original file line number Diff line number Diff line change
Expand Up @@ -22,12 +22,7 @@ def getModelList(tp_num):
new_model_list = []
for model in model_list:
if model['backend'] == 'pytorch':
if 'Llama-3' in model['model'] and '70B' in model['model']:
model[
'extra'] = '--max-batch-size 256 --cache-max-entry-count 0.65' # noqa: E501
else:
model[
'extra'] = '--max-batch-size 256 --cache-max-entry-count 0.7' # noqa: E501
model['extra'] = '--max-batch-size 256 --cache-max-entry-count 0.7'
elif 'Llama-2' in model['model']:
model[
'extra'] = '--max-batch-size 256 --cache-max-entry-count 0.95'
Expand Down
5 changes: 1 addition & 4 deletions autotest/utils/benchmark_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -248,10 +248,7 @@ def create_multi_level_directory(path):

def get_max_cache_entry(model, backend):
if backend == 'pytorch':
if 'Llama-3' in model and '70B' in model:
return '--cache-max-entry-count 0.65'
else:
return '--cache-max-entry-count 0.7'
return '--cache-max-entry-count 0.7'
if 'Llama-2' in model:
return '--cache-max-entry-count 0.95'
elif 'internlm2' in model:
Expand Down

0 comments on commit 8483e3f

Please sign in to comment.