Skip to content

Commit

Permalink
Merge pull request #454 from containers/llama-simple-chat
Browse files Browse the repository at this point in the history
Switch to llama-simple-chat
  • Loading branch information
rhatdan authored Nov 15, 2024
2 parents 1b5eb20 + 1db401d commit 8a61a4a
Show file tree
Hide file tree
Showing 4 changed files with 5 additions and 16 deletions.
2 changes: 1 addition & 1 deletion container-images/cuda/Containerfile
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
# Base image with CUDA for compilation
FROM docker.io/nvidia/cuda:12.6.2-devel-ubi9 AS builder

ARG LLAMA_CPP_SHA=1329c0a75e6a7defc5c380eaf80d8e0f66d7da78
ARG LLAMA_CPP_SHA=af148c9386da825a60c7038549c121c35ca56b50
# renovate: datasource=git-refs depName=ggerganov/whisper.cpp packageName=https://github.com/ggerganov/whisper.cpp gitRef=master versioning=loose type=digest
ARG WHISPER_CPP_SHA=f19463ece2d43fd0b605dc513d8800eeb4e2315e

Expand Down
2 changes: 1 addition & 1 deletion container-images/ramalama/Containerfile
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
FROM registry.access.redhat.com/ubi9/ubi:9.4-1214.1729773476

ARG LLAMA_CPP_SHA=1329c0a75e6a7defc5c380eaf80d8e0f66d7da78
ARG LLAMA_CPP_SHA=af148c9386da825a60c7038549c121c35ca56b50
# renovate: datasource=git-refs depName=ggerganov/whisper.cpp packageName=https://github.com/ggerganov/whisper.cpp gitRef=master versioning=loose type=digest
ARG WHISPER_CPP_SHA=f19463ece2d43fd0b605dc513d8800eeb4e2315e

Expand Down
15 changes: 2 additions & 13 deletions ramalama/model.py
Original file line number Diff line number Diff line change
Expand Up @@ -267,19 +267,8 @@ def run(self, args):
if not args.container:
exec_model_path = model_path

exec_args = ["llama-cli", "-m", exec_model_path, "--in-prefix", "", "--in-suffix", ""]

if not args.debug:
exec_args += ["--no-display-prompt"]

exec_args += [
"-p",
prompt,
] + self.common_params

if not args.ARGS and sys.stdin.isatty():
exec_args.append("-cnv")

exec_args = ["llama-simple-chat", "-m", exec_model_path]
exec_args += self.common_params
if args.gpu:
exec_args.extend(self.gpu_args())

Expand Down
2 changes: 1 addition & 1 deletion test/system/030-run.bats
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@ load helpers
is "$output" ".*${image} /bin/sh -c" "verify image name"
else
run_ramalama --dryrun run ${model}
is "$output" 'llama-cli -m /path/to/model --in-prefix --in-suffix --no-display-prompt -p.*' "dryrun correct"
is "$output" 'llama-simple-chat -m /path/to/model.*' "dryrun correct"

run_ramalama 1 run --name foobar tiny
is "${lines[0]}" "Error: --nocontainer and --name options conflict. --name requires a container." "conflict between nocontainer and --name line"
Expand Down

0 comments on commit 8a61a4a

Please sign in to comment.