feat(ai): add brainstorm model

This commit is contained in:
Ayo Ayco 2025-08-30 13:21:23 +02:00
parent f935279644
commit 2a74d2e273
3 changed files with 69 additions and 2 deletions

60
ai-brainstorm.sh Executable file
View file

@ -0,0 +1,60 @@
# Load config
. ${HOME}/ayo.conf
model=$brainstorm
modelfile=$brainstorm_modelfile
host=$ollama_remote_host
# Initialize variables
typora_flag=false
other_args=""
# Process arguments to handle -t flag and collect other args
while [[ $# -gt 0 ]]; do
case $1 in
-t)
typora_flag=true
shift
;;
*)
other_args="$other_args $1"
shift
;;
esac
done
# Set other_args to the first argument if it exists, otherwise empty string
if [[ -n "$other_args" ]]; then
# Remove leading space
other_args="${other_args# }"
IFS=' ' read -ra args_array <<< "$other_args"
if [[ ${#args_array[@]} -gt 1 ]]; then
# Remove first element and rejoin remaining elements
other_args="${args_array[*]:1}"
else
# If there's only one argument, set other_args to empty string
other_args=""
fi
fi
if ! [ "$other_args" = "" ]; then
if [ "$other_args" = "sleep" ]; then
OLLAMA_HOST=$host ollama stop $model
elif [ "$other_args" = "init" ]; then
OLLAMA_HOST=$host ollama create "$model" -f "$modelfile"
else
# If -t flag is set, use typora to display output
if [ "$typora_flag" = true ]; then
tempfile="$(mktemp)"
OLLAMA_HOST=$host ollama run $model "$other_args" --hidethinking > $tempfile
typora $tempfile
else
# If no -t flag, just run the command normally
OLLAMA_HOST=$host ollama run $model "$other_args" --hidethinking
fi
fi
else
OLLAMA_HOST=$host ollama run $model --hidethinking
fi

3
ayo.sh
View file

@ -75,6 +75,9 @@ case $1 in
ai-coder)
. ${scripts_dir}/ai-coder.sh "$@"
;;
ai-brainstorm)
. ${scripts_dir}/ai-brainstorm.sh "$@"
;;
ai-helper)
. ${scripts_dir}/ai.sh "$@"
;;

View file

@ -4,11 +4,15 @@ archive_dir="${HOME}/Notes/Archive"
scripts_dir="${HOME}/Projects/scripts"
editor="vim"
# TODO: Add a flag for force using local?
# Ollama host URL or IP; for fully local, leave empty or comment out
ollama_remote_host="http://remote-ollama-server"
ollama_remote_host="192.168.0.123"
coder="coder:latest"
coder_modelfile="${HOME}/Projects/llm-prompts/Modelfile-coder-lite"
coder_modelfile="${HOME}/Projects/llm-prompts/Modelfile-coder"
helper="helper:latest"
helper_modelfile="${HOME}/Projects/llm-prompts/Modelfile-helper"
brainstorm="brainstorm:latest"
brainstorm_modelfile="${HOME}/Projects/llm-prompts/Modelfile-brainstorm"