cli 도움말. show는 모델정보 보기 run은 실행하는거.. stop 하면 그러면 날릴 수 있나?
| C:\>ollama --help Large language model runner Usage: ollama [flags] ollama [command] Available Commands: serve Start Ollama create Create a model show Show information for a model run Run a model stop Stop a running model pull Pull a model from a registry push Push a model to a registry signin Sign in to ollama.com signout Sign out from ollama.com list List models ps List running models cp Copy a model rm Remove a model launch Launch the Ollama menu or an integration help Help about any command Flags: -h, --help help for ollama --nowordwrap Don't wrap words to the next line automatically --verbose Show timings for response -v, --version Show version information Use "ollama [command] --help" for more information about a command. |
cli 내부 명령어는 faq에 인터넷 검색으로 찾은 정도가 전부인거 같은데 내부에서 보면 제법 먼가 있긴하다.
| How can I specify the context window size? By default, Ollama uses a context window size of 4096 tokens. This can be overridden with the OLLAMA_CONTEXT_LENGTH environment variable. For example, to set the default context window to 8K, use: OLLAMA_CONTEXT_LENGTH=8192 ollama serve To change this when using ollama run, use /set parameter: /set parameter num_ctx 4096 |
[링크 : https://docs.ollama.com/faq]
눈에 띄는건 컨텍스트 길이와 gpu 갯수..인가?
| C:\>ollama run gemma3:1b >>> /set Available Commands: /set parameter ... Set a parameter /set system <string> Set system message /set history Enable history /set nohistory Disable history /set wordwrap Enable wordwrap /set nowordwrap Disable wordwrap /set format json Enable JSON mode /set noformat Disable formatting /set verbose Show LLM stats /set quiet Disable LLM stats /set think Enable thinking /set nothink Disable thinking >>> /set parameter Available Parameters: /set parameter seed <int> Random number seed /set parameter num_predict <int> Max number of tokens to predict /set parameter top_k <int> Pick from top k num of tokens /set parameter top_p <float> Pick token based on sum of probabilities /set parameter min_p <float> Pick token based on top token probability * min_p /set parameter num_ctx <int> Set the context size /set parameter temperature <float> Set creativity level /set parameter repeat_penalty <float> How strongly to penalize repetitions /set parameter repeat_last_n <int> Set how far back to look for repetitions /set parameter num_gpu <int> The number of layers to send to the GPU /set parameter stop <string> <string> ... Set the stop parameters |
'프로그램 사용 > ai 프로그램' 카테고리의 다른 글
| ollama 데이터 파일 구조 및 내용 (0) | 2026.04.12 |
|---|---|
| ollama /set parameter num_gpu 테스트 (0) | 2026.04.12 |
| ollama cpu only (0) | 2026.04.11 |
| ollama context len (0) | 2026.04.11 |
| ollama for windows (0) | 2026.04.10 |
