Releases: containers/ramalama
Releases · containers/ramalama
v0.6.1
What's Changed
- chore: use absolute link for the RamaLama logo by @benoitf in #781
- Reuse Ollama cached image when available by @kush-gupt in #782
- Add env var RAMALAMA_GPU_DEVICE to allow for explicit declaration of the GPU device to use by @cgruver in #773
- Change RAMALAMA_GPU_DEVICE to RAMALAMA_DEVICE for AI accelerator device override by @cgruver in #786
- Add Security information to README.md by @rhatdan in #787
- Fix exiting on llama-serve when user hits ^c by @rhatdan in #785
- Check if file exists before sorting them into a list by @kush-gupt in #784
- Add ramalama run --keepalive option by @rhatdan in #789
- Stash output from container_manager by @rhatdan in #790
- Install llama.cpp for mac and nocontainer tests by @rhatdan in #792
- _engine is set to None or has a value by @ericcurtin in #793
- Only run dnf commands on platforms that have dnf by @ericcurtin in #794
- Add ramalama rag command by @rhatdan in #501
- Attempt to use build_llama_and_whisper.sh by @rhatdan in #795
- Change --network-mode to --network by @ericcurtin in #800
- Add some more gfx values to the default list by @ericcurtin in #806
- Update registry.access.redhat.com/ubi9/ubi Docker tag to v9.5-1739449058 by @renovate in #808
- Prepare containers to run with ai-lab-recipes by @rhatdan in #803
- If ngl is not specified by @ericcurtin in #802
- feat: add ramalama labels about the execution on top of container by @benoitf in #810
- Add run and serve arguments for --device and --privileged by @cgruver in #809
- chore: rewrite readarray function to make it portable by @benoitf in #815
- chore: replace RAMALAMA label by ai.ramalama by @benoitf in #814
- Upgrade from 6.3.1 to 6.3.2 by @ericcurtin in #816
- Removed error wrapping in urlopen by @engelmi in #818
- Encountered a bug where this function was returning -1 by @ericcurtin in #817
- Align runtime arguments with run, serve, bench, and perplexity by @cgruver in #820
- README: fix inspect command description by @kush-gupt in #826
- Pin dev dependencies to major version and improve formatting + linting by @engelmi in #824
- README: Fix typo by @bupd in #827
- Switch apt-get to apt by @ericcurtin in #832
- Update registry.access.redhat.com/ubi9/ubi Docker tag to v9.5-1739751568 by @renovate in #834
- Add entrypoint container images by @rhatdan in #819
- HuggingFace Cache Implementation by @kush-gupt in #833
- Make serve by default expose network by @ericcurtin in #830
- Fix up man page help verifacation by @rhatdan in #835
- Fix handling of --privileged flag by @rhatdan in #821
- chore: fix links of llama.cpp repository by @benoitf in #841
- Unify CLI options (verbosity, version) by @mkesper in #685
- Add system tests to pull from the Hugging Face cache by @kush-gupt in #846
- Just one add_argument call for --dryrun/--dry-run by @ericcurtin in #847
- Fix ramalama info to display NVIDIA and amd GPU information by @rhatdan in #848
- Remove LICENSE header from gpu_detector.py by @ericcurtin in #850
- Allowing modification of pull policy by @rhatdan in #843
- Include instructions for installing on Fedora 42+ by @stefwalter in #849
- Bump to 0.6.1 by @rhatdan in #851
New Contributors
- @benoitf made their first contribution in #781
- @bupd made their first contribution in #827
- @mkesper made their first contribution in #685
- @stefwalter made their first contribution in #849
Full Changelog: v0.6.0...v0.6.1
v0.6.0
What's Changed
- fix error on macOS for M1 pro by @volker48 in #687
- This should be a global variable by @ericcurtin in #703
- Update registry.access.redhat.com/ubi9/ubi Docker tag to v9.5-1736404036 by @renovate in #702
- Update install.sh to include "gpu_detector.py" by @graystevens in #704
- add --ngl to specify the number of gpu layers, and --keep-groups so podman has access to gpu by @khumarahn in #659
- We are displaying display driver info, scope creep by @ericcurtin in #710
- Use CODEOWNERS file for autoassign by @dougsland in #706
- common: general improvements by @dougsland in #713
- Fix macOS emoji compatibility with Alacritty by @ericcurtin in #716
- Makelint by @dougsland in #715
- Adding slp, engelmi, also by @ericcurtin in #711
- Report error when huggingface-cli is not available by @rhatdan in #719
- Add --network-mode option by @rhjostone in #674
- README: add convert to commands list by @kush-gupt in #723
- Revert "Add --network-mode option" by @ericcurtin in #731
- Check for apple,arm-platform in /proc by @ericcurtin in #730
- Packit: downstream jobs for EPEL 9,10 by @lsm5 in #728
- Add logic to build intel-gpu image to build_llama_and_whisper.sh by @cgruver in #724
- Add --network-mode option by @rhatdan in #734
- Honor RAMALAMA_IMAGE if set by @rhatdan in #733
- ramalama container: Make it possible to build basic container on all RHEL architectures by @jcajka in #722
- Add docs for using podman farm to build multi-arch images by @cgruver in #735
- Update registry.access.redhat.com/ubi9/ubi Docker tag to v9.5-1738643550 by @renovate in #729
- modify container_build.sh to add capability to use podman farm for multi-arch images by @cgruver in #736
- There's a comma in the list of files in install.sh by @ericcurtin in #739
- Make the default of ngl be -1 by @ericcurtin in #707
- github actions: ramalama install by @dougsland in #738
- [skip-ci] Update actions/checkout action to v4 by @renovate in #740
- On macOS this was returning an incorrect path by @ericcurtin in #741
- Begin process of packaging PRAGmatic by @rhatdan in #597
- Allow users to build RAG versus Docling images by @rhatdan in #744
- Update vLLM containers by @ericcurtin in #746
- Update README.md by @bmbouter in #748
- Update progress bar only once every 100ms by @ericcurtin in #717
- Remove reference to non-existent docs in CONTRIBUTING.md by @cgruver in #761
- Check if krunkit process is running with --all-providers by @ericcurtin in #763
- update_progress only takes one parameter by @ericcurtin in #764
- Detect Intel ARC GPU in Meteor Lake chipset by @cgruver in #749
- Drop all capablities and run with no-new-privileges by @rhatdan in #765
- Progress bar fixes by @ericcurtin in #767
- typo: Add quotes to intel-gpu argument in build llama and whisper script by @hanthor in #766
- chore(deps): update registry.access.redhat.com/ubi9/ubi docker tag to v9.5-1738814488 by @renovate in #771
- There would be one case where this wouldn't work by @ericcurtin in #768
- docs: update ramalama.1.md by @eltociear in #775
- Add community documents by @rhatdan in #777
- Parse https://ollama.com/library/ syntax by @ericcurtin in #648
- Use containers CODE-OF-CONDUCT.md by @rhatdan in #778
- Add model inspect cli by @engelmi in #776
- Cleanup READMEs and man pages. by @rhatdan in #780
- Bump to v0.6.0 by @rhatdan in #779
New Contributors
- @volker48 made their first contribution in #687
- @graystevens made their first contribution in #704
- @khumarahn made their first contribution in #659
- @rhjostone made their first contribution in #674
- @jcajka made their first contribution in #722
- @bmbouter made their first contribution in #748
- @hanthor made their first contribution in #766
- @eltociear made their first contribution in #775
Full Changelog: v0.5.5...v0.6.0
v0.5.5
What's Changed
- Add perplexity subcommand to RamaLama CLI by @ericcurtin in #637
- throwing an exception with there is a failure in http_client.init by @jhjaggars in #647
- Add container image to support Intel ARC GPU by @cgruver in #644
- Guide users to install huggingface-cli to login to huggingface by @pbabinca in #645
- Update intel-gpu Containerfile to reduce the size of the builder image by @cgruver in #657
- Look for configs also in /usr/local/share/ramalama by @jistr in #672
- remove ro as an option when mounting images by @kush-gupt in #676
- Add generated man pages for section 7 into gitignore by @jistr in #673
- Revert "Added --jinja to llama-run command" by @ericcurtin in #683
- Pull the source model if it isn't already in local storage for the convert and push functions by @kush-gupt in #680
- bump llama.cpp to latest release hash aa6fb13 by @maxamillion in #692
- Introduce a mode so one call install from git by @ericcurtin in #690
- Add ramalama gpu_detector by @dougsland in #670
- Bump to v0.5.5 by @rhatdan in #701
New Contributors
- @cgruver made their first contribution in #644
- @pbabinca made their first contribution in #645
- @jistr made their first contribution in #672
- @kush-gupt made their first contribution in #676
- @maxamillion made their first contribution in #692
- @dougsland made their first contribution in #670
Full Changelog: v0.5.4...v0.5.5
v0.5.4
What's Changed
- Attempt to install podman by @ericcurtin in #621
- Introduce ramalama bench by @ericcurtin in #620
- Add man page for cuda support by @rhatdan in #623
- Less verbose output by @ericcurtin in #624
- Avoid dnf install on OSTree system by @ericcurtin in #622
- Fix list in README - Credits section by @kubealex in #627
- added mac cpu only support by @bmahabirbu in #628
- Added --jinja to llama-run command by @engelmi in #625
- Update llama.cpp version by @ericcurtin in #630
- Add shortname for deepseek by @rhatdan in #631
- fixed rocm detection by adding gfx targets in containerfile by @bmahabirbu in #632
- Point macOS users to script install by @kubealex in #635
- Update docker.io/nvidia/cuda Docker tag to v12.8.0 by @renovate in #633
- feat: add argument to define amd gpu targets by @jobcespedes in #634
- Bump to v0.5.4 by @rhatdan in #641
New Contributors
- @kubealex made their first contribution in #627
- @engelmi made their first contribution in #625
- @jobcespedes made their first contribution in #634
Full Changelog: v0.5.3...v0.5.4
v0.5.3
What's Changed
- We no longer have python dependancies by @ericcurtin in #588
- container_build.sh works on MAC by @rhatdan in #590
- Added vllm cuda support by @bmahabirbu in #582
- Remove omlmd from OCI calls by @rhatdan in #591
- Build with curl support by @pepijndevos in #595
- Add model transport info to ramalama run/serve manpage by @rhatdan in #593
- Various README.md updates by @ericcurtin in #600
- code crashes for rocm added proper type cast for env var by @bmahabirbu in #602
- ROCm build broken by @ericcurtin in #605
- Cleaner output if a machine executes this command by @ericcurtin in #604
- Update to version that has command history by @ericcurtin in #603
- Remove these lines they are unused by @ericcurtin in #606
- Had to make this change for my laptop to suppor nvidia by @rhatdan in #609
- Start making vllm work with RamaLama by @rhatdan in #610
- Treat hf.co/ prefix the same as hf:// by @ericcurtin in #612
- We need the rocm libraries in here by @ericcurtin in #613
- A couple of cleanups in build_llama_and_whisper.sh by @rhatdan in #615
- Bump to v0.5.3 by @rhatdan in #614
New Contributors
- @pepijndevos made their first contribution in #595
Full Changelog: v0.5.2...v0.5.3
v0.5.2
What's Changed
- This is all dead code which isn't called by @ericcurtin in #574
- On ARM by default turn on GPU acceleration by @ericcurtin in #573
- Capitalize constants in python files (CONSTANT_CASE) by @swarajpande5 in #579
- Add flake by @jim3692 in #581
- Update llama.cpp to include minor llama-run by @ericcurtin in #580
- Simplify this comparison by @ericcurtin in #576
- Fix ramalama run on docker to work correctly by @rhatdan in #583
- granite-code models in Ollama are malformed by @ericcurtin in #584
- Bump to v0.5.2 by @rhatdan in #585
New Contributors
Full Changelog: v0.5.1...v0.5.2
v0.5.1
What's Changed
- When it debug mode show the exception for subcommand missing by @rhatdan in #568
- Revert "If podman machine is running use containers" by @ericcurtin in #570
- Check all files are adding to packaging by @ericcurtin in #569
- Add podman info or docker info to ramalama info output by @rhatdan in #567
- Bump to v0.5.1 by @rhatdan in #571
Full Changelog: v0.5.0...v0.5.1
v0.5.0
What's Changed
- Remove some more gfx9 stuff by @ericcurtin in #554
- Only use the X/Y parts of the version when pulling images by @rhatdan in #555
- If podman machine is running use containers by @ericcurtin in #557
- Update llama.cpp to contain llama-run fix by @ericcurtin in #556
- We are passing the same value in twice, causes exception by @rhatdan in #558
- Remove firefox from builds by @ericcurtin in #561
- Remove cruft added to redirect llama-cli 2>/dev/null by @rhatdan in #559
- run_cmd to exec_cmd by @ericcurtin in #564
- smollm:135m for testing purposes by @ericcurtin in #563
- Fix podman stop --all by @rhatdan in #562
- Bump to v0.5.0 by @rhatdan in #565
Full Changelog: v0.4.0...v0.5.0
v0.4.0
What's Changed
- ramalama rm should require at least one argument by @rhatdan in #516
- Fix handling of missing args for login/logout by @rhatdan in #517
- Add --version and fix sort order of commands by @rhatdan in #518
- Update shortnames.conf to alias new granite models by @ericcurtin in #519
- Prettify and improve download bars during model pulls by @swarajpande5 in #521
- Remove
tqdm
dependency by @swarajpande5 in #523 - Relocate
File
andHttpClient
to their respective modules by @swarajpande5 in #522 - Add missing files to install.sh script by @ericcurtin in #524
- Change pip to pipx in installation instructions by @ericcurtin in #520
- Pass ASAHI, CUDA, HIP, HSA prefixed env vars to container by @abn in #526
- Resolves - Debug option shows traceback when no subcommand is passed by @pghole in #527
- Skip codespell of
venv
by adding--skip
in the command by @swarajpande5 in #528 - Improved command handling and refactoring by @ericcurtin in #531
- Update README and model.py for Nvidia GPU support by @ericcurtin in #530
- Add UI support documentation by @ericcurtin in #529
- Add llama-run by @ericcurtin in #452
- Update container images to use latest by @ericcurtin in #534
- Update llama.cpp to get context size fix by @ericcurtin in #538
- If runtime is vllm, use image from modh quay.io by @ericcurtin in #533
- Remove huggingface warning by @ericcurtin in #540
- When using podman run add --pull=newer to update container image by @rhatdan in #541
- Build to build_rm by @ericcurtin in #543
- whisper.cpp now installs properly by @rhatdan in #544
- Re-add mkdir -p by @ericcurtin in #545
- This might temporarily fix an issue we are having by @ericcurtin in #547
- Attempt to use versioned container by @ericcurtin in #537
- Fix llama-run to be installed with correct library paths. by @rhatdan in #549
- Add CPU inferencing container build steps to CI by @ericcurtin in #551
- This should save some space by @ericcurtin in #552
- Bump to v0.4.0 by @rhatdan in #550
New Contributors
Full Changelog: v0.3.0...v0.4.0
v0.3.0
What's Changed
- Move man page README.md to full links by @rhatdan in #483
- Allow users to set ctx-size via command line by @rhatdan in #484
- Add --seed option by @rhatdan in #485
- Update install.sh by @jtligon in #493
- Take in fixes from @Churchyard to modernize spec file by @smooge in #494
- Fix up building and pushing OCI Images by @rhatdan in #492
- Fix handling of file_not_found errors by @rhatdan in #499
- Updated nv docs to align with latest WSL2 cuda setup by @bmahabirbu in #503
- Add ramalama convert command by @rhatdan in #500
- Stop checking if command is running in container by @rhatdan in #505
- Add initial CONTRIBUTING.md file by @rhatdan in #507
- Place image name just before command by @ericcurtin in #511
- Simplify install by @ericcurtin in #510
- Fix handling of README.md in docs directory by @rhatdan in #512
- Add installation steps for Podman 5 in CI workflows by @ericcurtin in #508
- Bump to v0.3.0 by @rhatdan in #513
New Contributors
Full Changelog: v0.2.0...v0.3.0