Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bump jinja2 from 3.1.3 to 3.1.4 #17

Open
wants to merge 35 commits into
base: master
Choose a base branch
from

Commits on Jul 18, 2024

  1. Configuration menu
    Copy the full SHA
    506d2fe View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    88289b5 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    3f7c4b9 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    9d5207b View commit details
    Browse the repository at this point in the history
  5. llama : wrap llama_new_context_with_model in try/catch

    This fixes a crash where ggml_vk_allocate fails in llama_kv_cache_init,
    but the exception is never caught.
    cebtenzzre committed Jul 18, 2024
    Configuration menu
    Copy the full SHA
    b6891bc View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    b80287e View commit details
    Browse the repository at this point in the history
  7. kompute : fix ggml_vk_allocate failure control flow

    The correct way to indicate an OOM condition is for alloc_buffer to
    return NULL. This fixes undefined behavior caused by passing an
    exception over the C boundary.
    
    The rest of the changes help fix VRAM leaks in GPT4All when model
    loading fails on GPU.
    cebtenzzre committed Jul 18, 2024
    Configuration menu
    Copy the full SHA
    dc7a50b View commit details
    Browse the repository at this point in the history
  8. kompute : disable GPU offload for Mixtral

    We haven't implemented the necessary GPU kernels yet.
    
    Fixes this crash:
    
    ggml_vk_graph_compute: error: unsupported op 'ARGSORT'
    GGML_ASSERT: /home/jared/src/forks/gpt4all/gpt4all-backend/llama.cpp-mainline/ggml-kompute.cpp:1508: !"unsupported op"
    cebtenzzre committed Jul 18, 2024
    Configuration menu
    Copy the full SHA
    c5014f6 View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    c76f5c3 View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    6ff4387 View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    12dcddc View commit details
    Browse the repository at this point in the history
  12. Configuration menu
    Copy the full SHA
    82b50e5 View commit details
    Browse the repository at this point in the history
  13. kompute : enable GPU support for 10 more model architectures

    These are Baichuan, Bert and Nomic Bert, CodeShell, GPT-2, InternLM,
    MiniCPM, Orion, Qwen, and StarCoder.
    cebtenzzre committed Jul 18, 2024
    Configuration menu
    Copy the full SHA
    a76f5f4 View commit details
    Browse the repository at this point in the history
  14. Configuration menu
    Copy the full SHA
    877851b View commit details
    Browse the repository at this point in the history
  15. Configuration menu
    Copy the full SHA
    729d661 View commit details
    Browse the repository at this point in the history
  16. Configuration menu
    Copy the full SHA
    2b8cb26 View commit details
    Browse the repository at this point in the history
  17. Configuration menu
    Copy the full SHA
    6e0b5d9 View commit details
    Browse the repository at this point in the history
  18. Configuration menu
    Copy the full SHA
    aea0abe View commit details
    Browse the repository at this point in the history
  19. Configuration menu
    Copy the full SHA
    535c7b1 View commit details
    Browse the repository at this point in the history
  20. vulkan : guard against multiple initialization

    This trades a late heap-use-after-free for an early abort, which feels
    more correct.
    cebtenzzre committed Jul 18, 2024
    Configuration menu
    Copy the full SHA
    2a91dbf View commit details
    Browse the repository at this point in the history
  21. Configuration menu
    Copy the full SHA
    ad1ab57 View commit details
    Browse the repository at this point in the history
  22. Configuration menu
    Copy the full SHA
    09058b1 View commit details
    Browse the repository at this point in the history
  23. Configuration menu
    Copy the full SHA
    b0ccbe1 View commit details
    Browse the repository at this point in the history
  24. Configuration menu
    Copy the full SHA
    74a41c6 View commit details
    Browse the repository at this point in the history
  25. Configuration menu
    Copy the full SHA
    f10326c View commit details
    Browse the repository at this point in the history
  26. Configuration menu
    Copy the full SHA
    e5c0df7 View commit details
    Browse the repository at this point in the history
  27. llama : use the correct buffer type when we choose not to load on GPU

    This fixes a regression in commit b2db03a ("llama: replace ngl=0 hack
    with llama_model_using_gpu").
    cebtenzzre committed Jul 18, 2024
    Configuration menu
    Copy the full SHA
    159235e View commit details
    Browse the repository at this point in the history
  28. Configuration menu
    Copy the full SHA
    c301b42 View commit details
    Browse the repository at this point in the history
  29. Configuration menu
    Copy the full SHA
    7d402b3 View commit details
    Browse the repository at this point in the history
  30. Configuration menu
    Copy the full SHA
    48a830c View commit details
    Browse the repository at this point in the history
  31. kompute : fix dangling references in ggml_vk_graph_kompute

    ggml_vk_get_tensor_aligned() returns a shared_ptr, not a reference, so
    we must copy the value.
    cebtenzzre committed Jul 18, 2024
    Configuration menu
    Copy the full SHA
    6e0ad3c View commit details
    Browse the repository at this point in the history
  32. kompute : avoid freeing device/instance until absolutely necessary

    Eagerly freeing the instance when we are done with it is simple, but
    incurs an overhead, and more importantly, causes test-backend-ops
    crashes on the current proprietary NVIDIA driver.
    
    Instead, we now only cleanup device resources without freeing the device
    unless we actually need to change devices. And even when we free the
    device, we do not free the instance. We only free the instance when both
    the backend and all buffers have been unreferenced.
    cebtenzzre committed Jul 18, 2024
    Configuration menu
    Copy the full SHA
    c3d5264 View commit details
    Browse the repository at this point in the history
  33. kompute : update ggml_vk_supports_op to fix false pos/neg

    test-backend-ops hit assertion failures in ggml_vk_graph_compute because
    of ops we do not yet support. Some of the checks have to be made more
    restrictive because of features that were added to llama.cpp.
    
    We also claimed to not support no-op operations on certain data types,
    even though they are actually supported on all data types. There are now
    243 passsing tests, instead of 150 without the fixes for false
    negatives. This also fixes complaints during LLM inference about
    unsupported NONE operations for the output tensor.
    cebtenzzre committed Jul 18, 2024
    Configuration menu
    Copy the full SHA
    561d0ce View commit details
    Browse the repository at this point in the history
  34. Configuration menu
    Copy the full SHA
    cd13f44 View commit details
    Browse the repository at this point in the history
  35. Bump jinja2 from 3.1.3 to 3.1.4

    Bumps [jinja2](https://github.com/pallets/jinja) from 3.1.3 to 3.1.4.
    - [Release notes](https://github.com/pallets/jinja/releases)
    - [Changelog](https://github.com/pallets/jinja/blob/main/CHANGES.rst)
    - [Commits](pallets/jinja@3.1.3...3.1.4)
    
    ---
    updated-dependencies:
    - dependency-name: jinja2
      dependency-type: indirect
    ...
    
    Signed-off-by: dependabot[bot] <[email protected]>
    dependabot[bot] committed Jul 18, 2024
    Configuration menu
    Copy the full SHA
    84c5f62 View commit details
    Browse the repository at this point in the history