diff --git a/gpt4all-chat/CHANGELOG.md b/gpt4all-chat/CHANGELOG.md index 49c0734fd6ed..a3119d0c5597 100644 --- a/gpt4all-chat/CHANGELOG.md +++ b/gpt4all-chat/CHANGELOG.md @@ -4,7 +4,7 @@ All notable changes to this project will be documented in this file. The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.1.0/). -## [Unreleased] +## [3.6.0] - 2024-12-19 ### Added - Automatically substitute chat templates that are not compatible with Jinja2Cpp in GGUFs ([#3327](https://github.com/nomic-ai/gpt4all/pull/3327)) @@ -239,6 +239,7 @@ The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.1.0/). - Fix several Vulkan resource management issues ([#2694](https://github.com/nomic-ai/gpt4all/pull/2694)) - Fix crash/hang when some models stop generating, by showing special tokens ([#2701](https://github.com/nomic-ai/gpt4all/pull/2701)) +[3.6.0]: https://github.com/nomic-ai/gpt4all/compare/v3.5.3...v3.6.0 [3.5.3]: https://github.com/nomic-ai/gpt4all/compare/v3.5.2...v3.5.3 [3.5.2]: https://github.com/nomic-ai/gpt4all/compare/v3.5.1...v3.5.2 [3.5.1]: https://github.com/nomic-ai/gpt4all/compare/v3.5.0...v3.5.1 diff --git a/gpt4all-chat/CMakeLists.txt b/gpt4all-chat/CMakeLists.txt index 7947b2d4b91c..d3cf83b8ce98 100644 --- a/gpt4all-chat/CMakeLists.txt +++ b/gpt4all-chat/CMakeLists.txt @@ -3,10 +3,10 @@ cmake_minimum_required(VERSION 3.25) # for try_compile SOURCE_FROM_VAR include(../common/common.cmake) set(APP_VERSION_MAJOR 3) -set(APP_VERSION_MINOR 5) -set(APP_VERSION_PATCH 4) +set(APP_VERSION_MINOR 6) +set(APP_VERSION_PATCH 0) set(APP_VERSION_BASE "${APP_VERSION_MAJOR}.${APP_VERSION_MINOR}.${APP_VERSION_PATCH}") -set(APP_VERSION "${APP_VERSION_BASE}-dev0") +set(APP_VERSION "${APP_VERSION_BASE}") project(gpt4all VERSION ${APP_VERSION_BASE} LANGUAGES CXX C) diff --git a/gpt4all-chat/metadata/models3.json b/gpt4all-chat/metadata/models3.json index 7a16a2556a0a..3b83d5ac539a 100644 --- a/gpt4all-chat/metadata/models3.json +++ b/gpt4all-chat/metadata/models3.json @@ -5,7 +5,7 @@ "name": "Reasoner v1", "filename": "qwen2.5-coder-7b-instruct-q4_0.gguf", "filesize": "4431390720", - "requires": "3.5.4-dev0", + "requires": "3.6.0", "ramrequired": "8", "parameters": "8 billion", "quant": "q4_0",