diff --git a/README.md b/README.md index c07a9f2e..01113cb5 100644 --- a/README.md +++ b/README.md @@ -197,11 +197,11 @@ aicommits config set proxy= #### model -Default: `gpt-3.5-turbo` +Default: `gpt-4o-mini` The Chat Completions (`/v1/chat/completions`) model to use. Consult the list of models available in the [OpenAI Documentation](https://platform.openai.com/docs/models/model-endpoint-compatibility). -> Tip: If you have access, try upgrading to [`gpt-4`](https://platform.openai.com/docs/models/gpt-4) for next-level code analysis. It can handle double the input size, but comes at a higher cost. Check out OpenAI's website to learn more. +> Tip: If budget allows, consider upgrading to [`gpt-4o`](https://platform.openai.com/docs/models/gpt-4o) for next-level code analysis. It offers enhanced capabilities but comes at a higher cost. Check out OpenAI's website to learn more. #### timeout diff --git a/package.json b/package.json index 4eb26f13..111e2c7d 100644 --- a/package.json +++ b/package.json @@ -26,7 +26,7 @@ "prepack": "pnpm build && clean-pkg-json" }, "dependencies": { - "@dqbd/tiktoken": "^1.0.2" + "@dqbd/tiktoken": "^1.0.15" }, "devDependencies": { "@clack/prompts": "^0.7.0", diff --git a/pnpm-lock.yaml b/pnpm-lock.yaml index 4f5069d5..6aedd6d4 100644 --- a/pnpm-lock.yaml +++ b/pnpm-lock.yaml @@ -6,8 +6,8 @@ settings: dependencies: '@dqbd/tiktoken': - specifier: ^1.0.2 - version: 1.0.7 + specifier: ^1.0.15 + version: 1.0.15 devDependencies: '@clack/prompts': @@ -100,8 +100,8 @@ packages: bundledDependencies: - is-unicode-supported - /@dqbd/tiktoken@1.0.7: - resolution: {integrity: sha512-bhR5k5W+8GLzysjk8zTMVygQZsgvf7W1F0IlL4ZQ5ugjo5rCyiwGM5d8DYriXspytfu98tv59niang3/T+FoDw==} + /@dqbd/tiktoken@1.0.15: + resolution: {integrity: sha512-a6I67K1xUkuqcuwulobIJiLikkoE7egMaviI1Jg5bxSn2V7QGqXsGE3jTKr8UIOU/o74mAAd5TkeXFNBtaKF4A==} dev: false /@esbuild/android-arm64@0.18.20: diff --git a/src/utils/config.ts b/src/utils/config.ts index 0e07a59f..fd125451 100644 --- a/src/utils/config.ts +++ b/src/utils/config.ts @@ -82,7 +82,7 @@ const configParsers = { }, model(model?: string) { if (!model || model.length === 0) { - return 'gpt-3.5-turbo'; + return 'gpt-4o-mini'; } return model as TiktokenModel; diff --git a/tests/specs/openai/conventional-commits.ts b/tests/specs/openai/conventional-commits.ts index 7a3e43e6..90dbba1b 100644 --- a/tests/specs/openai/conventional-commits.ts +++ b/tests/specs/openai/conventional-commits.ts @@ -140,7 +140,7 @@ export default testSuite(({ describe }) => { } as ValidConfig; const commitMessages = await generateCommitMessage( OPENAI_KEY!, - 'gpt-3.5-turbo', + 'gpt-4o-mini', config.locale, gitDiff, config.generate,