Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Snyk] Security upgrade transformers from 4.30.2 to 4.36.0 #14

Open
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

abdulrahman305
Copy link
Owner

@abdulrahman305 abdulrahman305 commented Dec 21, 2023

This PR was automatically created by Snyk using the credentials of a real user.


Snyk has created this PR to fix one or more vulnerable packages in the `pip` dependencies of this project.

Changes included in this PR

  • Changes to the following files to upgrade the vulnerable dependencies to a fixed version:
    • gpt4all-training/requirements.txt
⚠️ Warning
torch 1.13.1 requires nvidia-cudnn-cu11, which is not installed.
torch 1.13.1 requires nvidia-cuda-nvrtc-cu11, which is not installed.
torch 1.13.1 requires nvidia-cuda-runtime-cu11, which is not installed.
torch 1.13.1 requires nvidia-cublas-cu11, which is not installed.
scikit-learn 1.0.2 requires scipy, which is not installed.
multiprocess 0.70.15 has requirement dill>=0.3.7, but you have dill 0.3.6.

Vulnerabilities that will be fixed

By pinning:
Severity Priority Score (*) Issue Upgrade Breaking Change Exploit Maturity
critical severity 873/1000
Why? Proof of Concept exploit, Recently disclosed, Has a fix available, CVSS 9.6
Deserialization of Untrusted Data
SNYK-PYTHON-TRANSFORMERS-6134594
transformers:
4.30.2 -> 4.36.0
No Proof of Concept
critical severity 843/1000
Why? Proof of Concept exploit, Recently disclosed, Has a fix available, CVSS 9
Deserialization of Untrusted Data
SNYK-PYTHON-TRANSFORMERS-6135747
transformers:
4.30.2 -> 4.36.0
No Proof of Concept

(*) Note that the real score may have changed since the PR was raised.

Some vulnerabilities couldn't be fully fixed and so Snyk will still find them when the project is tested again. This may be because the vulnerability existed within more than one direct dependency, but not all of the affected dependencies could be upgraded.

Check the changes in this PR to ensure they won't cause issues with your project.


Note: You are seeing this because you or someone else with access to this repository has authorized Snyk to open fix PRs.

For more information:
🧐 View latest project report

🛠 Adjust project settings

📚 Read more about Snyk's upgrade and patch logic


Learn how to fix vulnerabilities with free interactive lessons:

🦉 Deserialization of Untrusted Data

Summary by CodeRabbit

  • Refactor
    • Updated the minimum required version of the "transformers" package to ensure compatibility with the latest features and improvements.

Copy link

coderabbitai bot commented Dec 21, 2023

Walkthrough

The update involves a simple yet significant change in the version requirement for the "transformers" package, increasing the minimum version from 4.28.0 to 4.36.0. This change hints at the need to leverage newer features or fixes introduced in the later versions of the package.

Changes

File Path Change Summary
gpt4all-training/requirements.txt Updated the "transformers" package version requirement from ">=4.28.0" to ">=4.36.0".

🐇✨
To code anew, we hop with glee,
With transformers fresh, we're bug-free!
Version leaps, from old to bold,
In lines of code, our tales are told. 🌟📜

Thank you for using CodeRabbit. We offer it for free to the OSS community and would appreciate your support in helping us grow. If you find it useful, would you consider giving us a shout-out on X ?


Tips

Chat with CodeRabbit Bot (@coderabbitai)

  • You can reply to a review comment made by CodeRabbit.
  • You can tag CodeRabbit on specific lines of code or files in the PR by tagging @coderabbitai in a comment.
  • You can tag @coderabbitai in a PR comment and ask one-off questions about the PR and the codebase. Use quoted replies to pass the context for follow-up questions.

CodeRabbit Commands (invoked as PR comments)

  • @coderabbitai pause to pause the reviews on a PR.
  • @coderabbitai resume to resume the paused reviews.
  • @coderabbitai review to trigger a review. This is useful when automatic reviews are disabled for the repository.
  • @coderabbitai resolve resolve all the CodeRabbit review comments.
  • @coderabbitai help to get help.

Additionally, you can add @coderabbitai ignore anywhere in the PR description to prevent this PR from being reviewed.

CodeRabbit Configration File (.coderabbit.yaml)

  • You can programmatically configure CodeRabbit by adding a .coderabbit.yaml file to the root of your repository.
  • The JSON schema for the configuration file is available here.
  • If your editor has YAML language server enabled, you can add the path at the top of this file to enable auto-completion and validation: # yaml-language-server: $schema=https://coderabbit.ai/integrations/coderabbit-overrides.v2.json

Copy link

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Review Status

Actionable comments generated: 0

Configuration used: CodeRabbit UI

Commits Files that changed from the base of the PR and between 0e20f61 and 4cadbb8.
Files selected for processing (1)
  • gpt4all-training/requirements.txt (1 hunks)
Files skipped from review due to trivial changes (1)
  • gpt4all-training/requirements.txt

@gitauto-ai gitauto-ai bot added the gitauto label Oct 4, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants