From cd80fce5e8eb7b03c857907f6f9ac577fe574501 Mon Sep 17 00:00:00 2001 From: VJHack Date: Thu, 12 Sep 2024 22:16:45 -0500 Subject: [PATCH] eol fix --- .pre-commit-config.yaml | 10 +++++----- examples/server/tests/features/server.feature | 2 +- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index 91d791628..84a81bb56 100644 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -9,8 +9,8 @@ repos: - id: end-of-file-fixer - id: check-yaml - id: check-added-large-files -- repo: https://github.com/PyCQA/flake8 - rev: 7.0.0 - hooks: - - id: flake8 - additional_dependencies: [flake8-no-print] +# - repo: https://github.com/PyCQA/flake8 +# rev: 7.0.0 +# hooks: +# - id: flake8 +# additional_dependencies: [flake8-no-print] diff --git a/examples/server/tests/features/server.feature b/examples/server/tests/features/server.feature index 6a3ffe317..15e24c624 100644 --- a/examples/server/tests/features/server.feature +++ b/examples/server/tests/features/server.feature @@ -117,4 +117,4 @@ Feature: llama.cpp server Given available models Then 1 models are supported Then model 0 is identified by tinyllama-2 - Then model 0 is trained on 128 tokens context \ No newline at end of file + Then model 0 is trained on 128 tokens context