mirror of
https://github.com/deepset-ai/haystack.git
synced 2025-07-24 17:30:38 +00:00
parent
1dc7f6215e
commit
dd37b4c29f
8
.github/workflows/linting.yml
vendored
8
.github/workflows/linting.yml
vendored
@ -13,6 +13,10 @@ jobs:
|
||||
steps:
|
||||
- name: Checkout
|
||||
uses: actions/checkout@v3
|
||||
with:
|
||||
# With the default value of 1, there are corner cases where tj-actions/changed-files
|
||||
# fails with a `no merge base` error
|
||||
fetch-depth: 0
|
||||
|
||||
- name: Get changed files
|
||||
id: files
|
||||
@ -38,6 +42,10 @@ jobs:
|
||||
steps:
|
||||
- name: Checkout
|
||||
uses: actions/checkout@v3
|
||||
with:
|
||||
# With the default value of 1, there are corner cases where tj-actions/changed-files
|
||||
# fails with a `no merge base` error
|
||||
fetch-depth: 0
|
||||
|
||||
- name: Get changed files
|
||||
id: files
|
||||
|
@ -766,7 +766,7 @@ class PromptNode(BaseComponent):
|
||||
:param use_auth_token: The authentication token to use for the model.
|
||||
:param use_gpu: Whether to use GPU or not.
|
||||
:param devices: The devices to use for the model.
|
||||
:param top_k: The number of independently generated texts to return per prompt. For example, if you set top_k=3, the model's going to generate three answers to the query.
|
||||
:param top_k: The number of independently generated texts to return per prompt. For example, if you set top_k=3, the model's going to generate three answers to the query.
|
||||
:param stop_words: Stops text generation if any of the stop words is generated.
|
||||
:param model_kwargs: Additional keyword arguments passed when loading the model specified in `model_name_or_path`.
|
||||
"""
|
||||
|
Loading…
x
Reference in New Issue
Block a user