fix where the tokens are reset (#175)
This commit is contained in:
@@ -483,9 +483,6 @@ class WhisperModel:
|
|||||||
|
|
||||||
seek += segment_size
|
seek += segment_size
|
||||||
|
|
||||||
if not options.condition_on_previous_text or temperature > 0.5:
|
|
||||||
prompt_reset_since = len(all_tokens)
|
|
||||||
|
|
||||||
if options.word_timestamps:
|
if options.word_timestamps:
|
||||||
self.add_word_timestamps(
|
self.add_word_timestamps(
|
||||||
current_segments,
|
current_segments,
|
||||||
@@ -538,6 +535,9 @@ class WhisperModel:
|
|||||||
),
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
|
if not options.condition_on_previous_text or temperature > 0.5:
|
||||||
|
prompt_reset_since = len(all_tokens)
|
||||||
|
|
||||||
def encode(self, features: np.ndarray) -> ctranslate2.StorageView:
|
def encode(self, features: np.ndarray) -> ctranslate2.StorageView:
|
||||||
# When the model is running on multiple GPUs, the encoder output should be moved
|
# When the model is running on multiple GPUs, the encoder output should be moved
|
||||||
# to the CPU since we don't know which GPU will handle the next job.
|
# to the CPU since we don't know which GPU will handle the next job.
|
||||||
|
|||||||
Reference in New Issue
Block a user