File size: 2,633 Bytes
a5d2e48
 
04d664a
 
a5d2e48
 
 
23a3cae
 
a5d2e48
23a3cae
 
a5d2e48
 
 
 
 
04d664a
a5d2e48
 
 
04d664a
a5d2e48
 
 
 
 
 
04d664a
a5d2e48
 
 
23a3cae
 
a5d2e48
 
 
 
 
04d664a
a5d2e48
 
04d664a
a5d2e48
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
"""OpenAI's handling of `response_format` and `timestamp_granularities` is a bit confusing and inconsistent. This test module exists to capture the OpenAI API's behavior with respect to these parameters."""  # noqa: E501

from pathlib import Path

from openai import AsyncOpenAI, BadRequestError
import pytest

from faster_whisper_server.api_models import TIMESTAMP_GRANULARITIES_COMBINATIONS, TimestampGranularities


@pytest.mark.asyncio
@pytest.mark.requires_openai
@pytest.mark.parametrize("timestamp_granularities", TIMESTAMP_GRANULARITIES_COMBINATIONS)
async def test_openai_json_response_format_and_timestamp_granularities_combinations(
    actual_openai_client: AsyncOpenAI,
    timestamp_granularities: TimestampGranularities,
) -> None:
    file_path = Path("audio.wav")
    if "word" in timestamp_granularities:
        with pytest.raises(BadRequestError):
            await actual_openai_client.audio.transcriptions.create(
                file=file_path,
                model="whisper-1",
                response_format="json",
                timestamp_granularities=timestamp_granularities,
            )
    else:
        await actual_openai_client.audio.transcriptions.create(
            file=file_path, model="whisper-1", response_format="json", timestamp_granularities=timestamp_granularities
        )


@pytest.mark.asyncio
@pytest.mark.requires_openai
@pytest.mark.parametrize("timestamp_granularities", TIMESTAMP_GRANULARITIES_COMBINATIONS)
async def test_openai_verbose_json_response_format_and_timestamp_granularities_combinations(
    actual_openai_client: AsyncOpenAI,
    timestamp_granularities: TimestampGranularities,
) -> None:
    file_path = Path("audio.wav")

    transcription = await actual_openai_client.audio.transcriptions.create(
        file=file_path,
        model="whisper-1",
        response_format="verbose_json",
        timestamp_granularities=timestamp_granularities,
    )

    assert transcription.__pydantic_extra__
    if timestamp_granularities == ["word"]:
        # This is an exception where segments are not present
        assert transcription.__pydantic_extra__.get("segments") is None
        assert transcription.__pydantic_extra__.get("words") is not None
    elif "word" in timestamp_granularities:
        assert transcription.__pydantic_extra__.get("segments") is not None
        assert transcription.__pydantic_extra__.get("words") is not None
    else:
        # Unless explicitly requested, words are not present
        assert transcription.__pydantic_extra__.get("segments") is not None
        assert transcription.__pydantic_extra__.get("words") is None