o
    r)iS                     @  sD  d dl mZ d dlmZ d dlmZ d dlZddlmZ ddl	m
Z
mZ ddlmZmZmZmZmZmZmZ dd	lmZmZ dd
lmZ ddlmZmZ ddlmZmZ ddlm Z m!Z! ddl"m#Z# ddl$m%Z%m&Z& ddl'm(Z( ddgZ)G dd deZ*G dd deZ+G dd dZ,G dd dZ-G dd dZ.G dd dZ/dS )    )annotations)Optional)LiteralN   )_legacy_response)batch_list_paramsbatch_create_params)BodyOmitQueryHeadersNotGivenomit	not_given)maybe_transformasync_maybe_transform)cached_property)SyncAPIResourceAsyncAPIResource)to_streamed_response_wrapper"async_to_streamed_response_wrapper)SyncCursorPageAsyncCursorPage)Batch)AsyncPaginatormake_request_options)MetadataBatchesAsyncBatchesc                   @     e Zd ZdZed.ddZed/ddZeed	d	d	ed
d0ddZ	d	d	d	ed d1d"d#Z
eed	d	d	ed$d2d*d+Zd	d	d	ed d1d,d-Zd	S )3r   ;Create large batches of API requests to run asynchronously.returnBatchesWithRawResponsec                 C     t | S a  
        This property can be used as a prefix for any HTTP method call to return
        the raw response object instead of the parsed content.

        For more information, see https://www.github.com/openai/openai-python#accessing-raw-response-data-eg-headers
        )r"   self r'   \/var/www/html/voicebot/backend/venv/lib/python3.10/site-packages/openai/resources/batches.pywith_raw_response      zBatches.with_raw_responseBatchesWithStreamingResponsec                 C  r#   z
        An alternative to `.with_raw_response` that doesn't eagerly read the response body.

        For more information, see https://www.github.com/openai/openai-python#with_streaming_response
        )r+   r%   r'   r'   r(   with_streaming_response&      zBatches.with_streaming_responseNmetadataoutput_expires_afterextra_headersextra_query
extra_bodytimeoutcompletion_windowLiteral['24h']endpointLiteral['/v1/responses', '/v1/chat/completions', '/v1/embeddings', '/v1/completions', '/v1/moderations', '/v1/images/generations', '/v1/images/edits']input_file_idstrr0   Optional[Metadata] | Omitr1   -batch_create_params.OutputExpiresAfter | Omitr2   Headers | Noner3   Query | Noner4   Body | Noner5   'float | httpx.Timeout | None | NotGivenr   c       	   
   	   C  s2   | j dt|||||dtjt||||	dtdS )  
        Creates and executes a batch from an uploaded file of requests

        Args:
          completion_window: The time frame within which the batch should be processed. Currently only `24h`
              is supported.

          endpoint: The endpoint to be used for all requests in the batch. Currently
              `/v1/responses`, `/v1/chat/completions`, `/v1/embeddings`, `/v1/completions`,
              `/v1/moderations`, `/v1/images/generations`, and `/v1/images/edits` are
              supported. Note that `/v1/embeddings` batches are also restricted to a maximum
              of 50,000 embedding inputs across all requests in the batch.

          input_file_id: The ID of an uploaded file that contains requests for the new batch.

              See [upload file](https://platform.openai.com/docs/api-reference/files/create)
              for how to upload a file.

              Your input file must be formatted as a
              [JSONL file](https://platform.openai.com/docs/api-reference/batch/request-input),
              and must be uploaded with the purpose `batch`. The file can contain up to 50,000
              requests, and can be up to 200 MB in size.

          metadata: Set of 16 key-value pairs that can be attached to an object. This can be useful
              for storing additional information about the object in a structured format, and
              querying for objects via API or the dashboard.

              Keys are strings with a maximum length of 64 characters. Values are strings with
              a maximum length of 512 characters.

          output_expires_after: The expiration policy for the output and/or error file that are generated for a
              batch.

          extra_headers: Send extra headers

          extra_query: Add additional query parameters to the request

          extra_body: Add additional JSON properties to the request

          timeout: Override the client-level default timeout for this request, in seconds
        /batchesr6   r8   r:   r0   r1   r2   r3   r4   r5   bodyoptionscast_to)_postr   r   BatchCreateParamsr   r   
r&   r6   r8   r:   r0   r1   r2   r3   r4   r5   r'   r'   r(   create/   s    @
zBatches.createrE   batch_idc                C  s4   |s	t d|| jd| t||||dtdS )F  
        Retrieves a batch.

        Args:
          extra_headers: Send extra headers

          extra_query: Add additional query parameters to the request

          extra_body: Add additional JSON properties to the request

          timeout: Override the client-level default timeout for this request, in seconds
        7Expected a non-empty value for `batch_id` but received 	/batches/rE   rH   rI   
ValueError_getr   r   r&   rN   r2   r3   r4   r5   r'   r'   r(   retrieve   s   zBatches.retrieveafterlimitr2   r3   r4   r5   rY   
str | OmitrZ   
int | OmitSyncCursorPage[Batch]c                C  2   | j dtt t||||t||dtjdtdS a,  List your organization's batches.

        Args:
          after: A cursor for use in pagination.

        `after` is an object ID that defines your place
              in the list. For instance, if you make a list request and receive 100 objects,
              ending with obj_foo, your subsequent call can include after=obj_foo in order to
              fetch the next page of the list.

          limit: A limit on the number of objects to be returned. Limit can range between 1 and
              100, and the default is 20.

          extra_headers: Send extra headers

          extra_query: Add additional query parameters to the request

          extra_body: Add additional JSON properties to the request

          timeout: Override the client-level default timeout for this request, in seconds
        rC   )rY   rZ   )r2   r3   r4   r5   query)pagerH   model)_get_api_listr   r   r   r   r   BatchListParamsr&   rY   rZ   r2   r3   r4   r5   r'   r'   r(   list   "   !zBatches.listc                C  s6   |s	t d|| jd| dt||||dtdS )
  Cancels an in-progress batch.

        The batch will be in status `cancelling` for up to
        10 minutes, before changing to `cancelled`, where it will have partial results
        (if any) available in the output file.

        Args:
          extra_headers: Send extra headers

          extra_query: Add additional query parameters to the request

          extra_body: Add additional JSON properties to the request

          timeout: Override the client-level default timeout for this request, in seconds
        rP   rQ   /cancelrE   rR   rT   rJ   r   r   rV   r'   r'   r(   cancel   s   
zBatches.cancel)r!   r"   )r!   r+   r6   r7   r8   r9   r:   r;   r0   r<   r1   r=   r2   r>   r3   r?   r4   r@   r5   rA   r!   r   rN   r;   r2   r>   r3   r?   r4   r@   r5   rA   r!   r   )rY   r[   rZ   r\   r2   r>   r3   r?   r4   r@   r5   rA   r!   r]   __name__
__module____qualname____doc__r   r)   r-   r   r   rM   rW   rf   rk   r'   r'   r'   r(   r      <    	X$:c                   @  r   )3r   r    r!   AsyncBatchesWithRawResponsec                 C  r#   r$   )rt   r%   r'   r'   r(   r)      r*   zAsyncBatches.with_raw_response!AsyncBatchesWithStreamingResponsec                 C  r#   r,   )ru   r%   r'   r'   r(   r-     r.   z$AsyncBatches.with_streaming_responseNr/   r6   r7   r8   r9   r:   r;   r0   r<   r1   r=   r2   r>   r3   r?   r4   r@   r5   rA   r   c       	   
   	     s@   | j dt|||||dtjI dH t||||	dtdI dH S )rB   rC   rD   NrE   rF   )rJ   r   r   rK   r   r   rL   r'   r'   r(   rM     s"   @
zAsyncBatches.createrE   rN   c                  s<   |s
t d|| jd| t||||dtdI dH S )rO   rP   rQ   rE   rR   NrS   rV   r'   r'   r(   rW   c  s   zAsyncBatches.retrieverX   rY   r[   rZ   r\   -AsyncPaginator[Batch, AsyncCursorPage[Batch]]c                C  r^   r_   )rc   r   r   r   r   r   rd   re   r'   r'   r(   rf     rg   zAsyncBatches.listc                  s>   |s
t d|| jd| dt||||dtdI dH S )rh   rP   rQ   ri   rE   rR   Nrj   rV   r'   r'   r(   rk     s   
zAsyncBatches.cancel)r!   rt   )r!   ru   rl   rm   )rY   r[   rZ   r\   r2   r>   r3   r?   r4   r@   r5   rA   r!   rv   rn   r'   r'   r'   r(   r      rs   c                   @     e Zd ZdddZdS )	r"   batchesr   r!   Nonec                 C  B   || _ t|j| _t|j| _t|j| _t|j| _d S N)_batchesr   to_raw_response_wrapperrM   rW   rf   rk   r&   rx   r'   r'   r(   __init__     
zBatchesWithRawResponse.__init__Nrx   r   r!   ry   ro   rp   rq   r   r'   r'   r'   r(   r"         r"   c                   @  rw   )	rt   rx   r   r!   ry   c                 C  rz   r{   )r|   r   async_to_raw_response_wrapperrM   rW   rf   rk   r~   r'   r'   r(   r     r   z$AsyncBatchesWithRawResponse.__init__Nrx   r   r!   ry   r   r'   r'   r'   r(   rt     r   rt   c                   @  rw   )	r+   rx   r   r!   ry   c                 C  :   || _ t|j| _t|j| _t|j| _t|j| _d S r{   )r|   r   rM   rW   rf   rk   r~   r'   r'   r(   r        
z%BatchesWithStreamingResponse.__init__Nr   r   r'   r'   r'   r(   r+     r   r+   c                   @  rw   )	ru   rx   r   r!   ry   c                 C  r   r{   )r|   r   rM   rW   rf   rk   r~   r'   r'   r(   r     r   z*AsyncBatchesWithStreamingResponse.__init__Nr   r   r'   r'   r'   r(   ru     r   ru   )0
__future__r   typingr   typing_extensionsr   httpx r   typesr   r   _typesr	   r
   r   r   r   r   r   _utilsr   r   _compatr   	_resourcer   r   	_responser   r   
paginationr   r   types.batchr   _base_clientr   r   types.shared_params.metadatar   __all__r   r   r"   rt   r+   ru   r'   r'   r'   r(   <module>   s0   $ c c