
     `i                     `   d dl Z d dlZd dlmZ d dlmZ d dlmZ d dlm	Z	m
Z
 d dlZd dlZddlmZmZmZmZmZmZmZmZmZmZmZ ddlmZmZmZmZmZm Z   e            rqd dl!Z"d dl#Z"e"j$        j%        Z& e            rQd d	l'm(Z( e&j)        e(j*        e&j+        e(j+        e&j,        e(j,        e&j-        e(j-        e&j.        e(j.        e&j/        e(j/        iZ0ni Z0 e            rd dl1Z1 ej2        e3          Z4e
d
ej5        de6d
         e6ej5                 e6d         f         Z7 G d de          Z8 G d de          Z9 G d de          Z:e;e<e
e=e<e6e;         f         f         Z>d Z? G d de          Z@d ZAd ZBde6fdZCd ZDd ZEd ZFdej5        deGfdZHdUd e=de6e7         fd!ZI	 dUde
e6e7         e7f         d e=de7fd"ZJ	 dUde
e6e7         e7f         d e=de6e7         fd#ZKdej5        fd$ZL	 dVdej5        d%e	e
e=eMe=d&f         f                  de8fd'ZN	 dVdej5        d(e	e
e8e<f                  de=fd)ZOdVdej5        d*e	e8         deMe=e=f         fd+ZPd,eMe=e=f         d-e=d.e=deMe=e=f         fd/ZQd0e;e<e
e6eMf         f         deGfd1ZRd0e;e<e
e6eMf         f         deGfd2ZSd3ee;e<e
e6eMf         f                  deGfd4ZTd3ee;e<e
e6eMf         f                  deGfd5ZUdVde
e<d
f         d6e	eV         dd
fd7ZW	 dVde
e6eMe<d
f         d6e	eV         de
d
e6d
         e6e6d
                  f         fd8ZX	 	 	 	 	 	 	 	 	 	 	 	 	 dWd9e	eG         d:e	eV         d;e	eG         d<e	e
eVe6eV         f                  d=e	e
eVe6eV         f                  d>e	eG         d?e	e
e;e<e=f         e=f                  d@e	eG         dAe	e;e<e=f                  dBe	eG         dCe	e;e<e=f                  dDe	dE         dFe	dG         fdHZY G dI dJ          ZZdKe9dLeMe9d&f         d3e6e;         ddfdMZ[dNe6e<         dOe6e<         fdPZ\ edQR           G dS dT                      Z]dS )X    N)Iterable)	dataclass)BytesIO)OptionalUnion   )ExplicitEnumis_jax_tensoris_numpy_arrayis_tf_tensoris_torch_availableis_torch_tensoris_torchvision_availableis_vision_availableloggingrequires_backendsto_numpy)IMAGENET_DEFAULT_MEANIMAGENET_DEFAULT_STDIMAGENET_STANDARD_MEANIMAGENET_STANDARD_STDOPENAI_CLIP_MEANOPENAI_CLIP_STD)InterpolationModezPIL.Image.Imageztorch.Tensorc                       e Zd ZdZdZdS )ChannelDimensionchannels_firstchannels_lastN)__name__
__module____qualname__FIRSTLAST     l/home/jaya/work/projects/VOICE-AGENT/VIET/agent-env/lib/python3.11/site-packages/transformers/image_utils.pyr   r   Q   s        EDDDr%   r   c                       e Zd ZdZdZdS )AnnotationFormatcoco_detectioncoco_panopticN)r   r    r!   COCO_DETECTIONCOCO_PANOPTICr$   r%   r&   r(   r(   V   s        %N#MMMr%   r(   c                   >    e Zd Zej        j        Zej        j        ZdS )AnnotionFormatN)r   r    r!   r(   r+   valuer,   r$   r%   r&   r.   r.   [   s$        %4:N$28MMMr%   r.   c                 \    t                      ot          | t          j        j                  S N)r   
isinstancePILImageimgs    r&   is_pil_imager7   c   s!      EZSY_%E%EEr%   c                   "    e Zd ZdZdZdZdZdZdS )	ImageTypepillowtorchnumpy
tensorflowjaxN)r   r    r!   r3   TORCHNUMPY
TENSORFLOWJAXr$   r%   r&   r9   r9   g   s'        
CEEJ
CCCr%   r9   c                 N   t          |           rt          j        S t          |           rt          j        S t          |           rt          j        S t          |           rt          j        S t          |           rt          j
        S t          dt          |                      )NzUnrecognized image type )r7   r9   r3   r   r?   r   r@   r   rA   r
   rB   
ValueErrortypeimages    r&   get_image_typerH   o   s    E }u e E $##U }
=U==
>
>>r%   c                     t          |           p;t          |           p,t          |           pt          |           pt	          |           S r1   )r7   r   r   r   r
   r5   s    r&   is_valid_imagerJ   }   sG    vs 3 3vs7K7Kv|\_O`O`vdqrudvdvvr%   imagesc                 8    | ot          d | D                       S )Nc              3   4   K   | ]}t          |          V  d S r1   )rJ   .0rG   s     r&   	<genexpr>z*is_valid_list_of_images.<locals>.<genexpr>   s*      DDE.//DDDDDDr%   all)rK   s    r&   is_valid_list_of_imagesrS      s$    DcDDVDDDDDDr%   c                 ,   t          | d         t                    rd | D             S t          | d         t          j                  rt          j        | d          S t          | d         t
          j                  rt          j        | d          S d S )Nr   c                     g | ]	}|D ]}|
S r$   r$   )rO   sublistitems      r&   
<listcomp>z$concatenate_list.<locals>.<listcomp>   s%    CCC7CC4CCCCr%   axis)dim)r2   listnpndarrayconcatenater;   Tensorcat)
input_lists    r&   concatenate_listrc      s    *Q-&& ,CCJCCCC	JqM2:	.	. ,~jq1111	JqM5<	0	0 ,y++++, ,r%   c                     t          | t          t          f          r| D ]}t          |          s dS nt	          |           sdS dS )NFT)r2   r\   tuplevalid_imagesrJ   )imgsr6   s     r&   rf   rf      sd    $u&&  	 	C$$ uu	 D!! u4r%   c                 h    t          | t          t          f          rt          | d                   S dS )Nr   F)r2   r\   re   rJ   r5   s    r&   
is_batchedri      s/    #e}%% &c!f%%%5r%   rG   returnc                     | j         t          j        k    rdS t          j        |           dk    ot          j        |           dk    S )zV
    Checks to see whether the pixel values have already been rescaled to [0, 1].
    Fr   r   )dtyper]   uint8minmaxrF   s    r&   is_scaled_imagerp      s?     {bhu 6%==A4"&--1"44r%      expected_ndimsc           	      F   t          |           r| S t          |           r| gS t          |           rP| j        |dz   k    rt	          |           } n0| j        |k    r| g} n!t          d|dz    d| d| j         d          | S t          dt          |            d          )a  
    Ensure that the output is a list of images. If the input is a single image, it is converted to a list of length 1.
    If the input is a batch of images, it is converted to a list of images.

    Args:
        images (`ImageInput`):
            Image of images to turn into a list of images.
        expected_ndims (`int`, *optional*, defaults to 3):
            Expected number of dimensions for a single input image. If the input image has a different number of
            dimensions, an error is raised.
    r   z%Invalid image shape. Expected either z or z dimensions, but got z dimensions.ztInvalid image type. Expected either PIL.Image.Image, numpy.ndarray, torch.Tensor, tf.Tensor or jax.ndarray, but got .)ri   r7   rJ   ndimr\   rD   rE   rK   rr   s     r&   make_list_of_imagesrw      s     &  F xf ;.1,,,&\\FF[N**XFF.8J . .P^ . .K. . .   
	0 $V	0 	0 	0  r%   c                 R   t          | t          t          f          r>t          d | D                       r%t          d | D                       rd | D             S t          | t          t          f          rWt	          |           rHt          | d                   s| d         j        |k    r| S | d         j        |dz   k    rd | D             S t          |           r:t          |           s| j        |k    r| gS | j        |dz   k    rt          |           S t          d|            )a  
    Ensure that the output is a flat list of images. If the input is a single image, it is converted to a list of length 1.
    If the input is a nested list of images, it is converted to a flat list of images.
    Args:
        images (`Union[list[ImageInput], ImageInput]`):
            The input image.
        expected_ndims (`int`, *optional*, defaults to 3):
            The expected number of dimensions for a single input image.
    Returns:
        list: A list of images or a 4d array of images.
    c              3   N   K   | ] }t          |t          t          f          V  !d S r1   r2   r\   re   rO   images_is     r&   rP   z+make_flat_list_of_images.<locals>.<genexpr>   0      KK
8dE]33KKKKKKr%   c              3   :   K   | ]}t          |          p| V  d S r1   rS   r{   s     r&   rP   z+make_flat_list_of_images.<locals>.<genexpr>   2      YYh'11A\YYYYYYr%   c                     g | ]	}|D ]}|
S r$   r$   rO   img_listr6   s      r&   rX   z,make_flat_list_of_images.<locals>.<listcomp>   s%    ???h??s????r%   r   r   c                     g | ]	}|D ]}|
S r$   r$   r   s      r&   rX   z,make_flat_list_of_images.<locals>.<listcomp>   s%    CCCH(CC3CCCCCr%   z*Could not make a flat list of images from 	r2   r\   re   rR   rS   r7   ru   rJ   rD   rv   s     r&   make_flat_list_of_imagesr      sX   " 	6D%=))@KKFKKKKK@ YYRXYYYYY@
 @?F????&4-(( D-DV-L-L Dq	"" 	fQin&F&FM!9>^a///CCCCCCf   	6;.#@#@8O;.1,,,<<
J&JJ
K
KKr%   c                 >   t          | t          t          f          r4t          d | D                       rt          d | D                       r| S t          | t          t          f          rXt	          |           rIt          | d                   s| d         j        |k    r| gS | d         j        |dz   k    rd | D             S t          |           r<t          |           s| j        |k    r| ggS | j        |dz   k    rt          |           gS t          d          )as  
    Ensure that the output is a nested list of images.
    Args:
        images (`Union[list[ImageInput], ImageInput]`):
            The input image.
        expected_ndims (`int`, *optional*, defaults to 3):
            The expected number of dimensions for a single input image.
    Returns:
        list: A list of list of images or a list of 4d array of images.
    c              3   N   K   | ] }t          |t          t          f          V  !d S r1   rz   r{   s     r&   rP   z-make_nested_list_of_images.<locals>.<genexpr>	  r}   r%   c              3   :   K   | ]}t          |          p| V  d S r1   r   r{   s     r&   rP   z-make_nested_list_of_images.<locals>.<genexpr>
  r   r%   r   r   c                 ,    g | ]}t          |          S r$   )r\   rN   s     r&   rX   z.make_nested_list_of_images.<locals>.<listcomp>  s    444EDKK444r%   z]Invalid input type. Must be a single image, a list of images, or a list of batches of images.r   rv   s     r&   make_nested_list_of_imagesr      sF     	6D%=))KKFKKKKK YYRXYYYYY
  &4-(( 5-DV-L-L 5q	"" 	fQin&F&F8O!9>^a///44V4444 f " 	6;.#@#@H:;.1,,,LL>!
t
u
uur%   c                     t          |           st          dt          |                      t                      r3t	          | t
          j        j                  rt          j        |           S t          |           S )NzInvalid image type: )
rJ   rD   rE   r   r2   r3   r4   r]   arrayr   r5   s    r&   to_numpy_arrayr     sl    # =;S		;;<<< C!A!A x}}C==r%   num_channels.c                    ||nd}t          |t                    r|fn|}| j        dk    rd\  }}n9| j        dk    rd\  }}n(| j        dk    rd\  }}nt          d| j                   | j        |         |v r>| j        |         |v r/t
                              d	| j         d
           t          j        S | j        |         |v rt          j        S | j        |         |v rt          j	        S t          d          )a[  
    Infers the channel dimension format of `image`.

    Args:
        image (`np.ndarray`):
            The image to infer the channel dimension of.
        num_channels (`int` or `tuple[int, ...]`, *optional*, defaults to `(1, 3)`):
            The number of channels of the image.

    Returns:
        The channel dimension of the image.
    Nr   rq   rq   )r            )r   r   z(Unsupported number of image dimensions: z4The channel dimension is ambiguous. Got image shape z. Assuming channels are the first dimension. Use the [input_data_format](https://huggingface.co/docs/transformers/main/internal/image_processing_utils#transformers.image_transforms.rescale.input_data_format) parameter to assign the channel dimension.z(Unable to infer channel dimension format)
r2   intru   rD   shapeloggerwarningr   r"   r#   )rG   r   	first_dimlast_dims       r&   infer_channel_dimension_formatr   (  s/    $0#;<<L&0s&C&CUL??LzQ"	88	q"	88	q"	88PEJPPQQQ{9--%+h2G<2W2W K5;  K  K  K	
 	
 	
  %%	Y	<	/	/%%	X	,	.	.$$
?
@
@@r%   input_data_formatc                     |t          |           }|t          j        k    r
| j        dz
  S |t          j        k    r
| j        dz
  S t          d|           )a  
    Returns the channel dimension axis of the image.

    Args:
        image (`np.ndarray`):
            The image to get the channel dimension axis of.
        input_data_format (`ChannelDimension` or `str`, *optional*):
            The channel dimension format of the image. If `None`, will infer the channel dimension from the image.

    Returns:
        The channel dimension axis of the image.
    Nrq   r   Unsupported data format: )r   r   r"   ru   r#   rD   )rG   r   s     r&   get_channel_dimension_axisr   O  sf      :5AA,222zA~	.3	3	3zA~
D1BDD
E
EEr%   channel_dimc                     |t          |           }|t          j        k    r| j        d         | j        d         fS |t          j        k    r| j        d         | j        d         fS t          d|           )a  
    Returns the (height, width) dimensions of the image.

    Args:
        image (`np.ndarray`):
            The image to get the dimensions of.
        channel_dim (`ChannelDimension`, *optional*):
            Which dimension the channel dimension is in. If `None`, will infer the channel dimension from the image.

    Returns:
        A tuple of the image's height and width.
    Nr   )r   r   r"   r   r#   rD   )rG   r   s     r&   get_image_sizer   g  sz     4U;;&,,,{2B//	(-	-	-{2B//B[BBCCCr%   
image_size
max_height	max_widthc                     | \  }}||z  }||z  }t          ||          }t          ||z            }t          ||z            }	||	fS )a  
    Computes the output image size given the input image and the maximum allowed height and width. Keep aspect ratio.
    Important, even if image_height < max_height and image_width < max_width, the image will be resized
    to at least one of the edges be equal to max_height or max_width.

    For example:
        - input_size: (100, 200), max_height: 50, max_width: 50 -> output_size: (25, 50)
        - input_size: (100, 200), max_height: 200, max_width: 500 -> output_size: (200, 400)

    Args:
        image_size (`tuple[int, int]`):
            The image to resize.
        max_height (`int`):
            The maximum allowed height.
        max_width (`int`):
            The maximum allowed width.
    )rn   r   )
r   r   r   heightwidthheight_scalewidth_scale	min_scale
new_height	new_widths
             r&   #get_image_size_for_max_height_widthr     s_    , MFE&Le#KL+..IVi'((JEI%&&Iy  r%   
annotationc                     t          | t                    rfd| v rbd| v r^t          | d         t          t          f          r<t	          | d                   dk    s!t          | d         d         t                    rdS dS )Nimage_idannotationsr   TFr2   dictr\   re   lenr   s    r&   "is_valid_annotation_coco_detectionr     s    :t$$
*$$Z''z-04-@@ ( 
=)**a//:j>WXY>Z\`3a3a/ t5r%   c                    t          | t                    rjd| v rfd| v rbd| v r^t          | d         t          t          f          r<t	          | d                   dk    s!t          | d         d         t                    rdS dS )Nr   segments_info	file_namer   TFr   r   s    r&   !is_valid_annotation_coco_panopticr     s    :t$$*$$z)):%%z/2T5MBB & 
?+,,11Z
?@[\]@^`d5e5e1 t5r%   r   c                 4    t          d | D                       S )Nc              3   4   K   | ]}t          |          V  d S r1   )r   rO   anns     r&   rP   z3valid_coco_detection_annotations.<locals>.<genexpr>  s+      NN31#66NNNNNNr%   rQ   r   s    r&    valid_coco_detection_annotationsr     s    NN+NNNNNNr%   c                 4    t          d | D                       S )Nc              3   4   K   | ]}t          |          V  d S r1   )r   r   s     r&   rP   z2valid_coco_panoptic_annotations.<locals>.<genexpr>  s+      MM#055MMMMMMr%   rQ   r   s    r&   valid_coco_panoptic_annotationsr     s    MMMMMMMMr%   timeoutc                    t          t          dg           t          | t                    r[|                     d          s|                     d          rGt
          j                            t          t          j
        | |          j                            } nt          j                            |           r t
          j                            |           } n|                     d          r|                     d          d         } 	 t!          j        |                                           }t
          j                            t          |                    } nU# t&          $ r}t)          d|  d	|           d
}~ww xY wt          | t
          j        j                  st+          d          t
          j                            |           } |                     d          } | S )a3  
    Loads `image` to a PIL Image.

    Args:
        image (`str` or `PIL.Image.Image`):
            The image to convert to the PIL Image format.
        timeout (`float`, *optional*):
            The timeout value in seconds for the URL request.

    Returns:
        `PIL.Image.Image`: A PIL Image.
    visionzhttp://zhttps://r   zdata:image/,r   zIncorrect image source. Must be a valid URL starting with `http://` or `https://`, a valid path to an image file, or a base64 encoded string. Got z. Failed with NzuIncorrect format used for image. Should be an url linking to an image, a base64 string, a local path, or a PIL image.RGB)r   
load_imager2   str
startswithr3   r4   openr   requestsgetcontentospathisfilesplitbase64decodebytesencode	ExceptionrD   	TypeErrorImageOpsexif_transposeconvert)rG   r   b64es       r&   r   r     s    j8*---% 
I&& 	%*:*::*F*F 	 INN78<w+O+O+O+W#X#XYYEEW^^E"" 	INN5))EE.. ,C((+(88	ws||44     B  jo  B  B  @  B  B   sy// 
 D
 
 	
 L''..EMM%  ELs   AE! !
F+F  Fc                     t          | t          t          f          rMt          |           r0t          | d         t          t          f          rfd| D             S fd| D             S t	          |           S )a  Loads images, handling different levels of nesting.

    Args:
      images: A single image, a list of images, or a list of lists of images to load.
      timeout: Timeout for loading images.

    Returns:
      A single image, a list of images, a list of lists of images.
    r   c                 ,    g | ]}fd |D             S )c                 2    g | ]}t          |           S r   r   rO   rG   r   s     r&   rX   z*load_images.<locals>.<listcomp>.<listcomp>  s&    QQQEZw777QQQr%   r$   )rO   image_groupr   s     r&   rX   zload_images.<locals>.<listcomp>  s/    lllVaQQQQ[QQQlllr%   c                 2    g | ]}t          |           S r   r   r   s     r&   rX   zload_images.<locals>.<listcomp>  s&    KKK5Jug666KKKr%   r   )r2   r\   re   r   r   )rK   r   s    `r&   load_imagesr     s     &4-(( 3v;; 	L:fQi$?? 	LllllekllllKKKKFKKKK&'2222r%   
do_rescalerescale_factordo_normalize
image_mean	image_stddo_padpad_sizedo_center_crop	crop_size	do_resizesizeresamplePILImageResamplinginterpolationr   c                     | r|t          d          |r|t          d          |r||t          d          |r|t          d          ||t          d          |	r|
||t          d          dS dS dS )a  
    Checks validity of typically used arguments in an `ImageProcessor` `preprocess` method.
    Raises `ValueError` if arguments incompatibility is caught.
    Many incompatibilities are model-specific. `do_pad` sometimes needs `size_divisor`,
    sometimes `size_divisibility`, and sometimes `size`. New models and processors added should follow
    existing arguments when possible.

    Nz=`rescale_factor` must be specified if `do_rescale` is `True`.zgDepending on the model, `size_divisor` or `pad_size` or `size` must be specified if `do_pad` is `True`.zP`image_mean` and `image_std` must both be specified if `do_normalize` is `True`.z<`crop_size` must be specified if `do_center_crop` is `True`.zbOnly one of `interpolation` and `resample` should be specified, depending on image processor type.zO`size` and `resample/interpolation` must be specified if `do_resize` is `True`.)rD   )r   r   r   r   r   r   r   r   r   r   r   r   r   s                r&   validate_preprocess_argumentsr     s    .  Zn,XYYY 	
(" u
 
 	
  m+y/@klll Y)+WXXX X%9p
 
 	
  l$*0DHajkkkl l0D0DHaHar%   c                       e Zd ZdZd ZddZd Zdej        de	e
ef         dej        fd	ZddZd ZddZddZd Zd ZddZdS )ImageFeatureExtractionMixinzD
    Mixin that contain utilities for preparing image features.
    c                     t          |t          j        j        t          j        f          s/t          |          s"t          dt          |           d          d S d S )Nz	Got type zU which is not supported, only `PIL.Image.Image`, `np.ndarray` and `torch.Tensor` are.)r2   r3   r4   r]   r^   r   rD   rE   selfrG   s     r&   _ensure_format_supportedz4ImageFeatureExtractionMixin._ensure_format_supported>  sq    %#)/2:!>?? 	X]H^H^ 	&DKK & & &  	 	 	 	r%   Nc                    |                      |           t          |          r|                                }t          |t          j                  r|%t          |j        d         t          j                  }|j        dk    r&|j	        d         dv r|
                    ddd          }|r|dz  }|                    t          j                  }t          j                            |          S |S )a"  
        Converts `image` to a PIL Image. Optionally rescales it and puts the channel dimension back as the last axis if
        needed.

        Args:
            image (`PIL.Image.Image` or `numpy.ndarray` or `torch.Tensor`):
                The image to convert to the PIL Image format.
            rescale (`bool`, *optional*):
                Whether or not to apply the scaling factor (to make pixel values integers between 0 and 255). Will
                default to `True` if the image type is a floating type, `False` otherwise.
        Nr   rq   r   r   r      )r  r   r<   r2   r]   r^   flatfloatingru   r   	transposeastyperm   r3   r4   	fromarray)r   rG   rescales      r&   to_pil_imagez(ImageFeatureExtractionMixin.to_pil_imageE  s     	%%e,,,5!! 	"KKMMEeRZ(( 
	.$UZ]BK@@zQ5;q>V#;#;1a00 $LL**E9&&u---r%   c                     |                      |           t          |t          j        j                  s|S |                    d          S )z
        Converts `PIL.Image.Image` to RGB format.

        Args:
            image (`PIL.Image.Image`):
                The image to convert.
        r   )r  r2   r3   r4   r   r   s     r&   convert_rgbz'ImageFeatureExtractionMixin.convert_rgbc  sE     	%%e,,,%11 	L}}U###r%   rG   scalerj   c                 6    |                      |           ||z  S )z7
        Rescale a numpy image by scale amount
        )r  )r   rG   r  s      r&   r	  z#ImageFeatureExtractionMixin.rescaleq  s"     	%%e,,,u}r%   Tc                    |                      |           t          |t          j        j                  rt	          j        |          }t          |          r|                                }|%t          |j        d         t          j	                  n|}|r3| 
                    |                    t          j                  d          }|r"|j        dk    r|                    ddd          }|S )a  
        Converts `image` to a numpy array. Optionally rescales it and puts the channel dimension as the first
        dimension.

        Args:
            image (`PIL.Image.Image` or `np.ndarray` or `torch.Tensor`):
                The image to convert to a NumPy array.
            rescale (`bool`, *optional*):
                Whether or not to apply the scaling factor (to make pixel values floats between 0. and 1.). Will
                default to `True` if the image is a PIL Image or an array/tensor of integers, `False` otherwise.
            channel_first (`bool`, *optional*, defaults to `True`):
                Whether or not to permute the dimensions of the image to put the channel dimension first.
        Nr   p?rq   r   r   )r  r2   r3   r4   r]   r   r   r<   r  integerr	  r  float32ru   r  )r   rG   r	  channel_firsts       r&   r   z*ImageFeatureExtractionMixin.to_numpy_arrayx  s     	%%e,,,eSY_-- 	$HUOOE5!! 	"KKMME;B?*UZ]BJ777PW 	FLLbj!9!99EEE 	-UZ1__OOAq!,,Er%   c                     |                      |           t          |t          j        j                  r|S t	          |          r|                    d          }nt          j        |d          }|S )z
        Expands 2-dimensional `image` to 3 dimensions.

        Args:
            image (`PIL.Image.Image` or `np.ndarray` or `torch.Tensor`):
                The image to expand.
        r   rY   )r  r2   r3   r4   r   	unsqueezer]   expand_dimsr   s     r&   r  z'ImageFeatureExtractionMixin.expand_dims  sq     	%%e,,, eSY_-- 	L5!! 	2OOA&&EEN5q111Er%   Fc                    |                      |           t          |t          j        j                  r|                     |d          }n|rt          |t
          j                  r4|                     |                    t
          j	                  d          }n7t          |          r(|                     |                                d          }t          |t
          j                  rt          |t
          j                  s,t          j        |                              |j                  }t          |t
          j                  s,t          j        |                              |j                  }nt          |          rddl}t          ||j                  s;t          |t
          j                  r |j        |          }n |j        |          }t          ||j                  s;t          |t
          j                  r |j        |          }n |j        |          }|j        dk    r-|j        d         dv r||ddddf         z
  |ddddf         z  S ||z
  |z  S )a  
        Normalizes `image` with `mean` and `std`. Note that this will trigger a conversion of `image` to a NumPy array
        if it's a PIL Image.

        Args:
            image (`PIL.Image.Image` or `np.ndarray` or `torch.Tensor`):
                The image to normalize.
            mean (`list[float]` or `np.ndarray` or `torch.Tensor`):
                The mean (per channel) to use for normalization.
            std (`list[float]` or `np.ndarray` or `torch.Tensor`):
                The standard deviation (per channel) to use for normalization.
            rescale (`bool`, *optional*, defaults to `False`):
                Whether or not to rescale the image to be between 0 and 1. If a PIL image is provided, scaling will
                happen automatically.
        T)r	  r  r   Nrq   r   )r  r2   r3   r4   r   r]   r^   r	  r  r  r   floatr   rl   r;   r`   
from_numpytensorru   r   )r   rG   meanstdr	  r;   s         r&   	normalizez%ImageFeatureExtractionMixin.normalize  s6     	%%e,,,eSY_-- 	?''t'<<EE  	?%,, ?U\\"*%=%=yII '' ?U[[]]I>>eRZ(( 	,dBJ// :x~~,,U[99c2:.. 8hsmm**5;77U## 	,LLLdEL11 .dBJ// .+5+D11DD'5<--Dc5<00 ,c2:.. ,*%*3//CC&%,s++C:??u{1~77DD$//3qqq$}3EEEDLC''r%   c                    ||nt           j        }|                     |           t          |t          j        j                  s|                     |          }t          |t                    rt          |          }t          |t                    st          |          dk    r|r*t          |t                    r||fn|d         |d         f}n|j        \  }}||k    r||fn||f\  }}	t          |t                    r|n|d         }
||
k    r|S |
t          |
|	z  |z            }}|8||
k    rt          d| d|           ||k    rt          ||z  |z            |}}||k    r||fn||f}|                    ||          S )a  
        Resizes `image`. Enforces conversion of input to PIL.Image.

        Args:
            image (`PIL.Image.Image` or `np.ndarray` or `torch.Tensor`):
                The image to resize.
            size (`int` or `tuple[int, int]`):
                The size to use for resizing the image. If `size` is a sequence like (h, w), output size will be
                matched to this.

                If `size` is an int and `default_to_square` is `True`, then image will be resized to (size, size). If
                `size` is an int and `default_to_square` is `False`, then smaller edge of the image will be matched to
                this number. i.e, if height > width, then image will be rescaled to (size * height / width, size).
            resample (`int`, *optional*, defaults to `PILImageResampling.BILINEAR`):
                The filter to user for resampling.
            default_to_square (`bool`, *optional*, defaults to `True`):
                How to convert `size` when it is a single int. If set to `True`, the `size` will be converted to a
                square (`size`,`size`). If set to `False`, will replicate
                [`torchvision.transforms.Resize`](https://pytorch.org/vision/stable/transforms.html#torchvision.transforms.Resize)
                with support for resizing only the smallest edge and providing an optional `max_size`.
            max_size (`int`, *optional*, defaults to `None`):
                The maximum allowed for the longer edge of the resized image: if the longer edge of the image is
                greater than `max_size` after being resized according to `size`, then the image is resized again so
                that the longer edge is equal to `max_size`. As a result, `size` might be overruled, i.e the smaller
                edge may be shorter than `size`. Only used if `default_to_square` is `False`.

        Returns:
            image: A resized `PIL.Image.Image`.
        Nr   r   zmax_size = zN must be strictly greater than the requested size for the smaller edge size = )r   )r   BILINEARr  r2   r3   r4   r
  r\   re   r   r   r   rD   resize)r   rG   r   r   default_to_squaremax_sizer   r   shortlongrequested_new_short	new_shortnew_longs                r&   r   z"ImageFeatureExtractionMixin.resize  s   <  (3889K9T%%e,,,%11 	-%%e,,EdD!! 	;;DdC   	[CIINN  ['1$'<'<Td||47DQRGBT %
v16&ufoovuot.8s.C.C&Pdda#/// L&93?RUY?Y\a?a;b;b8	'#666(G( G G@DG G    (**.1(Y2F2Q.R.RT\8	05	8,,hPYEZ||D8|444r%   c                 V   |                      |           t          |t                    s||f}t          |          st          |t          j                  rN|j        dk    r|                     |          }|j        d         dv r|j        dd         n|j        dd         }n|j	        d         |j	        d         f}|d         |d         z
  dz  }||d         z   }|d         |d         z
  dz  }||d         z   }t          |t          j        j                  r|                    ||||f          S |j        d         dv }|sWt          |t          j                  r|                    ddd          }t          |          r|                    ddd          }|dk    r-||d         k    r!|dk    r||d         k    r|d||||f         S |j        dd         t          |d         |d                   t          |d         |d                   fz   }	t          |t          j                  rt	          j        ||	          }
n$t          |          r|                    |	          }
|	d         |d         z
  dz  }||d         z   }|	d	         |d         z
  dz  }||d         z   }||
d||||f<   ||z  }||z  }||z  }||z  }|
dt          d|          t%          |
j        d         |          t          d|          t%          |
j        d	         |          f         }
|
S )
a  
        Crops `image` to the given size using a center crop. Note that if the image is too small to be cropped to the
        size given, it will be padded (so the returned result has the size asked).

        Args:
            image (`PIL.Image.Image` or `np.ndarray` or `torch.Tensor` of shape (n_channels, height, width) or (height, width, n_channels)):
                The image to resize.
            size (`int` or `tuple[int, int]`):
                The size to which crop the image.

        Returns:
            new_image: A center cropped `PIL.Image.Image` or `np.ndarray` or `torch.Tensor` of shape: (n_channels,
            height, width).
        r   r   r   r   N.r   )r   r   )r  r2   re   r   r]   r^   ru   r  r   r   r3   r4   cropr  permutero   
zeros_like	new_zerosrn   )r   rG   r   image_shapetopbottomleftrightr  	new_shape	new_imagetop_pad
bottom_padleft_pad	right_pads                  r&   center_cropz'ImageFeatureExtractionMixin.center_crop#  sd    	%%e,,,$&& 	 $<D 5!! 	9Zrz%B%B 	9zQ((//-2[^v-E-E%+abb//5;WYXYWY?KK :a=%*Q-8K1~Q'A-tAwAa(Q.tAw eSY_-- 	:::tS%8999 A&0  	/%,, 11a00u%% /aA.. !88+a.00TQYY5KXYNCZCZc&j$u*455 K$DG[^(D(Dc$q'S^_`SaFbFb'cc	eRZ(( 	3e9===IIU## 	3	22IR=;q>1a7{1~-
bMKN2q8{1~-	AF	#wz)8I+==>w'Qs9?2#6???QPST]TcdfTginPoPoAoo
	 r%   c                     |                      |           t          |t          j        j                  r|                     |          }|dddddddf         S )a  
        Flips the channel order of `image` from RGB to BGR, or vice versa. Note that this will trigger a conversion of
        `image` to a NumPy array if it's a PIL Image.

        Args:
            image (`PIL.Image.Image` or `np.ndarray` or `torch.Tensor`):
                The image whose color channels to flip. If `np.ndarray` or `torch.Tensor`, the channel dimension should
                be first.
        Nr   )r  r2   r3   r4   r   r   s     r&   flip_channel_orderz.ImageFeatureExtractionMixin.flip_channel_ordern  s`     	%%e,,,eSY_-- 	/''..ETTrT111aaaZ  r%   r   c                     ||nt           j        j        }|                     |           t	          |t           j        j                  s|                     |          }|                    ||||||          S )a  
        Returns a rotated copy of `image`. This method returns a copy of `image`, rotated the given number of degrees
        counter clockwise around its centre.

        Args:
            image (`PIL.Image.Image` or `np.ndarray` or `torch.Tensor`):
                The image to rotate. If `np.ndarray` or `torch.Tensor`, will be converted to `PIL.Image.Image` before
                rotating.

        Returns:
            image: A rotated `PIL.Image.Image`.
        N)r   expandcenter	translate	fillcolor)r3   r4   NEARESTr  r2   r
  rotate)r   rG   angler   r<  r=  r>  r?  s           r&   rA  z"ImageFeatureExtractionMixin.rotate  s}      (3889J%%e,,,%11 	-%%e,,E||HVFicl  
 
 	
r%   r1   )NT)F)NTN)Nr   NNN)r   r    r!   __doc__r  r
  r  r]   r^   r   r  r   r	  r   r  r  r   r8  r:  rA  r$   r%   r&   r   r   9  s             <$ $ $RZ eSj0A bj       @  (2( 2( 2( 2(hA5 A5 A5 A5FI I IV! ! !"
 
 
 
 
 
r%   r   annotation_formatsupported_annotation_formatsc                     | |vrt          dt           d|           | t          j        u rt	          |          st          d          | t          j        u rt          |          st          d          d S d S )NzUnsupported annotation format: z must be one of zInvalid COCO detection annotations. Annotations must a dict (single image) or list of dicts (batch of images) with the following keys: `image_id` and `annotations`, with the latter being a list of annotations in the COCO format.zInvalid COCO panoptic annotations. Annotations must a dict (single image) or list of dicts (batch of images) with the following keys: `image_id`, `file_name` and `segments_info`, with the latter being a list of annotations in the COCO format.)rD   formatr(   r+   r   r,   r   )rD  rE  r   s      r&   validate_annotationsrH    s    
  <<<q6qqSoqqrrr,;;;/<< 	B   ,:::.{;; 	M   ;:	 	r%   valid_processor_keyscaptured_kwargsc                     t          |                              t          |                     }|r5d                    |          }t                              d| d           d S d S )Nz, zUnused or unrecognized kwargs: rt   )set
differencejoinr   r   )rI  rJ  unused_keysunused_key_strs       r&   validate_kwargsrQ    sp    o&&11#6J2K2KLLK L;//JJJJKKKKKL Lr%   T)frozenc                       e Zd ZU dZdZee         ed<   dZee         ed<   dZ	ee         ed<   dZ
ee         ed<   dZee         ed<   dZee         ed<   d	 ZdS )
SizeDictz>
    Hashable dictionary to store image size information.
    Nr   r   longest_edgeshortest_edger   r   c                 h    t          | |          rt          | |          S t          d| d          )NzKey z not found in SizeDict.)hasattrgetattrKeyError)r   keys     r&   __getitem__zSizeDict.__getitem__  s>    4 	&4%%%:c:::;;;r%   )r   r    r!   rC  r   r   r   __annotations__r   rU  rV  r   r   r\  r$   r%   r&   rT  rT    s           !FHSM   E8C="&L(3-&&&#'M8C=''' $J$$$#Ix}###< < < < <r%   rT  )rq   r1   )NNNNNNNNNNNNN)^r   r   collections.abcr   dataclassesr   ior   typingr   r   r<   r]   r   utilsr	   r
   r   r   r   r   r   r   r   r   r   utils.constantsr   r   r   r   r   r   	PIL.Imager3   PIL.ImageOpsr4   
Resamplingr   torchvision.transformsr   r@  NEAREST_EXACTBOXr  HAMMINGBICUBICLANCZOSpil_torch_interpolation_mappingr;   
get_loggerr   r   r^   r\   
ImageInputr   r(   r.   r   r   r   AnnotationTyper7   r9   rH   rJ   rS   rc   rf   ri   boolrp   rw   r   r   r   re   r   r   r   r   r   r   r   r   r  r   r   r   r   rH  rQ  rT  r$   r%   r&   <module>rr     s	    				 $ $ $ $ $ $ ! ! ! ! ! !       " " " " " " " "                                               --!! -<<<<<< &(9(G"$5$9'):)C&(9(A&(9(A&(9(A+
'' +-'  LLL 
	H	%	% rz>48I3JDQSQ[L\^bcq^rr

    |   
$ $ $ $ $| $ $ $
9 9 9 9 9\ 9 9 9
 c5c4:!5667F F F       ? ? ?w w wED E E E E, , ,	 	 	  52: 5$ 5 5 5 5$ $ $D<L $ $ $ $R #L #L$z"J./#L#L #L #L #L #LP $v $v$z"J./$v$v 
*$v $v $v $vN2:     NR$A $A:$A%-eCsCx4H.I%J$A$A $A $A $AP TXF F:F*259I39N3O*PFF F F F0D D"* D8<L3M DY^_bdg_gYh D D D D0!c3h!! ! 38_	! ! ! !>4U4;=O8O3P UY    $sE$+<N7N2O TX     O(4U4QV;EW@W;X2Y O^b O O O ON$sE$PU+DV?V:W1X N]a N N N N' 'eC!223 'huo 'Yj ' ' ' 'V TX3 3$s$5563AI%3
d#45tDAR<S7TTU3 3 3 3, "&&*#'6:59!59%)*. $%)/3371l 1l1lUO1l 4.1l ud5k123	1l
 eT%[0121l TN1l uT#s(^S0121l TN1l S#X'1l ~1l 4S>
"1l +,1l /01l 1l 1l 1lj\
 \
 \
 \
 \
 \
 \
 \
~
'"'(8#(="> d 
	   2L$s) Ld3i L L L L $< < < < < < < < < <r%   