§
    `ƒis  ã                   ód   — d dl Z d dlZd dlmc mc mc mZ  G d„ dej        j        ¦  «        Z	dS )é    Nc                   ó‚   — e Zd ZdZdej        j        dej        fd„Z	e
d„ ¦   «         Zd„ Zd„ Zd„ Zd	„ Zddefd„Zd„ ZdS )ÚPostLocalSGDOptimizeraì  
    Wraps an arbitrary :class:`torch.optim.Optimizer` and runs `post-local SGD <https://arxiv.org/abs/1808.07217>`_,
    This optimizer runs local optimizer at every step.
    After the warm-up stage, it averages parameters periodically after the local optimizer is applied.

    Args:
        optim: The local optimizer.
        averager: A model averager instance to run post-localSGD algorithm.

    Example::

        >>> # xdoctest: +SKIP("undefined variables")
        >>> import torch
        >>> import torch.distributed as dist
        >>> import torch.distributed.algorithms.model_averaging.averagers as averagers
        >>> import torch.nn as nn
        >>> from torch.distributed.optim import PostLocalSGDOptimizer
        >>> from torch.distributed.algorithms.ddp_comm_hooks.post_localSGD_hook import (
        >>>   PostLocalSGDState,
        >>>   post_localSGD_hook,
        >>> )
        >>>
        >>> model = nn.parallel.DistributedDataParallel(
        >>>    module, device_ids=[rank], output_device=rank
        >>> )
        >>>
        >>> # Register a post-localSGD communication hook.
        >>> state = PostLocalSGDState(process_group=None, subgroup=None, start_localSGD_iter=100)
        >>> model.register_comm_hook(state, post_localSGD_hook)
        >>>
        >>> # Create a post-localSGD optimizer that wraps a local optimizer.
        >>> # Note that ``warmup_steps`` used in ``PostLocalSGDOptimizer`` must be the same as
        >>> # ``start_localSGD_iter`` used in ``PostLocalSGDState``.
        >>> local_optim = torch.optim.SGD(params=model.parameters(), lr=0.01)
        >>> opt = PostLocalSGDOptimizer(
        >>>     optim=local_optim,
        >>>     averager=averagers.PeriodicModelAverager(period=4, warmup_steps=100)
        >>> )
        >>>
        >>> # In the first 100 steps, DDP runs global gradient averaging at every step.
        >>> # After 100 steps, DDP runs gradient averaging within each subgroup (intra-node by default),
        >>> # and post-localSGD optimizer runs global model averaging every 4 steps after applying the local optimizer.
        >>> for step in range(0, 200):
        >>>    opt.zero_grad()
        >>>    loss = loss_fn(output, labels)
        >>>    loss.backward()
        >>>    opt.step()
    ÚoptimÚaveragerc                 óD   — || _         | j         j        | _        || _        d S ©N)r   Úparam_groupsr   )Úselfr   r   s      úƒ/home/jaya/work/projects/VOICE-AGENT/VIET/agent-env/lib/python3.11/site-packages/torch/distributed/optim/post_localSGD_optimizer.pyÚ__init__zPostLocalSGDOptimizer.__init__:   s!   € ØˆŒ
Ø œJÔ3ˆÔØ ˆŒˆˆó    c                 ó   — | j         j        S r   )r   Ústate©r
   s    r   r   zPostLocalSGDOptimizer.state?   s   € àŒzÔÐr   c                 ó4   — | j                              ¦   «         S r   )r   Ú__repr__r   s    r   r   zPostLocalSGDOptimizer.__repr__C   s   € ØŒz×"Ò"Ñ$Ô$Ð$r   c                 óV   — | j                              ¦   «         }| j        j        |d<   |S )zí
        This is the same as :class:`torch.optim.Optimizer` :meth:`state_dict`,
        but adds an extra entry to record model averager's step to the checkpoint
        to ensure reload does not cause unnecessary warm up again.
        Ústep)r   Ú
state_dictr   r   )r
   Úoptim_state_dicts     r   r   z PostLocalSGDOptimizer.state_dictF   s.   € ð  œ:×0Ò0Ñ2Ô2ÐØ#'¤=Ô#5Ð˜Ñ ØÐr   c                 óª   — | j                              |¦  «         d|v r|d         | j        _        dS t	          j        d¦  «         d| j        _        dS )aW  
        This is the same as :class:`torch.optim.Optimizer` :meth:`load_state_dict`,
        but also restores model averager's step value to the one
        saved in the provided ``state_dict``.

        If there is no ``"step"`` entry in ``state_dict``,
        it will raise a warning and initialize the model averager's step to 0.
        r   z]Loaded state dict does not contain a step counter for an averager. Setting step counter to 0.r   N)r   Úload_state_dictr   r   ÚwarningsÚwarn)r
   r   s     r   r   z%PostLocalSGDOptimizer.load_state_dictP   se   € ð 	Œ
×"Ò" :Ñ.Ô.Ð.ØZÐÐØ!+¨FÔ!3ˆDŒMÔÐÐåŒMð-ñô ð ð "#ˆDŒMÔÐÐr   c                 óx   — | j                              ¦   «          | j                             | j        ¬¦  «         dS )zI
        Performs a single optimization step (parameter update).
        )ÚparamsN)r   r   r   Úaverage_parametersr	   r   s    r   r   zPostLocalSGDOptimizer.stepc   s8   € ð 	Œ
ŠÑÔÐØŒ×(Ò(°Ô0AÐ(ÑBÔBÐBÐBÐBr   TÚset_to_nonec                 ó<   — | j                              |¬¦  «         d S )N)r   )r   Ú	zero_grad)r
   r   s     r   r    zPostLocalSGDOptimizer.zero_gradj   s!   € ØŒ
×Ò¨ÐÑ5Ô5Ð5Ð5Ð5r   c                 ó:   — | j                              |¦  «         d S r   )r   Úadd_param_group)r
   Úparam_groups     r   r"   z%PostLocalSGDOptimizer.add_param_groupm   s   € ØŒ
×"Ò" ;Ñ/Ô/Ð/Ð/Ð/r   N)T)Ú__name__Ú
__module__Ú__qualname__Ú__doc__Útorchr   Ú	OptimizerÚ	averagersÚModelAveragerr   Úpropertyr   r   r   r   r   Úboolr    r"   © r   r   r   r      sÒ   € € € € € ð/ð /ðb!˜eœkÔ3ð !¸yÔ?Vð !ð !ð !ð !ð
 ð ð  ñ „Xð ð%ð %ð %ð ð  ð  ð#ð #ð #ð&Cð Cð Cð6ð 6 Tð 6ð 6ð 6ð 6ð0ð 0ð 0ð 0ð 0r   r   )
r   r(   Ú6torch.distributed.algorithms.model_averaging.averagersÚdistributedÚ
algorithmsÚmodel_averagingr*   r   r)   r   r.   r   r   ú<module>r3      s‡   ðà €€€à €€€Ø JÐ JÐ JÐ JÐ JÐ JÐ JÐ JÐ JÐ JÐ JÐ JÐ JÐ JÐ Jðf0ð f0ð f0ð f0ð f0˜EœKÔ1ñ f0ô f0ð f0ð f0ð f0r   