U
    Jc                     @   s   d dl Z d dlmZ d dlmZmZ d dlmZ ddlm	Z	m
Z
 ddlmZmZmZmZ edd	G d
d dZG dd de jZG dd de jZdS )    N)	dataclass)ListAny)Future   )MetadataMetadataIndex)LoadPlanSavePlanSavePlannerLoadPlannerT)frozenc                   @   s&   e Zd ZU eed< eed< eed< dS )WriteResultindexZsize_in_bytesZstorage_dataN)__name__
__module____qualname__r   __annotations__intr    r   r   O/tmp/pip-unpacked-wheel-gikjz4vx/torch/distributed/_shard/checkpoint/storage.pyr      s   
r   c                   @   s   e Zd ZdZejeddddZejeedddZ	eje
e e
e d	d
dZejeeee
e  dddZejee
e
e  ddddZdS )StorageWritera  
    Interface used by ``save_state_dict`` to write to storage.

    One StorageWriter instance acts as both the coordinator and the follower
    in a distributed checkpoint. As part of initialization, each instance
    is told its role.

    A subclass should expect the following sequence of calls.

    1) (all ranks) init()
    2) (all ranks) prepare_local_plan()
    3) (coordinator) prepare_global_plan()
    4) (all ranks) write_data()
    5) (coordinator) finish()
    N)is_coordinatorreturnc                 C   s   dS )z
        Initialize this instance.

        Args:
            is_coordinator (bool): Whether this instance is reponsible for coordinating
              the checkpoint.
        Nr   )selfr   r   r   r   init+   s    	zStorageWriter.initplanr   c                 C   s   dS )a  
        Perform storage-specific local planning.

        While this method can produce a completely different plan, the recomended
        way is to store storage specific data in SavePlan::storage_data.

        Args:
            plan (SavePlan): The local plan from the ``SavePlanner`` in use.

        Returns:
            A transformed ``SavePlan`` after storage local planning
        Nr   r   r   r   r   r   prepare_local_plan6   s    z StorageWriter.prepare_local_planplansr   c                 C   s   dS )a  
        Perform centralized planning of storage.

        This method is only called on the coordinator instance.

        While this method can produce a completely different plan, the prefered
        way is to store storage specific data in SavePlan::storage_data.

        Args:
            plans: A list of ``SavePlan`` instances, one for each rank.

        Returns:
            A list of transformed ``SavePlan`` after storage global planning
        Nr   r   r!   r   r   r   prepare_global_planF   s    z!StorageWriter.prepare_global_planr   plannerr   c                 C   s   dS )a  
        Write all items from ``plan`` using ``planner`` to resolve the data.

        A subclass should call ``SavePlanner::resolve_data`` on each item
        from the plan to get access to the underlying object to write.

        Subclasses should lazily call `resolve_data` as it can allocate memory.
        In case of tensors, make following assuptions:

        - They might be on any device, including not matching the one on ``WriteItem::tensor_data``
        - They might be views or not contiguous. Only the projection needs to be saved.

        Args:
            plan (SavePlan): The save plan to execute.
            planner (SavePlanner): Planner object to be used to resolve items to data.

        Returns:
            A future that completes to a list of WriteResult
        Nr   r   r   r%   r   r   r   
write_dataX   s    zStorageWriter.write_data)metadataresultsr   c                 C   s   dS )a  
        Writes the metadata and marks the current checkpoint as sucessful.

        The actual format/schema used for serializing `metadata` is an
        implemetation detail. The only requirement is that it's recoverable
        in to the same object graph.

        Args:
            metadata (Metadata): metadata for the new checkpoint
            results: A list of WriteResults from all ranks.

        Returns:
            None
        Nr   )r   r(   r)   r   r   r   finishs   s    zStorageWriter.finish)r   r   r   __doc__abcabstractmethodboolr   r
   r   r   r#   r   r   r   r'   r   r*   r   r   r   r   r      s   

r   c                   @   s   e Zd ZdZejedddZejeeddddZ	eje
e
d	d
dZejee
 ee
 dddZeje
eed dddZdS )StorageReadera  
    Interface used by ``load_state_dict`` to read from storage.

    One StorageReader instance acts as both the coordinator and the follower
    in a distributed checkpoint. As part of initialization, each instance
    is told its role.

    A subclass should expected the following sequence of calls by ``load_state_dict``:

    1) (all ranks) read_metadata()
    2) (all ranks) init
    3) (all ranks) prepare_local_plan
    4) (coordinator) prepare_global_plan
    5) (all ranks) read_data
    )r   c                 C   s   dS )z
        Reads the checkpoint metadata.

        Returns:
            The metatada object associated with the checkpoint being loaded.

        Nr   )r   r   r   r   read_metadata   s    	zStorageReader.read_metadataN)r(   r   r   c                 C   s   dS )z
        Initialize this instance.

        Args:
            metadata (Metadata): The metadata schema to use.
            is_coordinator (bool): Whether this instance is reponsible for coordinating
              the checkpoint.
        Nr   )r   r(   r   r   r   r   r      s    
zStorageReader.initr   c                 C   s   dS )a  
        Perform storage-specific local planning.

        While this method can produce a completely different plan, the recomended
        way is to store storage specific data in LoadPlan::storage_data.

        Args:
            plan (LoadPlan): The local plan from the ``LoadPlan`` in use.

        Returns:
            A transformed ``LoadPlan`` after storage local planning
        Nr   r   r   r   r   r      s    z StorageReader.prepare_local_planr    c                 C   s   dS )a  
        Perform centralized planning of storage loading.

        This method is only called on the coordinator instance.

        While this method can produce a completely different plan, the prefered
        way is to store storage specific data in LoadPlan::storage_data.

        Args:
            plans: A list of ``LoadPlan`` instances, one for each rank.

        Returns:
            A list of transformed ``LoadPlan`` after storage global planning
        Nr   r"   r   r   r   r#      s    z!StorageReader.prepare_global_planr$   c                 C   s   dS )a  
        Reads all items from ``plan`` using ``planner`` to resolve the data.

        A subclass should call ``LoadPlanner::load_bytes`` to deserialize a BytesIO
        object into the right place.

        A subclass should call ``LoadPlanner::resolve_tensor`` to get access to the
        tensors that in should load data into.

        It's the StorageLayer responsibility to properly schedule any cross device copies
        required.

        Args:
            plan (LoadPlan): The local plan to execute on
            planner (LoadPlanner): The planner object to use to resolve items.

        Returns:
            A future that completes once all reads are finished.
        Nr   r&   r   r   r   	read_data   s    zStorageReader.read_data)r   r   r   r+   r,   r-   r   r0   r.   r   r	   r   r   r#   r   r   r1   r   r   r   r   r/      s   
r/   )r,   Zdataclassesr   typingr   r   Ztorch.futuresr   r(   r   r   r%   r	   r
   r   r   r   ABCr   r/   r   r   r   r   <module>   s   k