
                            d Z ddlmZ ddlmZ ddlmZ ddlmZ ddlmZ ej                   ej                  ej                  j                  ej                  j                         G d dej                                      Zy	)
z'The command group for the profiles CLI.    )absolute_import)division)unicode_literals)base)utilc                   (    e Zd ZdZej
                  Zy)Profilesas  Quickstart engine for GKE AI workloads.

  The GKE Inference Quickstart helps simplify deploying AI inference on Google
  Kubernetes Engine (GKE). It provides tailored profiles based on
  Google's internal benchmarks. Provide inputs like your preferred open-source
  model (e.g. Llama, Gemma, or Mistral) and your application's performance
  target. Based on these inputs, the quickstart generates accelerator choices
  with performance metrics, and detailed, ready-to-deploy profiles for
  compute, load balancing, and autoscaling. These profiles are provided
  as standard Kubernetes YAML manifests, which you can deploy or modify.

  To visualize the benchmarking data that support these estimates, see the
  accompanying Colab notebook:
  https://colab.research.google.com/github/GoogleCloudPlatform/kubernetes-engine-samples/blob/main/ai-ml/notebooks/giq_visualizations.ipynb
  N)__name__
__module____qualname____doc__r   SDK_TOOLS_CATEGORYcategory     -lib/surface/container/ai/profiles/__init__.pyr	   r	      s      $$(r   r	   N)r   
__future__r   r   r   googlecloudsdk.callioper   #googlecloudsdk.command_lib.projectsr   UniverseCompatibleReleaseTracksReleaseTrackALPHAGAGroupr	   r   r   r   <module>r      sn     . &  ' ( 4 D%%++T->->-A-AB%tzz % C %r   