OXIESEC PANEL
- Current Dir:
/
/
opt
/
gsutil
/
gslib
/
commands
/
__pycache__
Server IP: 2a02:4780:11:1594:0:ef5:22d7:a
Upload:
Create Dir:
Name
Size
Modified
Perms
📁
..
-
02/11/2025 08:19:49 AM
rwxr-xr-x
📄
__init__.cpython-39.pyc
323 bytes
02/11/2025 08:19:49 AM
rw-r--r--
📄
acl.cpython-39.pyc
18.76 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
autoclass.cpython-39.pyc
6.01 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
bucketpolicyonly.cpython-39.pyc
6.75 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
cat.cpython-39.pyc
4.17 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
compose.cpython-39.pyc
4.57 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
config.cpython-39.pyc
40.97 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
cors.cpython-39.pyc
6.56 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
cp.cpython-39.pyc
42.21 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
defacl.cpython-39.pyc
11.25 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
defstorageclass.cpython-39.pyc
5.78 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
du.cpython-39.pyc
8.47 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
hash.cpython-39.pyc
7.87 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
help.cpython-39.pyc
6.41 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
hmac.cpython-39.pyc
12.39 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
iam.cpython-39.pyc
24.6 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
kms.cpython-39.pyc
14.88 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
label.cpython-39.pyc
10.46 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
lifecycle.cpython-39.pyc
6.37 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
logging.cpython-39.pyc
9.04 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
ls.cpython-39.pyc
18.81 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
mb.cpython-39.pyc
12.23 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
mv.cpython-39.pyc
5.06 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
notification.cpython-39.pyc
25.47 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
pap.cpython-39.pyc
6.53 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
perfdiag.cpython-39.pyc
62.18 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
rb.cpython-39.pyc
3.78 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
requesterpays.cpython-39.pyc
5.53 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
retention.cpython-39.pyc
20.02 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
rewrite.cpython-39.pyc
16.12 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
rm.cpython-39.pyc
10.86 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
rpo.cpython-39.pyc
5.96 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
rsync.cpython-39.pyc
53.86 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
setmeta.cpython-39.pyc
11.25 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
signurl.cpython-39.pyc
20.99 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
stat.cpython-39.pyc
5.04 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
test.cpython-39.pyc
17.73 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
ubla.cpython-39.pyc
6.96 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
update.cpython-39.pyc
12.26 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
version.cpython-39.pyc
5.25 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
versioning.cpython-39.pyc
5.51 KB
02/11/2025 08:19:49 AM
rw-r--r--
📄
web.cpython-39.pyc
7.63 KB
02/11/2025 08:19:49 AM
rw-r--r--
Editing: config.cpython-39.pyc
Close
a +(Wg�� � @ s@ d Z ddlmZ ddlmZ ddlmZ ddlmZ ddlZddlZddlZddl Z ddl Z ddlZddlZddl Z ddlZddlZddlZddlmZ ddlmZ ddlZdd lmZ ddlZdd lmZ ddlmZ ddlmZ dd lmZ ddl m!Z! ddl m"Z" ddl#m$Z$ ddl%m&Z& ddl'm(Z( ddl'm)Z) ddl*m+Z+ ddl*m,Z, ddl*m-Z- ddl*m.Z. ddl/m0Z0 ddl1m2Z2 ddl3m4Z4 dZ5de5 d Z6dZ7dZ8e0� \Z9Z:e9�r�d Z;d!Z<ne=e�>� d"�Z;d#Z<d$Z?d%Z@d&ZAd'ZBd(ZCd)ZDd*ZEd+ZFd,ZGd-e,e-e+e.e(jHe;e<e?e@e?e@eCeeeDeEd.� ZId/ZJG d0d1� d1e�ZKd2d3� ZLdS )4zJImplementation of config command for creating a gsutil configuration file.� )�absolute_import)�print_function)�division)�unicode_literalsN)�input)�ResponseNotReady)�Provider)�Command)�!DEFAULT_TASK_ESTIMATION_THRESHOLD)�MAX_COMPOSE_ARITY)� CredTypes��AbortException)�CommandException)�'CheckAndMaybePromptForAnalyticsEnabling)�RegisterSignalHandler)� constants)�system_util)�CHECK_HASH_ALWAYS)�CHECK_HASH_IF_FAST_ELSE_FAIL)�CHECK_HASH_IF_FAST_ELSE_SKIP)�CHECK_HASH_NEVER)�ShouldProhibitMultiprocessing)�ServerNotFoundError)� HAS_CRYPTOz/ gsutil [-D] config (-a|-e) [-n] [-o <file>] z <B>SYNOPSIS</B> a� <B>DESCRIPTION</B> The ``gsutil config`` command generally applies to users who have legacy standalone installations of gsutil. If you installed gsutil via the Cloud SDK, ``gsutil config`` fails unless you are specifically using the ``-a`` flag or have configured gcloud to not pass its managed credentials to gsutil (via the command ``gcloud config set pass_credentials_to_gsutil false``). For all other use cases, Cloud SDK users should use the ``gcloud auth`` group of commands instead, which configures OAuth2 credentials that gcloud implicitly passes to gsutil at runtime. To check if you are using gsutil from the Cloud SDK or as a legacy standalone, use ``gsutil version -l`` and in the output look for "using cloud sdk". Important: The default behavior for the ``gsutil config`` command is to obtain user account credentials for authentication. However, user account credentials are no longer supported for standalone gsutil. For this reason, running the default ``gsutil config`` command fails, and using any of the following flags causes the command to fail: ``-b``, ``-f``, ``-r``, ``--reauth``, ``-s``, ``-w``. When using standalone gsutil, it's recommended that you use service account credentials via the ``-e`` flag. The ``gsutil config`` command obtains access credentials for Cloud Storage and writes a `boto/gsutil configuration file <https://cloud.google.com/storage/docs/boto-gsutil>`_ containing the obtained credentials along with a number of other configuration- controllable values. Unless specified otherwise (see OPTIONS), the configuration file is written to ~/.boto (i.e., the file .boto under the user's home directory). If the default file already exists, an attempt is made to rename the existing file to ~/.boto.bak; if that attempt fails the command exits. A different destination file can be specified with the ``-o`` option (see OPTIONS). Because the boto configuration file contains your credentials you should keep its file permissions set so no one but you has read access. (The file is created read-only when you run ``gsutil config``.) <B>CONFIGURING SERVICE ACCOUNT CREDENTIALS</B> Service accounts are useful for authenticating on behalf of a service or application (as opposed to a user). If you use gsutil as a legacy stand-alone tool, you configure credentials for service accounts using the ``-e`` option: gsutil config -e Note that if you use gsutil through the Cloud SDK, you instead activate your service account via the `gcloud auth activate-service-account <https://cloud.google.com/sdk/gcloud/reference/auth/activate-service-account>`_ command. When you run ``gsutil config -e``, you are prompted for the path to your private key file and, if not using a JSON key file, your service account email address and key file password. To get this data, follow the instructions on `Service Accounts <https://cloud.google.com/storage/docs/authentication#generating-a-private-key>`_. Using this information, gsutil populates the "gs_service_key_file" attribute in the boto configuration file. If not using a JSON key file, gsutil also populates the "gs_service_client_id" and "gs_service_key_file_password" attributes. Note that your service account is NOT considered an Owner for the purposes of API access (see "gsutil help creds" for more information about this). See https://developers.google.com/identity/protocols/OAuth2ServiceAccount for further information on service account authentication. If you want to use credentials based on access key and secret (the older authentication method before OAuth2 was supported), see the ``-a`` option in the OPTIONS section. If you wish to use gsutil with other providers (or to copy data back and forth between multiple providers) you can edit their credentials into the [Credentials] section after creating the initial boto configuration file. <B>OPTIONS</B> -a Prompt for Google Cloud Storage access key and secret (the older authentication method before OAuth2 was supported) instead of obtaining an OAuth2 token. -e Prompt for service account credentials. This option requires that ``-a`` is not set. -n Write the configuration file without authentication configured. This flag is mutually exclusive with all flags other than ``-o``. -o <file> Write the configuration to <file> instead of ~/.boto. Use ``-`` for stdout. z)https://cloud.google.com/console#/projectaY # This file contains credentials and other configuration information needed # by the boto library, used by gsutil. You can edit this file (e.g., to add # credentials) but be careful not to mis-edit any of the variable names (like # "gs_access_key_id") or remove important markers (like the "[Credentials]" and # "[Boto]" section delimiters). # � � � � �0Z50MZ150MZ200M� Z2G� aK [Boto] # http_socket_timeout specifies the timeout (in seconds) used to tell httplib # how long to wait for socket timeouts. The default is 70 seconds. Note that # this timeout only applies to httplib, not to httplib2 (which is used for # OAuth2 refresh/access token exchanges). #http_socket_timeout = 70 # The following two options control the use of a secure transport for requests # to S3 and Google Cloud Storage. It is highly recommended to set both options # to True in production environments, especially when using OAuth2 bearer token # authentication with Google Cloud Storage. # Set 'https_validate_certificates' to False to disable server certificate # checking. The default for this option in the boto library is currently # 'False' (to avoid breaking apps that depend on invalid certificates); it is # therefore strongly recommended to always set this option explicitly to True # in configuration files, to protect against "man-in-the-middle" attacks. https_validate_certificates = True # 'debug' controls the level of debug messages printed for the XML API only: # 0 for none, 1 for basic boto debug, 2 for all boto debug plus HTTP # requests/responses. #debug = <0, 1, or 2> # 'num_retries' controls the number of retry attempts made when errors occur # during data transfers. The default is 6. # Note 1: You can cause gsutil to retry failures effectively infinitely by # setting this value to a large number (like 10000). Doing that could be useful # in cases where your network connection occasionally fails and is down for an # extended period of time, because when it comes back up gsutil will continue # retrying. However, in general we recommend not setting the value above 10, # because otherwise gsutil could appear to "hang" due to excessive retries # (since unless you run gsutil -D you won't see any logged evidence that gsutil # is retrying). # Note 2: Don't set this value to 0, as it will cause boto to fail when reusing # HTTP connections. #num_retries = <integer value> # 'max_retry_delay' controls the max delay (in seconds) between retries. The # default value is 60, so the backoff sequence will be 1 seconds, 2 seconds, 4, # 8, 16, 32, and then 60 for all subsequent retries for a given HTTP request. # Note: At present this value only impacts the XML API and the JSON API uses a # fixed value of 60. #max_retry_delay = <integer value> aB [GoogleCompute] # 'service_account' specifies the a Google Compute Engine service account to # use for credentials. This value is intended for use only on Google Compute # Engine virtual machines and usually lives in /etc/boto.cfg. Most users # shouldn't need to edit this part of the config. #service_account = default az0 [GSUtil] # 'resumable_threshold' specifies the smallest file size [bytes] for which # resumable Google Cloud Storage uploads are attempted. The default is 8388608 # (8 MiB). #resumable_threshold = %(resumable_threshold)d # 'rsync_buffer_lines' specifies the number of lines of bucket or directory # listings saved in each temp file during sorting. (The complete set is # split across temp files and separately sorted/merged, to avoid needing to # fit everything in memory at once.) If you are trying to synchronize very # large directories/buckets (e.g., containing millions or more objects), # having too small a value here can cause gsutil to run out of open file # handles. If that happens, you can try to increase the number of open file # handles your system allows (e.g., see 'man ulimit' on Linux; see also # http://docs.python.org/2/library/resource.html). If you can't do that (or # if you're already at the upper limit), increasing rsync_buffer_lines will # cause gsutil to use fewer file handles, but at the cost of more memory. With # rsync_buffer_lines set to 32000 and assuming a typical URL is 100 bytes # long, gsutil will require approximately 10 MiB of memory while building # the synchronization state, and will require approximately 60 open file # descriptors to build the synchronization state over all 1M source and 1M # destination URLs. Memory and file descriptors are only consumed while # building the state; once the state is built, it resides in two temp files that # are read and processed incrementally during the actual copy/delete # operations. #rsync_buffer_lines = 32000 # 'state_dir' specifies the base location where files that # need a static location are stored, such as pointers to credentials, # resumable transfer tracker files, and the last software update check. # By default these files are stored in ~/.gsutil #state_dir = <file_path> # gsutil periodically checks whether a new version of the gsutil software is # available. 'software_update_check_period' specifies the number of days # between such checks. The default is 30. Setting the value to 0 disables # periodic software update checks. #software_update_check_period = 30 # 'tab_completion_timeout' controls the timeout (in seconds) for tab # completions that involve remote requests (such as bucket or object names). # If tab completion does not succeed within this timeout, no tab completion # suggestions will be returned. # A value of 0 will disable completions that involve remote requests. #tab_completion_timeout = 5 # 'parallel_process_count' and 'parallel_thread_count' specify the number # of OS processes and Python threads, respectively, to use when executing # operations in parallel. The default settings should work well as configured, # however, to enhance performance for transfers involving large numbers of # files, you may experiment with hand tuning these values to optimize # performance for your particular system configuration. #parallel_process_count = %(parallel_process_count)d #parallel_thread_count = %(parallel_thread_count)d # 'parallel_composite_upload_threshold' specifies the maximum size of a file to # upload in a single stream. Files larger than this threshold will be # partitioned into component parts and uploaded in parallel and then composed # into a single object. # The number of components will be the smaller of # ceil(file_size / parallel_composite_upload_component_size) and # MAX_COMPOSE_ARITY. The current value of MAX_COMPOSE_ARITY is # %(max_compose_arity)d. # If 'parallel_composite_upload_threshold' is set to 0, then automatic parallel # uploads will never occur. # Setting an extremely low threshold is unadvisable. The vast majority of # environments will see degraded performance for thresholds below 80M, and it # is almost never advantageous to have a threshold below 20M. # 'parallel_composite_upload_component_size' specifies the ideal size of a # component in bytes, which will act as an upper bound to the size of the # components if ceil(file_size / parallel_composite_upload_component_size) is # less than MAX_COMPOSE_ARITY. # Values can be provided either in bytes or as human-readable values # (e.g., "150M" to represent 150 mebibytes) # # Note: At present parallel composite uploads are disabled by default, because # using composite objects requires a compiled crcmod (see "gsutil help crcmod"), # and for operating systems that don't already have this package installed this # makes gsutil harder to use. Google is actively working with a number of the # Linux distributions to get crcmod included with the stock distribution. Once # that is done we will re-enable parallel composite uploads by default in # gsutil. # # Note: Parallel composite uploads should not be used with NEARLINE, COLDLINE, # or ARCHIVE storage class buckets, as doing this incurs an early deletion # charge for each component object. # # Note: Parallel composite uploads are not enabled with Cloud KMS encrypted # objects as a source or destination, as composition with KMS objects is not yet # supported. #parallel_composite_upload_threshold = %(parallel_composite_upload_threshold)s #parallel_composite_upload_component_size = %(parallel_composite_upload_component_size)s # # 'parallel_composite_upload_bypass_kms_check' removes the object/bucket KMS checks # used to guard composition of KMS objects. #disable_parallel_composite_upload_kms_check = False # 'sliced_object_download_threshold' and # 'sliced_object_download_component_size' have analogous functionality to # their respective parallel_composite_upload config values. # 'sliced_object_download_max_components' specifies the maximum number of # slices to be used when performing a sliced object download. #sliced_object_download_threshold = %(sliced_object_download_threshold)s #sliced_object_download_component_size = %(sliced_object_download_component_size)s #sliced_object_download_max_components = %(sliced_object_download_max_components)s # Compressed transport encoded uploads buffer chunks of compressed data. When # running a composite upload and/or many uploads in parallel, compression may # consume more memory than available. This setting restricts the number of # compressed transport encoded uploads running in parallel such that they # don't consume more memory than set here. This is 2GiB by default. # Values can be provided either in bytes or as human-readable values # (e.g., "2G" to represent 2 gibibytes) #max_upload_compression_buffer_size = %(max_upload_compression_buffer_size)s # GZIP compression level, if using compression. Reducing this can have # a dramatic impact on compression speed with minor size increases. # This is a value from 0-9, with 9 being max compression. # A good level to try is 6, which is the default used by the gzip tool. #gzip_compression_level = %(gzip_compression_level)s # 'task_estimation_threshold' controls how many files or objects gsutil # processes before it attempts to estimate the total work that will be # performed by the command. Estimation makes extra directory listing or API # list calls and is performed only if multiple processes and/or threads are # used. Estimation can slightly increase cost due to extra # listing calls; to disable it entirely, set this value to 0. #task_estimation_threshold=%(task_estimation_threshold)s # 'use_magicfile' specifies if the 'file --mime <filename>' command should be # used to guess content types instead of the default filename extension-based # mechanism. Available on UNIX and macOS (and possibly on Windows, if you're # running Cygwin or some other package that provides implementations of # UNIX-like commands). When available and enabled use_magicfile should be more # robust because it analyzes file contents in addition to extensions. #use_magicfile = False # Service account emails for testing the hmac command. If these fields are not # populated with distinct service accounts the tests for the hmac command will # not be run. Primarily useful for tool developers. #test_hmac_service_account = #test_hmac_alt_service_account = #test_hmac_list_service_account = # Service account emails for testing impersonation credentials. If this field is # not populated with a service account the tests for service account # impersonation will not run. Primarily useful for tool developers. #test_impersonate_service_account = # 'content_language' specifies the ISO 639-1 language code of the content, to be # passed in the Content-Language header. By default no Content-Language is sent. # See the ISO 639-1 column of # http://www.loc.gov/standards/iso639-2/php/code_list.php for a list of # language codes. content_language = en # 'check_hashes' specifies how strictly to require integrity checking for # downloaded data. Legal values are: # '%(hash_fast_else_fail)s' - (default) Only integrity check if the digest # will run efficiently (using compiled code), else fail the download. # '%(hash_fast_else_skip)s' - Only integrity check if the server supplies a # hash and the local digest computation will run quickly, else skip the # check. # '%(hash_always)s' - Always check download integrity regardless of possible # performance costs. # '%(hash_never)s' - Don't perform download integrity checks. This setting is # not recommended except for special cases such as measuring download # performance excluding time for integrity checking. # This option exists to assist users who wish to download a GCS composite object # and are unable to install crcmod with the C-extension. CRC32c is the only # available integrity check for composite objects, and without the C-extension, # download performance can be significantly degraded by the digest computation. # This option is ignored for daisy-chain copies, which don't compute hashes but # instead (inexpensively) compare the cloud source and destination hashes. #check_hashes = if_fast_else_fail # 'encryption_key' specifies a single customer-supplied encryption key that # will be used for all data written to Google Cloud Storage. See # "gsutil help encryption" for more information # Encryption key: RFC 4648 section 4 base64-encoded AES256 string # Warning: If decrypt_key is specified without an encrypt_key, objects will be # decrypted when copied in the cloud. #encryption_key= # Each 'decryption_key' entry specifies a customer-supplied decryption key that # will be used to access and Google Cloud Storage objects encrypted with # the corresponding key. # Decryption keys: Up to 100 RFC 4648 section 4 base64-encoded AES256 strings # in ascending numerical order, starting with 1. #decryption_key1= #decryption_key2= #decryption_key3= # The ability to specify an alternative JSON API version is primarily for cloud # storage service developers. #json_api_version = v1 # Specifies the API to use when interacting with cloud storage providers. If the # gsutil command supports this API for the provider, it will be used instead of # the default API. Commands typically default to XML for S3 and JSON for GCS. # Note that if any encryption configuration options are set (see above), the # JSON API will be used for interacting with Google Cloud Storage buckets even # if XML is preferred, as gsutil does not currently support this functionality # when using the XML API. #prefer_api = json #prefer_api = xml # Disables the prompt asking for opt-in to data collection for analytics. #disable_analytics_prompt = True # The "test" command runs tests against regional buckets (unless you supply the # `-b` option). By default, the region used is us-central1, but you can change # the default region using this option. #test_cmd_regional_bucket_location = us-central1 # Tests for the "notification watchbucket" command require a notification URL. # If this option is not supplied, those tests will be skipped. #test_notification_url = https://yourdomain.url/notification-endpoint # Used in conjunction with --stet flag on cp command for end-to-end encryption. # STET binary path. If not specified, gsutil checks PATH for "stet". #stet_binary_path = <Path to binary "/usr/local/bin/stet"> # STET config path. If not specified, the STET binary will run with its default # settings. #stet_config_path = ~/.config/my_config.yaml # Adds an API call before parallel operations that triggers a reauth challenge. #trigger_reauth_challenge_for_parallel_operations = False )Zhash_fast_else_failZhash_fast_else_skipZhash_alwaysZ hash_never�resumable_threshold�parallel_process_count�parallel_thread_count�#parallel_composite_upload_threshold�(parallel_composite_upload_component_size� sliced_object_download_threshold�%sliced_object_download_component_size�%sliced_object_download_max_componentsZmax_compose_arity�task_estimation_threshold�"max_upload_compression_buffer_sizeZgzip_compression_levela0 [OAuth2] # This section specifies options used with OAuth2 authentication. # 'token_cache' specifies how the OAuth2 client should cache access tokens. # Valid values are: # 'in_memory': an in-memory cache is used. This is only useful if the boto # client instance (and with it the OAuth2 plugin instance) persists # across multiple requests. # 'file_system' : access tokens will be cached in the file system, in files # whose names include a key derived from the refresh token the access token # based on. # The default is 'file_system'. #token_cache = file_system #token_cache = in_memory # 'token_cache_path_pattern' specifies a path pattern for token cache files. # This option is only relevant if token_cache = file_system. # The value of this option should be a path, with place-holders '%(key)s' (which # will be replaced with a key derived from the refresh token the cached access # token was based on), and (optionally), %(uid)s (which will be replaced with # the UID of the current user, if available via os.getuid()). # Note that the config parser itself interpolates '%' placeholders, and hence # the above placeholders need to be escaped as '%%(key)s'. # The default value of this option is # token_cache_path_pattern = <tmpdir>/oauth2client-tokencache.%%(uid)s.%%(key)s # where <tmpdir> is the system-dependent default temp directory. # The following options specify the label and endpoint URIs for the OAUth2 # authorization provider being used. Primarily useful for tool developers. #provider_label = Google #provider_authorization_uri = https://accounts.google.com/o/oauth2/auth #provider_token_uri = https://oauth2.googleapis.com/token # 'oauth2_refresh_retries' controls the number of retry attempts made when # rate limiting errors occur for OAuth2 requests to retrieve an access token. # The default value is 6. #oauth2_refresh_retries = <integer value> # The following options specify the OAuth2 client identity and secret that is # used when requesting and using OAuth2 tokens. If not specified, a default # OAuth2 client for the gsutil tool is used; for uses of the boto library (with # OAuth2 authentication plugin) in other client software, it is recommended to # use a tool/client-specific OAuth2 client. For more information on OAuth2, see # http://code.google.com/apis/accounts/docs/OAuth2.html c @ s� e Zd ZdZejdg d�eddddgdddd� Zejdg d �d de i d�Z d d� Zdd� Zdd� Z dd� Zejdfdd�Zdd� ZdS )� ConfigCommandz(Implementation of gsutil config command.�config)�cfg�conf� configurer zabefhno:rs:w�reauthF) �command_name_aliases�usage_synopsis�min_args�max_args�supported_sub_args�supported_private_args�file_url_ok�provider_url_ok�urls_start_arg)r. r/ r0 �aws�s3�command_helpz0Obtain credentials and create configuration file)� help_name�help_name_aliases� help_type�help_one_line_summary� help_text�subcommand_help_textc C s~ t jt jB t jB }tt d�r&|t jO }zt �||d�}W n8 ttfyp } zt d||f ��W Y d}~n d}~0 0 t � |d�S )a Creates and opens a configuration file for writing. The file is created with mode 0600, and attempts to open existing files will fail (the latter is important to prevent symlink attacks). It is the caller's responsibility to close the file. Args: file_path: Path of the file to be created. Returns: A writable file object for the opened file. Raises: CommandException: if an error occurred when opening the file (including when the file already exists). �O_NOINHERITi� z!Failed to open %s for writing: %sN�w)�os�O_RDWR�O_CREAT�O_EXCL�hasattrrD �open�OSError�IOErrorr �fdopen)�self� file_path�flags�fd�e� rT �$/opt/gsutil/gslib/commands/config.py�_OpenConfigFileV s �zConfigCommand._OpenConfigFilec C s� t jr dS t�|�}ttjtjB |j@ �r�| j� d� t d�}|dv r�zt�|d� | j�d� W q� t y� } z| j� d� W Y d}~q�d}~0 0 n| j�d� dS ) aA Checks that the file has reasonable permissions for a private key. In particular, check that the filename provided by the user is not world- or group-readable. If either of these are true, we issue a warning and offer to fix the permissions. Args: file_path: The name of the private key file. Nz� Your private key file is readable by people other than yourself. This is a security risk, since anyone with this information can use your service account. zDWould you like gsutil to change the file permissions for you? (y/N) )�y�Y� z� The permissions on your file have been successfully modified. The only access allowed is readability by the user (permissions 0400 in chmod).z� We were unable to modify the permissions on your file. If you would like to fix this yourself, consider running: "sudo chmod 400 </path/to/key>" for improved security.z� You have chosen to allow this file to be readable by others. If you would like to fix this yourself, consider running: "sudo chmod 400 </path/to/key>" for improved security.)r � IS_WINDOWSrF �stat�bool�S_IRGRP�S_IROTH�st_mode�logger�warnr �chmod�info� Exception)rO rP �stZfix_it�_rT rT rU �_CheckPrivateKeyFilePermissionss s* ����z-ConfigCommand._CheckPrivateKeyFilePermissionsc C s* |sd| }d| }|� d||f � dS )a� Writes proxy name/value pair or comment line to config file. Writes proxy name/value pair if value is not None. Otherwise writes comment line. Args: config_file: File object to which the resulting config file will be written. name: The config variable name. value: The value, or None. desc: Human readable description (for comment). z#%sz<%s>z%s = %s N)�write)rO �config_file�name�value�descrT rT rU �_WriteConfigLineMaybeCommented� s z,ConfigCommand._WriteConfigLineMaybeCommentedc C s� t j}|�d� | �|d|�ddd�d� | �|d|�ddd�d� | �|d|�ddd�d � | �|d |�dd d�d� | �|d|�ddd�d � | �|d|�ddd�d� dS )z�Writes proxy section of configuration file. Args: config_file: File object to which the resulting config file will be written. a? # To use a proxy, edit and uncomment the proxy and proxy_port lines. # If you need a user/password with this proxy, edit and uncomment # those lines as well. If your organization also disallows DNS # lookups by client machines, set proxy_rdns to True (the default). # If you have installed gsutil through the Cloud SDK and have # configured proxy settings in gcloud, those proxy settings will # override any other options (including those set here, along with # any settings in proxy-related environment variables). Otherwise, # if proxy_host and proxy_port are not specified in this file and # one of the OS environment variables http_proxy, https_proxy, or # HTTPS_PROXY is defined, gsutil will use the proxy server specified # in these environment variables, in order of precedence according # to how they are listed above. �proxy�BotoNz proxy host� proxy_typez4proxy type (socks4, socks5, http) | Defaults to http� proxy_portz proxy port� proxy_userz proxy user� proxy_passzproxy password� proxy_rdnsFzLlet proxy server perform DNS lookups (True,False); socks proxy not supported)�botor- rh rm � get_value)rO ri r- rT rT rU �_WriteProxyConfigFileSection� s<