The expected configuration.toml has changed several keys to make it easier on users to find the right information in the Azure Management Console. The following keys have changed:
client_id
is nowbatch_application_id
principal_id
is nowbatch_object_id
application_id
is nowsp_application_id
Refer to the example_config.toml in the examples folder, found here to view the required keys/values needed in the configuration file.
The cfa_azure
python module is intended to ease the challenge of working with Azure via multiple Azure python modules which require the correct steps and many lines of code to execute. cfa_azure
simplifies many repeated workflows when interacting with Azure, Blob Storage, Batch, and more. For example, creating a pool in Azure may take different credentials and several clients to complete, but with cfa_azure
, creating a pool is reduced to a single function with only a few parameters.
The cfa_azure
module is composed of three submodules: clients
, automation
and helpers
. The module clients
contains what we call the AzureClient, which combines the multiple Azure Clients needed to interact with Azure and consolidates to a single client. The module helpers
contains more fine-grained functions which are used within the clients
module or independently for more control when working with Azure. The automation
module introduces a simplified way to upload files and submit jobs/tasks to Batch via another configuration toml file. For help getting started with the automation
module, please see this overview.
Classes:
- AzureClient: a client object used for interacting with Azure. It initializes based on a supplied configuration file and creates various Azure clients under the hood. It can be used to upload containers, upload files, run jobs, and more.
To customize the logging capabilities of cfa_azure, two environment variables can be set. These are LOG_LEVEL and LOG_OUTPUT.
LOG_LEVEL: sets the logging level. Choices are:
- debug
- info
- warning
- error
LOG_OUTPUT: sets the output of the logs. Choices are:
- file: saves log output to a file, nested within a ./logs/ folder
- stdout: saves log output to stdout
- both: saves log output to both file and stdout
Example:
Run the following in the terminal in which cfa_azure
will be run.
export LOG_LEVEL="info"
export LOG_OUTPUT="stdout"
Using Various Credential Methods
When instantiating a AzureClient object, there is an option to set which credential_method
to use. Previously, only a service principal could be used. Now, there are three an options to choose identity
, sp
, or env
.
identity
: Uses the managed identity associated with the VM where the code is running.sp
: Uses a service principal for the credential. The following values must be set in the configuration file: tenant_id, sp_application_id, and the corresponding secret fetched from Azure Key Vault.env
: Uses environment variables to create the credential. When choosingenv
, the following environment variables will need to be set:AZURE_TENANT_ID
,AZURE_CLIENT_ID
, andAZURE_CLIENT_SECRET
.
You can also use use_env_vars=True
to allow the configuration to be loaded directly from environment variables, which may be helpful in containerized environments.
By default, the managed identity option will be used. In whichever credential method is used, a secret is pulled from the key vault using the credential to create a secret client credential for interaction with various Azure services.
Example:
from cfa_azure.clients import AzureClient
# Using Managed Identity
client = AzureClient(config_path="./configuration.toml", credential_method="identity")
# Using Service Principal (credentials from the config file)
client = AzureClient(config_path="./configuration.toml", credential_method="sp")
# Using Environment Variables
import os
os.environ["AZURE_TENANT_ID"] = "your-tenant-id"
os.environ["AZURE_CLIENT_ID"] = "your-client-id"
os.environ["AZURE_CLIENT_SECRET"] = "your-client-secret" #pragma: allowlist secret
client = AzureClient(credential_method="env", use_env_vars=True)
Persisting stdout and stderr to Blob Storage
In certain situations, it is beneficial to save the stdout and stderr from each task to Blob Storage (like when using autoscale pools). It is possible to persist these to Blob Storage by specifying the blob container name in the save_logs_to_blob
parameter when using client.add_job()
. *Note that the blob container specified must be mounted to the pool being used for the job.
For example, if we would like to persist stdout and stderr to the blob container "input-test" for a job named "persisting_test", we would use the following code:
client.add_job("persisting_test", save_logs_to_blob = "input-test")
Availability Zones
To make use of Azure's availability zone functionality there is a parameter available in the set_pool_info()
method called availability_zones
. To use availability zones when building a pool, set this parameter to True. If you want to stick with the default Regional configuration, this parameter can be left out or set to False. Turn availability zone on like the following:
client.set_pool_info(
...
availability_zones = True,
...
)
Updated High Performance Compute Image
The default base Ubuntu image used for Azure Batch nodes is Ubuntu 20.04, which is nearing end of life on 4/22/2025. There is an option to use a high performance compute image using Ubuntu 22.04 as the base OS. It's important to use a compatible VM size with these HPC images. To implement a HPC image for Azure pools, set the parameter use_hpc_image
to True
in the AzureClient
method set_pool_info()
, like the following:
client.set_pool_info("autoscale",
timeout=60,
...,
use_hpc_image = True
)
create_pool
: creates a new Azure batch pool using default autoscale mode Example:client = AzureClient("./configuration.toml") client.create_pool("my-test-pool")
update_containers
: modifies the containers mounted on an existing Azure batch pool. It essentially recreates the pool with new mounts. Use force_update=True to recreate the pool without waiting for running tasks to complete.upload_files_to_container
: uploads files from a specified folder to an Azure Blob container. It also includes options likeforce_upload
to allow or deny large file uploads without confirmation. Example:
client.upload_files_to_container(
folder_names=["/path/to/folder"],
input_container_name="my-input-container",
blob_service_client=client.blob_service_client,
force_upload=True
)
-
update_scale_settings
: modifies the scaling mode (fixed or autoscale) for an existing pool Example:# Specify new autoscale formula that will be evaluated every 30 minutes client.scaling = "autoscale" client.update_scale_settings( pool_name="my-test-pool", autoscale_formula_path="./new_autoscale_formula.txt", evaluation_interval="PT30M" ) # Set the pool name property to avoid sending pool_name parameter on every call to update_scale_settings client.pool_name = "my-test-pool" # Use default 15 minute evaluation interval client.update_scale_settings(autoscale_formula_path="./new_autoscale_formula.txt") # Switch to fixed scaling mode with 10 on-demand EC2 nodes and requeuing of current jobs client.scaling = "fixed" client.update_scale_settings(dedicated_nodes=10, node_deallocation_option='Requeue') # Switch to fixed scaling mode with 15 spot EC2 nodes and forced termination of current jobs client.update_scale_settings(low_priority_nodes=15, node_deallocation_option='Terminate')
-
update_containers: modifies the containers mounted on an existing Azure batch pool. It essentially recreates the pool with new mounts. Example:
# First create a pool client = AzureClient("./configuration.toml") client.set_input_container("some-input-container") client.set_output_container("some-output-container") client.create_pool(pool_name="my-test-pool") # Now change the containers mounted on this pool client.update_containers( pool_name="my-test-pool", input_container_name="another-input-container", output_container_name="another-output-container", force_update=False )
If all the nodes in pool were idle when update_containers() method was invoked, Azure Batch service will recreate the pool with new containers mounted to /input and /output paths respectively. However, if any nodes in pool were in Running state, then the following error shall be displayed:
There are N compute nodes actively running tasks in pool. Please wait for jobs to complete or retry with
force_update=True
.As the message suggests, you can either wait for existing jobs to complete in the pool and retry the
update_containers()
operation. Or you can change theforce_update
parameter toTrue and re-run the
update_containers()` operation to immediately recreate the pool with new containers.
add_task
: adds task to existing job in pool. You can also specify which task it depends on. By default, dependent tasks will only run if the parent task succeeds. However, this behavior can be overridden by specifyingrun_dependent_tasks_on_fail=True
on the parent task. When this property is set to True, any runtime failures in parent task will be ignored. However, execution of dependent tasks will only begin after completion (regardless of success or failure) of the parent task.
Example: Run tasks in parallel without any dependencies.
task_1 = client.add_task(
"test_job_id",
docker_cmd=["some", "docker", "command"], # replace with actual command
input_files=["test_file_1.sh"]
)
task_2 = client.add_task(
"test_job_id",
docker_cmd=["some", "other", "docker", "command"], # replace with actual command
use_uploaded_files=False,
input_files=["test_file_2.sh"]
)
Example: Run tasks sequentially and terminate the job if parent task fails
parent_task = client.add_task(
"test_job_id",
docker_cmd=["some", "docker", "command"], # replace with actual command
use_uploaded_files=False,
input_files=["test_file_1.sh"]
)
child_task = client.add_task(
"test_job_id",
docker_cmd=["some", "other", "docker", "command"], # replace with actual command
use_uploaded_files=False,
depends_on=parent_task,
input_files=["test_file_2.sh"]
)
Example: Run tasks sequentially with 1-to-many dependency. Run the child tasks even if parent task fails.
parent_task = client.add_task(
"test_job_id",
docker_cmd=["some", "docker", "command"], # replace with actual command
use_uploaded_files=False,
run_dependent_tasks_on_fail=True,
input_files=["test_file_1.sh"]
)
child_task_1 = client.add_task(
"test_job_id",
docker_cmd=["some", "other", "docker", "command"], # replace with actual command
depends_on=parent_task,
input_files=["test_file_2.sh"]
)
child_task_2 = client.add_task(
"test_job_id",
docker_cmd=["another", "docker", "command"], # replace with actual command
depends_on=parent_task,
input_files=["test_file_3.sh"]
)
Example: Create many-to-1 dependency with 2 parent tasks that run before child task. Second parent task is optional: job should not terminate if it fails.
parent_task_1 = client.add_task(
"test_job_id",
docker_cmd=["some", "docker", "command"], # replace with actual command
input_files=["test_file_1.sh"]
)
parent_task_2 = client.add_task(
"test_job_id",
docker_cmd=["some", "other", "docker", "command"], # replace with actual command
run_dependent_tasks_on_fail=True,
input_files=["test_file_2.sh"]
)
child_task = client.add_task(
"test_job_id",
docker_cmd=["another", "docker", "command"], # replace with actual command
depends_on=(parent_task_1 + parent_task_2)
input_files=["test_file_2.sh"]
)
Please view this documentation on getting started with the automation
module.
The helpers
module provides a collection of functions that helps manage Azure resources and perform key tasks, such as interacting with Blob storage, Azure Batch, configuration management, and data transformations. Below is an expanded overview of each function.
Functions:
read_config
: reads in a configuration toml file and returns it as a Python dictionary
read_config("/path/to/config.toml")
create_container
: creates an Azure Blob container if it doesn't already exist
create_container("my-container", blob_service_client)
get_autoscale_formula
: finds and readsautoscale_formula.txt
from working directory or subdirectory
get_autoscale_formula(filepath="/path/to/formula.txt")
get_sp_secret
: retrieves the user's service principal secret from the key vault based on the provided config file
get_sp_secret(config, DefaultAzureCredential())
get_sp_credential
: retrieves the service principal credential
get_sp_credential(config)
get_blob_service_client
: creates a Blob Service Client for interacting with Azure Blob
blob_service_client = get_blob_service_client(config, DefaultAzureCredential())
get_batch_mgmt_client
: creates a Batch Management Client for interacting with Azure Batch, such as pools and jobs
batch_mgmt_client = get_batch_mgmt_client(config, DefaultAzureCredential())
create_blob_containers
: uses create_container() to create input and output containers in Azure Blob
create_blob_containers(blob_service_client, "input-container", "output-container")
get_batch_pool_json
: creates a dict based on config for configuring an Azure Batch pool
pool_config = get_batch_pool_json("input-container", "output-container", config)
create_batch_pool
: creates a Azure Batch Pool based on info using the provided configuration details
create_batch_pool(batch_mgmt_client, pool_config)
list_containers
: lists the containers in Azure Blob Storage Account
list_containers(blob_service_client)
upload_files_in_folder
: uploads all files in specified folder to the specified container
upload_files_in_folder("/path/to/folder", "container-name", blob_service_client)
get_batch_service_client
: creates a Batch Service Client object for interacting with Batch jobs
batch_client = get_batch_service_client(config, DefaultAzureCredential())
add_job
: creates a new job to the specified Azure Batch pool. By default, a job remains active after completion of enclosed tasks. You can optionally specify the mark_complete_after_tasks_run argument to True if you want job to auto-complete after completion of enclosed tasks.
add_job("job-id", "pool-id", True, batch_client)
add_task_to_job
: adds a task to the specified job based on user-input Docker command
add_task_to_job("job-id", "task-id", "docker-command", batch_client)
monitor_tasks
: monitors the tasks running in a job
monitor_tasks("example-job-id", batch_client)
list_files_in_container
: lists out all files stored in the specified Azure container
list_files_in_container(container_client)
df_to_yaml
: converts a pandas dataframe to yaml file, which is helpful for configuration and metadata storage
df_to_yaml(dataframe, "output.yaml")
yaml_to_df
: converts a yaml file to pandas dataframe
yaml_to_df("input.yaml")
edit_yaml_r0
: takes in a YAML file and produces replicate YAML files with ther0
changed based on the specified range (i.e. start, stop, and step)
edit_yaml_r0("input.yaml", start=1, stop=5, step=1)
get_user_identity
: retrieves the user identity based on the provided config information
get_user_identity(config)
get_network_config
: gets the network configuration based on the config information
get_network_config(config: str)
get_deployment_config
: retrieves deployment configuration for Azure Batch pool, including container registry settings and optional HPC image
get_deployment_config("container_image_name", "container_registry_url", "container_registry_server", config, DefaultAzureCredential())
check_blob_existence
: checks whether a blob exists in the specified container
check_blob_existence(c_client, "blob_name")
check_virtual_directory_existence
: checks whether any blobs exist with the specified virtual directory path
check_virtual_directory_existence(c_client, "vdir_path")
download_file
: downloads a file from Azure Blob storage to a specified location
download_file(c_client, "src_path", "dest_path")
list_blobs_flat
: lists all blobs in a specified container
list_blobs_flat("container_name", blob_service_client)
get_log_level
: retrieves the logging level from environment variables or defaults to debug
get_log_level()
delete_blob_snapshots
: deletes a blob and all its snapshots in a container
delete_blob_snapshots("blob_name", "container_name", blob_service_client)
delete_blob_folder
: deletes all blobs in a specified folder in a container
delete_blob_folder("folder_path", "container_name", blob_service_client)
format_extensions
: formats file extensions into a standard format for use
format_extensions([".txt", "jpg"])
check_autoscale_parameters
: checks which arguments are incompatible with the provided scaling mode
check_autoscale_parameters("autoscale", dedicated_nodes=5)
get_rel_mnt_path
: retrieves the relative mount path for a specified Blob container in an Azure Batch pool
get_rel_mnt_path("blob_name", "pool_name", "resource_group_name", "account_name", batch_mgmt_client)
get_pool_mounts
: lists all mounted Blob containers for a given Azure Batch pool
get_pool_mounts("pool_name", "resource_group_name", "account_name", batch_mgmt_client)
check_env_req
: checks if all necessary environment variables exist for the Azure client
check_env_req()
check_config_req
:checks if the provided configuration file contains all necessary components for the Azure client
check_config_req(config)
get_container_registry_client
: retrieves a Container Registry client for Azure
get_container_registry_client("endpoint", DefaultAzureCredential(), "audience")
check_azure_container_exists
: checks if a container with the specified name, repository, and tag exists in Azure Container Registry
check_azure_container_exists("registry_name", "repo_name", "tag_name", DefaultAzureCredential())
generate_autoscale_formula
: generates a generic autoscale formula for use based on a specified maximum number of nodes
generate_autoscale_formula(max_nodes=8)
format_rel_path
: formats a given relative path by removing the leading slash if present
format_rel_path("/path/to/resource")
get_timeout
: converts a given duration string (in ISO 8601 format) to minutes
get_timeout("PT1H30M")
check_job_exists
: checks whether a job with the specified ID exists in Azure Batch
check_job_exists("job_id", batch_client)
get_completed_tasks
: returns the number of completed tasks for the specified job
get_completed_tasks("job_id", batch_client)
check_job_complete
: checks if the specified job is complete
check_job_complete("job_id", batch_client)
get_job_state
: returns the state of the specified job, such as 'completed' or 'active'
get_job_state("job_id", batch_client)
package_and_upload_dockerfile
: packages a Dockerfile and uploads it to the specified registry and repo with the designated tag
package_and_upload_dockerfile("registry_name", "repo_name", "tag", use_device_code=True)
upload_docker_image
: uploads a Docker image to a specified Azure Container Registry repo with an optional tag
upload_docker_image("image_name", "registry_name", "repo_name", tag="latest", use_device_code=False)
check_pool_exists
: checks if a specified pool exists in Azure Batch
check_pool_exists("resource_group_name", "account_name", "pool_name", batch_mgmt_client)
get_pool_info
: gets the basic information for a specified Azure Batch pool
get_pool_info("resource_group_name", "account_name", "pool_name", batch_mgmt_client)
get_pool_full_info
: retrieves the full information of a specified Azure Batch pool
get_pool_full_info("resource_group_name", "account_name", "pool_name", batch_mgmt_client)
delete_pool
: deletes the specified pool from Azure Batch
delete_pool("resource_group_name", "account_name", "pool_name", batch_mgmt_client)
upload_blob_file
: uploads a specified file to Azure Blob storage
upload_blob_file("file_path", location="folder/subfolder", container_client=container_client, verbose=True)
- download_directory: downloads a directory using prefix matching from Azure Blob storage
download_directory("container_name", "src_path", "dest_path", blob_service_client, include_extensions=".txt", verbose=True)
Example Workflow: Uploading files to Blob Storage, creating an Azure Batch Pool, adding jobs, and monitoring tasks.
# Step 1: Read configuration
config = read_config("config.toml")
# Step 2: Create Blob containers
blob_service_client = get_blob_service_client(config, DefaultAzureCredential())
create_blob_containers(blob_service_client, "input-container", "output-container")
# Step 3: Upload files to the container
upload_files_in_folder("/path/to/folder", "input-container", blob_service_client)
# Step 4: Create an Azure Batch Pool
batch_mgmt_client = get_batch_mgmt_client(config, DefaultAzureCredential())
pool_config = get_batch_pool_json("input-container", "output-container", config)
create_batch_pool(batch_mgmt_client, pool_config)
# Step 5: Create a job and add tasks
batch_client = get_batch_service_client(config, DefaultAzureCredential())
add_job("job-id", "pool-id", True, batch_client)
add_task_to_job("job-id", "task-id", "docker command", batch_client)
# Step 6: Monitor the tasks
monitor_tasks("job-id", batch_client)
This repository constitutes a work of the United States Government and is not subject to domestic copyright protection under 17 USC § 105. This repository is in the public domain within the United States, and copyright and related rights in the work worldwide are waived through the CC0 1.0 Universal public domain dedication. All contributions to this repository will be released under the CC0 dedication. By submitting a pull request you are agreeing to comply with this waiver of copyright interest.
The repository utilizes code licensed under the terms of the Apache Software License and therefore is licensed under ASL v2 or later.
This source code in this repository is free: you can redistribute it and/or modify it under the terms of the Apache Software License version 2, or (at your option) any later version.
This source code in this repository is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the Apache Software License for more details.
You should have received a copy of the Apache Software License along with this program. If not, see http://www.apache.org/licenses/LICENSE-2.0.html
The source code forked from other open source projects will inherit its license.
This repository contains only non-sensitive, publicly available data and information. All material and community participation is covered by the Disclaimer and Code of Conduct. For more information about CDC's privacy policy, please visit http://www.cdc.gov/other/privacy.html.
Anyone is encouraged to contribute to the repository by forking or creating a new branch and submitting a pull request. (If you are new to GitHub, you might start with a basic tutorial.) By contributing to this project, you grant a world-wide, royalty-free, perpetual, irrevocable, non-exclusive, transferable license to all users under the terms of the Apache Software License v2 or later.
All comments, messages, pull requests, and other submissions received through CDC including this GitHub page may be subject to applicable federal law, including but not limited to the Federal Records Act, and may be archived. Learn more at http://www.cdc.gov/other/privacy.html.
Help make this package/repo more robust and stable by creating issues as you see fit. Please use the following issues template as an outline for your issue: issue template
This repository is not a source of government records, but is a copy to increase collaboration and collaborative potential. All government records will be published through the CDC web site.
Please refer to CDC's Template Repository for more information about contributing to this repository, public domain notices and disclaimers, and code of conduct.