Skip to content

Commit 9ffcb16

Browse files
committed
Update README.md
1 parent cf3145a commit 9ffcb16

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -55,7 +55,7 @@ The repository contains the following resources:
5555
- [**TensorFlow Script Mode Debug Training Script**](tensorflow_script_mode_debug_local_training): This example shows how to debug your training script running inside a prebuilt SageMaker Docker image for TensorFlow, on your local machine using SageMaker local mode.
5656
- [**TensorFlow Script Mode Deploy a Trained Model and inference on file from S3**](tensorflow_script_mode_local_model_inference): This example shows how to deploy a trained model to a SageMaker endpoint, on your local machine using SageMaker local mode, and inference with a file in S3 instead of http payload for the SageMaker Endpoint.
5757
- [**TensorFlow Script Mode Training and Batch Transform**](tensorflow_script_mode_california_housing_local_training_and_batch_transform): This example shows how to train your model and run Batch Transform job with TensorFlow and SageMaker script mode, on your local machine using SageMaker local mode.
58-
- [**TensorFlow Extend AWS TensorFlow Deep Learning Container Image**](tensorflow_extend_dlc_california_housing_local_training): In this example we show how to package a TensorFlow container, extending the SageMaker TensorFlow container, with a Python example which works with the California Housing dataset. By extending the SageMaker TensorFlow container we can utilize the existing training solution made to work on SageMaker, leveraging SageMaker TensorFlow `Estimator` object, with `entry_point` parameter, specifying your local Python source file which should be executed as the entry point to training. To make it work, we replace the `framework_version` and `py_version` parameters, with `image_uri` of the Docker Image we have created.
58+
- [**TensorFlow extend AWS TensorFlow Deep Learning Container Image**](tensorflow_extend_dlc_california_housing_local_training): In this example we show how to package a TensorFlow container, extending the SageMaker TensorFlow container, with a Python example which works with the California Housing dataset. By extending the SageMaker TensorFlow container we can utilize the existing training solution made to work on SageMaker, leveraging SageMaker TensorFlow `Estimator` object, with `entry_point` parameter, specifying your local Python source file which should be executed as the entry point to training. To make it work, we replace the `framework_version` and `py_version` parameters, with `image_uri` of the Docker Image we have created.
5959

6060
- **PyTorch resources:**
6161

0 commit comments

Comments
 (0)