- Generate example data to use in training
- Train the model to make predictions, or inferences
- Use an algorithm and example data to train the model
- Evaluate the model for inference accuracy
- Integrate model into your application to generate inferences in real time and at scale
Train the Model
Training Algorithm Options
- Use a SageMaker built-in algorithm
- Use the SageMaker debugger
- Use Apache Spark with SageMaker
- Use custom deep learning code
- Use your own algorithms
- Use an algorithm from the AWS Marketplace
Deploy the Model
Deploy in one of two ways
- Persistent endpoint to get one prediction at a time: SageMaker hosting services
- Get predictions for an entire dataset: SageMaker batch transform
SageMaker Hosting Services
SageMaker provides an HTTPS endpoint, making your model available for inference requests
- Create model in SageMaker
- Create an endpoint configuration
- Create HTTPS endpoint
SageMaker Batch Transform
SageMaker batch transform provides inferences for an entire dataset
- Create batch transform job using trained model and dataset
- Run the batch transform job
- SageMaker saves in a results S3 bucket
Evaluate the Model
- After training and deploying the model, evaluate it to determine performance and accuracy
- Often generate multiple models with different algorithms/hyperparameters and evaluate each
- Two different validation approaches
- Offline testing: use historical data to send requests to the model for inferences
- Online testing with live data: use production variants
- Options for offline evaluation
- Holdout set: set aside a subset of the data for evaluation after training
- K-fold validation: split the example dataset into k parts, treat each as a holdout set for k training runs
本文由
Oscaner
创作, 采用
知识共享署名4.0
国际许可协议进行许可
本站文章除注明转载/出处外, 均为本站原创或翻译, 转载前请务必署名
-
Previous
[MLS-C01] [Exploratory Data Analysis] Analyze and visualize data -
Next
[MLS-C01] [Modeling] Modeling Concepts