- Don't use the email you registered with GitHub for commits. Instead, GitHub provides you with a proxy email for this purpose. Just go to
'Settings - Emails'in your GitHub account, and you'll find the proxy email there. - Don't use your GitHub login password for commits. Instead, go to
'Settings - Developer Settings - Personal access tokens', create a token, and use that as your password for commits. SinceFine-grained tokensare still inPreview, I'm using a classic token for now.
- π EC2 Instances: Full User Control (Least Pre-built Content)
With EC2, you have complete control over the entire setup. You need to:- Start an EC2 instance (e.g., GPU-enabled for training deep learning models).
- Install dependencies manually (e.g., Python, ML libraries like PyTorch or TensorFlow).
- Copy or configure the training script, and handle the training data management (downloading data from S3 or other sources).
- Run the training process manually using your own code.
- Manage all aspects of the environment, scaling, and resource management.
To apply distributed training for the AWS SageMaker Linear Learner algorithm, you would typically rely on SageMaker's built-in distributed training capabilities. The Linear Learner algorithm supports distributed training by scaling across multiple instances and using multiple GPUs or CPU cores.
SageMaker Linear Learner algorithm provides a straightforward approach to use distributed training across multiple instances by setting the instance_count parameter to more than 1.
- WebDataset source code
https://github.com/webdataset/webdataset
Code snippets are from the following sources:
- β
Why I Chose WebDataset for Training on 50TB of Data?
Ahmad Sachal, May 22, 2023
-
Uninstall all VS Code extensions
DeleteC:\Users\*\.vscode\extensionsfolder
Reinstall extensions -
Remove Jupyter kernels
(base) PS D:\github\udacity-nd009t-capstone-starter> jupyter kernelspec list
Available kernels:β οΈ π’ Issue: training error
[1,mpirank:0,algo-1]<stderr>:../aten/src/ATen/native/cuda/Loss.cu:242: nll_loss_forward_reduce_cuda_kernel_2d: block: [0,0,0], thread: [0,0,0] Assertion `t >= 0 && t < n_classes` failed.
[1,mpirank:0,algo-1]<stderr>:../aten/src/ATen/native/cuda/Loss.cu:242: nll_loss_forward_reduce_cuda_kernel_2d: block: [0,0,0], thread: [6[1,mpirank:0,algo-1]<stderr>:,0,0] Assertion `t >= 0 && t < n_classes` failed.
[1,mpirank:0,algo-1]<stderr>:../aten/src/ATen/native/cuda/Loss.cu:242: nll_loss_forward_reduce_cuda_kernel_2d: block: [0,0,0], thread: [30,0,0] Assertion `t >= 0 && t < n_classes` failed.
...
[1,mpirank:1,algo-2]<stdout>: File "train.py", line 675, in <module>
[1,mpirank:1,algo-2]<stdout>: main(task)
[1,mpirank:1,algo-2]<stdout>: File "train.py", line 572, in main- Google Docs: 20250125_AWS SageMaker Input Mode, WebDataset
## example code for webdataset
import webdataset as wds
import io
import json(awsmle_py310) PS D:\github\udacity-cd13926-Building-Apps-Amazon-Bedrock-exercises\Experiments> aws bedrock invoke-model `
>> --model-id anthropic.claude-3-5-sonnet-20240620-v1:0 `
>> --body file://claude_input.json output.json
usage: aws [options] <command> <subcommand> [<subcommand> ...] [parameters]
To see help text, you can run:
- βοΈ Check my Google Docs
π’ROSLaunch that Gazebo world that I created via udacity_office.launch.
To export SAP HANA data in real time continuously, you can use several methods depending on the target system and the purpose. Here are some of the most common approaches:
- Use Case: Real-time data replication and transformation.
- How: SDI allows you to create real-time data replication tasks between SAP HANA and other systems. You can define data flows that continuously export data from HANA and send it to another system, such as another HANA instance or a non-HANA database.
- Steps:
- Set up a Data Provisioning Agent.
- Configure the SDI connection to the target system.