Discussed the role of cloud computing in supporting data processing, training, and deploying AI models at scale.
Cloud vs. On-premise:
Cloud (especially AWS) provides a strong foundation for the Data Science pipeline — from collection, storage, and processing to training and deploying AI models.
AWS organizes the AI ecosystem into three layers, helping users choose the right level of management based on skills and goals:
1. AI Services (Fully Managed Layer)
For users who want to apply AI without deep Machine Learning knowledge.
Fully managed AI services that have been pre-trained by AWS.
Users can call APIs to use them directly in applications.
Examples:
👉 Benefits: Fast deployment, no model training needed, cost scales with usage.
2. ML Services (Semi-managed Layer)
For Data Scientists and ML Engineers who want to build, train, and deploy ML models with more customization.
Amazon SageMaker is at the center of this layer: it provides tools to build, train, and deploy ML models.
Key features:
👉 Benefits: Full control over the ML pipeline, customizable algorithms, training environments, and deployment workflows.
3. AI Infrastructure (Self-managed Layer)
For organizations or experts who want to fully manage AI/ML infrastructure to optimize cost or performance.
Users can build training environments using core AWS infrastructure services:
👉 Benefits: Maximum flexibility and control, but requires higher technical expertise.
1. Amazon SageMaker
Integrated development environment (SageMaker Studio) for the full ML lifecycle:
Supports AutoML, GPU training, model monitoring, and CI/CD for AI models.
2. Amazon Comprehend
NLP service to analyze, understand, and classify natural language.
Main capabilities:
Use cases:
3. Amazon Translate
Neural machine translation service.
Supports over 75 languages with high accuracy and easy integration.
Applications:
4. Amazon Textract
Goal: Show how to build an AI training pipeline without heavy coding.
Tools: Amazon SageMaker Studio / SageMaker Canvas
Demo steps:
Prepare the dataset and upload it to Amazon S3.
Use SageMaker’s drag-and-drop interface to:
Visually monitor training progress and model results (accuracy, confusion matrix, metrics, etc.).
Key message: Students and developers can quickly create AI workflows without complex code, speeding up research and experimentation.
Goal: Demonstrate how to deploy an AI model so users can access it in practice.
Tools: Amazon SageMaker Endpoint, API Gateway, and Lambda.
Demo steps:
Key message: Demonstrates how AWS supports moving AI from research to production — easy to share, scale, and commercialize.
| Criteria | Cloud (AWS) | On-premise |
|---|---|---|
| Scalability | Easily scale resources as needed | Limited by fixed hardware |
| Cost | Pay-as-you-go | High upfront investment |
| Deployment | Automated, fast | Manual, time-consuming |
| Maintenance | Managed by AWS | User is responsible |
| Suitable for students | ✅ Free Tier available, easy to learn | ❌ Harder to access, costly |
Attending the workshop “AI Services on AWS for Data Science” was very valuable. It helped me better understand the role of cloud in Data Science and how AWS supports training, deploying, and accessing AI models.
