- Natural Language Processing with BERT
- ML Pipelines and ML Operations (MLOps)
- A/B Testing and Model Deployment
- Data Labeling at Scale
- Automated Machine Learning (AutoML)
- Statistical Data Bias Detection
- Multi-class Classification with FastText and BlazingText
- Data ingestion
- Exploratory Data Analysis
- ML Pipelines and MLOps
- Model Training and Deployment with BERT
- Model Debugging and Evaluation
Spécialisation Practical Data Science on the AWS Cloud
Become a cloud data science expert. Develop and scale your data science projects into the cloud using Amazon SageMaker
Offert par



Ce que vous allez apprendre
Prepare data, detect statistical data biases, perform feature engineering at scale to train models, & train, evaluate, & tune models with AutoML
Store & manage ML features using a feature store, & debug, profile, tune, & evaluate models while tracking data lineage and model artifacts
Build, deploy, monitor, & operationalize end-to-end machine learning pipelines
Build data labeling and human-in-the-loop pipelines to improve model performance with human intelligence
Compétences que vous acquerrez
À propos de ce Spécialisation
Projet d'apprentissage appliqué
By the end of this Specialization, you will be ready to:
• Ingest, register, and explore datasets
• Detect statistical bias in a dataset
• Automatically train and select models with AutoML
• Create machine learning features from raw data
• Save and manage features in a feature store
• Train and evaluate models using built-in algorithms and custom BERT models
• Debug, profile, and compare models to improve performance
• Build and run a complete ML pipeline end-to-end
• Optimize model performance using hyperparameter tuning
• Deploy and monitor models
• Perform data labeling at scale
• Build a human-in-the-loop pipeline to improve model performance
• Reduce cost and improve performance of data products
Working knowledge of ML & Python, familiarity with Jupyter notebook & stat, completion of the Deep Learning & AWS Cloud Technical Essentials courses
Working knowledge of ML & Python, familiarity with Jupyter notebook & stat, completion of the Deep Learning & AWS Cloud Technical Essentials courses
Comment fonctionne la Spécialisation
Suivez les cours
Une Spécialisation Coursera est une série de cours axés sur la maîtrise d'une compétence. Pour commencer, inscrivez-vous directement à la Spécialisation ou passez en revue ses cours et choisissez celui par lequel vous souhaitez commencer. Lorsque vous vous abonnez à un cours faisant partie d'une Spécialisation, vous êtes automatiquement abonné(e) à la Spécialisation complète. Il est possible de terminer seulement un cours : vous pouvez suspendre votre formation ou résilier votre abonnement à tout moment. Rendez-vous sur votre tableau de bord d'étudiant pour suivre vos inscriptions aux cours et vos progrès.
Projet pratique
Chaque Spécialisation inclut un projet pratique. Vous devez réussir le(s) projet(s) pour terminer la Spécialisation et obtenir votre Certificat. Si la Spécialisation inclut un cours dédié au projet pratique, vous devrez terminer tous les autres cours avant de pouvoir le commencer.
Obtenir un Certificat
Lorsque vous aurez terminé tous les cours et le projet pratique, vous obtiendrez un Certificat que vous pourrez partager avec des employeurs éventuels et votre réseau professionnel.

Cette Spécialisation compte 3 cours
Analyze Datasets and Train ML Models using AutoML
In the first course of the Practical Data Science Specialization, you will learn foundational concepts for exploratory data analysis (EDA), automated machine learning (AutoML), and text classification algorithms. With Amazon SageMaker Clarify and Amazon SageMaker Data Wrangler, you will analyze a dataset for statistical bias, transform the dataset into machine-readable features, and select the most important features to train a multi-class text classifier. You will then perform automated machine learning (AutoML) to automatically train, tune, and deploy the best text-classification algorithm for the given dataset using Amazon SageMaker Autopilot. Next, you will work with Amazon SageMaker BlazingText, a highly optimized and scalable implementation of the popular FastText algorithm, to train a text classifier with very little code.
Build, Train, and Deploy ML Pipelines using BERT
In the second course of the Practical Data Science Specialization, you will learn to automate a natural language processing task by building an end-to-end machine learning pipeline using Hugging Face’s highly-optimized implementation of the state-of-the-art BERT algorithm with Amazon SageMaker Pipelines. Your pipeline will first transform the dataset into BERT-readable features and store the features in the Amazon SageMaker Feature Store. It will then fine-tune a text classification model to the dataset using a Hugging Face pre-trained model, which has learned to understand the human language from millions of Wikipedia documents. Finally, your pipeline will evaluate the model’s accuracy and only deploy the model if the accuracy exceeds a given threshold.
Optimize ML Models and Deploy Human-in-the-Loop Pipelines
In the third course of the Practical Data Science Specialization, you will learn a series of performance-improvement and cost-reduction techniques to automatically tune model accuracy, compare prediction performance, and generate new training data with human intelligence. After tuning your text classifier using Amazon SageMaker Hyper-parameter Tuning (HPT), you will deploy two model candidates into an A/B test to compare their real-time prediction performance and automatically scale the winning model using Amazon SageMaker Hosting. Lastly, you will set up a human-in-the-loop pipeline to fix misclassified predictions and generate new training data using Amazon Augmented AI and Amazon SageMaker Ground Truth.
Offert par

deeplearning.ai
DeepLearning.AI is an education technology company that develops a global community of AI talent.

Amazon Web Services
Since 2006, Amazon Web Services has been the world’s most comprehensive and broadly adopted cloud platform. AWS offers over 90 fully featured services for compute, storage, networking, database, analytics, application services, deployment, management, developer, mobile, Internet of Things (IoT), Artificial Intelligence, security, hybrid and enterprise applications, from 44 Availability Zones across 16 geographic regions. AWS services are trusted by millions of active customers around the world — including the fastest-growing startups, largest enterprises, and leading government agencies — to power their infrastructure, make them more agile, and lower costs.
Foire Aux Questions
Quelle est la politique de remboursement ?
Puis-je m'inscrire à un seul cours ?
Une aide financière est-elle possible ?
Puis-je suivre le cours gratuitement ?
Ce cours est-il vraiment accessible en ligne à 100 % ? Dois-je assister à certaines activités en personne ?
What is the Practical Data Science Specialization about?
What are the benefits of running data science projects in the cloud?
What will I learn in the Practical Data Science Specialization?
What will I be able to do upon completing the Practical Data Science Specialization?
What background knowledge is necessary for the Practical Data Science Specialization?
Who is the Practical Data Science Specialization for?
How long does it take to complete the Practical Data Science Specialization?
Who created the Practical Data Science Specialization?
Is this a standalone course or a Specialization?
How do I take the Specialization?
Do I need to take the courses in a specific order?
How much does the Specialization cost?
Can I apply for financial aid?
Can I audit the Practical Data Science Specialization?
How do I get a receipt to get this reimbursed by my employer?
I want to purchase this Specialization for my employees. How can I do that?
Will I receive a certificate at the end of the Specialization?
Puis-je obtenir des crédits universitaires si je réussis la Spécialisation ?
D'autres questions ? Visitez le Centre d'Aide pour les Étudiants.