Gen AI Masters Program (31 Blogs) Become a Certified Professional

What is Data Augmentation? Use Cases & Examples

Published on Apr 01,2025 31 Views

MERN stack web developer with expertise in full-stack development. Skilled in React,... MERN stack web developer with expertise in full-stack development. Skilled in React, Node.js, Express, and MongoDB, building scalable web solutions.
image not found!image not found!image not found!image not found!Copy Link!

Data augmentation is critical for boosting the performance of machine learning models, particularly deep learning models. The quality, amount, and importance of training data are important for how well these models perform. One of the main problems with using machine learning in real life is not having enough data. Gathering the needed info can take a lot of time and money.

Businesses can use data augmentation to create more accurate machine learning models more quickly and to lessen their dependency on gathering and preparing training data.

What is Data Augmentation?

Data augmentation is the method of making altered copies of a dataset using current data, hence artificially augmenting the training set. It involves either using deep learning to create fresh data points or small dataset modifications.

Augmented vs. synthetic data

Augmented-vs.-synthetic-data

 

Augmented data: This means changing current data to make the dataset more diverse. For instance, in image processing, changing current images by rotating, flipping, or adjusting their colors can improve how well models work.

Synthetic dataThis means that researchers and developers use fake data to try and improve their algorithms. This way, they don’t put real data at risk regarding privacy or security.

Why is data augmentation important?

Data augmentation improves machine learning models by making the most of available data. It helps avoid overfitting, enhances accuracy, and adds variety to the training data, which is important when the datasets are small or uneven.

 

Using methods like rotating, cutting, flipping, and adding noise helps models better deal with changes in the real world. This not only helps them understand new information better but also lowers mistakes when working with data they haven’t seen before.

Data augmentation reduces the need for expensive and time-consuming data collection, making it a smart and affordable way to boost model performance.

When should you use data augmentation?

  • To help models learn better without just memorizing the training data.
  • When there isn’t enough data to train properly.
  • To make predictions more accurate and reliable.
  • To save time and effort spent on organizing and labeling data.

What are the use cases of data augmentation?

What-are-the-use-cases-of-data-augmentation

 

Data augmentation has many uses in different businesses and helps improve the performance of machine learning models in many areas.

  1. Healthcare:  Helps doctors detect diseases better by creating slightly different versions of medical scans like X-rays and MRIs. This way, they don’t need a huge number of real scans to train their systems.
  2. Automotive (Self-Driving Cars): Makes self-driving cars safer by showing them images of roads in different weather, lighting, and angles. This helps them recognize objects like pedestrians and traffic signs in all conditions.
  3. E-commerce – Improves online shopping by making product searches and recommendations more accurate. By slightly changing product images and reviews, websites understand what customers like better.
  4. Finance – Helps banks catch fraud by creating different versions of spending patterns. This makes it easier to spot unusual transactions while reducing mistakes in blocking real purchases.

How does data augmentation work?

 

How-does-data-augmentation-work

  1. Check your data first: Look at how it’s organized—image sizes, text style, or how balanced the data is. Watch for hidden biases (e.g., too many similar images or repetitive phrases).
  2. Pick the right tools:
    • For images: Try cropping, rotating, adjusting brightness, or adding noise.
    • For text: Swap words with synonyms, rephrase sentences, or change sentence structure.
  3. Create new versions: Apply these changes while keeping the original format consistent (e.g., file names, text tone).
  4. Review manually: Check if the new data looks natural.
  5. Combine everything: Mix the new data with the original to build a richer, more varied dataset. Avoid repeating earlier biases.
  6. Final check: Ensure the final dataset is balanced, realistic, and ready for training.
Limitations of data augmentation
  1. If the original dataset has biases, those biases will carry over into the augmented data, potentially affecting model performance.
  2. Ensuring high-quality augmented data requires significant effort and investment.
  3. Developing advanced systems for specific applications, such as generating high-resolution images with GANs, can be complex and demanding.
  4. Identifying the right data augmentation strategy is not always straightforward and may require extensive experimentation.

Data Augmentation Techniques

Data-Augmentation-Techniques

Computer Vision

Data augmentation in computer vision helps improve model performance by creating variations of existing images.

  1. Position Augmentation: Cropping, rotating, flipping, and resizing images to generate new versions.
  2. Color Augmentation: Adjusting brightness, contrast, and saturation to enhance diversity.

Audio Data Augmentation

Common techniques for audio files include adding background noise, changing playback speed, and altering pitch to create variations.

Text Data Augmentation

In NLP, text augmentation involves shuffling sentences, reordering words, replacing words with synonyms, adding new words, or removing certain words.

Neural Style Transfer

This technique extracts and combines style and content from images to generate multiple variations for training.

Adversarial Training

Small pixel-level modifications, like adding slight noise, test a model’s ability to recognize altered images and improve robustness.

Ethical Implications of Data Augmentation

Ethical-Implications-of-Data-Augmentation

Data augmentation helps improve machine learning models, but it also brings up important social issues that need to be addressed.

  1. Risk of reinforcing bias: If a dataset has biases, using methods to change it can make the problem worse and result in unfair outcomes, especially for groups that are not well-represented.
  2. Privacy challenges: When making fake data, there’s still a risk that private information from the original data may be kept, which could cause privacy issues.
  3. Authenticity concerns: If augmentations are not used carefully, they can create data that doesn’t truly reflect real-life situations, which could hurt the model’s trustworthiness.
  4. Need for transparency: It’s important to keep a record of any changes made to data so that users and partners understand how it has been handled.
  5. Fairness in representation: Augmented data should be varied and include many different types of people and events so that machine learning models work well for everyone.
  6. Compliance with laws and ethics: You must follow data security laws and ethical guidelines when using augmented data to avoid legal problems and keep trust.

To use data enhancement responsibly, it’s important to check the quality of the new data, try to eliminate bias, and follow legal and ethical guidelines.

Data Augmentation with Keras and TensorFlow

Having enough varied data is often a major hurdle when building computer vision models. In many cases, raw datasets might not capture the diverse scenarios your model will face in the real world.

This is where data augmentation comes in. By applying transformations such as flipping, rotation, random brightness, cropping, and more, you can expand your dataset without collecting new samples. The result is typically higher accuracy and better generalization.

In the examples below, we focus on Keras and TensorFlow to implement data augmentation on an image classification task. We’ll briefly walk through the workflow of loading a dataset, applying different augmentation techniques, and then training a simple neural network to see how augmented data can improve performance.

For this demonstration, we use TensorFlow’s built-in cats_vs_dogs dataset, which contains images of cats and dogs labeled for binary classification. We rely on:

  • TensorFlow (and its Keras API) for building and training models.
  • matplotlib for visualizing images.
1
2
3
4
5
6
7
8
9
10
11
<span style="font-weight: 400;">import matplotlib.pyplot as plt</span>
 
<span style="font-weight: 400;">import numpy as np</span>
 
<span style="font-weight: 400;">import tensorflow as tf</span>
 
<span style="font-weight: 400;">from tensorflow.keras import layers</span>
 
<span style="font-weight: 400;">from tensorflow.keras.models import Sequential</span>
 
<span style="font-weight: 400;">import tensorflow_datasets as tfds</span>

Data Loading

TensorFlow Datasets (tfds) makes it easy to load ready-to-use datasets. Here, we split the cats_vs_dogs data into three parts:

  • 80% for training
  • 10% for validation
  • 10% for testing
1
2
3
4
5
6
7
8
9
10
11
<span style="font-weight: 400;">(train_ds, val_ds, test_ds), metadata = tfds.load(</span>
 
<span style="font-weight: 400;">    'cats_vs_dogs',</span>
 
<span style="font-weight: 400;">    split=['train[:80%]', 'train[80%:90%]', 'train[90%:]'],</span>
 
<span style="font-weight: 400;">    with_info=True,</span>
 
<span style="font-weight: 400;">    as_supervised=True,</span>
 
<span style="font-weight: 400;">)</span>

You can iterate through a few samples to display them with matplotlib. This helps confirm that the dataset loaded correctly and the labels match what you see.

Data Analysis

The dataset has two classes: cat and dog. It’s always helpful to check the metadata before proceeding with any augmentation or model-building. You can then preview a handful of images to understand their size, orientation, and variety. This insight helps inform which augmentations make sense (e.g., flipping, rotation, color adjustments).

Data Augmentation with Keras Sequential Layers

Keras offers built-in layers specifically for data augmentation. You can define these transformations as part of a Sequential model or apply them in a preprocessing pipeline.

  1. Resizing and Rescaling

Often, you’ll resize images to a smaller dimension for faster processing and scale pixel values to a [0, 1] range by dividing by 255.

1
2
3
4
5
6
7
8
9
10
11
<span style="font-weight: 400;">IMG_SIZE = 180</span>
 
 
 
<span style="font-weight: 400;">resize_and_rescale = Sequential([</span>
 
<span style="font-weight: 400;">    layers.Resizing(IMG_SIZE, IMG_SIZE),</span>
 
<span style="font-weight: 400;">    layers.Rescaling(1./255)</span>
 
<span style="font-weight: 400;">])</span>

Applying resize_and_rescale(image) ensures every image is the same resolution and that raw pixel intensities are normalized.

  1. Random Flip and Rotate

You can stack augmentation layers in another Sequential model. In the example below, we define random flips and rotations to alter the images geometrically:

1
2
3
4
5
6
7
<span style="font-weight: 400;">data_augmentation = Sequential([</span>
 
<span style="font-weight: 400;">    layers.RandomFlip("horizontal_and_vertical"),</span>
 
<span style="font-weight: 400;">    layers.RandomRotation(0.4),</span>
 
<span style="font-weight: 400;">])</span>

When you pass your images through data_augmentation, it outputs a new, randomly transformed version each time. Plotting them side by side confirms how flipping and rotation can diversify your training data.

  1. Incorporating Augmentation Layers in the Model

There are two main approaches to include these preprocessing steps:

  1. i) Directly in the Model Definition
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
<span style="font-weight: 400;">model = Sequential([</span>
 
<span style="font-weight: 400;">    resize_and_rescale,</span>
 
<span style="font-weight: 400;">    data_augmentation,</span>
 
<span style="font-weight: 400;">    layers.Conv2D(16, 3, padding='same', activation='relu'),</span>
 
<span style="font-weight: 400;">    layers.MaxPooling2D(),</span>
 
<span style="font-weight: 400;">    layers.Flatten(),</span>
 
<span style="font-weight: 400;">    layers.Dense(128, activation='relu'),</span>
 
<span style="font-weight: 400;">    layers.Dense(64, activation='relu'),</span>
 
<span style="font-weight: 400;">    layers.Dense(1, activation='sigmoid')</span>
 
<span style="font-weight: 400;">])</span>

The augmentation layers only apply during training, not during inference (evaluate or predict).

  1. ii) Using a .map Function
1
<span style="font-weight: 400;">aug_ds = train_ds.map(lambda x, y: (data_augmentation(x, training=True), y))</span>

This approach creates a new dataset that includes transformed images for training.

Preprocessing and Model Training

Before feeding data into the model, it’s common to prepare and batch it. Below is a helper function that:

  1. Resizes and rescales every sample.
  2. Optionally shuffles the data.
  3. Batches the samples (e.g., 32 images per batch).
  4. (Optionally) applies augmentation.
  5. Prefetches to overlap data processing and model execution.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
<span style="font-weight: 400;">batch_size = 32</span>
 
<span style="font-weight: 400;">AUTOTUNE = tf.data.AUTOTUNE</span>
 
 
 
<span style="font-weight: 400;">def prepare(ds, shuffle=False, augment=False):</span>
 
<span style="font-weight: 400;">    ds = ds.map(lambda x, y: (resize_and_rescale(x), y),</span>
 
<span style="font-weight: 400;">                num_parallel_calls=AUTOTUNE)</span>
 
<span style="font-weight: 400;">    if shuffle:</span>
 
<span style="font-weight: 400;">        ds = ds.shuffle(1000)</span>
 
<span style="font-weight: 400;">    ds = ds.batch(batch_size)</span>
 
<span style="font-weight: 400;">    if augment:</span>
 
<span style="font-weight: 400;">        ds = ds.map(lambda x, y: (data_augmentation(x, training=True), y),</span>
 
<span style="font-weight: 400;">                    num_parallel_calls=AUTOTUNE)</span>
 
<span style="font-weight: 400;">    return ds.prefetch(buffer_size=AUTOTUNE)</span>
 
 
 
<span style="font-weight: 400;">train_ds = prepare(train_ds, shuffle=True, augment=True)</span>
 
<span style="font-weight: 400;">val_ds   = prepare(val_ds)</span>
 
<span style="font-weight: 400;">test_ds  = prepare(test_ds)</span>

Example Model:

A simple convolutional neural network might look like this:

1
2
3
4
5
6
7
8
9
10
11
12
13
<span style="font-weight: 400;">model = Sequential([</span>
 
<span style="font-weight: 400;">    layers.Conv2D(32, (3, 3), padding='same', activation='relu', input_shape=(180, 180, 3)),</span>
 
<span style="font-weight: 400;">    layers.MaxPooling2D((2, 2)),</span>
 
<span style="font-weight: 400;">    layers.Flatten(),</span>
 
<span style="font-weight: 400;">    layers.Dense(32, activation='relu'),</span>
 
<span style="font-weight: 400;">    layers.Dense(1, activation='softmax')  # For binary classification</span>
 
<span style="font-weight: 400;">])</span>

Compile the model:

1
2
3
4
5
6
7
8
9
<span style="font-weight: 400;">model.compile(</span>
 
<span style="font-weight: 400;">    optimizer='adam',</span>
 
<span style="font-weight: 400;">    loss='binary_crossentropy',</span>
 
<span style="font-weight: 400;">    metrics=['accuracy']</span>
 
<span style="font-weight: 400;">)</span>

Train for a given number of epochs:

1
2
3
4
5
6
7
8
9
<span style="font-weight: 400;">history = model.fit(</span>
 
<span style="font-weight: 400;">    train_ds,</span>
 
<span style="font-weight: 400;">    validation_data=val_ds,</span>
 
<span style="font-weight: 400;">    epochs=1</span>
 
<span style="font-weight: 400;">)</span>

Even with minimal tuning, you’ll notice a slight improvement in validation accuracy after applying data augmentation compared to training on only raw images.

Data Augmentation with tf.image

For finer-grained control, TensorFlow’s tf.image module provides functions to manipulate tensors directly. Common transformations include flipping, cropping, brightness adjustments, saturation changes, and rotation. Each operation is typically a single function call, such as:

1
2
3
4
5
<span style="font-weight: 400;">flipped = tf.image.flip_left_right(image)</span>
 
<span style="font-weight: 400;">gray    = tf.image.rgb_to_grayscale(image)</span>
 
<span style="font-weight: 400;">rotated = tf.image.rot90(image)</span>

To visualize the impact of each transformation, you can write a small helper function that plots the original image and the transformed version side by side.

Applying Randomness with tf.image.stateless_*

tf.image also provides stateless functions like tf.image.stateless_random_brightness, which use a seed parameter to ensure reproducibility. You can generate multiple variations of the same image, each with a different brightness level, simply by changing the seed.

Dataset-Wide Augmentation with .map

Just like Keras layers, you can apply tf.image transformations across the entire training set by defining an augment function and mapping it to your dataset:

[/python]

def augment(image, label):

    image = tf.cast(image, tf.float32)

    image = tf.image.resize(image, [IMG_SIZE, IMG_SIZE])

    image = image / 255.0

    image = tf.image.random_crop(image, size=[IMG_SIZE, IMG_SIZE, 3])

    image = tf.image.random_brightness(image, max_delta=0.5)

    return image, label

train_ds = (

    train_ds

    .shuffle(1000)

    .map(augment, num_parallel_calls=AUTOTUNE)

    .batch(batch_size)

    .prefetch(AUTOTUNE)

)

[/python]

This ensures every image fed to the model has some level of randomized transformation applied.

Data Augmentation with ImageDataGenerator

For users who prefer a simpler, more traditional Keras interface, ImageDataGenerator offers an easy way to apply common augmentations, especially when loading images from a directory or a NumPy array.

1
2
3
4
5
6
7
8
9
10
11
12
13
datagen = tf.keras.preprocessing.image.ImageDataGenerator(
 
<span style="font-weight: 400;">    rotation_range=20,</span>
 
<span style="font-weight: 400;">    width_shift_range=0.2,</span>
 
<span style="font-weight: 400;">    height_shift_range=0.2,</span>
 
<span style="font-weight: 400;">    horizontal_flip=True,</span>
 
<span style="font-weight: 400;">    validation_split=0.2</span>
 
<span style="font-weight: 400;">)</span>

You can then fit this generator on your training images, and it will yield batches of randomly augmented images. This is especially convenient for small to medium-scale projects.

Data augmentation can significantly improve a model’s generalization ability by experimenting with these tools, adjusting parameters, layering transformations, and systematically evaluating performance.

Data Augmentation Tools

Data-Augmentation-Tools

In this section, we’ll talk about some useful open-source tools that can help you perform different data augmentation techniques to boost your model’s performance.

Pytorch

Pytorch offers tools for image transformation through the torchvision.transforms module. You can apply transformations directly by adding them intorch.nn.Sequentialor apply them separately as functions on your dataset.

Augmentor

Augmentor is a Python tool for working with images. It allows you to perform several image transformations like rotating, cropping, mirroring, and applying elastic distortions. It also provides some basic pre-processing features for images.

Albumentations

Albumentations is a fast and flexible tool for image augmentation. It is widely used to improve deep learning models, especially convolutional neural networks. It’s known for being both quick and effective, making it popular in both research and industry.

Imgaug

Imgaug is an open-source tool that offers a wide range of image augmentation techniques, such as adding noise, adjusting contrast and sharpness, cropping, and flipping images. It’s easy to use and also supports advanced features like key points, bounding boxes, and heatmaps.

OpenCV

OpenCV is a powerful open-source library used for computer vision tasks and image processing. It’s great for building real-time applications, and it allows you to apply various image and video augmentations with ease.

Airbyte

Airbyte is a platform that helps move data from one place to another, particularly for unstructured and semi-structured data. It’s often used to send data into storage systems like data lakes or warehouses, making it useful for AI projects and machine learning.

LangChain

LangChain is a tool used for building AI applications powered by large language models (LLMs). It helps businesses incorporate their data into these models, making them more effective for a variety of applications.

Conclusion

Data augmentation is a game-changing technology in machine learning that enhances the generalizability of models by artificially expanding datasets. By employing techniques such as twisting, rotating, scaling, and color adjustment,

we can generate a variety of training samples, thereby reducing overfitting and enhancing real-world performance. Data augmentation expedites model training and reduces expenses, regardless of whether the application is healthcare, self-driving vehicles, or fraud detection.

Edureka’s Generative AI Masters Program is an excellent option for those who are enthusiastic about furthering their knowledge of AI and improving their proficiency in cutting-edge technologies such as generative AI.

This all-encompassing course provides a comprehensive understanding of AI fundamentals and advanced deep learning techniques, thereby equipping you with the necessary skills to apply AI in real-world scenarios.

FAQs

 

FAQ

Why use data augmentation in CNN?

Data augmentation in CNNs enhances real-world performance by adding variations like flipping, rotating, scaling, and color changes. This improves generalization, reduces overfitting, and increases stability.

What is augmentation with an example.

Augmentation expands datasets by transforming existing data (e.g., rotating or flipping images) to help CNNs learn better and generalize effectively.

What is the difference between data augmentation and preprocessing?

Data augmentation expands datasets with transformations like rotation and flipping to improve generalization.

Preprocessing modifies raw data (e.g., normalization) for consistency and better model performance without altering its meaning.

Is PCA used for data augmentation?

Principal Component Analysis (PCA) can be used to add to data, especially when handling images. PCA-based augmentation, also known as PCA jittering, changes the values of pixels along the principal components in a small way. This introduces variation while keeping important features, which helps make the model more stable.

What is the technique of data augmentation?

Data augmentation boosts performance by adding variations like flipping, rotating, and scaling, reducing overfitting and improving real-world adaptability.

Comments
0 Comments

Join the discussion

Browse Categories

webinar REGISTER FOR FREE WEBINAR
+91
  • India (भारत)+91
  • United States+1
  • United Kingdom+44
  • Afghanistan (‫افغانستان‬‎)+93
  • Albania (Shqipëri)+355
  • Algeria (‫الجزائر‬‎)+213
  • Andorra+376
  • Angola+244
  • Argentina+54
  • Armenia (Հայաստան)+374
  • Aruba+297
  • Australia+61
  • Austria (Österreich)+43
  • Azerbaijan (Azərbaycan)+994
  • Bahamas+1242
  • Bahrain (‫البحرين‬‎)+973
  • Bangladesh (বাংলাদেশ)+880
  • Barbados+1246
  • Belarus (Беларусь)+375
  • Belgium (België)+32
  • Belize+501
  • Benin (Bénin)+229
  • Bermuda+1441
  • Bhutan (འབྲུག)+975
  • Bolivia+591
  • Bosnia and Herzegovina (Босна и Херцеговина)+387
  • Botswana+267
  • Brazil (Brasil)+55
  • British Indian Ocean Territory+246
  • British Virgin Islands+1284
  • Brunei+673
  • Bulgaria (България)+359
  • Burkina Faso+226
  • Burundi (Uburundi)+257
  • Cambodia (កម្ពុជា)+855
  • Cameroon (Cameroun)+237
  • Canada+1
  • Cape Verde (Kabu Verdi)+238
  • Caribbean Netherlands+599
  • Cayman Islands+1345
  • Central African Republic (République centrafricaine)+236
  • Chad (Tchad)+235
  • Chile+56
  • China (中国)+86
  • Christmas Island+61
  • Cocos (Keeling) Islands+61
  • Colombia+57
  • Comoros (‫جزر القمر‬‎)+269
  • Congo (DRC) (Jamhuri ya Kidemokrasia ya Kongo)+243
  • Congo (Republic) (Congo-Brazzaville)+242
  • Cook Islands+682
  • Costa Rica+506
  • Côte d’Ivoire+225
  • Croatia (Hrvatska)+385
  • Cuba+53
  • Curaçao+599
  • Cyprus (Κύπρος)+357
  • Czech Republic (Česká republika)+420
  • Denmark (Danmark)+45
  • Djibouti+253
  • Dominican Republic (República Dominicana)+1
  • Ecuador+593
  • Egypt (‫مصر‬‎)+20
  • El Salvador+503
  • Equatorial Guinea (Guinea Ecuatorial)+240
  • Eritrea+291
  • Estonia (Eesti)+372
  • Ethiopia+251
  • Falkland Islands (Islas Malvinas)+500
  • Faroe Islands (Føroyar)+298
  • Fiji+679
  • Finland (Suomi)+358
  • France+33
  • French Guiana (Guyane française)+594
  • French Polynesia (Polynésie française)+689
  • Gabon+241
  • Gambia+220
  • Georgia (საქართველო)+995
  • Germany (Deutschland)+49
  • Ghana (Gaana)+233
  • Gibraltar+350
  • Greece (Ελλάδα)+30
  • Greenland (Kalaallit Nunaat)+299
  • Grenada+1473
  • Guadeloupe+590
  • Guatemala+502
  • Guernsey+44
  • Guinea (Guinée)+224
  • Guinea-Bissau (Guiné Bissau)+245
  • Guyana+592
  • Haiti+509
  • Honduras+504
  • Hong Kong (香港)+852
  • Hungary (Magyarország)+36
  • Iceland (Ísland)+354
  • India (भारत)+91
  • Indonesia+62
  • Iran (‫ایران‬‎)+98
  • Iraq (‫العراق‬‎)+964
  • Ireland+353
  • Isle of Man+44
  • Israel (‫ישראל‬‎)+972
  • Italy (Italia)+39
  • Jamaica+1876
  • Japan (日本)+81
  • Jersey+44
  • Jordan (‫الأردن‬‎)+962
  • Kazakhstan (Казахстан)+7
  • Kenya+254
  • Kiribati+686
  • Kosovo+383
  • Kuwait (‫الكويت‬‎)+965
  • Kyrgyzstan (Кыргызстан)+996
  • Laos (ລາວ)+856
  • Latvia (Latvija)+371
  • Lebanon (‫لبنان‬‎)+961
  • Lesotho+266
  • Liberia+231
  • Libya (‫ليبيا‬‎)+218
  • Liechtenstein+423
  • Lithuania (Lietuva)+370
  • Luxembourg+352
  • Macau (澳門)+853
  • Macedonia (FYROM) (Македонија)+389
  • Madagascar (Madagasikara)+261
  • Malawi+265
  • Malaysia+60
  • Maldives+960
  • Mali+223
  • Malta+356
  • Marshall Islands+692
  • Martinique+596
  • Mauritania (‫موريتانيا‬‎)+222
  • Mauritius (Moris)+230
  • Mayotte+262
  • Mexico (México)+52
  • Micronesia+691
  • Moldova (Republica Moldova)+373
  • Monaco+377
  • Mongolia (Монгол)+976
  • Montenegro (Crna Gora)+382
  • Morocco (‫المغرب‬‎)+212
  • Mozambique (Moçambique)+258
  • Myanmar (Burma) (မြန်မာ)+95
  • Namibia (Namibië)+264
  • Nauru+674
  • Nepal (नेपाल)+977
  • Netherlands (Nederland)+31
  • New Caledonia (Nouvelle-Calédonie)+687
  • New Zealand+64
  • Nicaragua+505
  • Niger (Nijar)+227
  • Nigeria+234
  • Niue+683
  • Norfolk Island+672
  • North Korea (조선 민주주의 인민 공화국)+850
  • Norway (Norge)+47
  • Oman (‫عُمان‬‎)+968
  • Pakistan (‫پاکستان‬‎)+92
  • Palau+680
  • Palestine (‫فلسطين‬‎)+970
  • Panama (Panamá)+507
  • Papua New Guinea+675
  • Paraguay+595
  • Peru (Perú)+51
  • Philippines+63
  • Poland (Polska)+48
  • Portugal+351
  • Puerto Rico+1
  • Qatar (‫قطر‬‎)+974
  • Réunion (La Réunion)+262
  • Romania (România)+40
  • Russia (Россия)+7
  • Rwanda+250
  • Saint Barthélemy+590
  • Saint Helena+290
  • Saint Martin (Saint-Martin (partie française))+590
  • Saint Pierre and Miquelon (Saint-Pierre-et-Miquelon)+508
  • Samoa+685
  • San Marino+378
  • São Tomé and Príncipe (São Tomé e Príncipe)+239
  • Saudi Arabia (‫المملكة العربية السعودية‬‎)+966
  • Senegal (Sénégal)+221
  • Serbia (Србија)+381
  • Seychelles+248
  • Sierra Leone+232
  • Singapore+65
  • Sint Maarten+1721
  • Slovakia (Slovensko)+421
  • Slovenia (Slovenija)+386
  • Solomon Islands+677
  • Somalia (Soomaaliya)+252
  • South Africa+27
  • South Korea (대한민국)+82
  • South Sudan (‫جنوب السودان‬‎)+211
  • Spain (España)+34
  • Sri Lanka (ශ්‍රී ලංකාව)+94
  • Sudan (‫السودان‬‎)+249
  • Suriname+597
  • Svalbard and Jan Mayen+47
  • Swaziland+268
  • Sweden (Sverige)+46
  • Switzerland (Schweiz)+41
  • Syria (‫سوريا‬‎)+963
  • Taiwan (台灣)+886
  • Tajikistan+992
  • Tanzania+255
  • Thailand (ไทย)+66
  • Timor-Leste+670
  • Togo+228
  • Tokelau+690
  • Tonga+676
  • Tunisia (‫تونس‬‎)+216
  • Turkey (Türkiye)+90
  • Turkmenistan+993
  • Tuvalu+688
  • Uganda+256
  • Ukraine (Україна)+380
  • United Arab Emirates (‫الإمارات العربية المتحدة‬‎)+971
  • United Kingdom+44
  • United States+1
  • Uruguay+598
  • Uzbekistan (Oʻzbekiston)+998
  • Vanuatu+678
  • Vatican City (Città del Vaticano)+39
  • Venezuela+58
  • Vietnam (Việt Nam)+84
  • Wallis and Futuna (Wallis-et-Futuna)+681
  • Western Sahara (‫الصحراء الغربية‬‎)+212
  • Yemen (‫اليمن‬‎)+967
  • Zambia+260
  • Zimbabwe+263
  • Åland Islands+358
webinar_success Thank you for registering Join Edureka Meetup community for 100+ Free Webinars each month JOIN MEETUP GROUP

Subscribe to our Newsletter, and get personalized recommendations.

image not found!
image not found!

What is Data Augmentation? Use Cases & Examples

edureka.co

preload imagepreload image