What is learning?
+ Increasing knowledge
+ Memorising wha has been learn
+ Applying and using knowledge
+ Undersanding wha has o be learned
+ Seeing hings differenly
+ Building social compeence
+ Embracing learning as somehing which is lie-long and lie-
wide
+ Changing as a person
Arficial Inelligence (AI), Machine Learning (ML), and Deep Learning (DL) are inerconneced
fields in compuer science, each wih disnc roles:
• Arficial Inelligence (AI) is he broades concep reerring o sysems or machines
designed o simulae human inelligence, including reasoning, learning, percepon, and
problem-solving, decision-making. I encompasses any echnique enabling compuers o
emulae human behavior, such as reasoning, planning, or naural language processing.
Examples include chabos, recommendaon sysems, or auonomous vehicles.
• Machine Learning (ML) is a branch o arficial inelligence (AI) ha ocuses on building
sysems ha can learn rom daa, ideny paerns, and make decisions wih minimal
human inervenon. In essence, machine learning allows compuers o "learn" rom pas
experiences (daa) and improve heir perormance over me wihou being explicily
programmed or every possible scenario. Common echniques include supervised
learning (e.g., regression, classificaon), unsupervised learning (e.g., clusering), and
reinorcemen learning. Applicaons include spam deecon or image recognion.
• Deep Learning (DL) is a specialized subse o ML ocused on neural neworks wih many
layers (deep neural neworks). DL excels a complex asks like image recognion, naural
language processing, and auonomous driving due o is abiliy o learn hierarchical
eaure represenaons rom large daases.
Machine Learning: Applicaons
Image Recognion:
Facial Recognion: Used in smarphones, social media plaorms (e.g., Facebook phoo agging),
and securiy sysems. ML algorihms can deec and recognize human aces in images or videos.
Medical Imaging: ML is used in analyzing X-rays, MRIs, and CT scans o ideny abnormalies
such as umors, racures, or oher diseases.
Objec Deecon: Used in auonomous vehicles, reail (e.g., Amazon Go sores), and securiy
surveillance sysems.
Speech and Voice Recognion:
Virual Assisans: ML powers virual assisans like Siri, Alexa, and Google Assisan, enabling
hem o recognize and respond o voice commands.
Speech-o-Tex: Applicaons like Google Voice Typing and auomaed ranscripon services
conver spoken words ino wrien ex.
Voice Biomerics: Used in securiy sysems or idenying individuals based on heir voice.
Naural Language Processing (NLP):
Chabos and Cusomer Suppor: ML-driven chabos can undersand and respond o cusomer
queries, improving cusomer service or companies like banks, e-commerce plaorms, and
service providers.
Language Translaon: Google Translae and oher language ranslaon apps use ML models o
ranslae ex and speech beween differen languages.
Senmen Analysis: Businesses use ML o analyze cusomer reviews, social media poss, and
oher ex daa o deermine public senmen or opinions abou producs or services.
Recommendaon Sysems:
E-commerce: Plaorms like Amazon and Alibaba use ML o recommend producs based on a
user’s browsing hisory, purchase behavior, and preerences.
Sreaming Services: Nelix, Spoy, and YouTube recommend movies, music, and videos by
analyzing user ineracons, viewing hisory, and preerences.
Social Media: Facebook, Twier, and Insagram use recommendaon algorihms o sugges
riends, poss, or conen ha align wih a user’s ineress.
Fraud Deecon:
Banking and Finance: ML algorihms analyze ransacon paerns o deec raudulen acvies,
such as credi card raud, money laundering, or accoun akeovers.
Insurance: Insurers use ML o deec raudulen insurance claims by spotng unusual paerns or
behaviors in claim submissions.
Auonomous Vehicles:
Self-Driving Cars: Companies like Tesla, Waymo, and Uber use ML o help auonomous vehicles
inerpre heir surroundings (e.g., deecng pedesrians, road signs, and oher cars), make driving
decisions, and navigae saely.
Drone Navigaon: ML is used in drones or roue planning, objec deecon, and collision
avoidance.
Healhcare and Medical Diagnosis:
Disease Predicon: ML models help predic he likelihood o diseases such as diabees, hear
disease, or cancer by analyzing paen daa.
Personalized Medicine: Machine learning enables he developmen o personalized reamen
plans based on a paen's genecs, medical hisory, and curren healh daa.
Drug Discovery: ML acceleraes he process o discovering new drugs by analyzing daa rom
clinical rials and idenying poenal drug candidaes.
• Finance and Sock Marke Analysis
• Markeng and Adversing
• Supply Chain Opmizaon
• Gaming, Robocs, Cybersecuriy
• Personalizaon
• Energy Efficiency
ill-Posed Problems in Machine Learning
An ill-posed problem, in he conex o ML, reers o a problem ha violaes one or more o he
condions or being well-posed, as defined by Jacques Hadamard:
1. A soluon exiss.
2. The soluon is unique.
3. The soluon is sable (small changes in inpu lead o small changes in oupu).
In ML, ill-posed problems ofen arise when he daa or model seup leads o ambiguiy, non-
uniqueness, or insabiliy in soluons, making i hard o achieve reliable predicons.
Relevance o ML (Cancer Deecon Example)
In cancer deecon rom medical imaging:
• Exisence: A soluon may exis, bu he complexiy o medical images (e.g., suble umor
paerns) makes i hard o guaranee he model will find i.
• Uniqueness: Mulple models or parameer setngs migh produce similar resuls, bu i’s
unclear which is opmal. For insance, differen CNN archiecures may deec umors
wih comparable accuracy bu ocus on differen image eaures.
• Sabiliy: Small changes in inpu images (e.g., noise, variaons in imaging equipmen, or
paen demographics) can lead o drascally differen predicons i he model isn’
robus.
Causes of Ill-Posed Problems in ML
1. Insufficient Data:
o Small or non-representative datasets fail to capture the full variability of the
problem. For instance, a dataset of 1,000 mammograms may not include enough
cases of rare tumor types, leading to incomplete learning.
o Lack of diversity (e.g., images from a single demographic) can cause the model to
miss generalizable patterns.
2. High Dimensionality:
o ML problems often involve high-dimensional data (e.g., medical images with
millions of pixels). This increases the risk of overfitting or finding multiple
solutions that fit the data equally well.
o In cancer detection, the high number of features (pixels) compared to the number
of samples can make the problem underdetermined.
3. Noisy or Ambiguous Data:
o Real-world data often contains noise (e.g., artifacts in medical images from
equipment) or ambiguous patterns (e.g., benign and malignant tissues with similar
appearances), making it hard to define a unique solution.
o For example, subtle differences between cancerous and non-cancerous regions
may be indistinguishable without additional context.
4. Ill-Defined Problem Formulation:
o Poorly defined objectives or labels can exacerbate ill-posedness. For instance, if
"cancer" labels in a dataset are inconsistent due to human error, the model may
learn incorrect patterns.
o In regression tasks (e.g., predicting tumor size), the relationship between inputs
and outputs may be inherently ambiguous due to biological variability.
5. Non-Linear and Complex Relationships:
o Many ML problems involve complex, non-linear relationships that are difficult to
model accurately. In cancer detection, the relationship between pixel patterns and
malignancy is highly non-linear, increasing the risk of instability.
Overfitng in Machine Learning
Overfitng is a crical challenge in machine learning (ML) where a model learns he raining daa
oo well, including is noise and ouliers, resulng in poor perormance on new, unseen daa. This
leads o a model ha is overly complex and ails o generalize o real-world scenarios
Relevance o ML (Cancer Deecon Example)
In cancer deecon:
• A CNN migh memorize specific pixel paerns in he raining mammograms, including
irrelevan deails like imaging aracs or paen-specific noise.
• When esed on new images (e.g., rom a differen hospial or machine), he model may
ail o correcly classiy umors, leading o low sensiviy or specificiy.
Causes of Overfitng
1. Complex Models:
o Models wih oo many parameers (e.g., deep neural neworks wih millions o
weighs) can fi he raining daa excessively, capuring noise insead o general
paerns.
o Example: A CNN wih 50 layers migh overfi a small daase o 1,000
mammograms by learning irrelevan pixel paerns.
2. Insufficien Daa:
o Small or non-diverse daases limi he model’s abiliy o learn generalizable
paerns.
o Example: I a cancer deecon daase has ew malignan cases or is sourced rom
one hospial, he model may overfi o specific imaging condions.
3. Noisy Daa:
o Real-world daa ofen conains noise (e.g., imaging aracs in medical scans or
mislabeled daa), which he model may misakenly rea as meaningul.
o Example: A CNN migh learn o associae scanner-specific noise wih cancer,
leading o alse posives on cleaner images.
4. Lack of Regularizaon:
o Wihou consrains, models can become overly flexible, fitng he raining daa
oo closely.
o Example: A CNN wihou dropou or weigh penales may priorize irrelevan
eaures in mammograms.
5. Imbalanced Daa:
o I he daase is skewed (e.g., 90% benign and 10% malignan cases), he model
may overfi o he majoriy class, neglecng minoriy paerns.
o Example: A model migh achieve high accuracy by predicng mos cases as benign,
missing crical malignan cases.
6. Overraining:
o Training a model or oo many epochs can cause i o memorize he raining daa
raher han learn general paerns.
o Example: Afer 50 epochs, a CNN migh sar fitng noise in mammograms,
reducing is abiliy o generalize.
Implicaons of Overfitng
• Poor Generalizaon: The model perorms well on raining daa bu poorly on new daa,
liming is praccal uliy.
o Example: A cancer deecon model wih 95% raining accuracy bu 70% es
accuracy ails o reliably deec umors in clinical setngs.
• Reduced Trus: In applicaons like healhcare, overfitng can lead o alse posives or
negaves, eroding confidence in he model.
• Resource Wase: Overfied models require more compuaonal resources and me o
rain, wih diminishing reurns on perormance.
Soluons o Migae Overfitng
1. Regularizaon:
o Adds consrains o reduce model complexiy and preven overfitng o noise.
o Techniques:
▪ L1/L2 Regularizaon: Penalizes large weighs in he model (e.g., adding a
erm o he loss uncon o discourage overly complex CNNs).
▪ Dropou: Randomly deacvaes a racon o neurons during raining,
orcing he model o learn robus eaures.
o Example: Applying dropou (e.g., 50% neuron dropou rae) o a CNN or cancer
deecon prevens reliance on specific pixel paerns.
2. Daa Augmenaon:
o Increases daase size and diversiy by generang synhec variaons o he
raining daa.
o Example: For mammograms, augmenaons like roaon, flipping, zooming, or
adjusng brighness help he model generalize o varied imaging condions.
3. More Daa:
o Collecng larger, more diverse daases ensures he model learns represenave
paerns.
o Example: Including mammograms rom mulple hospials, demographics, and
imaging devices reduces he risk o overfitng o specific condions.
4. Cross-Validaon:
o Splis daa ino k-olds (e.g., 5-old cross-validaon) o evaluae model
perormance on mulple subses, ensuring robusness.
o Example: A cancer deecon model esed across five olds o diverse daa is less
likely o overfi han one rained on a single spli.
5. Early Sopping:
o Moniors validaon loss during raining and sops when i no longer improves,
prevenng he model rom memorizing raining daa.
o Example: Halng raining afer 10 epochs i validaon loss increases while raining
loss connues o drop.
6. Simpler Models:
o Using less complex archiecures reduces he risk o overfitng, especially wih
small daases.
o Example: Swiching rom a 50-layer CNN o a 10-layer CNN or using ranser
learning wih a pre-rained model like ResNe.
7. Transfer Learning:
o Fine-unes pre-rained models (e.g., rained on ImageNe) on specific asks,
leveraging general eaures o reduce overfitng.
o Example: Fine-uning a pre-rained ResNe on a small mammogram daase
improves generalizaon compared o raining rom scrach.
8. Daa Cleaning:
o Removing or correcng noisy or mislabeled daa improves he qualiy o he
raining se.
o Example: Veriying labels in a cancer daase o ensure malignan/benign cases are
accuraely annoaed.
9. Balanced Daases:
o Addressing class imbalance (e.g., using oversampling, undersampling, or synhec
daa generaon like SMOTE) ensures he model learns paerns rom all classes.
o Example: Generang synhec malignan mammograms o balance a daase wih
ew posive cases.