SlideShare a Scribd company logo
Novelty generation with deep learning
Presented by : Cherti Mehdi
joint work with Balázs Kégl and Akın Kazakçı
• Research questions
• Why machine learning and deep learning ?
• Generating new types of objects (C) using
previously acquired knowledge (K)
• Evaluating novelty with out-of-class generation
metrics
• Perspectives
Roadmap
Summary
• Recently, generative models have gained
momentum. But such models are almost
exclusively used in a prediction pipeline.
• Our objective is to study
• a) whether such models can be used to generate
novelty
• b) how to evaluate their capacity for generating
novelty
Research questions
• What is meant by the generation of novelty?
• How can novelty be generated?
• How can a model generating novelty be evaluated?
Why machine learning and deep learning ?
• Knowledge is important : machine learning enable
the study of creativity in relation with knowledge
• Generative modeling: we want to generate objects
• Composition of features is important : deep
learning models can automatically learn a
hierarchy of features of growing abstraction
from raw data
I focus my thesis on deep generative models.
Generating new types of objects
In Kazakçı et al. 2016:
• We show that symbols of new types can be
generated by carefully tuned autoencoders
• We make a first step of defining the conceptual
and experimental framework of novelty
generation
Generating new types of objects:
autoencoders
• Autoencoders have existed for a long
time (Kramer 1991)
• Deep variants are more recent (Hinton,
Salakhutdinov, 2006; Bengio 2009)
• A deep autoencoder learns
successive transformations that
decompose and then recompose a
set of training objects
• The depth allows learning a hierarchy
of transformations
• Two ways of learning an autoencoder :
undercomplete (bottleneck) and
overcomplete representation
Slide adapted from from Kazakçı et al. 2016
Generating new types of objects:
autoencoders with undercomplete
representation
Reconstruction
Input (dim 625)
Bottleneck
Encode
Decode
Deep autoencoder with a bottleneck from Hinton, G. E., & Salakhutdinov, R. R. (2006).
Generating new types of objects:
autoencoders with overcomplete
representation
• Autoencoders can also be learned using an
overcomplete representation
• Problem : Risk of learning the identity function
• One solution : constrain the representation to be
“simple”
• Example : enforce sparsity of the representation with
sparse autoencoders
Generating new types of objects:
autoencoders with overcomplete
representation
• What does a sparse autoencoder
end up learning ?
• Detect features with the encode
function
• Superpose the detected features
in the reconstructed image with
the decode function
• Benefits of overcomplete
representation with sparsity: for
each image only a small fraction
of features are used but different
images use a different subset of
features
k is the sparsity rate in %
Figure taken from Makhzani, A., & Frey, B. (2013)
Generating new types of objects:
experimental setup
• Training data : MNIST, 70000
images of handwritten digits of
size 28x28
• We use a sparse convolutional
autoencoder trained to:
• Encode : take an image and
transform it to a sparse code
• Decode : take the sparse code
and reconstruct the image
• Training objective is to minimize
the reconstruction error
Slide adapted from from Kazakçı et al. 2016
Generating new types of objects:
generating new symbols
• We use an iterative method to build symbols the net has never seen
(inspired by Bengio et al. (2013) but we don’t try to avoid spurious
samples):
• Start with a random image
• force the network to construct (i.e. interpret)
• , until convergence, f(x) = decode(encode(x))
Slide adapted from from Kazakçı et al. 2016
Generating new types of objects:
generating new symbols
• What does the iterative
generation procedure do ?
• It’s a non-linear path on the input
space defined by the
autoencoder (encode + decode)
function
• It converges to fixed points
defined by the autoencoder
Figure taken from Alain and Bengio (2013)
Generating new types of objects:
Visualization of the structure of generated
images
• Colored clusters are original
digits (classes from 0 to 9)
• The gray dots are newly
generated objects
• New objects form new
clusters
• Using a clustering algorithm,
we recover coherent sets of
new symbols
Slide adapted from from Kazakçı et al. 2016
Generating new types of objects
In Kazakçı et al. 2016:
• We show that symbols of new types can be generated by
carefully tuned autoencoders
• We make a first step of defining a conceptual and
experimental framework of novelty generation
• However, we make no attempt to design evaluation metrics
A set of types (clusters) discovered by the model
Evaluating novelty
In “Out-of-class novelty generation: an experimental
foundation” :
• We design an experimental framework based on hold-
out classes
• We review and analyze the most common evaluation
techniques from the point of view of measuring
“out-of-distribution novelty” and propose new ones
• We run a large-scale experimentation to study the
capacity for generating novelty of a wide set of
generative models
Evaluating novelty
Experimental framework
• We contrast two main concepts : in-class and out-of-
class generation
• in-class generation: can a model re-generate the types
already seen in the dataset ? (traditional objective)
• out-of-class generation : can a model generate an
unseen (hold-out) set of types ? (a proxy to measure the
capacity of a model to generate novelty)
• setup : we train models on a set of types(in), we seek
for models that generate a hold-out set of types(out)
Evaluating novelty
Evaluation metrics
• In our experiments:
• We train models on
digits
• We seek for models that
generate letters
in-class:
out-of-class:
• We pre-train a
• digit classifier (0 to 9)
• a letter classifier (a to z)
• a classifier on a mixture of digits and letters
• Our evaluation metrics report a score for a set of
generated objects by a model
Evaluating novelty
Evaluation metrics
Given a set of images, out-of-class objectness is high if:
• the letter classifier is highly confident for each
image being one of the letters (a to z)
• we define in-class objectness similarly but using the
digit classifier
Evaluating novelty
Evaluation metrics
Given a set of images, out-of-class max and out-of-class
count are high if:
• the mixture of digits and letters classifier is highly
confident for each image being one of the letters (a
to z)
• we define in-class max and in-class count similarly
but for digits
Evaluating novelty
Evaluation metrics
• We do a large scale experiment where we train
~1000 models by varying their parameters
• from each model, we generate 1000 images, then
we evaluate the model using our proposed metrics
• We collect a total of ~1.000.000 generated images
Experiments
• We evaluate the
evaluators with
human assessment
• We build an
annotation tool to
check whether the
models selected by
our evaluation
metrics are
effectively good
Experiments
Evaluating the evaluators
• we found that selecting models for in-class generation will make
them memorize the classes they are trained to sample from
• we did succeed to find models which lead to out-of-class novelty
• Pangram obtained from the above model:
Experiments
Results
Experiments
Results
• The main focus was setting up the experimental
pipeline and to analyze various quality metrics,
designed to measure out-of-distribution novelty
• The immediate next goal is to analyze the models
in a systematic way
Perspectives
Thank you for listening!

More Related Content

PDF
Slides, thesis dissertation defense, deep generative neural networks for nove...
PDF
Salient KeypointSelection for Object Representation
PDF
Tutorial on Deep Generative Models
PDF
Bitácora Travesía San José de Chiquitos
PPTX
Compras difiniciones
PPTX
Spike jonze
PPTX
Blumler and katz
DOC
Accountant
Slides, thesis dissertation defense, deep generative neural networks for nove...
Salient KeypointSelection for Object Representation
Tutorial on Deep Generative Models
Bitácora Travesía San José de Chiquitos
Compras difiniciones
Spike jonze
Blumler and katz
Accountant

Viewers also liked (8)

PDF
전략영업-화장품분야
DOCX
Bev Umali, CPA
PPT
ESWC2008 Relational2RDF - Mapping Relational Databases to RDF with OpenLink V...
PDF
Dossier-Trazos de lo Apatente /Trama
DOC
Erika_Lempke_Resume
PDF
A Beginner's Guide to Deep Learning
PDF
Project Proposal on Stress Management
DOCX
Resume-DouglasJVanTornhout 2016
전략영업-화장품분야
Bev Umali, CPA
ESWC2008 Relational2RDF - Mapping Relational Databases to RDF with OpenLink V...
Dossier-Trazos de lo Apatente /Trama
Erika_Lempke_Resume
A Beginner's Guide to Deep Learning
Project Proposal on Stress Management
Resume-DouglasJVanTornhout 2016
Ad

Similar to Novelty generation with deep learning (20)

PPTX
Evolving a Medical Image Similarity Search
PDF
Creating new classes of objects with deep generative neural nets
PDF
Information to Wisdom: Commonsense Knowledge Extraction and Compilation - Part 3
PPTX
CPP13 - Object Orientation
PPTX
YU CS Summer 2021 Project | TensorFlow Street Image Classification and Object...
PPTX
Object Oriented Programming Class and Objects
PDF
Creativity through deep learning
PPTX
Java 102 intro to object-oriented programming in java
PPTX
Lecture_16_Self-supervised_Learning.pptx
PPTX
Learning a Joint Embedding Representation for Image Search using Self-supervi...
PDF
The Search for a New Visual Search Beyond Language - StampedeCon AI Summit 2017
PDF
A Unified Framework for Computer Vision Tasks: (Conditional) Generative Model...
PPTX
brief Introduction to Different Kinds of GANs
PPT
CBIR_white.ppt
PPTX
12_oop templa.pptx
PPTX
[RSS2023] Local Object Crop Collision Network for Efficient Simulation
PDF
Lec 1-2 ssdsdffffsssssfsdfsdfstGenAI.pdf
PDF
Crowdsourced Object Segmentation with a Game
PPTX
Computer vision-nit-silchar-hackathon
PDF
Introduction to Machine Learning with SciKit-Learn
Evolving a Medical Image Similarity Search
Creating new classes of objects with deep generative neural nets
Information to Wisdom: Commonsense Knowledge Extraction and Compilation - Part 3
CPP13 - Object Orientation
YU CS Summer 2021 Project | TensorFlow Street Image Classification and Object...
Object Oriented Programming Class and Objects
Creativity through deep learning
Java 102 intro to object-oriented programming in java
Lecture_16_Self-supervised_Learning.pptx
Learning a Joint Embedding Representation for Image Search using Self-supervi...
The Search for a New Visual Search Beyond Language - StampedeCon AI Summit 2017
A Unified Framework for Computer Vision Tasks: (Conditional) Generative Model...
brief Introduction to Different Kinds of GANs
CBIR_white.ppt
12_oop templa.pptx
[RSS2023] Local Object Crop Collision Network for Efficient Simulation
Lec 1-2 ssdsdffffsssssfsdfsdfstGenAI.pdf
Crowdsourced Object Segmentation with a Game
Computer vision-nit-silchar-hackathon
Introduction to Machine Learning with SciKit-Learn
Ad

Recently uploaded (20)

PPTX
POULTRY PRODUCTION AND MANAGEMENTNNN.pptx
PPTX
A powerpoint on colorectal cancer with brief background
PPTX
endocrine - management of adrenal incidentaloma.pptx
PPTX
Understanding the Circulatory System……..
PPT
veterinary parasitology ````````````.ppt
PPTX
perinatal infections 2-171220190027.pptx
PDF
The Land of Punt — A research by Dhani Irwanto
PDF
Assessment of environmental effects of quarrying in Kitengela subcountyof Kaj...
PPT
Heredity-grade-9 Heredity-grade-9. Heredity-grade-9.
PDF
Is Earendel a Star Cluster?: Metal-poor Globular Cluster Progenitors at z ∼ 6
PPTX
SCIENCE 4 Q2W5 PPT.pptx Lesson About Plnts and animals and their habitat
PPTX
ap-psych-ch-1-introduction-to-psychology-presentation.pptx
PPTX
Welcome-grrewfefweg-students-of-2024.pptx
PPTX
Fluid dynamics vivavoce presentation of prakash
PPT
Presentation of a Romanian Institutee 2.
PDF
Worlds Next Door: A Candidate Giant Planet Imaged in the Habitable Zone of ↵ ...
PDF
lecture 2026 of Sjogren's syndrome l .pdf
PDF
Unit 5 Preparations, Reactions, Properties and Isomersim of Organic Compounds...
PDF
Communicating Health Policies to Diverse Populations (www.kiu.ac.ug)
PDF
Looking into the jet cone of the neutrino-associated very high-energy blazar ...
POULTRY PRODUCTION AND MANAGEMENTNNN.pptx
A powerpoint on colorectal cancer with brief background
endocrine - management of adrenal incidentaloma.pptx
Understanding the Circulatory System……..
veterinary parasitology ````````````.ppt
perinatal infections 2-171220190027.pptx
The Land of Punt — A research by Dhani Irwanto
Assessment of environmental effects of quarrying in Kitengela subcountyof Kaj...
Heredity-grade-9 Heredity-grade-9. Heredity-grade-9.
Is Earendel a Star Cluster?: Metal-poor Globular Cluster Progenitors at z ∼ 6
SCIENCE 4 Q2W5 PPT.pptx Lesson About Plnts and animals and their habitat
ap-psych-ch-1-introduction-to-psychology-presentation.pptx
Welcome-grrewfefweg-students-of-2024.pptx
Fluid dynamics vivavoce presentation of prakash
Presentation of a Romanian Institutee 2.
Worlds Next Door: A Candidate Giant Planet Imaged in the Habitable Zone of ↵ ...
lecture 2026 of Sjogren's syndrome l .pdf
Unit 5 Preparations, Reactions, Properties and Isomersim of Organic Compounds...
Communicating Health Policies to Diverse Populations (www.kiu.ac.ug)
Looking into the jet cone of the neutrino-associated very high-energy blazar ...

Novelty generation with deep learning

  • 1. Novelty generation with deep learning Presented by : Cherti Mehdi joint work with Balázs Kégl and Akın Kazakçı
  • 2. • Research questions • Why machine learning and deep learning ? • Generating new types of objects (C) using previously acquired knowledge (K) • Evaluating novelty with out-of-class generation metrics • Perspectives Roadmap
  • 3. Summary • Recently, generative models have gained momentum. But such models are almost exclusively used in a prediction pipeline. • Our objective is to study • a) whether such models can be used to generate novelty • b) how to evaluate their capacity for generating novelty
  • 4. Research questions • What is meant by the generation of novelty? • How can novelty be generated? • How can a model generating novelty be evaluated?
  • 5. Why machine learning and deep learning ? • Knowledge is important : machine learning enable the study of creativity in relation with knowledge • Generative modeling: we want to generate objects • Composition of features is important : deep learning models can automatically learn a hierarchy of features of growing abstraction from raw data I focus my thesis on deep generative models.
  • 6. Generating new types of objects In Kazakçı et al. 2016: • We show that symbols of new types can be generated by carefully tuned autoencoders • We make a first step of defining the conceptual and experimental framework of novelty generation
  • 7. Generating new types of objects: autoencoders • Autoencoders have existed for a long time (Kramer 1991) • Deep variants are more recent (Hinton, Salakhutdinov, 2006; Bengio 2009) • A deep autoencoder learns successive transformations that decompose and then recompose a set of training objects • The depth allows learning a hierarchy of transformations • Two ways of learning an autoencoder : undercomplete (bottleneck) and overcomplete representation Slide adapted from from Kazakçı et al. 2016
  • 8. Generating new types of objects: autoencoders with undercomplete representation Reconstruction Input (dim 625) Bottleneck Encode Decode Deep autoencoder with a bottleneck from Hinton, G. E., & Salakhutdinov, R. R. (2006).
  • 9. Generating new types of objects: autoencoders with overcomplete representation • Autoencoders can also be learned using an overcomplete representation • Problem : Risk of learning the identity function • One solution : constrain the representation to be “simple” • Example : enforce sparsity of the representation with sparse autoencoders
  • 10. Generating new types of objects: autoencoders with overcomplete representation • What does a sparse autoencoder end up learning ? • Detect features with the encode function • Superpose the detected features in the reconstructed image with the decode function • Benefits of overcomplete representation with sparsity: for each image only a small fraction of features are used but different images use a different subset of features k is the sparsity rate in % Figure taken from Makhzani, A., & Frey, B. (2013)
  • 11. Generating new types of objects: experimental setup • Training data : MNIST, 70000 images of handwritten digits of size 28x28 • We use a sparse convolutional autoencoder trained to: • Encode : take an image and transform it to a sparse code • Decode : take the sparse code and reconstruct the image • Training objective is to minimize the reconstruction error Slide adapted from from Kazakçı et al. 2016
  • 12. Generating new types of objects: generating new symbols • We use an iterative method to build symbols the net has never seen (inspired by Bengio et al. (2013) but we don’t try to avoid spurious samples): • Start with a random image • force the network to construct (i.e. interpret) • , until convergence, f(x) = decode(encode(x)) Slide adapted from from Kazakçı et al. 2016
  • 13. Generating new types of objects: generating new symbols • What does the iterative generation procedure do ? • It’s a non-linear path on the input space defined by the autoencoder (encode + decode) function • It converges to fixed points defined by the autoencoder Figure taken from Alain and Bengio (2013)
  • 14. Generating new types of objects: Visualization of the structure of generated images • Colored clusters are original digits (classes from 0 to 9) • The gray dots are newly generated objects • New objects form new clusters • Using a clustering algorithm, we recover coherent sets of new symbols Slide adapted from from Kazakçı et al. 2016
  • 15. Generating new types of objects In Kazakçı et al. 2016: • We show that symbols of new types can be generated by carefully tuned autoencoders • We make a first step of defining a conceptual and experimental framework of novelty generation • However, we make no attempt to design evaluation metrics A set of types (clusters) discovered by the model
  • 16. Evaluating novelty In “Out-of-class novelty generation: an experimental foundation” : • We design an experimental framework based on hold- out classes • We review and analyze the most common evaluation techniques from the point of view of measuring “out-of-distribution novelty” and propose new ones • We run a large-scale experimentation to study the capacity for generating novelty of a wide set of generative models
  • 17. Evaluating novelty Experimental framework • We contrast two main concepts : in-class and out-of- class generation • in-class generation: can a model re-generate the types already seen in the dataset ? (traditional objective) • out-of-class generation : can a model generate an unseen (hold-out) set of types ? (a proxy to measure the capacity of a model to generate novelty) • setup : we train models on a set of types(in), we seek for models that generate a hold-out set of types(out)
  • 18. Evaluating novelty Evaluation metrics • In our experiments: • We train models on digits • We seek for models that generate letters in-class: out-of-class:
  • 19. • We pre-train a • digit classifier (0 to 9) • a letter classifier (a to z) • a classifier on a mixture of digits and letters • Our evaluation metrics report a score for a set of generated objects by a model Evaluating novelty Evaluation metrics
  • 20. Given a set of images, out-of-class objectness is high if: • the letter classifier is highly confident for each image being one of the letters (a to z) • we define in-class objectness similarly but using the digit classifier Evaluating novelty Evaluation metrics
  • 21. Given a set of images, out-of-class max and out-of-class count are high if: • the mixture of digits and letters classifier is highly confident for each image being one of the letters (a to z) • we define in-class max and in-class count similarly but for digits Evaluating novelty Evaluation metrics
  • 22. • We do a large scale experiment where we train ~1000 models by varying their parameters • from each model, we generate 1000 images, then we evaluate the model using our proposed metrics • We collect a total of ~1.000.000 generated images Experiments
  • 23. • We evaluate the evaluators with human assessment • We build an annotation tool to check whether the models selected by our evaluation metrics are effectively good Experiments Evaluating the evaluators
  • 24. • we found that selecting models for in-class generation will make them memorize the classes they are trained to sample from • we did succeed to find models which lead to out-of-class novelty • Pangram obtained from the above model: Experiments Results
  • 26. • The main focus was setting up the experimental pipeline and to analyze various quality metrics, designed to measure out-of-distribution novelty • The immediate next goal is to analyze the models in a systematic way Perspectives
  • 27. Thank you for listening!