SlideShare a Scribd company logo
Azure Brain: 4th paradigm, scientific
discovery & (really) big data (REC201)
Gabriel Antoniu
Senior Research Scientist, Inria
Head of the KerData Project-Team, Inria Rennes – Bretagne Atlantique
Radu Tudoran
PhD student, ENS Cachan – Brittany
KerData Project-Team, Inria Rennes – Bretagne Atlantique
Feb. 12, 2013
INRIA’s strategy in Cloud Computing
INRIA is among the leaders in Europe in the area of distributed computing and HPC
• Long history of researches around distributed systems, HPC, Grids
• Now several activities virtualized environments/cloud infrastructures
• Culture of multidisciplinary research
• Culture of exploration tools (owner of massively parallel machines since 1987, large scale
testbeds such as Grid’5000)
• Strong involvement in national, European and international collaborative projects
• Strong collaboration history with industry (Joint Microsoft Research – Inria Centre, IBM,
EDF, Bull, etc.)
- 2
Clouds: where within Inria ?
1
2
Networks, Systems and Services,
Distributed Computing3
Perception, Cognition, Interaction
4
5
Applied Mathematics, Computation
and Simulation
Algorithmics, Programming,
Software and Architecture
Computational Sciences
for Biology, Medicine and
the Environment
- 3
Some project-teams involved in Cloud Computing
INRIA Nancy
Grand Est
INRIA Grenoble
Rhône-Alpes
INRIA Sophia Antipolis
Méditerranée
INRIA Rennes
Bretagne Atlantique
INRIA Bordeaux
Sud-Ouest
INRIA Lille
Nord Europe
INRIA
Saclay
Île-de-France
INRIA Paris
Rocquencourt
KERDATA: Data Storage and Processing
MYRIADS: Autonomous Distributed Systems
ASCOLA: Languages and virtualization
CEPAGE: task management
AVALON: middleware & programming
MESCAL: models & tools
REGAL: Large Scale dist. systems
ALGORILLE: algorithms & models
OASIS: programming
ZENITH: Scientific Data Management
- 4
Initiatives to support Cloud Computing and HPC within Inria
Why dedicated initiatives to support HPC/Clouds ?
• Project-teams are geographically dispersed
• Project-teams belong to different domains
• Researchers from scientific computing need access to the latest research results
related to tools, libraries, runtime systems, …
• Researchers from “computer science” need access to applications to test their
ideas as well as to find new ideas !
Concept of “Inria Large Scale Initiatives”
• Enable ambitious projects linked with the strategic plan
• Promote an interdisciplinary approach
• Mobilizing expertise of Inria researchers around key challenges
- 5
CLOUD COMPUTING@
INRIA RENNES BRETAGNE ATLANTIQUE
- 6
Some Research Focus Areas
Software architecture and infrastructure for cloud computing
• Autonomic service management, resource management, SLA, sky
computing: Myriads
• Big Data storage and management, MapReduce: KerData
• Hybrid Cloud and P2P systems, privacy: ASAP
Advanced usage for specific application communities
• Bioinformatics: GENSCALE
• Cloud for medical imaging: EasyMed project (IRT B-Com): Visages
- 7
Some Research Focus Areas
Software architecture and infrastructure for cloud computing
• Autonomic service management, resource management, SLA, sky
computing: Myriads
• Big Data storage and management, MapReduce: KerData
• Hybrid Cloud and P2P systems, privacy: ASAP
Advanced usage for specific application communities
• Bioinformatics: GENSCALE
• Cloud for medical imaging: EasyMed project (IRT B-Com): Visages
- 8
Contrail EU project
Goal: develop an integrated approach to virtualization offering
• services for federating IaaS clouds
• elastic PaaS services on top of federated clouds
Overview: provide tools for
• managing federation of multiple heterogeneous IaaS clouds
• offering a secure yet usable platform for end users through federated identity management
• supporting SLAs and quality of service (QoS) for satisfying stringent business requirements for
using the cloud
Resource
Provider
Federa&on)API)
+)Fed.)core)
Resource
Provider
Storage(
Provider( Public(
Cloud(
Storage(
Provider(Network(
Provider(
A) A)A) A)
Applica&on)
Applica&on)
Applica&on)
Federa&on)API)
+)Fed.)core)
Federa&on)API)
+)Fed.)core)
Contrail is an open source cloud computing
software stack compliant with cloud
standards
http://contrail-project.eu
- 9
Contrail EU project
http://contrail-project.eu
http://contrail.projects.ow2.org/xwiki/bin/view/Main/WebHome
- 10
Open source software under the GNU GPLv2
license
http://snooze.inria.fr
Other Research Activities on Cloud Computing
Snooze: an autonomic energy-
efficient IaaS management system
Scalability
• Distributed VM management system
• Self-organizing & self-healing hierarchy
Energy conservation
• Idle nodes in power-saving mode
• Holistic approach to favor idle nodes
VM management algorithms
• Energy-efficient VM placement
• Under-load / overload mitigation
• Automatic node power-cycling and wake-
up
Resilin: Elastic MapReduce on
multiple clouds (sky computing)
Goals
• Creation of MapReduce execution
platforms on top of multiple clouds
• Elasticity of the platforms
• Support all kinds of Hadoop jobs
• Support different Hadoop versions
Interfaces
• Amazon EMR for users
• Libcloud with underlying IaaS providers
Open source software under GNU Affero
GPL license
http://resilin.inria.fr
- 11
KerData: Dealing with the Data Deluge
Deliver the capability to mine,
search and analyze this data in
near real time
Science itself is evolving
Credits: Microsoft
12- 12
Last
few decades
The Data Science:
The 4th Paradigm for Scientific Discovery
Thousand
years ago
Today and the
Future
Last few
hundred years
2
2
2.
3
4
a
cG
a
a










Simulation of
complex phenomena
Newton’s laws,
Maxwell’s equations…
Description of natural
phenomena
Unify theory, experiment
and simulation with
large multidisciplinary
Data
Using data exploration
and data mining
(from instruments,
sensors, humans…)
Distributed Communities
Crédits: Dennis Gannon
13
Last
few decades
The Data Science:
The 4th Paradigm for Scientific Discovery
Thousand
years ago
Today and the
Future
Last few
hundred years
2
2
2.
3
4
a
cG
a
a










Simulation of
complex phenomena
Newton’s laws,
Maxwell’s equations…
Description of natural
phenomena
Unify theory, experiment
and simulation with
large multidisciplinary
Data
Using data exploration
and data mining
(from instruments,
sensors, humans…)
Distributed Communities
14
Research Focus:
How to efficiently store, share and process data
for new-generation, data-intensive applications?
• Scientific challenges
• Massive data (1 object = 1 TB)
• Geographically distributed
• Fine-grain access (MB) for reading and writing
• High concurrency (10³ concurrent clients)
• Without locking
- Major goal: high-throughput under heavy concurrency
- Our contribution
Design and implementation of distributed algorithms
Validation with real apps on real platforms with real users
• Applications
• Massive data analysis: clouds (e.g. MapReduce)
• Post-Petascale HPC simulations: supercomputers
- 15
BlobSeer: A Software Platform for Scalable,
Distributed BLOB Management
Started in 2008, 6 PhD theses (Gilles Kahn/SPECIF PhD Thesis Award in 2011)
Main goal: optimized for concurrent accesses under heavy concurrency
Three key ideas
•Decentralized metadata management
•Lock-free concurrent writes (enabled by versioning)
- Write = create new version of the data
•Data and metadata “patching” rather than updating
A back-end for higher-level data management systems
•Short term: highly scalable distributed file systems
•Middle term: storage for cloud services
Our approach
•Design and implementation of distributed algorithms
•Experiments on the Grid’5000 grid/cloud testbed
•Validation with “real” apps on “real” platforms: Nimbus, Azure, OpenNebula clouds…
http://blobseer.gforge.inria.fr/
16- 16
Impact of BlobSeer: MapReduce
BlobSeer improves Hadoop
• Gain (execution time) : 35%
ANR MapReduce Project (2010-2014)
• Lead: G. Antoniu (KerData)
• Partners: INRIA (AVALON), Argonne National Lab, U. Illinois Urbana-Champaign, IBM,
JLPC, IBCP, MEDIT
• Strong collaboration with the Nimbus team from Argonne National Lab
- BlobSeer integrated with the Nimbus cloud toolkit
- BlobSeer used for efficient VM deployment and snapshotting
• Validation : Grid’5000 with Nimbus, FutureGrid (USA), Open Cirrus (USA)
http://mapreduce.inria.fr
- 17
The A-Brain Project: Data-Intensive Processing on
Microsoft Azure Clouds
Application
• Large-scale joint genetic and
neuroimaging data analysis
Goal
• Assess and understand the
variability between individuals
Approach
• Optimized data processing on
Microsoft’s Azure clouds
Inria teams involved
• KerData (Rennes)
• Parietal(Saclay)
Framework
• Joint MSR-Inria Research Center
• MS involvement: Azure teams,
EMIC
18
Genetic information: SNPs
G G
T G
T T
T G
G G
MRI brain images
Clinical / behaviour
The Imaging Genetics Challenge:
Comparing Heterogeneous Information
THere we focus
on this link
- 19
Neuroimaging-genetics: The Problem
 Several brain diseases have a genetic
origin, or their occurrence/severity related
to genetic factors
 Genetics important to understand & predict
response to treatment
 Genetic variability captured in
DNA micro-array data
p( )|
Gene→Image
geneticimage
20
Imaging Genetics Methodological Issues
Genetic dataBrain image
Y
~105-106
~2000
X
~105-106
– Anatomical MRI
– Functional MRI
– Diffusion MRI
– DNA array (SNP/CNV)
– gene expression data
– others...
- 21
A BIG DATA Challenge …
Azure can help…
Data:
double
permutation
voxels
SNPs
5%-10%
useful
Computation:
Estimate timespan
on single machine
Estimation for A-Brain on Azure (350 cores)
Storage capacity estimations (350 cores)
Imaging Genetics Methodological Issues
 Multivariate methods:
predict brain characteristic with many
genetic variables
 Elastic net regularization:
combination of ℓ1 and ℓ2 penalties →
sparse loadings
 parameters setting:
internal cross-validation/bootstrap
 Performance evaluated using
permutations
23
A-Brain as Map-Reduce Processing
- 24
A-Brain as Map-Reduce Data Processing
25
Efficient Procedures for Statistics
Example : voxelwise Genome Wide Association Studies (vGWAS)
 740 subjects
 ~ 50,000 voxels
 ~ 500,000 SNPs
 10,000 permutations
→ ~ 12,000 hours of computation
→ ~ 1.8 Po of statistical scores
- 26
Efficient Procedures for Statistics
Example : Ridge regression with cross-validation loops
 Some costly computations
(SVD ~ 60 sec) are used 1-2
millions of times and cannot be
kept in memory.
~ 60-120 x 106 sec / SVD
(1.9-3.8 years / SVD)
→ An efficient distributed cache
can achieve huge speedup!
- 27
TomusBlobs approach
- 28
Requirements for a cloud storage / data management
High throughput under heavy concurrency
Fine grain access
Scalability / Elasticity
Data availability
Transparency
Design principles
Data locality – use the local storage
No modification on the cloud middleware
Loose coupling between storage and applications
Storage hierarchy
- 29
TomusBlobs - Architecture
- 30
Computation nodes
Architecture contd.
System components
Initiator
- Cloud specific
- Generic stub
- Properties: Scaling; Self configuration
Distributed Storage
- Aggregates the virtual disks
- Not depending on a specific solution
Client API
- Cloud specific API
- Expose the operation transparently
Initiator
Local
Disk
Application
Client API
TB
entity
VM snapshot
Customizable
Environment
- 31
TomusBlobs Evaluation
• Scenario: Single reader / writer
• Data transfer from memory to storage
• Metric: Client IO throughput
TomusBlobs Evaluation
Cumulative read throughput Cumulative write throughput
• Scenario: Multiple readers / writers
• Throughput limited by bandwidth
• Read 4X ; Write 5X
- 34
TomusBlobs as a Storage Backend for
Sharing Application Data in MapReduce
App
API
App App App App
API API API API
TomusBlobs
- 35
TomusMapReduce Evaluation
• Scenario: Increase the problem size
• Optimize computation by managing better intermediate data
- 36
Iterative MapReduce - Daytona
 Merge Step
 In-Memory Caching of static data
 Cache aware hybrid scheduling using Queues as well as
using a bulletin board (special table)
Reduce
Reduce
Merge
Add
Iteration?
No
Map Combine
Map Combine
Map Combine
Data
Cache
Yes
Hybrid scheduling of the new iteration
Job Start
Job Finish
Crédits: Dennis Gannon
- 37
Beyond MapReduce
• Unique result with parallel reduce
phase
• No central control entity
• No synchronization barrier
Map
Reducer
Map
Map
Map
Map
Reducer
- 38
Zoom on the Reduction Ratio
• Compute the minimum of a set of large matrixes (7.5 GB)
using 30 mappers
- 39
Azure integration
- 40
The Most Frequent Words benchmark
•Input data size varies from 3.2 GB to 32 GB
•ReductionRatio = 5
- 41
Execution times for A-Brain
•Increasing number of map jobs = increasing size of data
(5 GB to 50 GB)
- 42
Beyond Single Site
processing
• Data movements across geo-distributed
deployments is costly
• Minimize the size and number of transfers
• The overall aggregate must collaborate
towards reaching the goal
• The deployments work as independent
services
• The architecture can be used for scenarios
in which data is produced in different
locations
- 43
Towards a Geo-distributed
TomusBlobs approach
• TomusBlobs for intra-
deployment data management
• Public Storage (Azure
Blobs/Queues) for intra-
deployment communication
• Iterative Reduce technique for
minimizing number of transfers
(and data size)
• Balance the network bottleneck
from single data center
- 44
Multi-Site MapReduce
• 3 deployments (NE,WE,NUS)
• 1000 CPUs
• ABrain execution across multiple sites
- 45
Beyond MapReduce -
Workflow Processing
- 46
Data access patterns for workflows [1]
[1] Vairavanathan et al.
A Workflow-Aware Storage
System: An Opportunity Study
http://ece.ubc.ca/~matei/paper
s/ccgrid2012.pdf
Pipeline
Caching
Data informed workflow
Input
Output
Broadcast
Replication
Data size
Input
Output
Reduce/Gather
Co-placement of all data
Data informed workflow
Input
Output
Scatter
File size awareness
Data informed workflow
Input
Output
- 47
eScience Central
(Newcastle University)
- 48
Generic Worker Walkthrough
(Microsoft ATLE)
Local
storage
Client
code
Researcher
Job
Management
Service
Algorithm
HD
GW Driver
Pluggable
Runtime
Environment
Runtime
Business Logic
Job Details
Table
Job Index
Table
Notification Listeners
(Accounting, Status
Change, etc..)
BLOB Storage
Notification
Service
Scaling
Service
OGF
BES VM
SOAP WS–*
Use of interoperable standard
protocols and data schemas!
OGF
JSDL
Application Code
GW Services & SDKs
Existing Components
Input Files
Output Files
Shared
Storage
- 49
Credits: Microsoft
Defining the scope
ID
Files
Batch
jobs
Assumptions
about the
workflows
Workflows are
composed of
batch jobs with
well-defined
data passing
schemas
The input and
the output of the
batch jobs are
files The batch jobs
and their inputs
and outputs can
be uniquely
identified in the
system
Most workflows fit in this
subclass
Idea: manage files inside
the deployment
- 50
The Concept
File Name Locations
F1 VM1
F2 VM1,VM2
F3 VM2
VM 1 VM 2
Local Disk Local Disk
F1F2 F3
Transfer
Module
File Metadata Registry
(1) Register (F1,VM1)
(2) GetLocation(F1)
(3) DownloadFile(F1)
F1
F2
• Metadata Registry
• Transfer Module
Components
Transfer
Module
- 51
Characteristics of the components
Metadata Registry Transfer Module
Role
• Transfer files from one node to
another
Data type
• Files
Accessibility
• Each VM has such a module
• Applications access the local module
• The modules interact across nodes
Solutions
• FTP; Torrent; InMemory, HTTP etc.
Role
•Hold the location of files within the
deployment
Data type
•Key-value pairs –
(file identification; retrieval information)
Accessibility
•Accessible by all nodes
Solutions
•Azure Caching Preview, Azure Tables,
InMemory DB
Idea:
Adopt multiple transfer solutions
Adapt to the context: select the one that fits best
- 52
Transfer methods
Method Observations
InMemory • Caching data
• InMemory data offers fast access
• GBs of memory capacity per
deployment
• Small files
BitTorrent • Replicas for file dissemination
• Collaborative reads
• New way of stage-in data
FTP • TCP transfer
• Medium and large files
• Potential of inter-operability
- 53
VM Snapshot
VMMemory
MetaDataRegistry
Adaptive
Storage
FTP
Torrent
InMemory
Transfer Module Services
Replication Queue
Replication
FTP
Tracker
Peer
Local
Disk
- 54
F1
Azure Caching
Adaptive
Storage
Adaptive
Storage
App App
F1
Create
Upload(F1)
GetMetadata
Read(F1)
Memory Memory
Local Storage Local Storage
Read (F1)
WriteMetadata
Write(F1)
Download (F1)
API
API
- 55
Scenario 2 – Large files ; replication enabled
0
10
20
30
40
50
60
70
80
50 100 150 200 250
Time(sec)
Size of a single file (MB)
DirectLink Torrent Adaptive AzureBlobs
• Torrents are superior for broadcast when replicas are used
• DirectLink is faster for pipeline (reduction tree)
• Adaptive storage can chose each time the best strategy
- 56
NCBI Blast for Azure
Seamless Experience
• Evaluate data and invoke computational models
from Excel.
• Computationally heavy analysis done close to
large database of curated data.
• Scalable for large, surge computationally heavy
analysis.
selects DBs and
input sequence
Web Role Input Splitter
Worker Role
BLAST
Execution
Worker
Role #n….
Combiner
Worker Role
Genome
DB 1
Genome
DB K
BLAST DB
Configuration
Azure Blob
Storage
BLAST
Execution
Worker
Role #1
Crédits: Dennis Gannon
- 57
BLAST analysis – data management
component
0
30
60
90
120
5 10 15 25 35 40 50 60
Time(sec)
Number of BLAST jobs
Download Adaptive Download AzureBlobs Upload Adaptive Upload AzureBlobs
• Database files – 1.6 GB
• Input size – 800 MB
• 50 nodes
- 58
Scalable Storage on Clouds: Open Issues
Understanding price-performance trade-offs
• Consistency, availability, performance, cost, security,
quality of service, energy consumption
• Autonomy, adaptive consistency
• Dynamic elasticity
• Trade-offs exposed to the user
High performance variability
- Understand it, model it, cope with it
Deployment/application launching time is high
Latency of data accesses is still an issue
Data movements are expensive
Cope with tightly-coupled applications
Cope with various cloud programming models
Virtualization overhead
Benchmarking
Performance modeling
Self-optimization for cost reduction
- Elastic scale down
Security and privacy
- 59
Extreme scale does matter BUT not only
Other focus areas
– Affordability and usability of intermediate size systems
– Pervasiveness of usage across the entire industry, including Small and
Medium Enterprises (SMEs) and ISVs
– New HPC deployments (e.g. Big Data, HPC in Clouds)
– HPC and Cloud usage expansion, fostering the development of consultancy,
expertise and service business / end-user support
– Facilitating the creation of start-ups and the development of the SME sector
(hw/sw supply side)
– Education and training (inc. engineering skills for industry)
Cloud Computing@INRIA
Strategic Research Agenda
- 60
- 61
Azure Brain: 4th paradigm, scientific
discovery & (really) big data (REC201)
Gabriel Antoniu
Senior Research Scientist, Inria
Head of the KerData Project-Team, Inria Rennes – Bretagne Atlantique
Radu Tudoran
PhD student, ENS Cachan – Brittany
KerData Project-Team, Inria Rennes – Bretagne Atlantique
Contacts: Gabriel.Antoniu@inria.fr, Radu.Tudoran@inria.fr

More Related Content

PPTX
Data-intensive bioinformatics on HPC and Cloud
PPTX
Data-intensive applications on cloud computing resources: Applications in lif...
PPTX
From IoT Devices to Cloud
PPTX
Challenges and Issues of Next Cloud Computing Platforms
PDF
IDB-Cloud Providing Bioinformatics Services on Cloud
PDF
PLNOG 18 - Dr Marek Michalewicz - InfiniCortex: Superkomputer wielki jak świat
DOCX
Grid computing assiment
PPTX
Jorge gomes
Data-intensive bioinformatics on HPC and Cloud
Data-intensive applications on cloud computing resources: Applications in lif...
From IoT Devices to Cloud
Challenges and Issues of Next Cloud Computing Platforms
IDB-Cloud Providing Bioinformatics Services on Cloud
PLNOG 18 - Dr Marek Michalewicz - InfiniCortex: Superkomputer wielki jak świat
Grid computing assiment
Jorge gomes

What's hot (19)

PPTX
Grid'5000: Running a Large Instrument for Parallel and Distributed Computing ...
PDF
Virtualization for HPC at NCI
PDF
Building an Outsourcing Ecosystem for Science
PDF
Bergman Enabling Computation for neuro ML external
PDF
LambdaFabric for Machine Learning Acceleration
PPTX
The Pacific Research Platform Two Years In
PPT
TeraGrid Communication and Computation
PDF
What Are Science Clouds?
PPTX
Grid Computing by Mireille Raad
PPTX
Grid computing
PDF
Tutorial on Hybrid Data Infrastructures: D4Science as a case study
PPT
Grid Presentation
PPTX
Bionimbus - Northwestern CGI Workshop 4-21-2011
PDF
Using e-Infrastructures for Biodiversity Conservation
PPT
Cyberinfrastructure for Einstein's Equations and Beyond
PPT
Inroduction to grid computing by gargi shankar verma
PPTX
g-Social - Enhancing e-Science Tools with Social Networking Functionality
PPT
Gridcomputingppt
PPTX
SILECS/SLICES - Super Infrastructure for Large-Scale Experimental Computer Sc...
Grid'5000: Running a Large Instrument for Parallel and Distributed Computing ...
Virtualization for HPC at NCI
Building an Outsourcing Ecosystem for Science
Bergman Enabling Computation for neuro ML external
LambdaFabric for Machine Learning Acceleration
The Pacific Research Platform Two Years In
TeraGrid Communication and Computation
What Are Science Clouds?
Grid Computing by Mireille Raad
Grid computing
Tutorial on Hybrid Data Infrastructures: D4Science as a case study
Grid Presentation
Bionimbus - Northwestern CGI Workshop 4-21-2011
Using e-Infrastructures for Biodiversity Conservation
Cyberinfrastructure for Einstein's Equations and Beyond
Inroduction to grid computing by gargi shankar verma
g-Social - Enhancing e-Science Tools with Social Networking Functionality
Gridcomputingppt
SILECS/SLICES - Super Infrastructure for Large-Scale Experimental Computer Sc...
Ad

Viewers also liked (20)

PPTX
Mastering Brain Power
PPSX
Thetahealing
PDF
an introduction to Distem
PDF
A New Paradigm: Finding Knowledge Within
PDF
2013 ia summit
PPT
Telomeres: The Real Biologic Clock
PPTX
Paradigms
PDF
Data Science and the Fourth Paradigm by Torben Bach Pedersen
PPTX
The fourth paradigm in safety
PPT
Telomere seminar dec2012_new
PPT
Train the Mind Virtue as medicine through meditation
 
PPT
Sound: BINAURAL AUDIO
PPT
Binaural Audio
PPT
Linkedin Copy An Introduction To Silva Um Training
PPTX
ThetaHealing®Workshops
PPTX
Introduction to binaural beats
PPTX
The beta obsession
PPT
Ch7 alteredstates Reg. Psych
PPTX
Miracle grow your brain
Mastering Brain Power
Thetahealing
an introduction to Distem
A New Paradigm: Finding Knowledge Within
2013 ia summit
Telomeres: The Real Biologic Clock
Paradigms
Data Science and the Fourth Paradigm by Torben Bach Pedersen
The fourth paradigm in safety
Telomere seminar dec2012_new
Train the Mind Virtue as medicine through meditation
 
Sound: BINAURAL AUDIO
Binaural Audio
Linkedin Copy An Introduction To Silva Um Training
ThetaHealing®Workshops
Introduction to binaural beats
The beta obsession
Ch7 alteredstates Reg. Psych
Miracle grow your brain
Ad

Similar to Azure Brain: 4th paradigm, scientific discovery & (really) big data (20)

PPTX
NIST Big Data Public Working Group NBD-PWG
PPTX
PDF
Session19 Globus
PPTX
OpenNebulaConf 2013 - Keynote: Opening the Path to Technical Excellence by Jo...
PPTX
Opening the Path to Technical Excellence
PPTX
Progress of the Helix Nebula Science Cloud PCP Project
PPTX
Challenges for Standardization Cloud Computing and Big Data IOT
PPTX
information system.pptx
PPTX
SILECS: Super Infrastructure for Large-scale Experimental Computer Science
PDF
OpenNebulaConf2015 2.06 OpenNebula in the Wild - Ander Astudillo
PPTX
Cloud and Grid Computing
PPTX
Cloud and grid computing by Leen Blom, Centric
PDF
Practitioner's perspective on High Performance Computing services for innovat...
PDF
Grid is Dead ? Nimrod on the Cloud
PDF
Adoption of Cloud Computing in Scientific Research
PPTX
Utilising Cloud Computing for Research through Infrastructure, Software and D...
PDF
Big data and cloud computing 9 sep-2017
PPTX
Federated Cloud Computing
PPTX
Cloud Programming Models: eScience, Big Data, etc.
NIST Big Data Public Working Group NBD-PWG
Session19 Globus
OpenNebulaConf 2013 - Keynote: Opening the Path to Technical Excellence by Jo...
Opening the Path to Technical Excellence
Progress of the Helix Nebula Science Cloud PCP Project
Challenges for Standardization Cloud Computing and Big Data IOT
information system.pptx
SILECS: Super Infrastructure for Large-scale Experimental Computer Science
OpenNebulaConf2015 2.06 OpenNebula in the Wild - Ander Astudillo
Cloud and Grid Computing
Cloud and grid computing by Leen Blom, Centric
Practitioner's perspective on High Performance Computing services for innovat...
Grid is Dead ? Nimrod on the Cloud
Adoption of Cloud Computing in Scientific Research
Utilising Cloud Computing for Research through Infrastructure, Software and D...
Big data and cloud computing 9 sep-2017
Federated Cloud Computing
Cloud Programming Models: eScience, Big Data, etc.

More from Microsoft Technet France (20)

PDF
Automatisez, visualisez et améliorez vos processus d’entreprise avec Nintex
PPTX
Comment réussir votre déploiement de Windows 10
PPTX
OMS log search au quotidien
PPTX
Fusion, Acquisition - Optimisez la migration et la continuité des outils col...
PPTX
Wavestone déploie son portail Powell 365 en 5 semaines
PPTX
Retour d’expérience sur le monitoring et la sécurisation des identités Azure
PPTX
Scénarios de mobilité couverts par Enterprise Mobility + Security
PPTX
SharePoint Framework : le développement SharePoint nouvelle génération
PPTX
Stockage Cloud : il y en aura pour tout le monde
PPTX
Bien appréhender le concept de Windows As a Service
PPTX
Protéger vos données avec le chiffrement dans Azure et Office 365
PPTX
Protéger votre patrimoine informationnel dans un monde hybride avec Azure Inf...
PPTX
Comprendre la stratégie identité de Microsoft
PPTX
Vous avez dit « authentification sans mot de passe » : une illustration avec ...
PPTX
Sécurité des données
PPTX
Déploiement hybride, la téléphonie dans le cloud
PPTX
Supervisez la qualité des appels Skype for Business Online à l'aide de Call Q...
PPTX
SharePoint 2016 : architecture, déploiement et topologies hybrides
PPTX
Gestion de Windows 10 et des applications dans l'entreprise moderne
PPTX
Office 365 dans votre Système d'Informations
Automatisez, visualisez et améliorez vos processus d’entreprise avec Nintex
Comment réussir votre déploiement de Windows 10
OMS log search au quotidien
Fusion, Acquisition - Optimisez la migration et la continuité des outils col...
Wavestone déploie son portail Powell 365 en 5 semaines
Retour d’expérience sur le monitoring et la sécurisation des identités Azure
Scénarios de mobilité couverts par Enterprise Mobility + Security
SharePoint Framework : le développement SharePoint nouvelle génération
Stockage Cloud : il y en aura pour tout le monde
Bien appréhender le concept de Windows As a Service
Protéger vos données avec le chiffrement dans Azure et Office 365
Protéger votre patrimoine informationnel dans un monde hybride avec Azure Inf...
Comprendre la stratégie identité de Microsoft
Vous avez dit « authentification sans mot de passe » : une illustration avec ...
Sécurité des données
Déploiement hybride, la téléphonie dans le cloud
Supervisez la qualité des appels Skype for Business Online à l'aide de Call Q...
SharePoint 2016 : architecture, déploiement et topologies hybrides
Gestion de Windows 10 et des applications dans l'entreprise moderne
Office 365 dans votre Système d'Informations

Recently uploaded (20)

PDF
Reach Out and Touch Someone: Haptics and Empathic Computing
PPTX
MYSQL Presentation for SQL database connectivity
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PDF
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
PPTX
KOM of Painting work and Equipment Insulation REV00 update 25-dec.pptx
PDF
Electronic commerce courselecture one. Pdf
PDF
The Rise and Fall of 3GPP – Time for a Sabbatical?
PPTX
20250228 LYD VKU AI Blended-Learning.pptx
PDF
Dropbox Q2 2025 Financial Results & Investor Presentation
PDF
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
PDF
Bridging biosciences and deep learning for revolutionary discoveries: a compr...
PDF
NewMind AI Weekly Chronicles - August'25 Week I
PDF
Encapsulation theory and applications.pdf
PDF
CIFDAQ's Market Insight: SEC Turns Pro Crypto
PPTX
A Presentation on Artificial Intelligence
PPTX
Cloud computing and distributed systems.
PPT
“AI and Expert System Decision Support & Business Intelligence Systems”
PDF
Unlocking AI with Model Context Protocol (MCP)
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PPTX
PA Analog/Digital System: The Backbone of Modern Surveillance and Communication
Reach Out and Touch Someone: Haptics and Empathic Computing
MYSQL Presentation for SQL database connectivity
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
KOM of Painting work and Equipment Insulation REV00 update 25-dec.pptx
Electronic commerce courselecture one. Pdf
The Rise and Fall of 3GPP – Time for a Sabbatical?
20250228 LYD VKU AI Blended-Learning.pptx
Dropbox Q2 2025 Financial Results & Investor Presentation
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
Bridging biosciences and deep learning for revolutionary discoveries: a compr...
NewMind AI Weekly Chronicles - August'25 Week I
Encapsulation theory and applications.pdf
CIFDAQ's Market Insight: SEC Turns Pro Crypto
A Presentation on Artificial Intelligence
Cloud computing and distributed systems.
“AI and Expert System Decision Support & Business Intelligence Systems”
Unlocking AI with Model Context Protocol (MCP)
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PA Analog/Digital System: The Backbone of Modern Surveillance and Communication

Azure Brain: 4th paradigm, scientific discovery & (really) big data

  • 1. Azure Brain: 4th paradigm, scientific discovery & (really) big data (REC201) Gabriel Antoniu Senior Research Scientist, Inria Head of the KerData Project-Team, Inria Rennes – Bretagne Atlantique Radu Tudoran PhD student, ENS Cachan – Brittany KerData Project-Team, Inria Rennes – Bretagne Atlantique Feb. 12, 2013
  • 2. INRIA’s strategy in Cloud Computing INRIA is among the leaders in Europe in the area of distributed computing and HPC • Long history of researches around distributed systems, HPC, Grids • Now several activities virtualized environments/cloud infrastructures • Culture of multidisciplinary research • Culture of exploration tools (owner of massively parallel machines since 1987, large scale testbeds such as Grid’5000) • Strong involvement in national, European and international collaborative projects • Strong collaboration history with industry (Joint Microsoft Research – Inria Centre, IBM, EDF, Bull, etc.) - 2
  • 3. Clouds: where within Inria ? 1 2 Networks, Systems and Services, Distributed Computing3 Perception, Cognition, Interaction 4 5 Applied Mathematics, Computation and Simulation Algorithmics, Programming, Software and Architecture Computational Sciences for Biology, Medicine and the Environment - 3
  • 4. Some project-teams involved in Cloud Computing INRIA Nancy Grand Est INRIA Grenoble Rhône-Alpes INRIA Sophia Antipolis Méditerranée INRIA Rennes Bretagne Atlantique INRIA Bordeaux Sud-Ouest INRIA Lille Nord Europe INRIA Saclay Île-de-France INRIA Paris Rocquencourt KERDATA: Data Storage and Processing MYRIADS: Autonomous Distributed Systems ASCOLA: Languages and virtualization CEPAGE: task management AVALON: middleware & programming MESCAL: models & tools REGAL: Large Scale dist. systems ALGORILLE: algorithms & models OASIS: programming ZENITH: Scientific Data Management - 4
  • 5. Initiatives to support Cloud Computing and HPC within Inria Why dedicated initiatives to support HPC/Clouds ? • Project-teams are geographically dispersed • Project-teams belong to different domains • Researchers from scientific computing need access to the latest research results related to tools, libraries, runtime systems, … • Researchers from “computer science” need access to applications to test their ideas as well as to find new ideas ! Concept of “Inria Large Scale Initiatives” • Enable ambitious projects linked with the strategic plan • Promote an interdisciplinary approach • Mobilizing expertise of Inria researchers around key challenges - 5
  • 6. CLOUD COMPUTING@ INRIA RENNES BRETAGNE ATLANTIQUE - 6
  • 7. Some Research Focus Areas Software architecture and infrastructure for cloud computing • Autonomic service management, resource management, SLA, sky computing: Myriads • Big Data storage and management, MapReduce: KerData • Hybrid Cloud and P2P systems, privacy: ASAP Advanced usage for specific application communities • Bioinformatics: GENSCALE • Cloud for medical imaging: EasyMed project (IRT B-Com): Visages - 7
  • 8. Some Research Focus Areas Software architecture and infrastructure for cloud computing • Autonomic service management, resource management, SLA, sky computing: Myriads • Big Data storage and management, MapReduce: KerData • Hybrid Cloud and P2P systems, privacy: ASAP Advanced usage for specific application communities • Bioinformatics: GENSCALE • Cloud for medical imaging: EasyMed project (IRT B-Com): Visages - 8
  • 9. Contrail EU project Goal: develop an integrated approach to virtualization offering • services for federating IaaS clouds • elastic PaaS services on top of federated clouds Overview: provide tools for • managing federation of multiple heterogeneous IaaS clouds • offering a secure yet usable platform for end users through federated identity management • supporting SLAs and quality of service (QoS) for satisfying stringent business requirements for using the cloud Resource Provider Federa&on)API) +)Fed.)core) Resource Provider Storage( Provider( Public( Cloud( Storage( Provider(Network( Provider( A) A)A) A) Applica&on) Applica&on) Applica&on) Federa&on)API) +)Fed.)core) Federa&on)API) +)Fed.)core) Contrail is an open source cloud computing software stack compliant with cloud standards http://contrail-project.eu - 9
  • 11. Open source software under the GNU GPLv2 license http://snooze.inria.fr Other Research Activities on Cloud Computing Snooze: an autonomic energy- efficient IaaS management system Scalability • Distributed VM management system • Self-organizing & self-healing hierarchy Energy conservation • Idle nodes in power-saving mode • Holistic approach to favor idle nodes VM management algorithms • Energy-efficient VM placement • Under-load / overload mitigation • Automatic node power-cycling and wake- up Resilin: Elastic MapReduce on multiple clouds (sky computing) Goals • Creation of MapReduce execution platforms on top of multiple clouds • Elasticity of the platforms • Support all kinds of Hadoop jobs • Support different Hadoop versions Interfaces • Amazon EMR for users • Libcloud with underlying IaaS providers Open source software under GNU Affero GPL license http://resilin.inria.fr - 11
  • 12. KerData: Dealing with the Data Deluge Deliver the capability to mine, search and analyze this data in near real time Science itself is evolving Credits: Microsoft 12- 12
  • 13. Last few decades The Data Science: The 4th Paradigm for Scientific Discovery Thousand years ago Today and the Future Last few hundred years 2 2 2. 3 4 a cG a a           Simulation of complex phenomena Newton’s laws, Maxwell’s equations… Description of natural phenomena Unify theory, experiment and simulation with large multidisciplinary Data Using data exploration and data mining (from instruments, sensors, humans…) Distributed Communities Crédits: Dennis Gannon 13
  • 14. Last few decades The Data Science: The 4th Paradigm for Scientific Discovery Thousand years ago Today and the Future Last few hundred years 2 2 2. 3 4 a cG a a           Simulation of complex phenomena Newton’s laws, Maxwell’s equations… Description of natural phenomena Unify theory, experiment and simulation with large multidisciplinary Data Using data exploration and data mining (from instruments, sensors, humans…) Distributed Communities 14
  • 15. Research Focus: How to efficiently store, share and process data for new-generation, data-intensive applications? • Scientific challenges • Massive data (1 object = 1 TB) • Geographically distributed • Fine-grain access (MB) for reading and writing • High concurrency (10³ concurrent clients) • Without locking - Major goal: high-throughput under heavy concurrency - Our contribution Design and implementation of distributed algorithms Validation with real apps on real platforms with real users • Applications • Massive data analysis: clouds (e.g. MapReduce) • Post-Petascale HPC simulations: supercomputers - 15
  • 16. BlobSeer: A Software Platform for Scalable, Distributed BLOB Management Started in 2008, 6 PhD theses (Gilles Kahn/SPECIF PhD Thesis Award in 2011) Main goal: optimized for concurrent accesses under heavy concurrency Three key ideas •Decentralized metadata management •Lock-free concurrent writes (enabled by versioning) - Write = create new version of the data •Data and metadata “patching” rather than updating A back-end for higher-level data management systems •Short term: highly scalable distributed file systems •Middle term: storage for cloud services Our approach •Design and implementation of distributed algorithms •Experiments on the Grid’5000 grid/cloud testbed •Validation with “real” apps on “real” platforms: Nimbus, Azure, OpenNebula clouds… http://blobseer.gforge.inria.fr/ 16- 16
  • 17. Impact of BlobSeer: MapReduce BlobSeer improves Hadoop • Gain (execution time) : 35% ANR MapReduce Project (2010-2014) • Lead: G. Antoniu (KerData) • Partners: INRIA (AVALON), Argonne National Lab, U. Illinois Urbana-Champaign, IBM, JLPC, IBCP, MEDIT • Strong collaboration with the Nimbus team from Argonne National Lab - BlobSeer integrated with the Nimbus cloud toolkit - BlobSeer used for efficient VM deployment and snapshotting • Validation : Grid’5000 with Nimbus, FutureGrid (USA), Open Cirrus (USA) http://mapreduce.inria.fr - 17
  • 18. The A-Brain Project: Data-Intensive Processing on Microsoft Azure Clouds Application • Large-scale joint genetic and neuroimaging data analysis Goal • Assess and understand the variability between individuals Approach • Optimized data processing on Microsoft’s Azure clouds Inria teams involved • KerData (Rennes) • Parietal(Saclay) Framework • Joint MSR-Inria Research Center • MS involvement: Azure teams, EMIC 18
  • 19. Genetic information: SNPs G G T G T T T G G G MRI brain images Clinical / behaviour The Imaging Genetics Challenge: Comparing Heterogeneous Information THere we focus on this link - 19
  • 20. Neuroimaging-genetics: The Problem  Several brain diseases have a genetic origin, or their occurrence/severity related to genetic factors  Genetics important to understand & predict response to treatment  Genetic variability captured in DNA micro-array data p( )| Gene→Image geneticimage 20
  • 21. Imaging Genetics Methodological Issues Genetic dataBrain image Y ~105-106 ~2000 X ~105-106 – Anatomical MRI – Functional MRI – Diffusion MRI – DNA array (SNP/CNV) – gene expression data – others... - 21
  • 22. A BIG DATA Challenge … Azure can help… Data: double permutation voxels SNPs 5%-10% useful Computation: Estimate timespan on single machine Estimation for A-Brain on Azure (350 cores) Storage capacity estimations (350 cores)
  • 23. Imaging Genetics Methodological Issues  Multivariate methods: predict brain characteristic with many genetic variables  Elastic net regularization: combination of ℓ1 and ℓ2 penalties → sparse loadings  parameters setting: internal cross-validation/bootstrap  Performance evaluated using permutations 23
  • 24. A-Brain as Map-Reduce Processing - 24
  • 25. A-Brain as Map-Reduce Data Processing 25
  • 26. Efficient Procedures for Statistics Example : voxelwise Genome Wide Association Studies (vGWAS)  740 subjects  ~ 50,000 voxels  ~ 500,000 SNPs  10,000 permutations → ~ 12,000 hours of computation → ~ 1.8 Po of statistical scores - 26
  • 27. Efficient Procedures for Statistics Example : Ridge regression with cross-validation loops  Some costly computations (SVD ~ 60 sec) are used 1-2 millions of times and cannot be kept in memory. ~ 60-120 x 106 sec / SVD (1.9-3.8 years / SVD) → An efficient distributed cache can achieve huge speedup! - 27
  • 29. Requirements for a cloud storage / data management High throughput under heavy concurrency Fine grain access Scalability / Elasticity Data availability Transparency Design principles Data locality – use the local storage No modification on the cloud middleware Loose coupling between storage and applications Storage hierarchy - 29
  • 30. TomusBlobs - Architecture - 30 Computation nodes
  • 31. Architecture contd. System components Initiator - Cloud specific - Generic stub - Properties: Scaling; Self configuration Distributed Storage - Aggregates the virtual disks - Not depending on a specific solution Client API - Cloud specific API - Expose the operation transparently Initiator Local Disk Application Client API TB entity VM snapshot Customizable Environment - 31
  • 32. TomusBlobs Evaluation • Scenario: Single reader / writer • Data transfer from memory to storage • Metric: Client IO throughput
  • 33. TomusBlobs Evaluation Cumulative read throughput Cumulative write throughput • Scenario: Multiple readers / writers • Throughput limited by bandwidth • Read 4X ; Write 5X - 34
  • 34. TomusBlobs as a Storage Backend for Sharing Application Data in MapReduce App API App App App App API API API API TomusBlobs - 35
  • 35. TomusMapReduce Evaluation • Scenario: Increase the problem size • Optimize computation by managing better intermediate data - 36
  • 36. Iterative MapReduce - Daytona  Merge Step  In-Memory Caching of static data  Cache aware hybrid scheduling using Queues as well as using a bulletin board (special table) Reduce Reduce Merge Add Iteration? No Map Combine Map Combine Map Combine Data Cache Yes Hybrid scheduling of the new iteration Job Start Job Finish Crédits: Dennis Gannon - 37
  • 37. Beyond MapReduce • Unique result with parallel reduce phase • No central control entity • No synchronization barrier Map Reducer Map Map Map Map Reducer - 38
  • 38. Zoom on the Reduction Ratio • Compute the minimum of a set of large matrixes (7.5 GB) using 30 mappers - 39
  • 40. The Most Frequent Words benchmark •Input data size varies from 3.2 GB to 32 GB •ReductionRatio = 5 - 41
  • 41. Execution times for A-Brain •Increasing number of map jobs = increasing size of data (5 GB to 50 GB) - 42
  • 42. Beyond Single Site processing • Data movements across geo-distributed deployments is costly • Minimize the size and number of transfers • The overall aggregate must collaborate towards reaching the goal • The deployments work as independent services • The architecture can be used for scenarios in which data is produced in different locations - 43
  • 43. Towards a Geo-distributed TomusBlobs approach • TomusBlobs for intra- deployment data management • Public Storage (Azure Blobs/Queues) for intra- deployment communication • Iterative Reduce technique for minimizing number of transfers (and data size) • Balance the network bottleneck from single data center - 44
  • 44. Multi-Site MapReduce • 3 deployments (NE,WE,NUS) • 1000 CPUs • ABrain execution across multiple sites - 45
  • 45. Beyond MapReduce - Workflow Processing - 46
  • 46. Data access patterns for workflows [1] [1] Vairavanathan et al. A Workflow-Aware Storage System: An Opportunity Study http://ece.ubc.ca/~matei/paper s/ccgrid2012.pdf Pipeline Caching Data informed workflow Input Output Broadcast Replication Data size Input Output Reduce/Gather Co-placement of all data Data informed workflow Input Output Scatter File size awareness Data informed workflow Input Output - 47
  • 48. Generic Worker Walkthrough (Microsoft ATLE) Local storage Client code Researcher Job Management Service Algorithm HD GW Driver Pluggable Runtime Environment Runtime Business Logic Job Details Table Job Index Table Notification Listeners (Accounting, Status Change, etc..) BLOB Storage Notification Service Scaling Service OGF BES VM SOAP WS–* Use of interoperable standard protocols and data schemas! OGF JSDL Application Code GW Services & SDKs Existing Components Input Files Output Files Shared Storage - 49 Credits: Microsoft
  • 49. Defining the scope ID Files Batch jobs Assumptions about the workflows Workflows are composed of batch jobs with well-defined data passing schemas The input and the output of the batch jobs are files The batch jobs and their inputs and outputs can be uniquely identified in the system Most workflows fit in this subclass Idea: manage files inside the deployment - 50
  • 50. The Concept File Name Locations F1 VM1 F2 VM1,VM2 F3 VM2 VM 1 VM 2 Local Disk Local Disk F1F2 F3 Transfer Module File Metadata Registry (1) Register (F1,VM1) (2) GetLocation(F1) (3) DownloadFile(F1) F1 F2 • Metadata Registry • Transfer Module Components Transfer Module - 51
  • 51. Characteristics of the components Metadata Registry Transfer Module Role • Transfer files from one node to another Data type • Files Accessibility • Each VM has such a module • Applications access the local module • The modules interact across nodes Solutions • FTP; Torrent; InMemory, HTTP etc. Role •Hold the location of files within the deployment Data type •Key-value pairs – (file identification; retrieval information) Accessibility •Accessible by all nodes Solutions •Azure Caching Preview, Azure Tables, InMemory DB Idea: Adopt multiple transfer solutions Adapt to the context: select the one that fits best - 52
  • 52. Transfer methods Method Observations InMemory • Caching data • InMemory data offers fast access • GBs of memory capacity per deployment • Small files BitTorrent • Replicas for file dissemination • Collaborative reads • New way of stage-in data FTP • TCP transfer • Medium and large files • Potential of inter-operability - 53
  • 53. VM Snapshot VMMemory MetaDataRegistry Adaptive Storage FTP Torrent InMemory Transfer Module Services Replication Queue Replication FTP Tracker Peer Local Disk - 54
  • 54. F1 Azure Caching Adaptive Storage Adaptive Storage App App F1 Create Upload(F1) GetMetadata Read(F1) Memory Memory Local Storage Local Storage Read (F1) WriteMetadata Write(F1) Download (F1) API API - 55
  • 55. Scenario 2 – Large files ; replication enabled 0 10 20 30 40 50 60 70 80 50 100 150 200 250 Time(sec) Size of a single file (MB) DirectLink Torrent Adaptive AzureBlobs • Torrents are superior for broadcast when replicas are used • DirectLink is faster for pipeline (reduction tree) • Adaptive storage can chose each time the best strategy - 56
  • 56. NCBI Blast for Azure Seamless Experience • Evaluate data and invoke computational models from Excel. • Computationally heavy analysis done close to large database of curated data. • Scalable for large, surge computationally heavy analysis. selects DBs and input sequence Web Role Input Splitter Worker Role BLAST Execution Worker Role #n…. Combiner Worker Role Genome DB 1 Genome DB K BLAST DB Configuration Azure Blob Storage BLAST Execution Worker Role #1 Crédits: Dennis Gannon - 57
  • 57. BLAST analysis – data management component 0 30 60 90 120 5 10 15 25 35 40 50 60 Time(sec) Number of BLAST jobs Download Adaptive Download AzureBlobs Upload Adaptive Upload AzureBlobs • Database files – 1.6 GB • Input size – 800 MB • 50 nodes - 58
  • 58. Scalable Storage on Clouds: Open Issues Understanding price-performance trade-offs • Consistency, availability, performance, cost, security, quality of service, energy consumption • Autonomy, adaptive consistency • Dynamic elasticity • Trade-offs exposed to the user High performance variability - Understand it, model it, cope with it Deployment/application launching time is high Latency of data accesses is still an issue Data movements are expensive Cope with tightly-coupled applications Cope with various cloud programming models Virtualization overhead Benchmarking Performance modeling Self-optimization for cost reduction - Elastic scale down Security and privacy - 59
  • 59. Extreme scale does matter BUT not only Other focus areas – Affordability and usability of intermediate size systems – Pervasiveness of usage across the entire industry, including Small and Medium Enterprises (SMEs) and ISVs – New HPC deployments (e.g. Big Data, HPC in Clouds) – HPC and Cloud usage expansion, fostering the development of consultancy, expertise and service business / end-user support – Facilitating the creation of start-ups and the development of the SME sector (hw/sw supply side) – Education and training (inc. engineering skills for industry) Cloud Computing@INRIA Strategic Research Agenda - 60
  • 60. - 61 Azure Brain: 4th paradigm, scientific discovery & (really) big data (REC201) Gabriel Antoniu Senior Research Scientist, Inria Head of the KerData Project-Team, Inria Rennes – Bretagne Atlantique Radu Tudoran PhD student, ENS Cachan – Brittany KerData Project-Team, Inria Rennes – Bretagne Atlantique Contacts: Gabriel.Antoniu@inria.fr, Radu.Tudoran@inria.fr