BIG DIVE reaches its 8th edition!
BIG DIVE reaches its 8th edition!
BIG DIVE is improving every year: learning from past experiences, carefully integrating past alumni feedback and staying always up to date on the latest innovations and topics.
This is the reason why, beyond keep offering great teaching contents and top-class lectures from the most valuable experts, transferable for any discipline, we have also chosen to focus on one of the most crucial topic of our time.
Data from space to Earth
The present challenges, from climate change to sustainability, from fighting poverty to improving the quality of life, raise the importance to monitor our planet Earth from any point of view, including from space with satellite images.
For this reasons we decided to partner with successful companies in the field that already got a glimpse of the potential in using Data Engineering, Network Science, Machine and Deep Learning to leverage the data acquired from instruments and sensors in order to come up with new insights and to improve existing solutions or to create new ones aimed at preserving our planet, being reactive in case of emergencies and optimising resources consumptions.
Furthermore, the above studies have also many industrial applications, such as mobility, real estate and agriculture.
A complete data science program
BIG DIVE 8 will cover:
Real cases that prepare you to work with data
The highest value is that you learn through hands-on real cases, this means you can get insights from data-driven companies and research institutions on how they managed, analyzed and inferred on their own data and you have the possibility to apply and play with their methodologies asking directly to data scientists, researchers and developers.
This is a unique opportunity that you can’t get anywhere else!
Click on the boxes below for a preview of the program:
Speaker: Pierre-Philippe Mathieu — ESA
The view from space has forever changed our vision on our home planet, revealing its beauty while pointing at the same time to its inherent fragility. This new perspective from above contributed to the emergence of the concept of Sustainable Development (SD), by convincing many of the need to (better) manage our (rapidly depleting) resources in a sustainable manner that would “meet the needs of the present without compromising the ability of future generations to meet their own needs”.
Over the last decades, the principles of SD were progressively adopted by world leaders on the occasion of a series of Earth Summits. One of the key challenges to implement SD however lies in one’s ability to measure it. As stated by Lord Kelvin, “if you cannot measure it, you cannot improve it”. The challenge is further compounded by the inherent global nature of the problem, which calls for global data sets.
Earth Observation (EO) satellites can play a key role to meet this challenge, as they uniquely placed to monitor the state of our environment, in a global and consistent manner, ensuring sufficient resolution to capture the footprint of man-made activities.
The world of EO is rapidly changing driven by very fast advances in sensor and digital technologies. The speed of change has no historical precedent. Recent decades have witnessed extraordinary developments in ICT, including the Internet and cloud computing, and technologies such as Artificial Intelligence (AI), leading to radically new ways to collect, distribute and analyze data about our planet. This digital revolution is also accompanied by a sensing revolution providing an unprecedented amount of data on the state of our planet and its changes.
Europe is leading this sensing revolution in space through the EO missions of the European Space Agency (ESA), a new generation of meteo missions for Eumetsat, and especially the Copernicus initiative led by the European Union (EU). The latter is centered around the development of a family of Sentinel missions by ESA for the EU so as to enable global monitoring of our planet on an operational and sustained basis over the coming decades. In addition, a new trend, referred to as “New Space” in the US or “Space 4.0” in Europe, is now rapidly emerging through the increasing commoditization and commercialization of space. In particular, with the rapidly dropping costs of small sat building, launching and data processing, new EO actors including startups and ICT giants, in particular across the Atlantic, are now massively entering the space business, resulting in new constellations of small-sats delivering a new class of data on our planet with high spatial resolution and increased temporal frequency.
These new global data sets derived from space lead to a far more comprehensive picture of our planet, thereby enabling the monitoring of SD progress. In this context, this talk will briefly present some elements of the ESA EO programs and missions, and their evolution, highlighting their related scientific and societal applications, in particular regarding how space can help in supporting SD.
#ESA #EarthObservation #space #sustainability #development
Speaker: Stefano T. Chiadò — Vastalla
Artificial Intelligence is everywhere. Almost every startup doing financing rounds claims to be built on strong AI foundations.
But is it really so? Do we all agree on what Artificial Intelligence is? What is the state of the art in AI? What is its supposed long term evolution?
During the talk we will explore if you are a Padawan, a Knight or a Master in practicing AI. We will also brainstorm about the importance to gain the level required for the desired results. We will talk about efficiency and the importance of not reinventing the wheel making use of what is available on the market (i.e. AWS tools, Google code, etc.)
Prerequisites: common sense and open mind.
#AI #ML #brainstorming #AWS
Teacher: Fabio Franchino — TODO
Immersive lecture on the key elements and concepts behind data visualization.
Teacher: Stefania Delprete — TOP-IX
Interactive lessons using Jupyter Notebooks on Python and its most used libraries for data science: NumPy, Pandas, Matplotlib, and an initial Scikit-learn exposure. Plus you’ll get clear on what’s inside the Anaconda and SciPy ecosystems.
This session will include insights of the history and future of the open source libraries, how to contribute and participate to the community events.
Prerequisites: Exposure to Python and Jupyter Notebooks.
#datascience #python #numpy #pandas #matplotlib #seaborn
Teachers: Cristiano Nattero, Paolo Campanella — WASDI
Satellite Earth-Observation images are large files and complex objects: downloading them requires a delay, analyzing them takes a toll on processing capabilities, and the upload of the results can also introduce an additional lag. Batch processing of several images can be a daunting task.
WASDI (Web Advanced Space Developer Interface) is an open source web application that eases the job, by letting the users concentrate on the conceptual analysis, rather than on the infrastructural issues. This advantage is obtained by moving the processors to the data, rather than the other way around, and exploiting the computational power of the same cloud where the images are stored.
The lecture will show – hands on! – how to get existing images, how to plan the acquisition of new images, how to manipulate and elaborate them using library functions, and how to deploy your own code to run your own custom processing on the cloud.}
Prerequisites: Chrome, Python for scripting
#EOdata #Earth-Observation #datascience #WASDI #ESA
Teacher: Maurizio Napolitano — Fondazione Bruno Kessler
The workshop is an introduction to the geospatial technologies and everything needed to create maps and analyze geographical data.
As data sources will be used several open data sources including OpenStreetMap.
Prerequisites: Python and a complete installation of QGIS 3.6+
#geospatial #map #opendata #osm #qgis #geopandas
Teacher: Nicolò Bidotti — AgileLab
Big Data analysis is a hot trend and one of its major roles is to give new value to enterprise data. However data and information lose value as they become old, so it is important in a lot of contexts to do near real-time analysis of incoming data flows. Apache Spark is a major actor in the big data scenario and with its Streaming module aims to solve the main challenges in real-time data processing at scale in distributed environments.
This session aims to show the potential of streaming data analysis and how to leverage on Apache Spark with Structured Streaming to extract value from it without taking care of common problems of streaming processing at scale already solved by Apache Spark.
#bigdata #dataengineering #dataframework #apachespark
Teacher: Francesco Tarasconi — CELI
NLP lecture on the relations between Artificial Intelligence, Machine Learning and the mission of understanding natural language. Unstructured data as a potential asset, but also as a great challenge. Success in reaching state-of-the-art performance does not automatically translate into success in real-world problems. Overview of distributional methods and word embedding. Recent trends and breakthroughs in language modeling, with a focus on its practical applications.
Case study: I-REACT – Improving Resilience to Emergencies through Advanced Cyber Technologies.
I-REACT is an innovation project funded by the European Commission. The proposed system targets public administration authorities, private companies, as well as citizens in order to provide increased resilience to natural disasters, effective and fast emergency response, increased awareness and citizen engagement.
CELI leads two specific tasks: “Linked Data and Semantic structure” and “Social Media Data Engine”.
#datascience #NLP #python #pytorch
Teacher: Alessandro Molina — AXANT
Nowadays more and more data is generated by companies and software products, especially in the IoT world records are saved with a throughput of thousands per second.
That requires solutions able to scale writes and perform real time cleanup and analysis of thousands of records per second and MongoDB is getting wildly used in those
environments in the role of what’s commonly named “speed layers” to perform fast analytics over the most recent data and adapt or cleanup incoming records.
This session aims to show how MongoDB can be used as a primary storage for your data, scaling it to thousand of records and thousand of writes per second while also acting as a real-time analysis and visualization channel thanks to change streams and as a flexible analytics tool thanks to the aggregation pipeline and MapReduce.
#mongodb #realtime #scaling #mapreduce
Machine Learning techniques are a fundamental tool for automated decision systems and recommenders that substitute or support experts in a high number of decisions and fields (e.g., ranging from automated resume screening to credit score systems to criminal justice support systems).
In such a context, an increasing number of scientific studies and journalistic investigations has shown that such data-driven decision systems may have discriminating behaviors and amplify inequalities in society. In this talk we provide an overview of the problem, and we present preliminary approaches for measuring and possibly achieving fairness in ML-driven decision systems.
Prerequisites: Knowledge of R or Python, basics of probability.
#fair ML #data bias #decision systems #algorithmic discrimination
Teachers: Andrè Panisson, Alan Perotti — ISI Foundation
This in-depth part of the course allows to build an appealing and diversified Machine Learning portfolio. It starts with a Machine Learning introduction and application with Scikit-learn, and continues with Neural Networks and backpropagation lectures where you’ll start exploring Computer Vision techniques on a dataset of images.
Deep Learning methods. You’ll be challenged to use TensorFlow and Keras on a image classification real cases (such as distracted drivers, healthcare or plant diseases). The workshop ends with lessons in Transfer Learning and one last project building your data set by scraping Google images and practicing everything you learned.
Prerequisites: Python, Pandas, Statistics, exposure to Machine Learning is welcome.
#machinelearning #deeplearning #neuralnetworks #scikitlearn #tensorflow
Teacher: Cristiano De Nobili – Harman-Samsung
This lecture is intended to be an advanced Deep Learning lecture on NLP.
In the first part, we touch some relevant concepts in NLP, such as word and character embedding. In addition, we review with the ‘eye of a physicist‘ a few Information Theory quantities which are fundamental in machine learning.
During the second part, we understand how to build a Seq2seq (encoder/decoder) algorithm and how to train it. This architecture is at the core of many state-of-the-art NLP applications, such as language translators.
Case study: we will build a spell checker which is able to correct spelling mistakes in a sentence.
This is a simple model which takes advantage of the attention mechanism. We will also devote some time to build our dataset. This exercise is also thought to be an opportunity to test and learn the latest versions of TensorFlow (1.13 and 2.0).
Prerequisites: Curiosity (a lot), Python (a bit), TensorFlow (a bit), and Deep Learning basics.
#deeplearning #NLP #tensorflow #python #machinelearning
Speaker: Marco Aldinucci — Coordinator of Turin’s competence center in HPC for Artificial Intelligence and National delegate (Italy) at the EuroHPC governing board
In this talk we’ll meet and explore the Turin’s High-Performance Centre for Artificial Intelligence.
The University of Turin and Polytechnic University of Turin have joined forces to create a federated competence centre on High-Performance Computing (HPC), Artificial Intelligence (AI) and Big Data Analytics (BDA). A centre capable to collaborate with entrepreneurs to boost their ability to innovate on data-driven technologies and applications.
The first goal of HPC4AI is to establish a large and modern laboratory to co-design with industries and SMEs research and technology transfer projects. HPC4AI has been co-funded by Regione Piemonte via EU POR-FESR 2014-2020 with 4.5M€ and will ready for service at beginning of 2019. (e
Stay tuned! More modules coming up soon…
During the four weeks of the BIG DIVE 8 the lessons and activities took place from Monday to Friday from 9:30am to 4:30pm. Additional time was reserved for special lectures, exercises and “homeworks”.
The last week of the course was dedicated to a final project.
Here’s a timeline of what happened:
||BIG DIVE 8 – Registrations opening|
|April 7||Early-bird expiring|
|June 2||BIG DIVE 8 – Registration closing|
|From June 17 to July 12||BIG DIVE 8|
|July 12||BIG DIVE 8 closing event|
The application process started with a self-evaluation of the prerequisites (mostly related to your programming skills and Maths background) needed to access and fully enjoy the course. The course is offered in English, that’s why a fluent English is included in the prerequisites. Optional skills were taken in consideration to create a balanced classroom. You can download here a preview of the questions and requirements of the official application form.
In the form you can tell us more about you, your previous experiences and why we should choose you. We strongly encourage you to make a short video to stand out among the other candidates!
When you finally press ‘send’ it comes the waiting part. After two weeks from registrations opening our team started screening the received applications. Candidates might be contacted by the organizers and asked to provide more information about skills or to attend an interview (in person or using a remote audio-video communication tool).
The selection process continued till the official registration closure in order to create progressively a class of maximum 20 Divers.
Applicants selected before the official end of registration were asked to pay a deposit (20% of the total due fee – according to the profile). In case of missing deposit (deadline is one week after the request) the candidate lose the priority in the selection queue.
In case a selected candidate renounces to participate, a new Diver is selected.
The deposit is not refunded in case of waiver communicated after May 19.
All the news about selection, exclusion and deposit request are communicated by email through the email address inserted in the application form.
We understand the need to organise your working schedule and Summer activities in advance, so we do our best to get back to you with the final response as soon as possible.
The price includes
The price does not include
The price differs based on the category you fit in:
|With proof of full-time student status at the time of application (Phd Students are also eligible for this profile)||With proof of full-time / part-time employment at a non-profit organization at the time of application||If you are not eligible for Student or non-profit profiles|
|Full Price (till June 2)
Full Price (till June 2)
Full Price (till June 2)
(*) if you are eligible for VAT deduction (generally speaking, if you have a VAT Number), above prices will not include VAT.
A pleasure not only for the mind…
The venue for the course is a recently renovated space in a beautiful historical building in the center of Turin close to the river.
Turin is one hour from the Alps and two hours from the seaside.
By train you can easily traveling among Milan, Florence, Rome, Venice to celebrate after (or during) the BIG DIVE 8 with your classmates!
A copy of your University ID card or any certificate proving you are a student at the time of application. You can send it at firstname.lastname@example.org after you filled up the form online.
Must-have requirements to be admitted are:
 Background in Descriptive Statistics (mean, median and mode, standard deviation,…).
 Python fundamentals (syntax, control flow, functions,…).
 Command line basic experience.
 Basics of database management.
 Git version control system (a personal account on GitHub is required).
 As BIG DIVE will be taught in English, proper conversation and writing skills are required.
Please write us to email@example.com for any clarifications.
You video is strongly suggested for selection purposes, the letters of recommendation are not mandatory.
We encourage you do add them to your application to get know you and your motivation better and better evaluate your profile.
The course is designed to be fully-attended in Turin, Italy.
During the first three weeks you’ll focus on studying real cases and tutorials, while in the final week you’ll be part of a group to develop the final data science dedicated to a project.
Yes, at the end of the course you’ll receive a certificate of attendance if you take part at more than 85% of the lesson and complete the final project with your group.