A View From Above

Analyse real satellite data like professionals

Brief Description

Students investigate how satellite images obtained at different wavelengths help to identify Earth surface features like vegetation and open water areas by using a specially designed software package, LEO Works. Students inspect and analyse real satellite data to produce colour images and maps of spectral indices and learn how to interpret them and their uses.

This resource is part of the educational kit "The Climate Box". You can read more about the kit in the presentation attached. Find all related resources selecting the category "Our fragile planet" and "secondary level".


Students will get an insight into how multispectral satellite images can be diagnostic in deciphering Earth surface features like vegetation and the degree thereof as well as open water areas. They will get a hands-on understanding of how real remote sensing satellite data are being analysed. This will be done via a specially designed educational software package (LEO Works) which permits close to professional treatment of up to date satellite data. Students will understand the importance of such data for the lives of billions of people around the Earth and maybe grow interest in working in this field. Finally, the students will produce images and maps that are needed for the analysis. In the end, the students will be confident analysing satellite data on their own.

Learning Objectives

  • Students will inspect and analyse real satellite data at a close to professional standard.
  • Students will combine datasets to produce colour images and maps of spectral indices.
  • Students will answer questions and identify different surface features, such as vegetation and open water, by interpreting the maps of spectral indices.
  • Students will answer questions to discuss the importance of satellite data when dealing with issues like disaster management and climate change.


The major part of this activity is analysing satellite images. The products created during this exercise are images generated by combining the images in a certain way. The success can be evaluated by comparing the maps and images with the ones provided with this material. In addition, students will answer questions that will show how well they understood the importance of satellite data for various aspects.


  • Worksheet for students (needs to contain background information and activity steps)
  • Computer (the software needed is independent of the operating system)
  • Software installed: LEO Works 4, download at: http://leoworks.terrasigna.com
  • For the extension for advanced students: Landsat satellite data files: Venice_Landsat_ETM_multispectral_Jan2002.tif Venice_Landsat_ETM_multispectral_Jul2002.tif

Background Information

Remote sensing

The term remote sensing indicates a measurement technique that probes and analyses the Earth from outer space. Besides classical in-situ methods like weather stations, field surveying or taking samples, satellite based measurements are becoming an increasingly important source of data. The advantage is the fast and complete coverage of large areas. However, satellite data are not always easy to interpret and need substantial treatment.

The most abundant remote sensing devices are weather satellites. By employing suitable sensors, they provide information about cloud coverage, temperature distributions, wind speed and directions, water levels and snow thickness. Keeping the evolving climate change in mind, those data play an increasingly important role in disaster management during draughts and floods, climate simulations, atmospheric gas content and vegetation monitoring. In addition, urban and landscape management benefit from satellite data.

Figure 1: Overview of Landsat remote sensing satellites of NASA (NASA, https://www.usgs.gov/media/images/landsat-program).

The first weather satellites were launched by NASA as early as 1960. In the beginning of the 1970s, NASA started their earth observation programme using Landsat satellites (Figure 5). In Europe, France was first using their SPOT satellite fleet. They were followed by the remaining European countries in the 1990s after the foundation of ESA, the European Space Agency.

The Copernicus Programme

Already since 1997, the USA and NASA have been building a large programme for exploring the Earth, labelled the Earth Observation System, which consists of a large number of different satellites. Starting in 1998, the European equivalent, the Global Monitoring for Environment and Security (GMES) is being developed. In 2012, the programme was renamed to Copernicus. Information products for six applications are being derived from the satellite data: ocean, land and atmosphere monitoring, emergency response, security and climate change. The data products are offered to everyone free of charge. They are supplied via two branches: space based remote sensing devices (satellite component) as well as airborne, ground and marine probing (in-situ component). The core of the satellite component is the fleet of Sentinel satellites that have been and are being built exclusively for the Copernicus projects. They are supplemented by other domestic and commercial partner missions. The first Sentinel satellite (Sentinel 1-A) was launched 2014. Sentinel-2A and 3-A followed 2015 and 2016, respectively.

Figure 2: Computer model of the Sentinel-2A satellite launched on 23 June 2015 (Credit: ESA/ATG medialab, http://www.esa.int/spaceinimages/Images/2014/07/Sentinel-2_brings_land_into_focus).

Electromagnetic Spectrum

The kind of radiation that the human eye can see and interpret is called light. However, the full range of electromagnetic radiation (the spectrum) is much bigger. The part that is invisible to us can be detected by special cameras, such as the ones put on astronomical telescopes and satellites. A good overview on the different kinds of radiation is provided in Figure 3.

Figure 3: The spectrum of electromagnetic radiation. The visible light is only a very small part inside the full range (Inductiveload, https://commons.wikimedia.org/wiki/File:EM_Spectrum_Properties_reflected.svg, “EM Spectrum Properties reflected“, cropped by Markus Nielbock, https://creativecommons.org/licenses/by-sa/3.0/legalcode).

Multi-spectral Imaging

One of the core purposes of earth observation and remote sensing is taking and analysing pictures. Similar to modern astronomy, taking images with different spectral filters is very diagnostic when identifying and analysing terrestrial surface features. For this kind of data acquisition, the cameras rely on the sunlight that illuminates the Earth’s surface. Hence, they receive the portion of the sunlight that is reflected by the various surface features. Compared to the incident sunlight, the reflected light is modified by brightness and spectral composition.

Table 1: Spectral bands of the MSI camera of the Sentinel-2A satellite (Source: Sentinel Online).

Band Central wavelength (µm) Bandwidth (µm) Spatial resolution (m)
1 0.443 0.020 60
2 0.490 0.065 10
3 0.560 0.035 10
4 0.665 0.030 10
5 0.705 0.015 20
6 0.740 0.015 20
7 0.783 0.020 20
8 0.842 0.115 10
8a 0.865 0.020 20
9 0.945 0.020 60
10 1.380 0.030 60
11 1.610 0.090 20
12 2.190 0.180 20

The spectral bands of the camera “Multi-Spectral Instrument (MSI)” of the Sentinel-2A satellite is given as an example in Table 1. For example, band 2 covers a wavelength range of 0.065 µm centred on a wavelength of 0.490 µm. The smallest feature that could be seen in this band would be 10 m across. Those bands cannot be chosen arbitrarily because of the wavelength dependent transparency of the Earth atmosphere (grey area in Figure 4). They are referred to as spectral windows. The main culprit for the wavelength ranges, where the atmosphere blocks external radiation, is water vapour. Therefore, observations with cameras have to be designed in a way that only those wave bands are used, where the radiation is transmitted well enough to receive a good signal. Thus, these ranges are the ones the optical filters of the cameras are designed for.

Figure 4: Graphical representation of the spectral bands of MSI/Sentinel-2A compared to the cameras of the Landsat 7 and 8 satellites. The axes depict the wavelength in nanometres (1 nm = 10-3 µm = 10-9 m) and the terrestrial atmospheric transmission (grey) in percent (Credit: NASA, https://landsat.gsfc.nasa.gov/sentinel-2a-launches-our-compliments-our-complements/).

A proper choice of optical filters not only permits distinguishing between water and landscape, but also allows deciphering the state of vegetation or surface conditions. For instance, it indicates a noticeable difference between the reflective spectra of fresh and dry grass. The main reason for this is the absorption power of chlorophyll. In particular, the transition between the red (band 4) and the infrared ranges (bands 7 to 9) sees a sudden jump in the spectrum of fresh, green grass, while the spectrum of dry grass remains rather constant. When subtracting the signals of the bands, one can distinguish between the two states.

Figure 5: Reflective spectra of fresh (green curve) and dry (brown curve) grass in a wavelength range covered by the MSI/Sentinel-2 bands (yellow curves). There is a strong jump in the green grass spectrum between band 4 and band 7 (Credit: USGS Spectral Viewer, NASA).

Satellite images contain pixel values that represent the brightness or intensity of the reflected light in a given optical band. They are usually displayed in greyscale. Combining those images according to the rules of additive mixture of colour stimuli leads to the construction of coloured images. When selecting the images of the spectral bands representing the colours red, green, and blue, the resulting RGB image displays the colours in a realistic way (Figure 6, left).

Figure 6: Images obtained with MSI/Sentinel-2A. Left: Realistic RGB coloured image of the city of Milan; right: false colour visualisation of the area around the river Po, Italy. The colour red represents the near infrared band which is sensitive for green vegetation (Source: Copernicus data 2015/ESA).

Spectral index

By merging data from different optical bands, much can be learnt about vegetation or construction areas in a qualitative way (Figure 6, right). If quantitative information is required, a more detailed analysis is needed. An established tool is a spectral index. This is a number that is calculated from data obtained at different wavelengths and allows comparing the relative brightness of different wavelengths of light that is reflected by the Earth’s surface.

Normalised Differenced Vegetation Index (NDVI)

An important spectral index used for identifying healthy vegetation is the Normalised Differenced Vegetation Index (NDVI). It is calculated from the measured intensities obtained in the red (R) and near infrared (NIR) spectral regimes. As mentioned, the transition between those bands is diagnostic in distinguishing between green vegetation from other features (Figure 5). It is calculated as follows.

R: Intensity/brightness of reflected light in the red filter (ca. 0.6 – 0.7 µm)
NIR: Intensity/brightness of reflected light in the near infrared filter (ca. 0.8 – 0.9 µm)

Figure 7: NDVI world map of November 2007 based on data of the "Resolution Imaging Spectroradiometer (MODIS)" of the NASA Terra satellite (Credit: NASA).

They are provided by the bands 4 and 8 of the Sentinel-2 MSI camera (Table 1). The difference between NIR and R is normalised by their sum resulting in a range of values between -1 and +1. Negative values indicate water areas. A value between 0 and 0.2 represents nearly vegetation free surfaces, while a value close to +1 hints to a high coverage of green vegetation.

Normalised Differenced Moisture Index (NDMI)

Another spectral index is the Normalised Differenced Moisture Index (NDMI) or Normalised Differenced Water Index (NDWI). It is sensitive for humid vegetation and open wetland. It supplements the NDVI.

NIR: Intensity/brightness of reflected light in the near infrared filter (ca. 0.8 – 0.9 µm)
SWIR: Intensity/brightness of reflected light in the shortwave infrared filter (ca. 1.5 – 1.8 µm)

The NDMI helps distinguishing between dry and wet areas.

Modified Normalised Differenced Water Index (MNDWI)

The Modified Normalised Differenced Water Index (MNDWI) is regarded as an improvement of the NDMI. It helps identifying open wetland and excludes artificial buildings, vegetation and agricultural areas.

G: Intensity/brightness of reflected light in the green filter (ca. 0.5 – 0.6 µm)
SWIR: Intensity/brightness of reflected light in the shortwave infrared filter (ca. 1.5 – 1.8 µm)

Open wetland attains higher positive values than with the NDWI, while other landmarks like buildings, vegetation and crop land have negative values.

The software LEO Works 4

The European Space Agency (ESA) has developed an educational tool for teaching and learning the basic steps of analysing satellite data. The latest version 4 is being developed by Terrasigna in Romania. Since it is based on Java, it is independent of the operating system. It will be used for carrying out this activity.

Figure 8: Launch window of LEO Works 4.0, a software for treating and analysing satellite data for educational purposes. It can be downloaded at http://leoworks.terrasigna.com and runs on a wide variety of operating systems.


Gao, B.-C. (1996). A normalized difference water index for remote sensing of vegetation liquid water from space. Remote Sensing of Environment, 257-266

McFeeters, S. K. (1996). The use of the Normalized Difference Water Index (NDWI) in the delineation of open water features. International Journal of Remote Sensing, Vol. 17, Issue 7, 1425-1432

Rouse, J., Haas, R., Scheel, J., & Deering, D. (1974). Monitoring Vegetation Systems in the Great Plains with ERTS. 3rd Earth Resource Technology Satellite (ERTS) Symposium, (pp. 48-62)

Full Activity Description


Make printed or digital copies of the worksheet available to students. This contains the information in the background information which is needed to successfully analyse the data.

Install the LEO Works 4 software http://leoworks.terrasigna.com and make it available on the students' computers. It is required in order to perform this activity.


Introduce the topic by asking students what they know about Earth observing. How can we observe the Earth and what is remote sensing? What information can we collect by remote sensing and what are their applications? The most obvious answers should include weather satellites.

Ask the students, if they knew where the images in Google Maps or Earth come from. The source of the images is mentioned at the bottom of the screen. They might find names like SPOT or Landsat. Ask students to choose one of these satellite campaigns to research. Let them compile information on satellite launch dates, their orbits and countries of origin.

Hands-on activity

The activity is set up as a step-by-step instruction to analyse real satellite data. The exercise is interspersed with questions to evaluate the students’ understanding as well as to point to the relevance of the satellite data. Some tasks contain very similar and repeating The activity is set up as a step-by-step instruction to analyse real satellite data. The exercise is interspersed with questions to evaluate the students’ understanding as well as to point to the relevance of the satellite data. Some tasks contain very similar and repeating procedures that are used to reinforce the steps used in the analysis.

Analysis of satellite imagery data using LEO Works 4

This activity introduces basic tasks for processing and analysing remote sensing satellite data.

The installed version already contains some example data sets that can be used for exercise purposes. They are stored in the leoworks.data folder. When using MS Windows, it can be found in the user directory. From the existing data sets, the one labelled Venice will be used.

Reading the data

After launching, the software presents its workspace as shown in Figure 9. Open the file Venice_Landsat_ETM_multispectral.tif by clicking on the first icon in the menu bar or via the menu File -> Open -> Single File Dataset(s). A window appears from which the file is selected (Figure 10).

Figure 9: LEO Works 4 workspace. The menu bar contains procedures and tolls for displaying and analysing the data. There are three windows below that provide a list of the loaded data sets and image displays.

The file contains seven individual images obtained in seven bands of the camera “Enhanced Thematic Mapper Plus (ETM+)” of NASA’s Landsat 7 satellite (Table 2) covering the vicinity around the city of Venice in Italy. When the window Specify Subset appears, acknowledge by clicking OK.

Figure 10: Window for file selection.

The data automatically appear in the window to the upper left. The element Bands can be expanded by clicking on it to show the list of the seven images. They are labelled band_1 to band_7 and correspond to the spectral bands of Table 2.

Table 2: List of the seven spectral bands of the “Enhanced Thematic Mapper Plus (ETM+)” camera of the Landsat 7 satellite (Source: NASA, column with colours is not revealed to students).

Landsat 7 Wavelength (µm) Resolution (m) Colour
Band 1 0.45 - 0.52 30 Blue
Band 2 0.52 - 0.60 30 Green
Band 3 0.63 - 0.69 30 Red
Band 4 0.77 - 0.90 30 NIR
Band 5 1.55 - 1.75 30 SWIR
Band 6 10.40 - 12.50 60* (30) Thermal IR
Band 7 2.09 - 2.35 30 IR

*The data were obtained with a spatial resolution of 60 m and scaled to a 30 m resolution.

Action: Fill in the column labelled “Colour” of Table 2 for bands 1 to 5. Use the information provided with the introduction of the spectral indices.

Figure 11: List of loaded data.

Image display

A double-click on the band name issues a command that displays the image.

Action: Do this for band 1 first.

You will see an image of the city of Venice and its surroundings. It consists of different shades of grey, a greyscale display, that correspond to the brightness or intensity measured at a given spot (pixel) in the image. The contrast is a quite poor and should be adjusted using the tool Interactive Stretching.

Action: Find the corresponding button or menu item.

You can explore the meaning of the different buttons when moving the mouse pointer above them. After clicking, a new window appears as shown in Figure 12.

Figure 12: Windows for adjusting the contrast levels using Interactive Stretching. A window contains two graphs showing the distribution of pixel values in the image and the ones used for display, respectively. Adjustment is done by moving the flags. The setting is adopted by clicking Apply. Left: Distribution before adjustment; middle: after adopting the adjustment; right: the same shown in logarithmic scale, acquired by clicking the bottom left icon to the right.

The scaling of the contrast is accomplished by moving the flags. The window provides additional tools like displaying the data in a logarithmic scale.

Figure 13: Image of band 1 before (left) and after (right) adjusting contrast scaling.

Action: Display the seven images and adjust their scaling.

Creating a realistically coloured image

After having adjusted the contrast settings, a colour picture can be produced by superposing three images. A bad contrast will lead to shallow colours. For a realistic impression, the three bands representing blue, green, and red have to be selected.

Action: Find the corresponding bands in Table 2. If you need help assigning colours to wavelengths, research the missing information on the internet.

Select View -> New RGB View. A new window appears (Figure 14). Choose the matching bands for red, green, blue and click on OK.

Figure 14: Window for selecting the bands to be used for constructing an RGB image.

A new colour image appears. If necessary, you can adjust the colours with Interactive Stretching.

Actions: Inspect the result and try to identify landscape elements (buildings, water, soil, vegetation).

Find the airport.

Figure 15: Three-colour image (RGB) created from satellite data of Venice.

Creating a false colour image

You have just produced an RGB image that corresponds to the natural impression of colours how humans see it. It consists of the colours red, green, and blue. Imagine other species like bees or snakes. They can see other parts of the electromagnetic spectrum like the ultraviolet (UV) or the infrared (IR). We can simulate such kind of vision skills by combining different spectral bands than red, green and blue. The resulting colours do not match the natural ones we can see with our eyes, but they can help making interesting details visible.

Use the knowledge that the chlorophyll in green plants absorbs red light but reflects infrared radiation.

Actions: Produce a three-colour image from the near infrared (ca. 0.8 µm), red (ca. 0.65 µm), and green (ca. 0.5 µm).

What are the corresponding bands?

Put the infrared band in the red channel, the red band in the green channel and the green band in the blue channel of the RGB image.

Compare this image with Figure 14. Where do you find green vegetation?

Can you distinguish between green crops and green water (algae)?

What does uncultivated land look like?

Figure 16: False colour image produced by combining the green, red and infrared bands.

Analysis via NDVI

You have already seen in the information section that the NDVI is a colour or spectral index

that is particularly sensitive to green vegetation. The index provides a number that objectively reflects the degree of vegetation. Remember that there is a jump in the spectrum of green vegetation between the red (R) and the infrared (NIR) range (Figure 5). You will now construct a map that contains the NDVI for every image pixel. LEO Works provides a tool for this.

Action: Find the NDVI tool.

After activating that tool, a new window pops up. You select the dataset at the top. The next line contains the name of the image to be constructed and how it appears in the list of data. A name is already suggested. Select the suitable bands in the following rows below.

Action: What are the bands to be selected here? The answer can be found in the section about the NDVI and Table 2.

The formula is shown below. In the beginning, the variables show "null" as long as no band is selected. It is automatically updated as soon as you select the band corresponding to the NIR and the R bands. The NDVI map is created by clicking OK. A suitable false colour representation is chosen automatically, which helps identify green vegetation. However, the scaling of the colour table must be adjusted.

Figure 17: Window of the NDVI computing tool.

The tool Color manipulation is used for this. Move the flag of the maximum value to the upper end of the distribution. Then move the flag of the minimum value until the first green coloured flag reaches a value of 0.2 (Figure 17). The new setting is adopted after clicking Apply.

Figure 18: Window that allows adjusting the colour table.

The result should look similar to Figure 19. You see large white zones with alternating yellow and green areas in between.

Action: Compare the NDVI map with the previously produced images. What can you say about the degree of vegetation in the green and yellow areas?

Would you be able to detect a seasonal change, if the images were taken at a monthly rate?

What would be the situation during a draught?

Figure 19: Map of the NDVI in the vicinity of Venice, based on Landsat 7 satellite data.

Analysis via MNDWI

You will now use the satellite data to identify open wetland with the MNDWI.

Especially small ponds and narrow rivers are not easily found on naturally coloured images. The MNDWI can theoretically be constructed using the NDVI tool. However, the correct assignment of the corresponding bands can be confusing. LEO Works provides a generic tool to do all kinds of mathematical operations. The procedure is called Band arithmetic.

Action: Find the tool in the tool bar or in the menu and open it.

Similar to the tool for calculating the NDVI, you first select the dataset and the name of the image to construct (Figure 20, left). Then click Edit expression … for opening a new window (Figure 20, right). This is where you enter the formula for calculating the spectral index.

Figure 20: Window for doing mathematical operations on the spectral band images.

Action: Find out what bands are needed to calculate the MNDWI.

From the formula of this index you see that you divide the difference of the intensities of the reflected light measured in two spectral bands by their sum. Be careful with placing operators and brackets according to the formula.

After confirming the formula, it also appears in the first window. The procedure is executed by clicking OK.

The resulting image presents the values of the index in greyscale. To improve the readability of the map, you can assign colours to certain values via the Color manipulation tool. A colour table is assigned by clicking on the symbol as shown in Figure 20.

Figure 21: Colours can be assigned to image values to improve the readability of the map.

Action: Select the file gradient_red_white_blue.cpd.

Adjust the flags such that the values are well covered and the central flag represents the value 0.

What is the colour coding of water?

Compare the MNDWI map with the previously produced images. Would you be able to find wetland also on the naturally coloured image?

Can you imagine situations for which the identification of water levels can be important or even life-saving?

What would the image look like, if the water level rises?

If you have time, produce a map of the NDMI. Compare it with the other results.

Figure 21: Map of the MNDWI in the vicinity of Venice based on Landsat 7 satellite data.

For advanced students

Two additional datasets are provided that show the same area in January and July 2002. The already analysed dataset is from August 2001.

**Action: Load the two additional datasets like the previous one.

Produce naturally coloured RGB images.

Produce images of the NDVI distributions.

Compare the results from the three datasets obtained at different dates during the year. Indicate, how the vegetation changes.

In light of the results, describe and explain the advantage of satellite remote sensing.**


Space Awareness curricula topics (EU and South Africa)

Our fragile planet, satellites, surface

Country Level Subject Exam Board Section
UK KS 3 Geography - Geographical skills and fieldwork
use Geographical Information Systems (GIS) to view, analyse and interpret places and data
UK GCSE (2016) Geography AQA Skills 3.4.5: Use of qualitative and quantitative data from both primary and secondary sources to obtain, illustrate, communicate, interpret, analyse and evaluate geographical information. Including:
- geo-spatial data presented in a geographical information system (GIS) framework
- satellite imagery

Maps in association with photographs:
- be able to compare maps
- photographs: use and interpret ground, aerial and satellite photographs
- describe human and physical landscapes (landforms, natural vegetation, land-use and settlement) and geographical phenomena from photographs.
UK GCSE (2016) Geography Edexcel Cartographic skills
describe and interpret geo-spatial data presented in a GIS framework framework (e.g. analysis of flood hazard using the interactive maps on the Environment Agency website)
UK GCSE (2016) Geography OCR A and B Geographical skills

1.6. Describe, interpret and analyse geo-spatial data presented in a GIS framework.

4.1. Deconstruct, interpret, analyse and evaluate visual images including photographs, cartoons, pictures and diagrams.
UK GCSE Geography WJEC A and B (2016) Cartographic skills

3.4 Describe and interpret geo-spatial data presented in a GIS framework
UK AS/ A level Geography AQA (2016) ICT skills

- Use of remotely sensed data (as described in Core skills)

3.5.1 Quantitative data: understanding of what makes data geographical and the geospatial technologies (e.g. GIS) that are used to collect, analyse and present geographical data
UK A level Geography Edexcel Geographical Skills: 2. Quantitative data

a) understand what makes data geographical and the geospatial technologies (e.g. GIS) that are used to collect, analyse and present geographical data

b) demonstrate an ability to collect and to use digital, geo-located data, and to understand a range of approaches to the use and analysis of such data

c) use, interpret and analyse geographical information including …, linear and logarithmic scales, …, satellite images, GIS.
UK A level Geography OCR Geographical skills: the use of technology, e.g. GIS, remote sensing, etc. as research tools.
UK A level Geography WJEC Geographical skills non-numerical:

5. Cartographical information for landscape system identification

7. Digital and geo-located data: geospatial technologies including aerial photographs, digital images, satellite images, geographic information systems (GIS), global positioning systems (GPS), databases.

Additional Information

Suitable image material from other areas can be downloaded via the ESA Eduspace image server at: http://www.esa.int/SPECIALS/Eduspace_EN/SEMLK0F1EHH_0.html

Another source of suitable satellite data: https://earthexplorer.usgs.gov


The students used the LEO Works software to inspect and analyse real satellite data at a close to professional standard. They combined datasets to produce colour images and maps of spectral indices and learnt how to interpret them. Students should understand the importance of satellite data when dealing with issues like disaster management and climate change.

This resource was developed by Markus Nielbock, Haus der Astronomie, Heidelberg, Germany. This resource is under peer-review, proof reading, and will be updated and improved in the coming year.

Curriculum topic
satellites, surface
Big idea of science
remote sensing, Earth observation, vegetation, climate, satellites, satellite imagery, Copernicus, Sentinel, Landsat, light spectrum, spectral index
Age range
14 - 19
Education level
Secondary, University, Informal
Group size
Supervised for safety
Low Cost
Small Indoor Setting (e.g. classroom)
Core skills
Asking questions, Developing and using models, Analysing and interpreting data, Using mathematics and computational thinking, Constructing explanations, Engaging in argument from evidence, Communicating information
Type of learning activity
Markus Nielbock, Haus der Astronomie
Related resources