Digital technologies / Year 7 and 8 / Digital Technologies Processes and Production Skills

Curriculum content descriptions

Define and decompose real-world problems taking into account functional requirements and economic, environmental, social, technical and usability constraints (ACTDIP027)

Elaborations
  • determining the factors that influence proposed solution ideas, for example user age affects the language used for instructions, dexterity affects the size of buttons and links, hearing or vision loss influence captioned or audio-described multimedia as alternative ways that common information is presented on a website
  • investigating types of environmental constraints of solutions, for example reducing energy consumption and on-screen output of solutions
  • identifying that problems can be decomposed into sub elements, for example creating a decision tree to represent the breakdown and relationships of sub elements to the main problem or identifying the elements of game design such as characters, movements, collisions and scoring
  • starting from a simplified system, gradually increase complexity until a model of a real-world system is developed, and record the difficulties associated with each stage of implementation
General capabilities
  • Literacy Literacy
  • Critical and creative thinking Critical and creative thinking
  • ICT capability Information and Communication Technology (ICT) capability
  • Ethical understanding Ethical understanding
Cross-curriculum priorities
ScOT terms

Information and communication technologies,  Problem solving,  Functionality,  Usability

Online

Computational thinking

Find out about Computational thinking. Use this topic from the Digital Technologies Hub to learn more, get ideas about how to teach about it, find out what other schools are doing and use the applications and games in the classroom.

Online

Design thinking

Find out about Design thinking. Use this topic from the Digital Technologies Hub to learn more, get ideas about how to teach about it, find out what other schools are doing and use the applications and games in the classroom.

Online

Data bias in AI

Artificial intelligence can sometimes be biased to certain shapes or colours. When such AI systems are applied to situations that involve people, then this bias can manifest itself as bias against skin colour or gender. This lesson explores bias in AI, where it comes from and what can be done to prevent it.

Online

Project Quantum: Online assessment system

Project Quantum helps computing teachers check their students’ understanding, and support their progress, by providing free access to an online multiple-choice assessment system and question bank. To use Project Quantum, you will need to create a free account.

Online

Anti-bullying AI

Sometimes we write and post things on social media in a hurry. Such posts can hurt people and even make them feel bullied. Wouldn't it be great if an Artificial Intelligence application could check our posts as we write them, and warn us if they were potentially hurtful?

Online

Home automation with AI

Home automation is all the rage. You talk to your mobile phone to control the lights, the fan, the air conditioner, or your pool pump. But how does it work? In this lesson, we explore the AI that could power a home automation system.

Online

Recognising AI

Use the tasks in this lesson to introduce concepts that underpin artificial intelligence (AI). The majority of the tasks are unplugged (do not require a digital device). Use the downloadable AI cards with your students to explore what they know about AI.

Online

Scratch Creative Computing Guide

There is also a series of units comprising learning activities, paired with assessment activities and templates that can be used to support use of the Scratch (MIT) platform. The Scratch Creative Computing Guide supports assessment activities with visual programming environments.

Online

AI image recognition - exploring limitations and bias

A hands-on activity to practise training and testing an artificial intelligence (AI) model, using cartoon faces, including a discussion about sources of potential algorithmic bias and how to respond to these sources.