A.I and the future of education
At ELSE we dedicate one day in ten to R&D and self-initiated projects with varying themes. The following piece is one of them. The only direction our workgroup was given was that it needed to be a speculative design exercise.
During the early days of this project, all I remember is that we talked a lot (and most of the times in circles). However what was undeniable, was the shared sense of excitement within the group for being given the opportunity to work on something completely undefined and carve it ourselves.
Before settling on a topic, we flip-flopped around many possible themes, yet there was always a recurring point of convergence in our conversations: How artificial intelligence will have a leading role to play in our future.
So having put a pin on A.I we needed to find an angle that would allow us to frame our exploration and thinking. We talked about the future state of work, politics, the environment… but education seemed to be the space that sparked something in all of us.
So we finally set out to explore: A.I and the future of education.
What’s more evident today than ever before is that learning is not a linear process – it can happen anytime, anywhere. Yet the traditional education model is typically rigid, calling for children to learn in a specific way, in a specific place and assess their knowledge at specific times.
Through this R&D project we take a look at how A.I. has the potential to radically change the landscape of education. From drivers and catalysts of change, to hyper-personalised curriculums, real-time holistic evaluations, and A.I teaching assistants – let’s step into the classroom of 2049.
N.B. As this is an ongoing piece of (unscientific) work there’s bound to be a number of non sequiturs and some things might make more sense than others. Nevertheless we hope you can see past those and that and take this for what it is… a fun piece of work.
A.I and the future of education
The road to 2049
Drivers & Catalysts
Having settled on a title, we needed to set the scene for our future scenario. Why have things changed? How did we get there? And finally what does that future look like?
And if you are wondering about the when, we chose 2049 because it simply felt distant enough for change to take effect but still close enough to create a plausible scenario.
Drivers of change
The western education model has come into a lot of scrutiny lately and there are several signs showing that it is a system reaching a tipping point.
We hear about schools having a recruitment crisis, about shrinking catchment areas and slipping school standards. We also hear about a widening educational gap between the advantaged and the disadvantaged – with data suggesting the latter being on average a year and a half behind their richer peers in the UK.
If we look at the situation globally, the inequality is even greater.
Abigail Cox in her article on EdSurge Independent says: “The educational system was built with a bias; bias in deciding where to direct funds, bias in the material we teach, and bias in where the school is located. It makes all the difference in the world if your neighbourhood is wealthy or poor. The bias is dependent upon “the have and have nots” mentality, and this bias determines student success.”
Limited access to good teachers, quality provisions and opportunities is what we considered being the first driver of change.
Schools have also been criticised for not preparing young people for the world of work effectively. A survey conducted in 2015 by the British Chambers of Commerce shows that most business leaders believe that young adults are ill prepared to get on the career ladder, lacking fundamental skills.
The main reason behind this, is the fact that the national curriculum has always been behind the curve of the demands of the job market and does very little to prepare young adults for tomorrow’s job world.
In desperate need of modernisation, an outdated curriculum was therefore what we considered being the second driver of change.
Finally, the current exam-based methodology of assessing comprehension and mastery of a certain body of knowledge can never be truly representative of one’s ability.
Some people just test well, while others don’t. One day you might get lucky in an exam – and get the topics you studied – and the other day not. But more importantly, exams and the qualifications associated with them, don’t provide the full picture. They are only representative of one’s performance at a given day — like fixed keyframes in a timeline.
So the third area we thought needed transformation was the current ineffective assessment method.
Catalysts of change
The next exercise we went through to help us paint the picture of a possible 2049, was to speculate how technology and mindsets – we thought were relevant to this field – would evolve in the next 20 to 30 years.
To help us do that, we first looked at how today’s technology was inspired by yesterday’s fiction and then we time-reversed the process to see how current nascent and even fictional tech could evolve, mature and eventually fall into mainstream in the future.
The technological roadmap (present-2049)
1. Artificial Intelligence
Since the conception of the term in 1950, it has taken A.I eight decades to have a noticeable impact in our daily lives. NLP voice-assistants, intelligent recommendation engines, and self-driving cars, were all once science-fiction.
Today the use of some of these applications has become so mainstream, some would say it’s even verging on the banal. But while current A.I applications provide certain utility, they don’t yet allow us to seamlessly mix our cognitive and analytical capabilities with that of machines.
Our vision for the future of education was for people to be able to use A.I as a means to enhance inner discourse and to do that, technology needs to develop much further.
Here is how the field of A.I could evolve:
In 2025, Google releases Live View – an A.I building block that recognises objects, places and actions in real-time video streams and camera feeds. Facebook follows soon after with their own proprietary version.
In 2028, Amazon acquires the Piero sports graphic system from Ross Video and two years later, it offers its capabilities to Prime Video subscribers for personalised use in their live sports streams.
In 2035 Amazon collaborates with Open A.I to develop alternative uses for Piero and by 2040 they co-launch the first multi-purpose real-time graphic system. It enables augmented reality developers to create applications allowing users to observe and analyse situations in physical environments.
In 2043, in an effort to accelerate the development of artificial general intelligence (A.G.I), some A.I companies join forces and agree to release free versions of their software to the open-source community for study and experimentation.
Two years later, a global team of academics comes a step closer to emulating A.G.I, by cobbling together the very first platform that allows A.I capabilities from different providers to be stacked together and work in unison — They call it Athena.
In order to maximise data access and accelerate machine learning scope, project Athena continued development with a focus on backwards compatibility with legacy devices.
By 2049, while not be most feature packed solution, Athena is the most accessible and widely used system owning 48.11% of the global market.
Our vision for the future of education was for people to be able to use A.I as a means to enhance inner discourse
2. Augmented, Virtual & Mixed Reality
During the two decades preceding 2050 we are seeing huge leaps in terms of technological development in the augmented, virtual and mixed reality space.
At the higher end, Google, Microsoft, Apple, HTC and Amazon Valve are still sharing the biggest portion of the pie but there are now tens of other independent companies that offer alternative and more accessible solutions in the hardware market.
The average entry-level price of hardware has lowered to the point where the technology becomes as ubiquitous as smartphones are today.
In the early 30’s, the proliferation of reality devices, has prompted the W3C to create a new ontology language extending their semantic web technology stack to the physical world — World Ontology Language (OWL)
The ultimate goal was to make real world objects more accessible to programs. This innovation becomes the catalyst for the birth of a new trend
Artists are the first to experiment with object-tagging but it doesn’t take long for adoption to explode. Businesses, marketers, academics and influencers enter the race and reality is rapidly becoming saturated with digital noise.
At the end of the decade, Google, Microsoft and Athena launch a new type of A.R browsers that allows users to filter digital content in the physical world based on their preferences, behaviour, affiliations, and social relationships.
Accessing augmented reality information becomes as easy as asking Google a question or Alexa to play a song.
1. Alternative education models
The limitations of the western educational model became increasingly pronounced as the world population grew steadily to reach 9.8 billion by 2049.
Over the three decades preceding the half century mark, increasing demand on formal classroom teaching reached a point where it was no longer serviceable.
As a consequence, alternative education models like home schooling saw a dramatic rise in popularity and new models arose like sponsored learning and Sit(uated)-Net(worked) learning – forms of A.I assisted education that were both officially accepted by governing bodies and employers alike.
2. The post-classroom era
With ever increasing classroom sizes and teacher-to-pupil ratios plummeting well below the recommended guidelines, teachers worked progressively harder while results declined.
As a result, private tuition demand saw a big rise in the late 20’s with the market reaching an estimated value of 3bn.
But the private tuition model had a scalability limitation for tutors that kept session prices high, which in turn translated into an affordability problem for households.
With prices ranging between £20 to £70 an hour for a single session, those who could afford private tuition, had to find the right compromise between price, quality and frequency.
To illustrate the point, let’s take two families from a similar socio-economic background: the Smiths and the Sergeants.
Both families have a need for their children to complement their education with private tutoring and have a fixed available budget of £70 a week. Ben Smith, needs some help with maths, while Alexia Sergeant needs help with both maths and physics. That means that while Ben’s parents can choose their preferred local maths tutor who charges £70 an hour, Alexia’s parents have to either compromise, or prioritise.
In the early 30’s, inspired by platforms like Masterclass and Peloton, Amazon introduced a new A.I-enabled service that promised to deliver the benefits of one-to-one tuition for learners, while ensuring maximum reach for teachers.
By using Alexa as the underlying technology, teachers could create their unique voice assistant profiles, trained to simulate their personality, tone, voice and teaching patterns.
That meant that if a user had a profile installed, s/he could always access personal teaching assistance without the need for a teacher to actually be online.
By removing the one learner per session limitation, the Amazon Education Network provided the first marketplace that enabled teachers to reach a global audience.
And with the scalability problem solved, private tutoring became accessible to households of all income brackets.
Historic interactions with personal A.I assistants could reveal much more complex aspects of one’s personality
3. The new quantified self
While A.I helped people to observe and analyse the outside world in new ways, it also provided an esoteric porthole that allowed them to better understand themselves and others.
In the same way that one’s browsing history can tell you a lot about a person, historic interactions with A.I assistants could reveal much more complex aspects of one’s personality — like sensitivities, biases and mindset.
But more importantly, by looking at how one acted as a consequence with their interactions with A.I, the technology itself could extrapolate and quantify previously intangible attributes like creativity, emotional intelligence, mental elasticity and critical thinking.
This was a game changer for employers as it allowed them to have more insightful qualification criteria for candidates besides academic achievements.
4. Hyper-personal learning
Last but not least, the most notable paradigm shift in education was the evolution of the definition, structure and delivery of the curriculum.
Learners were no longer restricted neither by a single delivery method nor by a designated linear process. So in order to learn, it was no longer needed to attend a specific classroom at a specific time taught by a specific teacher in a very specific way.
Personal A.I would adapt to the needs, stimuli, learning patterns and preferences of the learner and suggest optimised learning plans with dynamic lesson provisions that could be accessed at any teachable moment.
One could learn aspects of trigonometry while playing football or concepts of political science while watching Homeland. Every moment had the potential to be a teachable moment and every context could be turned into a situational learning playground.
The only thing you needed to be was curious.
This lead to an overhaul of the assessment system and instead of evaluating knowledge periodically at fixed moments in time (through exams), learners would be evaluated continuously in real-time through their queries, interactions and learning progress.
Structured courses didn’t disappear entirely though. But rather than being a set compulsory modules that one could either pass or fail, they were used as classes in the back-end to provide lesson provisions and pathway recommendations.
As a result, professional orientation instead of being associated loosely to degree types, would be algorithmically calculated by actual knowledge and skills acquired and mastered.
The future of education
A teacher’s perspective
Having speculated what a possible future landscape could look like, the next step was to take a stab at loosely illustrating the user experience for our two core user groups: Teachers & Learners
For teachers, we wanted to explore a synergistic relationship between technology and people, and more specifically look into the areas where A.I could augment human capability.
Having interviewed several teachers at different levels, we found out that while they love teaching, some aspects of the job were consuming a lot time and effort, resulting to a negative experience.
These pain points provided us with the key opportunity areas where A.I could step in and take the load of teachers and allow them to focus on doing what they loved: Teach.
There were two forms of planning that we looked into. One was planning curriculum and provisions for the day, term and academic year and the other was what is known as “In The Moment Planning’ (ITMP)— a method used specifically in the Early Years Foundation Stage.
In the moment planning is a very simple idea – observing and interacting with children as they pursue their own interests and also assessing and moving the learning on in that moment. The written account of some of these interactions becomes a learning journeyQuote from the Early Years Foundation Stage Forum
Theoretically, ITMP is a more effective teaching method because by constantly taking into account context and adapting the lesson to the current moment, it ensures that the imparted knowledge is directly applicable and coherent.
But the concept also has critics – the pragmatists.
This is what an Early years Foundation stage teacher and ITMP advocate told us in an interview: ‘It takes a lot of hard work and good knowledge of the children… many in EYFS don’t think it works because you need to be really on it.’
This is where we think personal A.I teaching assistants could step in to address the issue. Instead of relying on a teacher’s ability and capacity to observe and adapt lessons, personal A.I could do that for every learner in real-time.
Since personal A.I is aware of the user’s current activity, it can therefore adapt the lesson, choose the most effective provisions, look at outcomes and send data to the network which would in turn create connections between learners pursuing similar goals.
Observation is another important part of a teacher’s job, but it is also one that is prone to failures – especially when the group is beyond a certain size or when teaching happens remotely.
Due to the fact that a teacher’s attention is limited, teachable moments can be missed, and certain group members can be underserved.
But with a network constantly analysing learners’ data and relationships, teachers could potentially evaluate both group and individual situations more effectively and choose in which situations they need to intervene.
For example teachers could be asking the system questions such as:
- show me who has had the most interactions”
- show me the who has the strongest connections”
- show me which learners are isolated” – etc…
Of course these questions can also be answered today – whether it’s in a real classroom environment through observations and notes, or in a virtual teaching platform through analytics and video feedback. But current systems have long lags between brief, execution and feedback. In the future the difference will be in the time it takes to capture and process that information. Future technology would make that process instant and remote teaching more effective.
Creating reports is another time-consuming process that is in desperate need of modernisation. Teachers have to go through multiple records – analog and digital, synthesise and generate reports for different purposes.
Here is a list of some types of reports that teachers need to generate at a regular basis:
- Parent reports
- Special needs funding request for the local authority
- Vulnerable children case studies
- Results and performance reports
- School archives
- Discipline specific reports
- Moderation reports
- Baseline assessment reports
- Ofsted evidence data
What usually happens in the above scenarios, is that teachers will be looking at archived information, extract what they need, repurpose it to generate reports for different purposes and then distribute it to the relevant parties.
A smart archival system could turn this repetitive manual overhead obsolete by allowing an automated generation of reports. The system would handle query requests depending on permissions so the only responsibility for teachers would be to maintain data integrity.
Below is an illustration of how a distributed hybrid smart ledger could be used to store learner data in a private network while allowing different parties to access data permissible to them in a public network.
A snapshot from the future
So from a teacher’s perspective, how would an A.I-enabled virtual classroom could look like in 2049? Here is a short video we put together showing ATHENA’s core features:
Well, there you have it folks, that’s how far we got to in the short amount we had to bring this idea to life. Going forward, we would love to explore what a learner’s point-of view would be as well as the potential dangers of this technology as we know that there is always a dark side to technology that as designers we need to be mindful of. But that’s for another Friday.
To be continued…