Results

 » Scientific Events  » Project Reports  » Scientific Publications  » Downloads
 

 

Scientific Events

 


Workshop: 4th International Workshop on Teaching Analytics

As opposed to the focus of the workshop at LAK'16, this workshop emphasized the importance of tailoring learning analytics to the needs of teachers and the aspect of improving teaching by analytics. Details and the proceedings booklet are available here.

 

 


 

 

Workshop: Learning Analytics for Learners

In the context of LAK'17 in Edinburgh, UK, Lea's Box organized a workshop that aimed at emphasizing the opening and communicating of learnign analytcis results to the actual learners. Details and the proceedings booklet are available here.

 

 

Learning Analytics Summer Camp 2015

The Lea's Box project was organizing a Learning Analytics Summer Camp in Prague on 31 July 2015. The summer camp is based on an in-depth presentation of the projects and research directions and the search for a common ground and alignment of common strengths. This event is not so much a mere exchange of knowledge and solutions, but much more a joint effort to bring all that we have in the pockets to a concrete and broad application in the European educational landscape. Essentially, the major aim of the summer camp is to find joint solutions to the ‘impact’ we are expected to make. We believe that streamlining our expertise and already existing solutions can bring us a big leap forward. This is not trivial however, since it requires a team that is big enough to comprise major knowledge and a team that is small enough to be able to really accomplish our ambitious, but so important goal. Together we can form a team that is strong enough to accomplish this. Check out the workshop website with the results of the workshop!

 

 

Workshop: Learning Analytics for an in Serious Games

Joint workshop of the GALA Network of Excellence and the LEA’s BOX project at EC-TEL 2014, September 17, 2014, Graz, Austria. Topic is to present and discuss approaches to improve serious games by techniques of learning analytics and educational data mining. Details and the proceedings booklet are available here.

 

 

 

 

Project Reports

 


Final Project Report [Dec. 2016]

The final report summarizes all project activities and achievements over the entire project duration. This report is particularly interesting since it elaborates on how Lea’s Box advanced the state of the art in learning analytics. In this public version of the report, specific management-related and confidential pages have been removed.


 

 

Deliverables D1.3


Delivery Date M34 (Dec. 2016)
» PDF (1.5 MB)


Training and Dissemination Report 2 [Dec. 2016]

Deliverables 5.8 present the workshops and other activities conducted in the light of providing end users with training and information about the solutions of the project. Training and dissemination activities, being not isolated from all the other project´s activities, are closely connected to all other forms of interactions with end-users, be it piloting and evaluation, co designing, advisory, or exploitation. In the course of the project we undertook numerous activities in the partner countries Austria, the Czech Republic, Turkey, and the UK.


 

 

Deliverables D5.8– Training and Dissemination Report 2


Delivery Date M34 (Dec. 2016)
» PDF (1.5 MB)


Piloting and Evaluation Report 3 [Dec. 2016]

This report is the final collection of the results of evaluation studies and experiences from pilot studies within the project. The deliverable summarizes the results and insights and provides the evidence for the validity and suitability of the Lea’s Box solutions.
The main objective of piloting and evaluation activities was to use the Lea’s Box system and its components in real schools, to evaluate key aspects of the system and subsequently to modify the tools based on the feedback received. Earlier deliverables from work package 5 were very practically oriented, for example D5,4 presented different educational scenarios and classroom issues, while D5.5, described specific tools offer by the Lea’s Box system. Based on the feedback from received in year 1 and 2, the final release of the system was shaped and the integration of its individual components was considerably improved. This deliverable thus presents the main results of piloting and evaluation studies related to the whole Lea’s Box system.

All test instruments and questionnaires are available as zip archive (10 MB).


 

 

Deliverables D5.6 – Piloting and Evaluation Report 3

Delivery Date M34 (Dec. 2016)
» PDF (2 MB)

 


Final System Release [Sep. 2016]

In M31 of the project, i.e., September 2016, w released the final release of the system. This release is a fully integrated version of the main interface, processing, analytics, and visualization features, including the open learner modelling (OLM) system. The document bundle includes a technical description and manuals of the sub-components. Please note that the Lea’s Box system will be developed and advanced beyond the project and beyond the final deliverable in order to exploit the project results. This document bundle includes not only D2.6, the actual system release report, but also deliverable D3.4, third release of LA/EDM services and algorithms as well as deliverable D4.5, the final release of visualisation and OLM web services and tools.

This is the link to Lea's Portal. For access credentials please contact us!

 


 

 

Deliverables D2.6, D3.4, D4.5 – Final System and Tool resleases

Delivery Date M31 (Sep. 2016)
» ZIP (5.5 MB)

 

Dissemination and Exploitation Plan [August 2016]

LEA’s BOX contains a multitude of tools, enabled by a central executive to carry out multi-source, competency based learning analytics. Competency maps can be introduced using FCA and competency states can be calculated using CbKST, based on the evidences from received data. The tools use this capability to support a large set of pedagogical scenarios, including but not limited to formative assessment, self-evaluation, personalized course planning, and tracking. This report details a plan on the exploitation of the project outcomes. .

 


 

Deliverable D 6.5 – Dissem. and Exploitation Plan

Delivery Date M30 (Aug. 2016)
» PDF (1.5 MB)

 

Piloting and Evaluation Report 2 [March 2015]

Deliverable 5.5 presents the second stage of piloting of the LEA’s Box tools. Piloting is one of the most important parts of the LEA’ Box project: after all, no matter how successful research results may seem, it is always crucial to try out the tools and applications developed “in a lab” in the real world. This deliverable is a continuation of D5.3, which was very practically oriented, unlike deliverables D5.1, which focused on the piloting methodology, D5.2, which proposed activities for pilot studies, and D5.4, which described different scenarios and classroom issues. Deliverable D5.3 described several pilot activities and tools, such as the myClass tool or the mind mapping tool. In the meantime, the tools developed within the project were substantially improved, new tools were built on the existing ones and new use-cases were proposed and tested. These activities were preceded by a thorough evaluation of the feedback generated during the first phase of piloting. Therefore, this deliverable describes new extensions of the LEA’s Box tools and new use cases and explains what changes have been made and why, i.e. what steps we took in order to reflect teachers’ and students’ needs in the best possible way.

 


 

Deliverable D 5.5 – Piloting and Evaluation Report 2

Delivery Date M24 (Feb. 2016)
» PDF (3.5 MB)

 

 

Second System Release [December 2015]

In the very heart of WP2 is the design and development of a Learning Analytics web portal. Based on the requirements specification document, the focus groups and design studies, we designed, planned, and developed this platform and have a second, completely revsied release in December 2015.

The platform itself consists of various tools and underlying algorithms. The corresponding deliverables are listed in the box on the left. Along with the basic system we released also the second versopn of the CbKST/FCA-based Learning Analytics services as well as visualiztation and OLM tools.

This is the link to Lea's Portal. For access credentials please contact us!

 


 

Deliverable D 2.5– Second System Release
» PDF (1 MB)

Deliverable D 3.3– Second Release of LA/EDM
Services and Algorithms

» PDF (2 MB)

Deliverable D 4.3– Second Release of Visualisation
and OLM Services and Tools

» PDF (7 MB)

Educational Data Visualisation Approaches and Open Learner
Modelling [August 2015]

Visualizing partly very complex information is one of the key pillars of learning analytics. However, finding the rights techniques to make users (teachers, students, parents, ...) understand what they should understand is not an easy task. Lea's Box invests in a broad spectrum of research and development to identify the right solutions.

This report covers the research in visual analytics more generally, to situate the work on learning visualisations. We then consider learning analytics portals and dashboards, which most commonly visualise activity, performance or interaction data. We end by looking at open learner models, which typically show visualisations of competencies, understanding or skills, based on inferences about learning rather than the raw countable data. In between we consider other kinds of educational data visualisation, that are nether activity counts nor inferences about knowledge, but rather, ways of using students’ products to generate some information about their learning.

This could be viewed as intermediate between the other types of data that are visualised. Our main concerns in LEA’s Box are with learning analytics visualisations and open learner models. It will be seen that the former, mostly numerically-based data, more often tends to be presented as the familiar graphs, charts, plots, etc., while the latter, usually knowledge or skill-based information has a lower tendency to use these standard methods of conveying data, instead using a variety of simple and complex, sometimes structured, visualisations. This Deliverable provides an overview of some of the similarities and differences within and across types of learning visualisation, with a need for further evaluation of the relative utility of different visualisations during the remainder of the project.

 


 

 

Deliverable D 4.1 – Educational Data Visualisation Approaches and Open Learner Modelling

Delivery Date M18 (August 2015)
» PDF (2 MB)

 

 

Revised focus group and design Report [June 2015]

Piloting at schools is a vital part of the LEA’s Box project as it enables us to verify the functionality of the tools, their user-friendliness and comprehensibility. The previous deliverables focused on the description of the theoretical framework of piloting: D5.1 contained a plan of the piloting methodology, D5.4 focused on different scenarios at schools and on problems faced by current teachers, and D5.2 suggested particular steps and activities that were going to be done during the pilot studies. All these deliverables served as a background for planning the piloting activities, the outcomes of which will now be summarized.

The piloting included creating virtual classrooms, working with mind maps, observing students doing different activities and assessing their performance and behaviour. Feedback from teachers was gathered from online survey and focus groups held in the Czech Republic, Turkey and Austria. This report will contain some relevant information concerning the current situation at schools and a summary of teachers’ experience with LEA’s BOX.



 

Deliverable D5.3: Revised focus group and design Report

Delivery Date M16 (June 2015)
» PDF (2 MB)

 

 

Survey of Learning Styles and Cognitive Styles [April 2015]

The purpose of this document is to review the state of the art in the field of learning styles (LS) and to extract the “quintessence” of this wide-spread and diverse research area which offers a huge amount of models and theories. To extract this quintessence of the learning styles literature, existing models and theories which have been considered as important and representative for the whole LS research field have been identified, selected and described. The Formal Concept Analysis (FCA) has been applied formally describe and to cluster the learning styles suggested by the selected models and theories. The FCA took 70 learning styles as objects and 48 attributes (properties of learning resources, learning activities, etc.) into account.

Aiming for a non-invasive approach to measure a learner´s dominant learning style, a Competencebased Knowledge Space Theory (CbKST)- assessment procedure has been outlined. Afterwards, an excursus on the so-called “matching-hypothesis” is given. The matching-hypothesis suggests that the instructional style or the nature of the learning resources to be consumed (e.g. “concrete” or “visual”) should be aligned with the learner´s dominant learning style. Finally, we will conclude and outline future steps on research and development activities in LEA´s BOX related to learning styles. Based on the review provided by this deliverable, conceptual research on the LS approach will continue and will be translated into the implementation of a set of analytics and data mining services and their integration in the LEA’s BOX platform.

 


 

Deliverable D3.5: Survey of LearningStyles and Cognitive Styles

Delivery Date M14 (April 2015) Revised version from June 2016:
» PDF (800 kB)

 

 

First System Release [February 2015]

In the very heart of WP2 is the design and development of a Learning Analytics web portal. Based on the requirements specification document, the focus groups and design studies, we designed, planned, and developed such platform and have a first release at the end of year 1.

This web platform, of course, is a first initial step and will be filled with contents and not least users subsequently. The design and functionality mirrors all the experiences and demands of teachers that were involved in the studies. Basically, the key requirements are simplicity and individual design. Therefore, we do not offer a monolithic powerful tool, but a simple approach with many different tools that not necessarily haven even the same look and feel. Instead of a “Microsoft Word” we developed and released a number of “Text Editors” and individual “Apps”.

The platform itself consists of various tools and underlying algorithms. The corresponding deliverables are listed in the box on the left.

This is the link to Lea's Portal. For access credentials please contact us!


 

Deliverable D2.4: First System Release
» PDF (1 MB)

Doucmentation Bundle
» ZIP (1,5 MB)

Deliverable D3.2: First Release of LA/EDM Services and Algorithms
» PDF (1,5 MB)

Deliverable D4.2: First Release of Visualisation and OLM Services and Tools
» PDF (1,3 MB)


Learning Analytics and Data Mining require a solid basis of ethics and data protection considerations [December 2014]

A range of general guidelines, model codes, and principles for appropriate data and privacy protection exist that may serve the consideration of these topics in a learning analytics context. Approaches of defining frameworks for dealing with ethics and privacy issues specifically in the field of big data, in general, and learning analytics, in particular, adapt fundamental ethical principles of privacy and data protection to these analytics domains and adjusting it to the specific application field. Since ethics and data protection do not only need to be translated into an adequate general privacy policy and information practice, but also needs to be technically reflected by according system functionality, provisions, and data structures, the use of privacy by design principles provide a valuable starting point to ensure that data protection is embedded into the design and architecture of learning analytics.
       A general awareness of the importance and significance of data protection is also reflected in national and international laws and directives, where data protection is usually considered as a fundamental right. An overview on national regulations in Czech Republic, Turkey, Austria, and the United Kingdom, where LEA’s BOX is carrying out pilot and evaluation studies and collecting and processing personal data. In addition, the European data protection regulations are summarized, which mainly aim in establishing common rules for data protection in member states and keeping a balance between a high level of protection of individual privacy and the movement of personal data within the European Union.

 

A indepth overview of Learning Analytics methods and tools [October 2014]

In this document, a summary of the different categories and sources of data is given and types of indicators used for LA are presented. The analytics methodologies employed in a concrete LA project depend on the kind of data collected, as well as on the target stakeholder group(s) and their objectives. The most common methods that are used to extract meaningful patterns from educational data are presented in this document. An increasing number of tools exist that implement these methods and provide support in pre-processing, analysing, and visualising data. A systematic overview of different categories of tools and their main purposes and characteristics is given and examples of each type are provided. This is complemented by a summary of the most recent trends in LA technologies. Game-based learning and virtual worlds are emerging technologies that are acknowledged for its positive impact on learners. The application of LA in these research areas constitutes another trend in the field of LA and is summarised as an excursus section in this deliverable.
      Although much progress has been achieved in LA in the last years, there are still a number of great challenges to be addressed in future research. The existing research and practice gap is probably the most pressing one, which is related to a set of more specific challenges, including data integration from different sources and the implementation of meaningful and intuitive tools for teachers and learners. In addition, there is an urgent need for convincing empirical evidence on the positive impact and added value of LA for learning and teaching, to foster acceptance and adoption of LA technologies in educational practice.

 

LEA’s BOX concluded the first round of focus groups and design studies with teachers [August 2014]

Briefly summarized, teachers prefer using digital devices as a productive tool in their classrooms. Individual data from students comes not from learning applications but from classroom micromanagement led by teachers. Because of lack of accessible tools for formative assessment teachers are mostly focused on summative assessment. Teachers have a need for tools for learning analytics and formative assessment but they must be very easily manageable and time efficient. Based on experience it will be very hard to implement formative assessment tools to the classroom with typical frontal teaching strategy. More suitable for formative assessment approach is project-based learning where a teacher can work as a facilitator and a mentor. In this scenario the teacher has more time for micro-management and micro-assessment (in our case the teacher is gathering and creating data for follow-up analysis. The most preferred visualizations of student progress were Hasse diagrams and Radar charts.
       One thing however, is clear, as a large scale survey with over 2000 responses show: A wide majority of teachers acknowledge and appreciate the potential value of learning analytics tools to support a formatively inspired assessment and teaching. Only the tools must be available and suit the classroom reality. Systems such as TU Graz’ myClass or SEBIT’s Vitamin platform may provide clear advantages!
       Further useful information about the uptake of technology and its context conditions in Europe can be found on the European Commission Digital Agenda website; a similar report for the US is provided by PewResearch. For the detailed results of our studies, you may download the full document.

 


 

 

 

Deliverable D2.3: Privacy and Data Protection Policy

Delivery Date M10 (Dec. 2014)
» PDF (600 kB)

 

 

 

 

 

 

 

 

 

 

Deliverable D3.1: Review Article about LA and EDM Approaches
Delivery Date M8 (October 2014)
» PDF (3,7 MB)

 

 

 

 

 

 

 

 

 

Deliverable D5.2: User groups and design studies report

Delivery Date M6 (August 2014)
» PDF (1,5 MB)

 

 

 

 

   

 

    

leas-box@tugraz.at
+43 316 87330636
http://kti.tugraz.at/css

Attribution should be by the named
authors and  "LEA's  BOX  Project"
at URL: http://www.leas-box.eu

LEA's BOX is supported by the
European Commission  under
the 7th framework programme;
grant no. 619762

© 2014 TU Graz