Conference Information
L@S 2018: Annual ACM Conference on Learning at Scale
Submission Date:
Notification Date:
Conference Date:
London, UK
Viewed: 7050   Tracked: 0   Attend: 0

Call For Papers
Learning at Scale investigates large-scale, technology-mediated learning environments. Large-scale learning environments are incredibly diverse: massive open online courses, intelligent tutoring systems, open learning courseware, learning games, citizen science communities, collaborative programming communities, community tutorial systems, and the countless informal communities of learners are all examples of learning at scale. These systems either depend upon large numbers of learners, or they are enriched through use of data from previous use by many learners. They share a common purpose--to increase human potential--and a common infrastructure of data and computation to enable learning at scale.

Investigations of learning at scale naturally bring together two different research communities. Since the purpose of these environments is the advancement of human learning, learning scientists are drawn to study established and emerging forms of knowledge production, transfer, modeling, and co-creation. Since large-scale learning environments depend upon complex infrastructures of data storage, transmission, computation, and interface, computer scientists are drawn to the field as powerful site for the development and application of advanced computational techniques. At its very best, the Learning at Scale community supports the interdisciplinary investigation of these important sites of learning and human development.

The ultimate aim of the Learning at Scale community is the enhancement of human learning. In emerging education technology genres (such as intelligent tutors in the 1980s or MOOCs circa 2012), researchers often use a variety of proxy measures for learning, including measures of participation, persistence, completion, satisfaction, and activity. In the early stages of investigating a technological genre, it is entirely appropriate to begin lines of research by investigating these proxy outcomes. As lines of research mature, however, it is important for the community of researchers to hold each other to increasingly high standards and expectations for directly investigating thoughtfully-constructed measures of learning. In the early days of research on MOOCs, for instance, many researchers documented correlations between measures of activity (videos watched, forums posted, clicks) and other measures of activity, and between measures of activity and outcome proxies including participation, persistence, and completion. As MOOC research matures, additional studies that document these kinds of correlations should give way to more direct measures of student learning and of evidence of instructional techniques, technological infrastructures, learning habits, and experimental interventions that improve learning. As a community, we believe that that the very best of our early papers define a foundation to build upon but are not an established standard to aspire to.

We encourage diverse topical submissions to our conference, and example topics include but are not limited to the following topics. In all topics, we encourage a particular focus on contexts and populations that have been historically not well served.

    Novel assessments of learning, drawing on computational techniques for automated, peer, or human-assisted assessment
    New methods for validating inferences about human learning from established measures, assessments, or proxies.
    Experimental interventions in large-scale learning environments that show evidence of improved learning outcomes
        Evidence of heterogenous treatment effects in large experiments that point the way towards potential personalized or adaptive interventions
        Domain independent interventions inspired by social psychology, behavioral economics, and related fields with the potential to benefit learners in diverse fields and disciplines
        Domain specific interventions inspired by discipline-based educational research that have the potential to advance teaching and learning of specific ideas, misconceptions, and theories within a field
    Methodological papers that address challenges emerging from the “replication crisis” and “new statistics” in the context of Learning at Scale research:
        Best practices in open science, including pre-planning and pre-registration
        Alternatives to conducting and reporting null hypothesis significance testing
        Best practices in the archiving and reuse of learner data in safe, ethical ways
        Advances in differential privacy and other methods that reconcile the opportunities of open science with the challenges of privacy protection
    Tools or techniques for personalization and adaptation, based on log data, user modeling, or choice.
    The blended use of large-scale learning environments in specific residential or small-scale learning communities, or the use of sub-groups or small communities within large-scale learning environments
    The application of insights from small-scale learning communities to large-scale learning environments
    Usability studies and effectiveness studies of design elements for students or instructors, including:
        Status indicators of student progress
        Status indicators of instructional effectiveness
        Tools and pedagogy to promote community, support learning, or increase retention in at-scale environments
    Log analysis of student behavior, e.g.:
        Assessing reasons for student outcome as determined by modifying tool design
        Modeling students based on responses to variations in tool design
        Evaluation strategies such as quiz or discussion forum design
        Instrumenting systems and data representation to capture relevant indicators of learning.
    New tools and techniques for learning at scale, including:
        Games for learning at scale
        Automated feedback tools (for essay writing, programming, etc)
        Automated grading tools
        Tools for interactive tutoring
        Tools for learner modeling
        Tools for representing learner models
        Interfaces for harnessing learning data at scale
        Innovations in platforms for supporting learning at scale
        Tools to support for capturing, managing learning data
        Tools and techniques for managing privacy of learning data
Last updated by Dou Sun in 2017-10-29
Related Conferences
CCFCOREQUALISShortFull NameSubmissionNotificationConference
ICEECTInternational Conference on Electronic Engineering and Computer Technology2017-07-19 2017-07-28
ba*a2KRInternational Conference on the Principles of Knowledge Representation and Reasoning2022-02-022022-04-152022-07-31
BMSBInternational Symposium on Broadband Multimedia Systems and Broadcasting2015-01-082015-03-022015-06-17
WITSInternational Conference on WIreless Technologies, embedded and intelligent Systems2018-12-152019-02-012019-04-03
IIRWIEEE International Integrated Reliability Workshop2016-07-25 2016-10-09
APCASEAsia-Pacific Conference on Computer Aided System Engineering2015-05-032015-05-252015-07-14
SmartWorldIEEE Smart World Congress2022-09-012022-10-012022-12-16
ICTC'Information Communication Technologies Conference2022-03-052022-03-302022-05-06
CCWCIEEE Annual Computing and Communication Workshop and Conference2023-11-092023-11-302024-01-08
PEPSCPower Electronics and Power System Conference2023-10-252023-11-102023-11-24
Related Journals
CCFFull NameImpact FactorPublisherISSN
IEEE Transactions on Learning Technologies2.315IEEE1939-1382
aJournal of Machine Learning Research Microtome Publishing1532-4435
Networking Science Springer2076-0310
IEEE Transactions on Technology and SocietyIEEE2637-6415
Language Learning & Technology2.571University of Hawaii Press1094-3501
International Journal of Mobile Learning and OrganisationInderscience1746-725X
bMachine Learning2.940Springer0885-6125
cIEEE Transactions on Big DataIEEE2332-7790
Human-centric Computing and Information Sciences Springer2192-1962