Within our organisation we use Evolve alongside several other authoring tools. We would like to use it as our primary tool, but have to rely on using other tools when we have customer requirements around diving into learner response data.
For the xAPI publishing method, the xAPI statements that are produced by Evolve provide some basic bits of information about the question, but does not provide the minimum required amount of information of what the learner has actually answered. With a little more information added in to the statements, it would unlock a great deal of value.
It is my understanding that the Exceed publishing method can provide this information if the content is uploaded to Intellum, but as our customers all have their own LMS systems already in place this is not an option.
Are there any active plans in the roadmap for enhancing the xAPI data to include more information around the learner’s responses?
Another enhancement we would like on xAPI score settings and publishing would the possibility to NOT submit a score. Evolve can do this with SCORM, but not with xAPI.
This is causing issues where we have LMS-based scored evaluations, and not within the xAPI module itself. In this case since the xAPI module still reports a score of zero, it skewing the user score reports when it averages it with the actual result.
Hi, I posted this is another thread asking about interaction tracking, bit it’s relevant here also:
“We are considering working on reporting from a more holistic standpoint (rather than targeting one or two particular standards), but it would be a significant undertaking in development time, so we don’t have a definite timescale for this work yet. When we do the work, we would undertake research first to find out more about the kind of problems we would be trying to solve.”
did anyone try testing Evolve xAPI courses in ScormCloud?
When I checked my Evolve xAPI course in ScormCloud, I wasn’t able to get a started-but-not-finished statement. I also used some variables and logic to set a course score throughout the course, but it looks like ScormCloud didn’t receive those either.
Testing the same course in our LMS produced the same results, so I suspect that Evolve’s xAPI courses don’t send those statements.
Can someone confirm?
We really need to know if learners got started on a course - maybe someone has a workaround for this?