Survey Results

Creativity Support Index (CSI)

CSI Items and Process

There are 6 orthogonal factors (exploration, expressiveness, enjoyment, immersion, collaboration, results worth effort (RWE)) with 2 questions each (total of 12 questions). When evaluating more than one system, questions are repeated for each system:

  1. I was satisfied with what I got out of the system or tool. (RWE)

  2. It was easy for me to explore many different ideas, options, designs, or outcomes, using this system or tool. (Exploration)

  3. The system or tool allowed other people to work with me easily (mark 0 if this question is not relevant for the task). (Collaboration)

  4. I would be happy to use this system or tool on a regular basis. (Enjoyment)

  5. I was able to be very expressive and creative while doing the activity inside this system or tool. (Expressiveness)

  6. My attention was fully tuned to the activity, and I forgot about the system or tool that I was using. (Immersion)

  7. I enjoyed using this system or tool. (Enjoyment)

  8. The system or tool was helpful in allowing me to track different ideas, outcomes, or possibilities. (Exploration)

  9. What I was able to produce was worth the effort I had to exert to produce it. (RWE)

  10. The system or tool allowed me to be very expressive. (Expressiveness)

  11. It was really easy to share ideas and designs with other people inside this system or tool (mark 0 if this question is not relevant for the task). (Collaboration)

  12. I became so absorbed in the activity that I forgot about the system or tool that I was using. (Immersion)

There are also 15 pairwise factor questions, which is used to rank the 6 factors. When evaluating more than one system, this set of questions is only asked once:

  1. When doing this task, it’s most important that I’m able to...

    1. Explore many different ideas, outcomes, or possibilities. (Exploration)

    2. Work with other people. (Collaboration)

  2. When doing this task, it’s most important that I’m able to...

    1. Be creative and expressive. (Expressivity)

    2. Produce results that are worth the effort I put in. (RWE)

  3. When doing this task, it’s most important that I’m able to...

    1. Enjoy using the system or tool. (Enjoyment)

    2. Become immersed in the activity. (Immersion)

  4. When doing this task, it’s most important that I’m able to...

    1. Become immersed in the activity. (Immersion)

    2. Produce results that are worth the effort I put in. (RWE)

  5. When doing this task, it’s most important that I’m able to...

    1. Work with other people. (Collaboration).

    2. Enjoy using the system or tool. (Enjoyment)

  6. When doing this task, it’s most important that I’m able to...

    1. Produce results that are worth the effort I put in. (RWE)

    2. Explore many different ideas, outcomes, or possibilities. (Exploration)

  7. When doing this task, it’s most important that I’m able to...

    1. Be creative and expressive. (Expressivity)

    2. Become immersed in the activity. (Immersion)

  8. When doing this task, it’s most important that I’m able to...

    1. Work with other people. (Collaboration)

    2. Produce results that are worth the effort I put in. (RWE)

  9. When doing this task, it’s most important that I’m able to...

    1. Be creative and expressive. (expressivity)

    2. Enjoy using the system or tool. (enjoyment)

  10. When doing this task, it’s most important that I’m able to...

    1. Explore many different ideas, outcomes, or possibilities. (Exploration)

    2. Become immersed in the activity. (Immersion)

  11. When doing this task, it’s most important that I’m able to...

    1. Work with other people. (Collaboration)

    2. Be creative and expressive. (Expressivity)

  12. When doing this task, it’s most important that I’m able to...

    1. Produce results that are worth the effort I put in. (RWE)

    2. Enjoy using the system or tool. (Enjoyment)

  13. When doing this task, it’s most important that I’m able to...

    1. Explore many different ideas, outcomes, or possibilities. (Exploration)

    2. Be creative and expressive. (Expressivity)

  14. When doing this task, it’s most important that I’m able to...

    1. Work with other people. (Collaboration)

    2. Become immersed in the activity. (Immersion)

  15. When doing this task, it’s most important that I’m able to...

    1. Explore many different ideas, outcomes, or possibilities. (Exploration)

    2. Enjoy using the system or tool. (Enjoyment)

The answers are formatted using a rating scale (Likert scale) that range from 0 to 10 (from Highly Disagree to Highly Agree).

For each participant/system, you end up with a CSI index (0-100) that indicates the level of creativity support afforded by the system.

CSI Calculation

An average is calculated first between the two sets of questions of the first block, ending with 6 average values one for each factor. Then, each average factor is multiplied by the weighted result for each factor based in the pairwise comparisons. The overall summation is divided into 1.5.

Results

Figure CSI1 indicates the overall results from the csi and SUS measures, showning that creativity was supported more than usability. There was a near significant correlation between csi and SUS across participants (Pearson’s r(8) = 0.62, p = 0.054). While both represent scores on a 0-100 scale, the two measures should not be considered orthogonal as an assessment of the interface. The specificity of the csi reflects weighted perceptions based on its factor structure, rather than the broader framing of the SUS. Addressing the overall average (unweighted) responses from users (see Figure [CSI2]), the two factors with the highest averages were Collaboration (M (SD) = 7.83 (1.68)) and Enjoyment (M (SD) = 8.21 (1.88)). In particular, the high average score for Enjoyment does not contribute towards the final csi rating for many users, as its weighted value tends towards the mid-lower range. Observing the differences across users in Figure [CSI2], Exploration and Expressiveness are the only two factors to feature in all users WAFS responses. This result indicates individual variation in user perception of their needs when working with sound and music, but also the importance of being able to work on the content at hand effectively; exploring ideas with suitable levels of sonic detail and immediacy of control. These results are coherent with feedback in the Thematic Analysis, where users discussed exploration, immersion, engagement, tool transparency and expressiveness.

Creativity Support Index (CSI) and System Usability Scale (SUS) results for Invoke study.

 

Creativity Support Index (CSI) results for Invoke study.

AttrakDiff

Participant Instruction

With the help of the word pairs please select the category you consider the most appropriate description for the system you just used.

Questions

All items were ordinal scales (1-7) with extremes labelled.

Pragmatic Quality (PQ)

technical/human

complicated/simple

impractical/practical

cumbersome/straightforward

unpredictable/predictable

confusing/clearly structured

unruly/manageable

Hedonic Quality Identification (HQI)

isolating/connective

unprofessional/professional

tacky/stylish

cheap/premium

alienating/integrating

separates me from people/brings me closer to people

unpresentable/presentable

Hedonic Quality Stimulation (HQS)

RWE

conventional/inventive

unimaginative/creative

cautious/bold

conservative/innovative

dull/captivating

undemanding/challenging

ordinary/novel

Attractiveness (ATT)

unpleasant/pleasant

ugly/attractive

disagreeable/likeable

rejecting/inviting

bad/good

repelling/appealing

discouraging/motivating

Results
Aesthetic Interaction Semantic Differential responses, grouped by factor; compared to the questionnaire presentation, some items are inverted to align positive attributes with higher scores.

Results indicate high factor values for Attractiveness (M (SD) = 5.48 (1.01)), the Hedonic Quality of Identity (M (SD) = 5.39 (1.18)), and the Hedonic Quality of Stimulation (M (SD) = 5.91 (0.88)), but only a moderately positive response to Pragmatic Quality (M (SD) = 4.01 (1.49)). The results indicate a predominantly positive attitude towards the Invoke system as a tool for spatial audio. In particular, the Hedonic Stimulation items show high mean scores for differentials such as ordinary/novel (M (SD) = 6.25 (0.75)), conservative/innovative ( M (SD) = 6.17 (0.58)), and dull/captivating (M (SD) = 5.83 (0.83)). Echoing the usability concerns discussed throughout analyses, it can be seen that the system suffers from issues related to confusion, complexity, practicality, and general usability. Despite these usability issues, the other more hedonic qualities indicate positive attitudes.

Collaborative Sonic Interaction Questionnaire Likert Responses

In addition to other survey data, a series of Likert scales were designed to gather data on a variety of factors. These items and results were not included for brevity. Adapting usability questionnaires, mutual engagement questionnaires, and group learning systems, the CoSIQ aims to gather broad feedback using open response and likert items, grouped into factors.

Unless specified otherwise, Likert ranges are: 0 N/A; 1 I do not agree at all; 2 Do not agree; 3 Neither agree or disagree; 4 Agree; 5 I fully agree. Each group contains a single open response item (in italics), to promote subjective impressions, these are listed on another page, and are dealt with in the Thematic Analysis section.

Exploration and co-construction of meaning

Questions
Results

Likert responses in this category showed a broad trend that users’ felt that exploration was a positive feature, with most items mean being placed in an agreement with statements (CQ04_04).

Exploration Likert responses.

Collaboration

Questions
Results

When looked at together all four items make a poor scale. But further reliability analysis indicated that two groups of Likert items had moderate to high agreement, grouping first two and last two Likert questions.

Collaboration Likert responses.

Satisfaction

Questions
Results

Likert responses showed a consensus that they enjoyed making the music composition (CQ03_02), and that it felt like a creative process (CQ03_03). Appreciation of the output of collaboration was ambiguous (CQ03_01).

Satisfaction Likert responses.

Flow

Questions
Results

The open-question for flow was a poor question, as it elicited non-explanatory responses. In general, items indicated a capacity for the system to support aspects of flow, with most users’ agreeing or strongly agreeing that they: enjoyed the situation (CQ06_03), lost track of time (CQ06_04), and were not bored (CQ06_06). An important trend was that most users did not feel distracted (CQ06_02), this could be said to detract from a state of flow .

Flow likert responses.

Social Presence

Questions

Variable code questions relate to nmmsp items in Appendix Table [app:measures:nmmsp].

Results

Within the Likert responses to social presence items, based on factors, the following trends were observed:

Social Presence Likert responses.

Sonic Interaction

Questions
Results

Three items within the likert responses showed positive agreement (CQ08_02, CQ08_03, CQ08_04, CQ08_07), suggesting that users’ perceived a level of understanding in sound action and engagement in process. More ambiguous responses were observed for the ability to plan and control sound actions with a level of detail (CQ08_05, CQ08_06). Also the level of embodiment of gesture and sound had ambiguous results (CQ08_01).

Sonic Interaction Likert responses.

System Usability Scale (SUS)

Questions

The answers are formatted using a rating scale (Likert scale) of five levels (Strongly disagree, Disagree, Neutral, Agree, Strongly agree) that range from 1 to 5.

SUS Calculation

To calculate the SUS score, first sum the score contributions from each item. Each item’s score contribution will range from 0 to 4. For items 1,3,5,7 and 9, the score contribution is the scale position minus 1. For items 2,4,6,8 and 10, the contribution is 5 minus the scale position. Multiply the sum of the scores by 2.5 to obtain the overall value of SUS. SUS scores have a range of 0 to 100, where the higher the value, the higher level of usability.

Results

Given that a standardized test was used for usability (SUS), the results are summarised in the general system score, fig. [fig:invoke:csisus], individual scores can be seen in [fig:invoke:cq07].

Usability likert responses.