ERIC Number: EJ1444186
Record Type: Journal
Publication Date: 2024
Pages: 14
Abstractor: As Provided
ISBN: N/A
ISSN: ISSN-1040-0419
EISSN: EISSN-1532-6934
Testing Computational Assessment of Idea Novelty in Crowdsourcing
Kai Wang; Boxiang Dong; Junjie Ma
Creativity Research Journal, v36 n4 p573-586 2024
In crowdsourcing ideation websites, companies can easily collect large amount of ideas. Screening through such volume of ideas is very costly and challenging, necessitating automatic approaches. It would be particularly useful to automatically evaluate idea novelty since companies commonly seek novel ideas. Four computational approaches were tested, based on Latent Semantic Analysis (LSA), Latent Dirichlet Allocation (LDA), term frequency -- inverse document frequency (TF-IDF), and Global Vectors for Word Representation (GloVe), respectively. These approaches were used on three set of ideas and the computed idea novelty scores, along with crowd evaluation, were compared with human expert evaluation. The computational methods do not differ significantly with regard to correlation coefficients with expert ratings, even though TF-IDF-based measure achieved a correlation above 0.40 in two out of the three tasks. Crowd evaluation outperforms all the computational methods. Overall, our results show that the tested computational approaches do not match human judgment well enough to replace it.
Descriptors: Novelty (Stimulus Dimension), Creativity, Concept Formation, Creative Thinking, Information Retrieval, Data Processing, Computer Uses in Education, Expertise, Comparative Testing, Evaluation Methods, Test Reviews
Routledge. Available from: Taylor & Francis, Ltd. 530 Walnut Street Suite 850, Philadelphia, PA 19106. Tel: 800-354-1420; Tel: 215-625-8900; Fax: 215-207-0050; Web site: http://www.tandf.co.uk/journals
Publication Type: Journal Articles; Reports - Research
Education Level: N/A
Audience: N/A
Language: English
Sponsor: N/A
Authoring Institution: N/A
Grant or Contract Numbers: N/A