developing a collaborative mooc learning environment utilizing video sharing with discussion summarization as added-value
abstract
with the fast-growing massive open online courses (mooc) community and the increase in the number of learning management systems (lmss) available online, the amount of shared information is massive. current lms - in particular mooc providers - offer many advanced content delivery techniques: interactive video, active retrieval practices, and quizzes to enhance the pedagogical process. the main knowledge creation assets within moocs are encapsulated in other tools such as discussion forums, blogs, and wikis. although these tools exist as separate entities within the platform, they still follow traditional techniques. we believe these tools need to be fully integrated to the main content and encourage spontaneous collaboration. from my experience with some moocs, the amount of collaboration and information-sharing is still overwhelming due to the massive number of participants and the limited range of collaborative tools. however, most of the shared information could be redundant or irrelevant. this information must be processed in order to provide the most concise knowledge. therefore, we need to summarize this information from the discussions, blogs, and wikis and include the most relevant data in the course content. this thesis addresses this shortcoming by suggesting a new system with two primary components to accomplish this task. in the first component, we link the discussion tools to the main course content. then, in the second component, we apply natural language processing (nlp) techniques to present a summary of all shared content. we use techniques such as term frequency - inverse document frequency (tf-idf), stemming algorithm, vector space model (vsm), and cosine similarity to rank the sentences. we then tune the tf-idf values and boost the sentence ranks using the main content by delegating the first component’s features. the next step involves choosing the most relevant sentence to build our summary. finally, we evaluate our result using recall-oriented understudy for gisting evaluation (rouge) system, which compares our automated summary to human extracted summaries. these results demonstrate that we can achieve high improvement summar y compared to the baseline and other similar techniques.