Multiple Lecture Video Annotation and Conducting Quiz Using Random Tree Classification
|International Journal of Engineering Trends and Technology (IJETT)||
|© 2014 by IJETT Journal|
|Year of Publication : 2014|
|Authors : V.Anusha , J.Shereen
V.Anusha , J.Shereen. "Multiple Lecture Video Annotation And Conducting Quiz Using Random Tree Classification", International Journal of Engineering Trends and Technology(IJETT), V8(10),522-525 February 2014. ISSN:2231-5381. www.ijettjournal.org. published by seventh sense research group
In the modern days, the institutes of education are showing more interest in the field of E-Learning or in the field of internet based educational services. The growth of video lecture annotation tools also plays an essential role in the education environment. At first, the CLAS (Collaborative Lecture Annotation System) was limited to a single study-test episode .To deal with challenge we are proposing MLVA (Multiple Lecture Video Annotation) tool, developed to make the extraction of important information. The primary of concept MLVA is straight forward. While watching a video captured lecture, each student indicates key points in the lecture by a simple button press. Each button press represents a point-based semantic annotation and indicates, “For this user somewhat important happened at this point in the lecture.” The system relies on semantically constrained annotation, post-annotation data amalgamation and transparent display of this amalgamated data. As a future enhancement focus on conducting Quiz to the students and Instructor.
R. Farzan and P. Brusilovsky, “AnnotatEd: A Social Navigation and Annotation Service for Web-Based Educational Resources, “New Rev. in Hypermedia and Multimedia, vol. 14, pp. 3-32, 2008.
 M. Ketterl,J. Emden, and O. Vornberger, “Using Social Navigation for Multimedia Content Suggestion,” Proc. IEEE Fourth Int’lConf. Semantic Computing, 2010.
 B. Hosack, C. Miller, and D. Ernst, “VideoANT: Extending Video Beyond Content Delivery through Annotation,” Proc. World Conf.E-Learning in Corporate, Govt., Healthcare, and Higher Education,pp. 1654-1658, 2009
X. Mu, “Towards Effective Video Annotation: An Approach to Automatically Link Notes with Video Content,” Computers Education, vol. 55, pp. 1752-1763, 2010.
M.J. Weal, D. Michaelides, K.R. Page, D.C. De Roure, E. Monger, and M. Gobbi, “Semantic Annotation of Ubiquitous Learning Environments,” IEEE Trans. Learning Technologies, vol. 5, no. 2, pp. 143-156, Apr.-June 2012.
S. Chandra, “Experiences in Personal Lecture Video Capture, ”IEEE Trans. Learning Technologies, vol. 4, no. 3, pp. 261-274, July-Sept. 2011.
J. Steimle, O. Brdiczka, and M. Muhlhauser, “CoScribe: Integrating Paper and Digital Documents for Collaborative Knowledge Work,” IEEE Trans. Learning Technologies, vol. 2, no. 3, pp. 174-188,July-Sept. 2009.
C. Hermann and T. Ottmann, “Electures-Wiki - Towards Engaging Students to Actively Work with Lecture Recordings,” IEEETrans. Learning Technologies, vol. 4, no. 4, pp. 315-326, Oct.-Dec.2011.
A. Clark, Natural-Born Cyborgs: Why Minds and Technologies Are Made to Merge. Oxford Univ., 2003.