Future direction for german digital education strategy
A new direction for the German digital education strategy?
The German government has organized on the 22nd of February 2021 an online-event regarding a future strategy on digital education. The event was a mix of political intention statements and some advertisement of existing initiatives. What I found remarkable was that there is still the vision, that any overarching OER platform would solve any problems with education despite this has failed on the European level already many years ago. The video of the discourse is embedded below.
Visualizing the discourse
# We first install the Youtubecaption library install.packages('youtubecaption') # Now we import the captions from the video library(youtubecaption) library(quanteda) library(tidyverse) url <- "https://www.youtube.com/watch?v=WAlP39a5LUU" caption <- get_caption(url, language ="de") caption # Now we need to make a document-term-matrix from the content my_corpus <- corpus(caption$text) summary(my_corpus) # Extending the stopwordlist with a more extended one from Github ger_stopwords <- read_lines("https://raw.githubusercontent.com/stopwords-iso/stopwords-de/master/stopwords-de.txt") custom_stopwords <- setdiff(ger_stopwords, stopwords("german")) # Constructing the meine.dfm <- dfm(my_corpus, remove_numbers = TRUE, remove_punct = TRUE, remove = c(stopwords("german"), custom_stopwords)) meine.dfm.trim <- dfm_trim(meine.dfm, min_docfreq = 1, min_nchar = 3) # Visualizing the matrix textplot_wordcloud(meine.dfm.trim, min_size = 1, max_size = 2, max_words = 100)
This code takes the full transcript from youtube, builds a corpus, removes stopwords and visualizes the most frequent words in the resulting tagloud.