\noindent From the figure, it is obvious that the data is split into three major clusters until the last few connections of the dendrogram.
\noindent From the figure, it is obvious that the data is split into three major clusters until the last few connections of the dendrogram. To understand these four major groups and get an idea of whether they are somewhat divided after tempo (a rigorous test of this is done in 1.3), examples from each group and the mean tempo group (1 being low tempo, 4 being high tempo) of the songs is shown.
\begin{itemize}
\item Green group: 17 songs. Mean tempo group: \(2.7\)\\
Examples: \textit{Pyramids, Night on Bald Mountain, Zombie, Who is He and What Is He to You}
\item Red group: 39 songs. Mean tempo group: \(2.6\)\\
Examples: \textit{Mozart: Requiem in D Minor, I'll Let You Know, Romance, For Wee Folks}
\item Turquoise group: 161 songs. Mean tempo group: \(3.2\)\\
Examples: \textit{Mask Off, Xanny Family, I've Seen Footage, Bouncin}
\item Purple group: 1800 songs. Mean tempo group: \(3.3\)\\
Examples: \textit{Redbone, Master of None, Parallel Lines, Sneakin'}
\end{itemize}
From this superficial comparison, it seems that there might be a slight connection between the last two groups and songs with higher tempo. From the song names, it is hard to see genre seperations as we don't have access to a genre attribute in the data set. It might seem that music close to a rock genre is prevalent in the first group while the second group could be linked to instrumental, slower classical music. The third group can be linked to rap with multiple of the above songs from the artist \textit{Future} while the fourth could be seen as pop music or just music which is not contained in the other groups.
\subsection{GMM and Component Estimation}
\textit{Cluster the data by the Gaussian Mixture Model and find the number of clusters by cross-validation. Interpret the cluster centers.}\\
...
...
@@ -45,6 +57,16 @@ In this section, a clustering is performed using the Gaussian Mixture Model and
\begin{figure}[H]
\centering
\includegraphics[width = 0.6\linewidth]{gmmloss}
\end{figure}\noindent
\noindent From the illustration, it can be concluded that the most appropiate number of components is 9 where \(-\log\mathcal L =12,178\). Meanwhile, the negative log likelihood increases when the number of clusters is either increased (rises slightly) or decreased from 9, which can be understood as respectively over- and underfitting.
@@ -54,15 +76,24 @@ In this section, a clustering is performed using the Gaussian Mixture Model and
\noindent From the illustration, it is obvious that the most appropiate number of cluster i 9. Meanwhile, the negative log likelihood increases when the number of clusters is either increased or decreased from 9, which is to be expected.
\subsection{Evaluation of GMM and Hierarchical Clustering}
\textit{Evaluate the quality of the clusterings using GMM label information and for hierarchical clustering with the same number of clusters as in the GMM.}
Hierachical:
Rand: 0.5328153158470461
Jaccard: 0.19233601406747694
NMI: 0.022295065087228606
GMM:
Rand: 0.5231302775613633
Jaccard: 0.12366747954105438
NMI: 0.018119058189570287
%\textit{Evaluate the quality of the clusterings using GMM label information and for hierarchical clustering with the same number of clusters as in the GMM.}
To evaluate if the cluterings are similar to the premade clusterings of the tempo-attribute, three different similarities measures are used. These are the following: Rand index, Jaccard and NMI. The Rand Index similarity will typically be very high if there are many clusters. This is intuitively due to the fact that there is a lot of pairs of observations in different clusters rather than in the same cluster. This results in a Rand Index similarity close to one. Therefore the Jaccard index is also used as a similarity measure which disregard the pairs of observation in different cluster. The third measure is the normalized mutual information which is similar to both Jaccard and Rand Index. This similarity has a more theoretical background from information theory. It is based on quantifying the amount of information one cluster(s) provides of the other cluster. The evaluation of the GMM and Hierachical Clustering are illustrated in the following table.