Thanks for the response! :D
Is it possible to share the MADDL source code as well?
Also, is there a way to get around class imbalance problem? My label distribution is skewed. When I run the algorithm, the more common class is always getting a higher score than the the less frequently appearing label. Are there some techniques that will work well here, like say, downsampling?
My graph is a bipartite one with vertex sets V1 and V2. Suppose there are two labels L1 and L2.
My task is to classify each vertex v1 in V1. Initially in my dataset was V1xV2 with V1 being the samples and V2 being the features. But given my domain, it makes sense to model it as a bipartite graph as well. Logistic Regression and Boosted trees give decent results (assuming only one label L.
L = 0 ==> L1=1 and L = 1 ==> L2=1)
However when I ran Junto, the predictions were as good as random. Its weird because going by the Logistic Regression results, the features clearly have predictive power. Can you think of any possible reason? Or suggest a line of thought for debugging?
I've read the paper so I have some understanding of how the algorithm works. But I'm confounded wrt the results I'm getting. Not sure how to investigate. I have experimented with various values for the hyper parameters. (Not exhaustive since the paper suggests the also is not very sensitive to the hyper parameters)