Now you say that you would like to know how the posts 'answered your original query.'
Before, I addressed data problems. Your query, with a question mark, was, What were you doing wrong. - Okay, please forgive my bluntness in what follows, but I figure I am addressing 1000 other readers, too, some of whom know much less than you know.
Here are some points, for what you did wrong, judging from what you posted.
Multivariable analysis with no rationale. Look at univariate EFFECTs. What do you hope to learn from the regression? "How many of these warning signs do you have?" is an alternate approach, simpler to apply in practice.
Stepwise procedure. These are often disparaged even when there is a rationale for doing steps. My impression is that you might have started out with a lot more than 15 variables offered, and after 15 entries/steps forward, 9 of them still met that 001 criterion.
Instead of stepwise, consider doing multiple analyses. Age and sex confound MOST disease prevalence rates: You have ample sample to explore by sex and age-decade, if you don't have any particular hypotheses in mind and want to data-dredge.
Being upset at achieving 70% classification, compared to the chance-result of 62.5%. Okay, that is not GREAT discrimination, but that's what you got. Maybe you can put CIs on how small some effects must be, but do that from looking at the range of OR outcomes
seen in (say) geographically-selected subsets.
Not looking at Odds ratios as measures of effect. In observational studies, an OR of 1.25 is barely suggestive and apt to be worthless, since ORs as large as 1.5 may be achieved by selection artifact, etc. (The classic error at 1.5 was estrogen treatment for
menopause, which looked good because the original survey sample, women taking estrogen, was loaded with women who lived longer because they paid a lot of attention to their health: retrospective conclusion after a large controlled study showed elevated mortality
from estrogen.) (Okay, I'm speaking from my reading, not from good personal experience. I would probably believe a 1.25 OR if the authors convinced me that the problem was simple enough and the analysis was smart enough.)
Here is more explanation of the Classification results.
The 3:1 ratio used for Predicted and Actual gives a random table that is (9,3; 3;1) in relative frequencies — which yields the 62.5% 'correct' that I cited, (9+1)/16 = 10/16 = 5/8 = 62.5%.
As it happens, moving '1' to those diagonal cells from the off-diagonals give exactly 75% accuracy, with (10,2; 2,2). That table has an Odds Ratio of 5:1. This table becomes significant with a total N of under 40: which defines a 'moderate effect size' in
Cohen's description of effects in social science research. That's a huge effect for most medical predictions, when seen for a single variable. Smoking/cancer has OR effects at 5:1 to 9:1, depending on other confounds. Smoking/heart disease is more like 2:1.
Second-hand smoke yielded ORs that were in that range of 1.25 to 1.50 which is questionable for observational studies, and thus led to controversy and the need for a LOT of confirmation. (Randomized-control studies can report smaller effects since they 'control'
for the confounds. Smaller than 1.25? - outside of what I remember reading.)
Your OR result at 70% (two different cutoffs) is a value like, say, 3.0 — My experience is NOT with huge surveys and data mining, so I do NOT know what to expect from sampling artifacts for your data, to be achieved at random as the result of 'best prediction'
from 15 (or more) variables.
Other data analysis notes:
You should have hypotheses. I'm sure your 15 (or whatever) potential predictors are NOT pre-judged as equally likely: Sex and age, for example: Consider an analysis of ONLY the ones suggested by experience and the literature. If you want to consider them
together, think about interactions (if you haven't already).
And when you look at your predictors that are quantities, think about, "Do these scores represent equal intervals for the criterion, that is, in increasing the likelihood of predicting prevalence?" Weight, for example, might look like a good distribution (or
its square root) without outliers: but morbidly obese and morbidly underweight are both unhealthy conditions.