There are countless personality tests online. Many are "nice enough," some are even entertaining — but only a few are truly scientifically robust. One of the best-known tests worldwide is even completely torn apart by researchers as "not better than a fortune cookie". Because when a test is used in coaching, personnel development, or even hiring decisions, "it feels right" is no longer enough. And we asked ourselves: why shouldn't we apply the same high standards to personal development?
For the Hartman Value Profile (HVP), which our personality test is also based on, there is extensive validity and reliability research. Over many years, scientists conducted these studies with several thousand participants.
The central question: does Robert S. Hartman's model hold up under scientific scrutiny? The short answer: yes — and remarkably well. The long answer tells an exciting story about what it even means for a test to be "scientifically validated."
What makes a test "scientific"?In technical terms, three big concepts matter:
- Validity: Does the test actually measure what it claims to measure?
- Reliability: Does it produce stable, repeatable results — or is it secretly rolling dice?
- Fairness / non-discrimination: Does the test disadvantage certain groups (for example by age, gender, or background)?
Exactly these three questions were investigated for the Hartman Value Profile in several blocks: how well does the test actually reflect Hartman's value theory? Is it fair in the sense of U.S. equal-opportunity guidelines (EEOC)? And can the results be used in practice — for example, to predict professional success?
The Hartman Value Profile works with 18 different "value combinations." The theory specifies very precisely which of these combinations should be rated higher or lower. So the researchers' question was: does this theoretical order actually show up in the responses of real people?
The answer: yes — and strikingly clearly. The test behaves the way the theory predicts. It doesn't measure "just some psychological thing," but exactly what it was built to measure: our ability to recognize values clearly, differentiate them, and judge them consistently.
Good is good, bad is bad — even in the test
A second, very simple check sounds banal, but it's crucial: do most people even recognize in the test what is meant to be "good" and what is "problematic"?
In the HVP there are "favorable" statements and "unfavorable" statements. The theory's expectation: good combinations should be rated more positively [i]on average[/i], and problematic combinations more negatively. That's exactly what the data show: across large samples, people sort "good" statements as good and "bad" statements as bad. Individual deviations are intended and interesting (they reveal personal style and possible distortions), but on average the model is confirmed.
In practice, that means: the test doesn't systematically confuse constructive with destructive — the fundamentals are sound.
Is it fair? The question of age, gender, and background
At the latest when a test is used for selection or development decisions, a sensitive question arises: does the procedure unconsciously disadvantage certain groups? In the U.S., this is tightly regulated by EEOC rules. The HVP data were analyzed through exactly that lens: do results differ systematically by age, gender, or ethnic background? And if so: to what extent?
The findings in brief:
- Age: Across the vast majority of scales, there are no relevant differences. Where differences do appear, they explain only a very small part of the variation — far too little to speak of disadvantage.
- Gender: Men and women perform virtually the same on average at all levels. Where differences are measurable, they are small and not capable of disadvantaging a group.
- Ethnicity: A similar picture emerges here as well: the major patterns are the same, and minimal differences explain only a small percentage of the variance.
The central conclusion of the studies is essentially: if there are differences in mean values, they are attributable to other factors — not to age, gender, or ethnic background.
How stable are the results? (Reliability)
A test can be brilliant in content — but if its results change completely from one day to the next, it's practically useless. That's why researchers checked: how similar are the results when the same person takes the HVP again at a later point in time? In one study, 200 people were selected who had taken the test twice. The result is impressive: agreement was extremely high, with a coefficient of 0.97 — well above the value (0.55) needed to say with high confidence: this is not random.
Does it align with other established tests?
The Hartman Value Profile wasn't only tested "within itself," but also compared with other psychological instruments — including the MMPI (Minnesota Multiphasic Personality Inventory, one of the best-known clinical personality tests) and the 16PF (Cattell's 16-factor personality model). Researchers found numerous statistically highly significant relationships between HVP scales and scales from these comparison tests — especially where stress, self-image, psychological strain, and self-destructive tendencies are involved.
Perhaps the most exciting question is: can meaningful predictions be derived from HVP results — for example, whether someone will do well in sales, leadership, or certain roles? To explore this, sales teams were examined and divided into three groups: non-salespeople, moderately successful salespeople, and top salespeople. From the HVP data, 49 factors had previously been derived that relate to sales success — such as empathy, handling rejection, self-motivation, stress processing, planning ability, and so on.
When comparing the three groups, it became clear: a large portion of these factors differed clearly between the groups. Top salespeople, on average, have different value patterns than moderately successful salespeople — and clearly different patterns than non-salespeople.
What matters here is an honest takeaway from the study: there is no universal "success profile" that looks the same everywhere. Success patterns vary by task and role. So the HVP doesn't deliver a one-size-fits-all "suitability seal" — but, when used correctly, it can help identify success-relevant value structures and make development potential visible.