How honest should I be in disclosing not-so-exciting results?
up vote
7
down vote
favorite
I'm a sociology undegrad working on an essay for a methods class. I'm also planning on submitting it as a sample for my application to grad school. I don't want to be too specific, but I believe that this work is quite original and my hypothesis would confirm previous literature, and all in all I think it would would make a good impression on the admissions committee.
So basically I've run the tests and I'm getting conflicting results. Using one dataset (which has more observations) gives me very significant results, while using another one (which would arguably be more accurate) doesn't give me anything. So here I am at a crossroads, and I've come up with three possible options as to what to do:
Only show the significant results. After all, this is just a ten-page essay, it's not supposed to be publishable or anything, right?
Only use the better dataset and admit that there just isn't much there - maybe blaming it on the small sample size or on the not-so-good dependent variable. Hopefully the committee would appreciate the honesty and the relatively advanced methods that I used.
Show results from both datasets, suggesting that the differences might be due to the sample size or maybe to chance.
As I type this I'm leaning more towards option 3, but I'd like to hear from people with more experience in academia. What should I do?
graduate-admissions research-undergraduate negative-results
New contributor
add a comment |
up vote
7
down vote
favorite
I'm a sociology undegrad working on an essay for a methods class. I'm also planning on submitting it as a sample for my application to grad school. I don't want to be too specific, but I believe that this work is quite original and my hypothesis would confirm previous literature, and all in all I think it would would make a good impression on the admissions committee.
So basically I've run the tests and I'm getting conflicting results. Using one dataset (which has more observations) gives me very significant results, while using another one (which would arguably be more accurate) doesn't give me anything. So here I am at a crossroads, and I've come up with three possible options as to what to do:
Only show the significant results. After all, this is just a ten-page essay, it's not supposed to be publishable or anything, right?
Only use the better dataset and admit that there just isn't much there - maybe blaming it on the small sample size or on the not-so-good dependent variable. Hopefully the committee would appreciate the honesty and the relatively advanced methods that I used.
Show results from both datasets, suggesting that the differences might be due to the sample size or maybe to chance.
As I type this I'm leaning more towards option 3, but I'd like to hear from people with more experience in academia. What should I do?
graduate-admissions research-undergraduate negative-results
New contributor
10
Contradictory results are the first step towards a discovery.
– henning
3 hours ago
6
@henning ...or a debunking of scientific credos. Embrace the contradiction.
– Captain Emacs
3 hours ago
3
"this work is quite original and my hypothesis would confirm previous literature" It confirms existing previous results, but it's original?
– Acccumulation
2 hours ago
1
+1 for asking. I strongly recommend you visit Andrew Gelman;s blog regularly for discussions of the proper way to do statistics, particularly in the social sciences, Here;s one example andrewgelman.com/?s=file+drawer
– Ethan Bolker
1 hour ago
add a comment |
up vote
7
down vote
favorite
up vote
7
down vote
favorite
I'm a sociology undegrad working on an essay for a methods class. I'm also planning on submitting it as a sample for my application to grad school. I don't want to be too specific, but I believe that this work is quite original and my hypothesis would confirm previous literature, and all in all I think it would would make a good impression on the admissions committee.
So basically I've run the tests and I'm getting conflicting results. Using one dataset (which has more observations) gives me very significant results, while using another one (which would arguably be more accurate) doesn't give me anything. So here I am at a crossroads, and I've come up with three possible options as to what to do:
Only show the significant results. After all, this is just a ten-page essay, it's not supposed to be publishable or anything, right?
Only use the better dataset and admit that there just isn't much there - maybe blaming it on the small sample size or on the not-so-good dependent variable. Hopefully the committee would appreciate the honesty and the relatively advanced methods that I used.
Show results from both datasets, suggesting that the differences might be due to the sample size or maybe to chance.
As I type this I'm leaning more towards option 3, but I'd like to hear from people with more experience in academia. What should I do?
graduate-admissions research-undergraduate negative-results
New contributor
I'm a sociology undegrad working on an essay for a methods class. I'm also planning on submitting it as a sample for my application to grad school. I don't want to be too specific, but I believe that this work is quite original and my hypothesis would confirm previous literature, and all in all I think it would would make a good impression on the admissions committee.
So basically I've run the tests and I'm getting conflicting results. Using one dataset (which has more observations) gives me very significant results, while using another one (which would arguably be more accurate) doesn't give me anything. So here I am at a crossroads, and I've come up with three possible options as to what to do:
Only show the significant results. After all, this is just a ten-page essay, it's not supposed to be publishable or anything, right?
Only use the better dataset and admit that there just isn't much there - maybe blaming it on the small sample size or on the not-so-good dependent variable. Hopefully the committee would appreciate the honesty and the relatively advanced methods that I used.
Show results from both datasets, suggesting that the differences might be due to the sample size or maybe to chance.
As I type this I'm leaning more towards option 3, but I'd like to hear from people with more experience in academia. What should I do?
graduate-admissions research-undergraduate negative-results
graduate-admissions research-undergraduate negative-results
New contributor
New contributor
New contributor
asked 3 hours ago
undergrad_dilemma
361
361
New contributor
New contributor
10
Contradictory results are the first step towards a discovery.
– henning
3 hours ago
6
@henning ...or a debunking of scientific credos. Embrace the contradiction.
– Captain Emacs
3 hours ago
3
"this work is quite original and my hypothesis would confirm previous literature" It confirms existing previous results, but it's original?
– Acccumulation
2 hours ago
1
+1 for asking. I strongly recommend you visit Andrew Gelman;s blog regularly for discussions of the proper way to do statistics, particularly in the social sciences, Here;s one example andrewgelman.com/?s=file+drawer
– Ethan Bolker
1 hour ago
add a comment |
10
Contradictory results are the first step towards a discovery.
– henning
3 hours ago
6
@henning ...or a debunking of scientific credos. Embrace the contradiction.
– Captain Emacs
3 hours ago
3
"this work is quite original and my hypothesis would confirm previous literature" It confirms existing previous results, but it's original?
– Acccumulation
2 hours ago
1
+1 for asking. I strongly recommend you visit Andrew Gelman;s blog regularly for discussions of the proper way to do statistics, particularly in the social sciences, Here;s one example andrewgelman.com/?s=file+drawer
– Ethan Bolker
1 hour ago
10
10
Contradictory results are the first step towards a discovery.
– henning
3 hours ago
Contradictory results are the first step towards a discovery.
– henning
3 hours ago
6
6
@henning ...or a debunking of scientific credos. Embrace the contradiction.
– Captain Emacs
3 hours ago
@henning ...or a debunking of scientific credos. Embrace the contradiction.
– Captain Emacs
3 hours ago
3
3
"this work is quite original and my hypothesis would confirm previous literature" It confirms existing previous results, but it's original?
– Acccumulation
2 hours ago
"this work is quite original and my hypothesis would confirm previous literature" It confirms existing previous results, but it's original?
– Acccumulation
2 hours ago
1
1
+1 for asking. I strongly recommend you visit Andrew Gelman;s blog regularly for discussions of the proper way to do statistics, particularly in the social sciences, Here;s one example andrewgelman.com/?s=file+drawer
– Ethan Bolker
1 hour ago
+1 for asking. I strongly recommend you visit Andrew Gelman;s blog regularly for discussions of the proper way to do statistics, particularly in the social sciences, Here;s one example andrewgelman.com/?s=file+drawer
– Ethan Bolker
1 hour ago
add a comment |
3 Answers
3
active
oldest
votes
up vote
23
down vote
In research, you don't set out to prove that something is true. You set out to discover whether or not it is true. This would be knowledge. The other is just propaganda.
Negative results are not a failure. They give you evidence just as do positive results. If you ignore, or obscure, results you are lying to yourself and others. If you design an "experiment" so that it is guaranteed a priori to produce positive results, it isn't research.
Hoping that something is true isn't evidence. Many researchers start out with that idea. I think this is true. I really want it to be true. But if it is false, it is just as valuable (possibly more so) to know that and to be able to investigate why.
Report all your results. Try to explain why different aspects lead you in different directions. Only then can your learning begin.
add a comment |
up vote
10
down vote
Omitting negative findings and selectively reporting only the positive findings would be a breach of research ethics. As a researcher you are supposed to uncover knowledge,* not to obscure it. Findings are often contradictory and in need of interpretation. By explaining how you obtained these contradictory results (i.e. your methods), you help others to avoid dead ends in the future and to make sense of what looks confusing today.
*Interestingly, the knowledge that research creates often takes the form of higher-level confusion rather than ultimate certainty.
+1 because research ethics aren't something that applies only when something is "publishable" (as in "After all, this is just a ten-page essay, it's not supposed to be publishable or anything, right?")
– De Novo
1 hour ago
add a comment |
up vote
0
down vote
Are your significant results a large effect size, or just a tiny change that is significant because of the large sample size?
Are your non-significant results similar in direction and magnitude to the significant results from the other dataset?
Consider how much the size of the dataset is impacting what you are seeing - you may be able to frame one study as confirming the results of the other if they are in agreement apart from significance. Look at more than just the p-values, especially if they are coming from a very large dataset.
add a comment |
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
up vote
23
down vote
In research, you don't set out to prove that something is true. You set out to discover whether or not it is true. This would be knowledge. The other is just propaganda.
Negative results are not a failure. They give you evidence just as do positive results. If you ignore, or obscure, results you are lying to yourself and others. If you design an "experiment" so that it is guaranteed a priori to produce positive results, it isn't research.
Hoping that something is true isn't evidence. Many researchers start out with that idea. I think this is true. I really want it to be true. But if it is false, it is just as valuable (possibly more so) to know that and to be able to investigate why.
Report all your results. Try to explain why different aspects lead you in different directions. Only then can your learning begin.
add a comment |
up vote
23
down vote
In research, you don't set out to prove that something is true. You set out to discover whether or not it is true. This would be knowledge. The other is just propaganda.
Negative results are not a failure. They give you evidence just as do positive results. If you ignore, or obscure, results you are lying to yourself and others. If you design an "experiment" so that it is guaranteed a priori to produce positive results, it isn't research.
Hoping that something is true isn't evidence. Many researchers start out with that idea. I think this is true. I really want it to be true. But if it is false, it is just as valuable (possibly more so) to know that and to be able to investigate why.
Report all your results. Try to explain why different aspects lead you in different directions. Only then can your learning begin.
add a comment |
up vote
23
down vote
up vote
23
down vote
In research, you don't set out to prove that something is true. You set out to discover whether or not it is true. This would be knowledge. The other is just propaganda.
Negative results are not a failure. They give you evidence just as do positive results. If you ignore, or obscure, results you are lying to yourself and others. If you design an "experiment" so that it is guaranteed a priori to produce positive results, it isn't research.
Hoping that something is true isn't evidence. Many researchers start out with that idea. I think this is true. I really want it to be true. But if it is false, it is just as valuable (possibly more so) to know that and to be able to investigate why.
Report all your results. Try to explain why different aspects lead you in different directions. Only then can your learning begin.
In research, you don't set out to prove that something is true. You set out to discover whether or not it is true. This would be knowledge. The other is just propaganda.
Negative results are not a failure. They give you evidence just as do positive results. If you ignore, or obscure, results you are lying to yourself and others. If you design an "experiment" so that it is guaranteed a priori to produce positive results, it isn't research.
Hoping that something is true isn't evidence. Many researchers start out with that idea. I think this is true. I really want it to be true. But if it is false, it is just as valuable (possibly more so) to know that and to be able to investigate why.
Report all your results. Try to explain why different aspects lead you in different directions. Only then can your learning begin.
answered 2 hours ago
Buffy
32.9k6100171
32.9k6100171
add a comment |
add a comment |
up vote
10
down vote
Omitting negative findings and selectively reporting only the positive findings would be a breach of research ethics. As a researcher you are supposed to uncover knowledge,* not to obscure it. Findings are often contradictory and in need of interpretation. By explaining how you obtained these contradictory results (i.e. your methods), you help others to avoid dead ends in the future and to make sense of what looks confusing today.
*Interestingly, the knowledge that research creates often takes the form of higher-level confusion rather than ultimate certainty.
+1 because research ethics aren't something that applies only when something is "publishable" (as in "After all, this is just a ten-page essay, it's not supposed to be publishable or anything, right?")
– De Novo
1 hour ago
add a comment |
up vote
10
down vote
Omitting negative findings and selectively reporting only the positive findings would be a breach of research ethics. As a researcher you are supposed to uncover knowledge,* not to obscure it. Findings are often contradictory and in need of interpretation. By explaining how you obtained these contradictory results (i.e. your methods), you help others to avoid dead ends in the future and to make sense of what looks confusing today.
*Interestingly, the knowledge that research creates often takes the form of higher-level confusion rather than ultimate certainty.
+1 because research ethics aren't something that applies only when something is "publishable" (as in "After all, this is just a ten-page essay, it's not supposed to be publishable or anything, right?")
– De Novo
1 hour ago
add a comment |
up vote
10
down vote
up vote
10
down vote
Omitting negative findings and selectively reporting only the positive findings would be a breach of research ethics. As a researcher you are supposed to uncover knowledge,* not to obscure it. Findings are often contradictory and in need of interpretation. By explaining how you obtained these contradictory results (i.e. your methods), you help others to avoid dead ends in the future and to make sense of what looks confusing today.
*Interestingly, the knowledge that research creates often takes the form of higher-level confusion rather than ultimate certainty.
Omitting negative findings and selectively reporting only the positive findings would be a breach of research ethics. As a researcher you are supposed to uncover knowledge,* not to obscure it. Findings are often contradictory and in need of interpretation. By explaining how you obtained these contradictory results (i.e. your methods), you help others to avoid dead ends in the future and to make sense of what looks confusing today.
*Interestingly, the knowledge that research creates often takes the form of higher-level confusion rather than ultimate certainty.
edited 2 hours ago
answered 3 hours ago
henning
17.2k45989
17.2k45989
+1 because research ethics aren't something that applies only when something is "publishable" (as in "After all, this is just a ten-page essay, it's not supposed to be publishable or anything, right?")
– De Novo
1 hour ago
add a comment |
+1 because research ethics aren't something that applies only when something is "publishable" (as in "After all, this is just a ten-page essay, it's not supposed to be publishable or anything, right?")
– De Novo
1 hour ago
+1 because research ethics aren't something that applies only when something is "publishable" (as in "After all, this is just a ten-page essay, it's not supposed to be publishable or anything, right?")
– De Novo
1 hour ago
+1 because research ethics aren't something that applies only when something is "publishable" (as in "After all, this is just a ten-page essay, it's not supposed to be publishable or anything, right?")
– De Novo
1 hour ago
add a comment |
up vote
0
down vote
Are your significant results a large effect size, or just a tiny change that is significant because of the large sample size?
Are your non-significant results similar in direction and magnitude to the significant results from the other dataset?
Consider how much the size of the dataset is impacting what you are seeing - you may be able to frame one study as confirming the results of the other if they are in agreement apart from significance. Look at more than just the p-values, especially if they are coming from a very large dataset.
add a comment |
up vote
0
down vote
Are your significant results a large effect size, or just a tiny change that is significant because of the large sample size?
Are your non-significant results similar in direction and magnitude to the significant results from the other dataset?
Consider how much the size of the dataset is impacting what you are seeing - you may be able to frame one study as confirming the results of the other if they are in agreement apart from significance. Look at more than just the p-values, especially if they are coming from a very large dataset.
add a comment |
up vote
0
down vote
up vote
0
down vote
Are your significant results a large effect size, or just a tiny change that is significant because of the large sample size?
Are your non-significant results similar in direction and magnitude to the significant results from the other dataset?
Consider how much the size of the dataset is impacting what you are seeing - you may be able to frame one study as confirming the results of the other if they are in agreement apart from significance. Look at more than just the p-values, especially if they are coming from a very large dataset.
Are your significant results a large effect size, or just a tiny change that is significant because of the large sample size?
Are your non-significant results similar in direction and magnitude to the significant results from the other dataset?
Consider how much the size of the dataset is impacting what you are seeing - you may be able to frame one study as confirming the results of the other if they are in agreement apart from significance. Look at more than just the p-values, especially if they are coming from a very large dataset.
answered 1 hour ago
APH
1465
1465
add a comment |
add a comment |
undergrad_dilemma is a new contributor. Be nice, and check out our Code of Conduct.
undergrad_dilemma is a new contributor. Be nice, and check out our Code of Conduct.
undergrad_dilemma is a new contributor. Be nice, and check out our Code of Conduct.
undergrad_dilemma is a new contributor. Be nice, and check out our Code of Conduct.
Thanks for contributing an answer to Academia Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Some of your past answers have not been well-received, and you're in danger of being blocked from answering.
Please pay close attention to the following guidance:
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2facademia.stackexchange.com%2fquestions%2f121400%2fhow-honest-should-i-be-in-disclosing-not-so-exciting-results%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
10
Contradictory results are the first step towards a discovery.
– henning
3 hours ago
6
@henning ...or a debunking of scientific credos. Embrace the contradiction.
– Captain Emacs
3 hours ago
3
"this work is quite original and my hypothesis would confirm previous literature" It confirms existing previous results, but it's original?
– Acccumulation
2 hours ago
1
+1 for asking. I strongly recommend you visit Andrew Gelman;s blog regularly for discussions of the proper way to do statistics, particularly in the social sciences, Here;s one example andrewgelman.com/?s=file+drawer
– Ethan Bolker
1 hour ago