The End of Theory: The Data Deluge Makes the Scientific Method Obsolete

In summary: Data in, correlation out" (theorist)---I think this article is very much on target. Basically all this article is talking about is Data Mining. And I would agree that it's becoming, more and more, applicable to science. As an example, in undergrad I worked a coop at a place that did medical imaging research and one of the things they were talking about is getting computers to diagnose disease through Machine Learning. The gyst of it is is say that you want to be able to give a computer an MRI image and have it tell you whether the person has cancer. The best way to go about that is not to try and define what properties of an image codifies one with cancer
  • #1
CINA
61
0
Saw this interesting article and wondered what PF thought of it since its close to home. Personally I don't see the scienctific method being replaced anytime soon.

http://www.wired.com/science/discoveries/magazine/16-07/pb_theory/#/

Edit: Because of its predictive nature that is.
 
Last edited by a moderator:
Physics news on Phys.org
  • #2
That article is complete garbage. It doesn't actually say anything or have a point besides "SOOO MANY BYTES ZOMG!1"

Whoever wrote it doesn't actually understand science. It's not about having data, it's about being able to explain the data. That's the whole goal, to understand what is going on, not just to see what is going on. Everybody knew that Maxwell's equations gave the correct answer regardless of reference frame. The problem was nobody knew why until Einstein came along and gave an explanation. If they'd just kept on collecting data, we wouldn't have gotten anywhere.
 
  • #3
No, I see some validity in the argument. (That doesn't mean I agree, it just means it's a valid argument.)

The idea is: why do we need to have models of weather if we have enough data to simply predict what will happen without needing to know why it happens? (Bascially, we've created a Farmer's Almanac, writ large.)



Of course the downside is...

one can forsee a Twilight Zone or Trekkian clichéd future where citizens continue to use the tools that have worked for centuries but are at a loss if anything changes or if anything breaks down.
 
  • #4
WarPhalange said:
That article is complete garbage. It doesn't actually say anything or have a point besides "SOOO MANY BYTES ZOMG!1"

Whoever wrote it doesn't actually understand science. It's not about having data, it's about being able to explain the data. That's the whole goal, to understand what is going on, not just to see what is going on. Everybody knew that Maxwell's equations gave the correct answer regardless of reference frame. The problem was nobody knew why until Einstein came along and gave an explanation. If they'd just kept on collecting data, we wouldn't have gotten anywhere.

I think its more to interpret the data in a new way, which the computer can't do. "Data in, data out" (computer)---or "Data in, correlation out" (theorist)---
 
  • #5
I think this article is very much on target. Basically all this article is talking about is Data Mining. And I would agree that it's becoming, more and more, applicable to science. As an example, in undergrad I worked a coop at a place that did medical imaging research and one of the things they were talking about is getting computers to diagnose disease through Machine Learning. The gyst of it is is say that you want to be able to give a computer an MRI image and have it tell you whether the person has cancer. The best way to go about that is not to try and define what properties of an image codifies one with cancer, instead the best way is to just give a statistical learning algorithm a million MRI's and say these 500,000 don't have cancer and these 500,000 do, and let it figure out what the difference is. A lot of people have a problem with the 'black box' aspect of it but I think it's going to become, more and more, a standard tool in an ever growing list of applications
 
  • #6
Data mining, right! Closely related to this post.

The deluge of studies, articles and records would enhance the tendency to select what seems to be supporting and to ignore what seems contradictory.
 
  • #7
maverick_starstrider said:
I think this article is very much on target. Basically all this article is talking about is Data Mining. And I would agree that it's becoming, more and more, applicable to science. As an example, in undergrad I worked a coop at a place that did medical imaging research and one of the things they were talking about is getting computers to diagnose disease through Machine Learning. The gyst of it is is say that you want to be able to give a computer an MRI image and have it tell you whether the person has cancer. The best way to go about that is not to try and define what properties of an image codifies one with cancer, instead the best way is to just give a statistical learning algorithm a million MRI's and say these 500,000 don't have cancer and these 500,000 do, and let it figure out what the difference is. A lot of people have a problem with the 'black box' aspect of it but I think it's going to become, more and more, a standard tool in an ever growing list of applications

Okay, now how does that apply to finding new science?
 
  • #8
rewebster said:
I think its more to interpret the data in a new way, which the computer can't do. "Data in, data out" (computer)---or "Data in, correlation out" (theorist)---

Not quite true; there are classes of computer algorithms (neural networks, genetic algorithms etc) that are very good at predicting the outcome of future experiments if you train them well (i.e. give them lots of examples first); the interesting thing is that they tend to work well even when "normal" mathematical methods are very difficult to use which is why they are now being used in e.g. social sciences.
Moreover, it is often very hard to understand WHY it works even when the algorithm has been trained and you can look at what it is actually doing, it is therefore unlikely that a human would ever stumble over the methods "developed" by these algorithms.
 
  • #9
WarPhalange said:
Okay, now how does that apply to finding new science?

Well don't forget, quantum mechanics' original development was guided by the 'correspondence principle' (basically let's find a framework that reproduce these new 'bizarre results' and reduces to classical mechanics at the macro level' and worry about the physical intuition and meaning later).

Plus instead of say attempting to deduce, for example, which combination of ingredients and such could potentially create a room temperature superconducter from physical intuition you could apply basically machine learning.

There was an article in Scientific American like a year ago about using evoluntionary algorithms (a type of machine learning) to develop circuits. Within a fraction of a second the algorithm was able to develop the ideal circuit for things like high pass and low pass filtering and within a couple hours was even able to develop special use circuit layouts that were only patented a couple years ago.
 
  • #10
maverick_starstrider said:
Well don't forget, quantum mechanics' original development was guided by the 'correspondence principle' (basically let's find a framework that reproduce these new 'bizarre results' and reduces to classical mechanics at the macro level' and worry about the physical intuition and meaning later).

Plus instead of say attempting to deduce, for example, which combination of ingredients and such could potentially create a room temperature superconducter from physical intuition you could apply basically machine learning.

Okay, but you'd have to give it a bunch of models and things we don't even understand yet. What I mean is, you'd have to give it test data. Wood = bad conductor. Iron = better. HgBa2Ca2Cu3O8 = awesome. But that's not enough data to make it deduce new materials.

Going back to the cancer analogy, you fed 1million different data points to the computer. Here you have a handful.

There was an article in Scientific American like a year ago about using evoluntionary algorithms (a type of machine learning) to develop circuits. Within a fraction of a second the algorithm was able to develop the ideal circuit for things like high pass and low pass filtering and within a couple hours was even able to develop special use circuit layouts that were only patented a couple years ago.

Yeah, that's pretty awesome, but like I said we don't fully understand why superconductors do what they do (at least the high-temp ones). Circuit design is using what we know very well to create new things. Trial and error will get you there eventually and you can test whether you are there or not because we have a very good understanding of circuits. We can't simply model a new material because we're not even sure why the ones we already have work the way they do.
 

Related to The End of Theory: The Data Deluge Makes the Scientific Method Obsolete

1. What is "The End of Theory: The Data Deluge Makes the Scientific Method Obsolete"?

"The End of Theory: The Data Deluge Makes the Scientific Method Obsolete" is a concept proposed by Chris Anderson, editor-in-chief of Wired magazine. It suggests that with the abundance of data available in the digital age, traditional scientific methods of forming and testing hypotheses may no longer be necessary.

2. Is the scientific method really becoming obsolete?

This is a highly debated topic. Some argue that the scientific method is still a crucial tool for understanding the world, while others believe that with the vast amount of data available, traditional hypothesis-driven science is no longer necessary.

3. What are the implications of "The End of Theory" for the future of science?

If the scientific method does become obsolete, it could have significant implications for the way we approach research and make discoveries. It may shift the focus from hypothesis-driven research to data-driven research, and potentially change the role of scientists and the way they conduct their work.

4. Are there any potential drawbacks to this shift towards data-driven science?

Some argue that relying solely on data without a hypothesis or theoretical framework could lead to misleading or even false conclusions. It may also reduce the need for creativity and critical thinking in the scientific process.

5. How can scientists adapt to this changing landscape in science?

Scientists may need to develop new skills in data analysis and interpretation in order to effectively utilize the abundance of data available. They may also need to collaborate with experts in fields such as computer science and statistics to make the most of the data deluge.

Similar threads

  • General Discussion
Replies
6
Views
1K
Replies
4
Views
996
  • General Discussion
Replies
12
Views
1K
  • General Discussion
Replies
12
Views
1K
Replies
9
Views
1K
  • General Discussion
Replies
14
Views
2K
Replies
1
Views
108
  • General Discussion
Replies
5
Views
4K
Replies
10
Views
2K
  • General Discussion
Replies
2
Views
2K
Back
Top