Why Is Data Mining Still A Frontier? 223
bbsguru writes "How much do we know that we still don't know? A story in The Register points out that little has changed since Francis Bacon proposed combining knowledge to learn new things 400 years ago, despite all the computer power we now have. Scientific (and other) data is still housed in unrelated collections, waiting for some enterprising Relational Database Programmer to unlock the keys to understanding. Is RDBMS still a Brave New Frontier, or will Google make the art obsolete once they finish indexing everything?"
Shot in the dark: (Score:5, Insightful)
a) There's not enough money in it to make it worthwhile
or
b) It doesn't work.
Re:Shot in the dark: (Score:5, Insightful)
c) our ability to produce data far outstrips our ability and/or willingness to analzye it
Re:Shot in the dark: (Score:5, Insightful)
Re:Shot in the dark: (Score:3, Insightful)
People eat more ice cream in summer.
Conclusion: ice cream causes polio.
This was actually something people believed for a brief time before the Salk vaccine. Its also a great example of the kind of facts data mining most frequently dredges up- accidents or correlation with no real common cause.
Re:Shot in the dark: (Score:2)
Amazing how many people still don't understand the difference between correlation and causation. I think everyone should take an intro economisc or statistics class just to realize the difference.
Re:Shot in the dark: (Score:2)
Who has time for for an economics class? Summer is just around the corner. We've got to do something about all that ice cream out there!!
Please, think of the children.
(Maybe we should start adding vaccine to the sprinkles. (Am I the only one who calls them jimmies?))
Re:Shot in the dark: (Score:3, Insightful)
People design systems for what they want to put into it, without consider what they want to get back out of it. That usually results in crappy query performance and all that crap because of undue care. When designing a system, engineers need to be aware of : 1) What do we want to store and how do we want to store it, 2) how do we want to put it in there, 3) What do we want to get back out of it.
Many people in designin
Re:Shot in the dark: (Score:2)
Re:Shot in the dark: (Score:2)
Re:Shot in the dark: (Score:2)
Think star schema, with the central table containing just numerical 'facts'. Each record's key links to every other table, and, for query optimization, we've got just one 'fact' per record. Payments, APR, Balances, they all get slapped in.
It's one of the best OLAP tools I've seen. A hell of a lot of work to do it 'right', like ten hours pr
Re:Shot in the dark: (Score:2)
A specific case comes to mind. A guy I worked with wanted to design a billing system. He had six tables represnting detail lines on the invoce. Each table had identical fields except for a few items. The data should not have been normalized because a report on invoicing would have required a six table union. Unacceptable IMO.
Re:Shot in the dark: (Score:3, Funny)
Re:Shot in the dark: (Score:2)
Re:Shot in the dark: (Score:2)
Many people in designing systems pass over 3.
Good. The desired results change too often to put them in the data model.
It's been my experience that the best database designs come from focusing on a layout that makes sense. Aside from recursive hierarchys (which are a special case because they don't fit into the relational model very well), you should only need to look much at the actual queries you expect to run when you're deciding on indexes.
Re:or... (Score:2)
Our ability to produce meaningful results, in most cases, is little more than a crapshoot.
Re:Shot in the dark: (Score:2)
c) our ability to produce data far outstrips our ability and/or willingness to analzye it
Wouldn't that be the same as b) it doesn't work?
Re:Shot in the dark: (Score:2)
Re:Shot in the dark: (Score:2)
Re:Shot in the dark: (Score:2)
Real World Example: This past summer, I went to Cyprus for a field survey (surface examination and collection, no digging involved). In three weeks of 15 people working 4 hours a day, we grabbed over 10,000 pieces of worked stone. A proper excavation will yield enough data for an academic lifetime, but only a small percentage will ever be thoroughly analyzed and published.
Clarification of your "b". (Score:2)
How about "It doesn't work the way the vendor/consultant/salesguy/magazine said it would."
The information you get out depends upon the data you put in.
The people looking to "find" information in the data are the same people who decided what data to collect in the first place. And from whom to collect it. Etc.
That means that you'll find out that 2004 was a banner year for bubblegum ice cream. But you won't know what will be popular in the summer of 2006.
Re:Shot in the dark: (Score:5, Informative)
Analogies like this are always dangerous, but I'd say data mining now is about where language development was in the mid-1950's, when FORTRAN was first being developed. IOW, we have a set of tools that kind of work, most of the time, for certain applications -- but we can pretty much guarantee that they're not the best possible tools, and that we will build better ones. Consider how much work is still going on in language development half a century later, and you can see how much room there is for further development.
Re:Shot in the dark: (Score:5, Informative)
I would suggest that, in practice, the real difficulty is that the problems that need to really be solved for data mining to be as effective as some people seem to wish it was are, when you actually get down to it, issues of pure mathematics. Research in pure mathematics (and pure CS which is awfully similar really) is just hard. Pretending that this is a new and growing field is actually somewhat of a lie. It's avery very old field which people have been working on for a very long time, to the point where the problems that remain to be solved are incredibly difficult. What is new is someone other than pure mathematicians taking much interest in these problems. Do a search for "non linear manifold learning" on Google and you'll see what I mean.
Jedidiah.
Re:Shot in the dark: (Score:2)
Don't forget that if you ask the wrong questions you get either:
A. Wrong Answers
or
B. Garbage
Having computers crunch data to look for relationships is all well and good, but you're almost always going to need someone to interpret the results to make sure they aren't A or
Re:Shot in the dark: (Score:4, Interesting)
That's part of the problem.
Another part is computational complexity. No, I'm not kidding. These things are often in like the second and third powers of the data set size. The data sets are often terabytes in size. We don't have computers that big, and by the time we do, we'll probably have bigger data sets. Contemporary data mining is an exercise in finding a fast enough approximation that is accurate enough to look convincing. We're not really sure how accurate they actually are - most of the time, there's no way to find out for certain. "Probably good enough" is the best you normally get. Some researchers can put a number on that 'probably' for you, eventually. Mostly they just compare the available approximations and tell you which one works the best.
The biggest problem is the inability to figure out intelligent things to do with it. Computers aren't smart. You can't just hand them a heap of data and say "find me the things I want to know". You have to work out what the patterns in the data are for yourself, then do pure math research to turn those patterns into a mathematical model. Then you have to come up with useful questions to ask that model. That's two major insights plus several years of work - and most researchers only have one major insight in their entire career. Just to figure out what question to ask. Data mining is then the process of repeatedly answering that question for all possible values of the parameters. And the answers you get out will only be as good as the model you invented. The current method for discovering usable patterns in data is trial and error.
I think that 'data mining' is more or less a frontier by definition. It's all the things we don't yet know about the data we currently have which would take a huge amount of effort to discover. Most unsolved problems in mathematics could probably be called 'data mining problems': if an answer exists, it can be derived from the existing body of theory. Most decisions that people make, from deciding whether to eat now or later, to deciding whether to invade a foreign nation, can also qualify. The sheer range of things it could cover means that there will probably always be vastly more unsolved problems than solved ones.
Re:Shot in the dark: (Score:2)
Semantic web (Score:2)
What if you could "exp
Re:Shot in the dark: (Score:2)
That's what I said. Nobody knows a way to process it that today's computers can handle. We *do* know several ways to process it that those computers *can't* handle.
As to your claim that "we're storing hundreds of terabytes of data, obviously we can handle it" - you're just storing data. The problem is computational complexity, not storage. The well-known 'right' answers to most data mining problems
Re:Shot in the dark: (Score:2)
Re:Shot in the dark: (Score:5, Insightful)
Consider WalM*rt. When the 2005 hurricanes were predicted, they mined their sales data for previous hurricanes. They found that in the last hurricane people stocked up on beer, pop tarts and peanut butter, so they sent trucks full of that stuff to the stores in the path of the hurricanes. They made lots of sales, and provided a valuable service to the communities. Capitalism at its finest.
Data mining worked very well in this case. The issue was "here's an obvious problem, and a clever solution involving data mining."
The big problem is that people expect the same golden results from non-obvious situations. "Hey, sales are down in the Wisconsin stores, let's do some data mining to figure out what they'll buy" makes no sense. Data mining worked well in the case of an obvious trigger event, but data mining by itself didn't reveal the trigger. You can't predict hurricanes based on the sales of pop tarts and beer, for example.
But, can you ever correlate pop tart and beer sales to an external event? You might be able to go back and say "here's a strange case where pop tarts and beer sold out quickly, why did this happen?" If you can tie this to external events, you'd think you'd be better prepared to react to the same events in the future.
Maybe correlating sales to Google News is the next step? Republican scandal == lower white bread sales; French riots + Senate bickering over immigration control reform == higher 'Peeps' sales; etc. p. Or maybe it's always been a bad idea to equate correlation with causality.
Re:Shot in the dark: (Score:2)
But that's not what I was talking about. I'm in the retail industry, and keep one eye facing Walmart (everyone in retail does.) The "beer and poptarts" story was one of those stories that circulated about the same time as the hurricane, so I can't quote exactly which source I got it from first (could have been at a department
Re:Shot in the dark: (Score:2)
There's the problem. "If". Let's say the data mining came up with a correlation between French riots and immigration legislation with the sale of Peeps. Next Easter you're going to disappoint a lot of shoppers when you don't have Peeps available; and next fall when rioting and legislation happen to hit the news at the same time, you're going to have a lot of wasted Peeps on your store shelves.
You may say "Of course riots h
Re:Shot in the dark: (Score:2)
The basic problem is that, as with any worthwhile CS question, doing it well is hard. It is very easy to come up with false connections between data. Sorting the wheat from the chaff in any kind of automated or even semi-automated fashion, OTOH, is an enormous challenge.
I'll respectfully disagree. There's a very large number of organisations that are using predicitive modelling through data mining to conduct various forms of customer scoring and analytical CRM activities. These are being used in a pro
Re:Shot in the dark: (Score:3, Insightful)
c) The title of this submission is inaccurate, as data mining tools are both useful and financially lucative in a wide variety of domains today, particularly bioinformatics, image analysis and text mining.
Of course, the title of this article is quite ambiguous and misleading: the article itself is concerned with RDBMS, rather than the statistical analysis of data.
Re:Shot in the dark: (Score:2)
If anything, the article is support for the use and continued development of datamining technologies.
M
Re:Shot in the dark: (Score:2)
Re:Shot in the dark: (Score:3, Insightful)
Re:Shot in the dark: (Score:3, Insightful)
I believe there are several fundamental factors required to make a data-mining project succesful:
1) A mathematically precise definition of what it is to be modeled (the response) as in the probability of purchasing
Re:Shot in the dark: (Score:2)
I think it's more a problem of access to the data for purpose of mining. In order to do any meaningful dataming you have to have a few barriers removed. Namely:
It has to be cheap to access. This is in terms of network costs, labor costs, and most importantly everyone believes that they can make a profit if they sell the access to their data. For data mining purposes, this becomes cost prohibitive. You have to Free the Data.
It has to be legal to access. As time goes on, the amount of data, or the typ
Re:Shot in the dark: (Score:2)
But the simple fact is that once you have enough data available, you can "mine" any result you want! Datamining is not about letting the data lead you to certain conclusions. It's all about trying to find things in the data that "hidden" - things that really aren't there when the data are properly analyzed.
Depends what you mean by "data mining". As the other reply has already said, bad statistics is bad statistics, regardless of the name. There's plenty of techniques in use to prevent spurious or misl
What does the article have to do with the subject? (Score:3)
This still looks like a basic, specialized database to me. Where's the great leap to "all your data are belong to us?"
Companies are doing it, but... (Score:4, Insightful)
Considering how up and arms people are about it, how long before we have people accusing others of "data profiling"?
Re:Companies are doing it, but... (Score:3, Interesting)
A snippet from the article:
the tools taken as routine in business are being overlooked in academia
I can't see anybody getting upset about scientific data mining
Re:Companies are doing it, but... (Score:2)
I tell you why (from a bioinformatics viewpoint) (Score:5, Insightful)
Re:I tell you why (from a bioinformatics viewpoint (Score:2)
Balence, restored.
-nB
Re:I tell you why (from a bioinformatics viewpoint (Score:3, Insightful)
Well, that's pretty much how it works in academia (+/- the real dba). Problem is that this is a lab by lab (or department) solution to problems that appear in hundreds or thousands of institutions. The wheel is reinvented over and over again because either commercial/free solutions suck or don't exist. The commercial versions suck because they ar
Re:I tell you why (from a bioinformatics viewpoint (Score:2)
The um...field I've been working in for the last 6 years.
Programming + Biology + Statistics + Algorhitm development.
Re:I tell you why (from a bioinformatics viewpoint (Score:2)
Re:I tell you why (from a bioinformatics viewpoint (Score:2)
Re:I tell you why (from a bioinformatics viewpoint (Score:2)
Re:I tell you why (from a bioinformatics viewpoint (Score:3, Interesting)
Re:I tell you why (from a bioinformatics viewpoint (Score:2)
Is it possible the developers are saying something like "It starts out with the biologists saying they need 30 TB of data available 24/7 with 99.999% uptime and 200-250 concurrent users, and goes downhill from there..."
Simply saying the developers are idiots because they suggest Oracle really doesn't make sense without more context. If more than one group of developers sugges
Re:I tell you why (from a bioinformatics viewpoint (Score:2)
Because Larry Ellison needs a new sub-woofer [bizjournals.com]?
It's all in the management. (Score:2)
Re:I tell you why (from a bioinformatics viewpoint (Score:2)
It is hard to mix the two. This is even more frustrating if you're marginally inclined to understand where things come from and how they are designed. Some of the earliest proponets of object oriented software prog
Re:I tell you why (from a bioinformatics viewpoint (Score:2)
Re:I tell you why (from a bioinformatics viewpoint (Score:2)
Data Mining != RDBMS (Score:2)
Aristotle (Score:2)
Re:Aristotle (Score:3, Funny)
Re:Aristotle (Score:2)
Re:Aristotle (Score:2)
Re: (Score:2)
Re:Semantic Web goodness (Score:2)
Too bad that the Semantic Web will always be a pipe dream, at least until the day comes when it's possible for a computer to understand the semantic content of a document with zero hinting from the author. The potential is there, but the willingness of humans to spend time explaining semantic structures to machines, when they're obvious enough to other humans, is lacking.
Re:Semantic Web goodness (Score:3, Insightful)
Well, XML is not really import. The problem lies in going from the infinite real world to a well defined ontology or whatever
Re:Semantic Web goodness (Score:2)
And the stuff they've produced is still academic level. The average high school kid isn't going to be hacking OWL into his web pages.
The average high school kid has an RSS feed on their blog.
The average high school kid listens to MP3s tagged with IDV3 metadata.
The average high school kid annotates thier photos on Flickr with semantic metadata.
The average web user may not know what the Semantic Web is but that doesn't mean they're not using it.
Privacy (Score:2)
Another thing is that it is only usefull for information we don't already know.
We don't exactly need data mining to realize that people that buy diapers also buy baby food.
Re:Privacy (Score:2)
Re:Privacy (Score:2)
The other side is that some places use loyalty cards which actually advertise and use the loss of privacy as a selling point: "This is a personal promotion just for you, PHILIP J. FRY!"
Some people are comfortable giving i
Re:Privacy (Score:2)
You sign up for a grocery datamining card. You give them your name, phone, address, and they give you a card to scan when you buy groceries. Now you use it to buy things. Among other things you buy:
a six pack of beer. Every day.
tampons, even though you are a man.
stop buying tampons, but pick up some penicillian at the pharmacy in the back.
These things are very, very personal. And they have your name, number, address.
Because it's not sexy (Score:5, Insightful)
Re:Because it's not sexy (Score:5, Interesting)
So how do you arrive at all those possible patterns and connections? If you think the number of different ways of slicing, considering, and analysing a given large dataset is anything but stupendously amazingly big then you're fooling yourself. Aside from millions of ways of slicing and dicing the data there are all kinds of useful ways to transform or reinterpret the data to find other connections: do fourier transforms to look at frequency spaces, view it as a directed graph or a lattice, perform some manner of clustering or classification against [insert random property here] and reinterpret, and so on, each of which expose whole new levels of slice and dice that can be done. If you'ev got subject matter experts working closely with you then you can at least make some constructive guesses as to some directions that will be profitable, and some directions that definitely will not be, but in between is a vast space where you simply cannot know. Data mining, right now, involves an awful lot of fumbling in the dark because there are simply so many ways to analyse the sort of volume of data we have collected, and the only real way to judge any analysis is to present it to human because our computers simply aren't as good at seeing understanding an interpreting patterns to trust with the job. Anytime a process has to route everything through humans you know it is going to be very very slow.
Jedidiah.
Chloe to the rescue (Score:2)
Re:Chloe to the rescue (Score:2)
I think Jack tells Tony to keep Chloe in reserve so he can play the hero more.
FTFA . . . (Score:2)
Hmm, I read recently that Darwin's grandfather was also a Naturalist, as was Chuck. So, I don't think Darwin made the "leap," so much as his family was already in that direction. Methinks the article p
Data mining is DIFFICULT (Score:5, Informative)
Consider the following boring but difficult task I was given: two large organizations were to merge, each with a portfolio of about 100,000 items. Each item had a short history, some descriptive information, and some data such as internal quality ratings or sector assignments. This data was available (for various reasons) as big CSV file dumps. Questions to answer were: (1) how much overlap did the portfolios have? (2) were the sector distributions similar?
These are very simple, concrete questions. But you can imagine that since the categorizations differed, and descriptors differed within the CSV files, let alone between the two, the questions were difficult to answer. It required a lot of approximate matching, governed intelligently (or so I flatter myself).
Contrast this situation with what people typically think of as data-mining: answering interesting questions, and you can appreciate that without a whole lot of intelligence, artificial or otherwise, those questions will be unanswerable.
Re:Data mining is DIFFICULT (Score:2)
Nothing to do with Technology (Score:4, Informative)
The ultimate problem, is that for most datasets, there are an infinite (at least), set of relations that can be induced from the data. This doesn't even address the issue, that the choice of available data is a human task. However, going back to assuming we have all the data possible, you still need to have a specific performance task in mind.
Think of this in terms of permutations. Lets say you have variable A, B, and C. They are all binary (have values 1 or 0). Now, you are given a set of these assigments (eg A=1, B=1,C=1, A=1,B=1, C=1, and so on). Now, try to tell me what the correct partition is. Sort them in to two sets of any size. See the problem ? I didn't tell you what I wanted as characteristics of those sets - so in effect, they are all possible good partitions.
So, data-mining ultimately relies on human's deciding what they want to read from the tea-leaves of the data.
Now, give it up, and start addressing issues of efficient algorithms given that you have a specific performance task
Winton
Re:Honest question from serious lackey- (Score:2)
What's the question? What are the barriers? (Score:2)
How do I write a multi-variable where clause?
How do I ask a multi-variable question and then hone it or drill into it along one or more parameters, unfolding detail but preserving multiple layers of an outline hierarchy?
So just there is the idea of a different presentation layer, hierarchy and tabular perhaps.
Then, what kind of barriers do I have to getting at the data? Privacy issues? Copyright or patent issues?
If you want to conne
title misleading (Score:2)
TFA (Score:2)
Like there aren't things to write about like the Open Archives Initiative Protocol [openarchives.org].. Geez.
Copyrights in the way? (Score:2)
Google and Self-joins (Score:2)
[...]or will Google make the art obsolete once they finish indexing everything?
Isn't the value of relational databases in the ability to "relate" indexed datasets? Google doesn't support a "join" syntax, as far as I know.
Even Google's fantastic text indexing doesn't break the data up into the discreet "fields" that would be needed to do any meaningful relating. It's sort of like having all of your data in a single column in a single table, and tryin
Easy answer (Score:2)
We don't know
The problem is both easier and more difficult (Score:4, Insightful)
Data, whether held in databases (usually nice and tidy) or in flatfiles, or random text files spread all over hell's half acre, is simply data, not the information required to link it to other data. Even meta data about the data held in any data store is not the information required to link it to other data.
One of the things I believe will help (possibly) is ODF (buzzword warning sounds) because it begins to help format data in a universally accepted manner. Though it is not the only way, universal access methods are required for accessible data. Second, the structure of the data must be presented in a universal manner. This second part allows query languages to support cognitive understanding of the structure, and thus (with some work) the value of data held in a storage location, where ever and whatever that location is, be it RDBMS, text files, or phone bills.
Indexing is simply not enough. The ability to retrieve and utilize the index with the most probability of having relevent data is what is needed. We all know that any search engine can get you too many 'hits' that contain useless data. Google or anyone else is helpless until there are accepted methods for applying metadata and data structure descriptions on all data.
When there is far more organization to data storage, there will be a great sucking sound of people actually using data from the internet in brand new ways.... until then, its all hit and miss.
How do Google do their queries? (Score:2)
I'd love to know how search engines do do it - anyone reading this worked for one?
42 (Score:4, Insightful)
"I checked it very thoroughly," said the computer, "and that quite definitely is the answer. I think the problem, to be quite honest with you, is that you've never actually known what the question is."-Hitchhiker's Guide to the Galaxy"
One must remember when undertaking to find answers in the data to first figure out the question. Otherwise the answer you find will be as useful to you as the answer 42.
Without context you only have a neat compilation of arranged meaningless facts.
On the small scale data mining is used daily by marketing people and the like to figure out who would be most receptive to their approach. Webmasters use it to optimize content and respond to user trends. In most large corporations data mining is used on some level.
Data mining on the scale discussed here may be practical at some point in the future once we determine the questions we wish answers to.
Let us hope the answer is more useful than 42.
(Machine Learning == Data Mining) does work ! (Score:2, Interesting)
what used to be called 'data-mining' in 80 and 90s is now machine learning in 21st century.. and there are several instances where machine learning has shown tremendous success (probably this is the only by-product of AI that has shown promising real world applications)
- The DARPA Grand Challenge [pbs.org] - Stanely, the winning robot from Stanford used 'Adaptive vision' which used some real-time learning algorithms
- Clustering and Micro-Array Analysis [google.com] - Once genetic-medicine will become a reality, the physicians
Disappointed (Score:2)
easy answer: (Score:2)
data mining will always be a frontier, because consolidaiton and standardization of data will always be a frontier, because simple entropy leads to fragmentation. furthermore, for various reasons, some good, some bad, some data will always be purposefully constrained from consolidation, only to be released into freer usage later, when data mining can commence
it's a permanent frontier
Let the monkeys mine that data (Score:2, Funny)
Too much assumption, too much possibilities, too little knowdledge, and not enough monkeys. You can never have enough friggin' monkeys.
Its long and hard, just to get started (Score:2)
Re:Its long and hard, just to get started (Score:2)
New Use for Google. (Score:2)
As I've Said Repeatedly (Score:2)
A decent (read, relatively effective and efficient) simulation of conceptual processing would change the entire world of computer use from development to databases to computer education to robotics. It is THE world-class issue that needs to be resolved and soon.
A biotech scientists point of view (Score:2)
What would change the field ?
In science, what usually changes peoples minds is a BIOLOGICAL results obtained with a new technique that could not be obtained (easily) another way.
this may just be restating the old truism that success breeds success, but to get biologists interested in large scale
Re:Scooty Puff Jr!! (Score:2)
Re:Scooty Puff Jr!! (Score:2)
Better hurry. Google's already indexed 805,000 pages on "Beavers mate for life".
Re:Please correct your terminology! (Score:2)
You are clearly unaware that the Standard Metric Fuck Ton(ne), which is stored in Paris, France has recently be found to be shrinking at a rate of "Shit-all Squared" per year.
The current US administration has jumpped on this as a pretext to move to the new "God-damned Freedom Ton" which is defined to be exactly equal to 1 original Metric Fuck Tonne, except it is not in any way connected with Fra