A new research collaboration between UC Berkeley and MIT on probabilistic query answering. It allows users to specify trade-off between error confidence bound vs time.
17 TB on 100 nodes? That a lot of nodes to hold 17 TB; on average 174 GB-ish of data each.
The speed is super impressive, but using 100 nodes makes this look more like a parallel processing achievement than "big data".
1G/node/sec * 100 nodes * 2 sec = 200G.
17TB is 85 times that number.
A new research collaboration between UC Berkeley and MIT on probabilistic query answering. It allows users to specify trade-off between error confidence bound vs time.