top Researchers Create First All Optical Nanowire NAND Gate
It's not that humans are not adaptable, it's that parallel computing is hard for humans to figure out. Linear execution lends itself to all kinds of easy abstractions: loops, branches, methods, etc. Parallel computing, not so much. Mutexes are awful. The best we've got is message passing and functional programming, but even that is hard to design correctly to be both understandable and exploit inherent parallelism.
Y'know what's even harder to design? Analog computing. Holy cow. Remember, digital computing was invented by Touring before we even had built a computer. It's easy to visualize how it works. My brain explodes though trying to imagine a fuzzy-logic analog equivalent of a touring machine.
I used to think that AI research combined with neuroscience would figure out a simple solution to this problem, but it's increasingly seeming like, no, it's even complicated in the brain.
So people can pine for analog memristor computation, and analog optical computing all they want, but the hardware is the easy part here. Get the software side solved, and if you build it they will come. But it's not because we aren't used to these problems, it's because these problems are really really hard.
about a year and a half ago
top Amazon Releases Kindle Source Code
From the GPL:
"a) You must cause the modified files to carry prominent notices stating that you changed the files and the date of any change."
So it is required for them to list their changes.
top Siri Protocol Cracked
Speech recognition isn't too CPU intensive, but it's *massively* memory intensive. It's not unreasonable for speech recognition engines to eat up a gig of ram, and the 4S only has 512mb. However, push it to a server with lots of ram and it can handle lots and lots of simultaneous speech recognition queries. It's tailor made to be a server-side task. At least until phones have gigs of free memory that aren't needed.
top Michio Kaku's Dark Prediction For the End of Moore's Law
Even with transistors the same size, there are so many avenues to explore in processor design.
Just off the top of my head, how about a memristor-based analog co-processor for tasks like facial detection or language/speech recognition.
How about processors with asynchronous clocks, or clockless designs. Sure, they're harder to build, but once transistor sizes fixate, might as well spend the effort because designs will have a much longer lifecycle.
top My Camera ...
There's a few things a DSLR will get you that no point and shoot has.
First, big form factor means big sensor which means good shots in low light/fast exposure. Point-and-shooters are a huge handicap at sporting events for this reason.
Secondly, big lenses allow you to get tight depths of field. With p&s cameras, generally everything in frame is in focus. Being able to use focus to pull your subject out and blur the background is hugely valuable.
top My Camera ...
If you can put a micro four-thirds camera in your pocket you've got some seriously big pockets. But yeah, I hear yah.
top Why Charles Stross Hates Star Trek
Agreed. Star Wars very well could have had a medieval setting and it would have made no real difference to the plot..
I think that movie was called
top Game Over For Sony and Open Source?
Let me preface this answer by revealing that I no longer work in the video game industry, as I did not enjoy it enough to stay.
A lot of people cut their teeth on writing Windows stuff for fun, maybe working on mods, but a fair amount of developers worked their way up from QA. At least where I worked, it seemed like there were way too many people wanting to get into the video games industry, and once they did get in, they worked their asses off. People would learn to code due to their love of games, not because they liked coding. There seemed to be a lot of very bright high-school guys who, instead of doing the whole computer-science thing at a university, would work QA, and then progress up to be a developer. These people were highly respected because of their commitment.
There was another group of people who formed the more senior developers who got started in academia. People who worked on the engines ususally had PHD's in computer science with an emphasis on graphics. I would think graduate work on game theory or AI would put you in this group.
Being an old-school linux hacker who cut his teeth by contributing to OSS projects, I felt a bit out of place. Most of the guys in the industry don't leave because the idea of working on something other than video games is distasteful. Me, I find lots of engineering problems satisfying.
top Game Over For Sony and Open Source?
I used to work for Sony developing PS2 games. The number of people I met that cut their teeth writing code on the linux kit before getting into the business was exactly 0. I might have been the only person I knew who even had a modchipped PS2, everybody else just didn't care since they had the PS2Tool on their desk to do development. Sony is probably discontinuing offering Linux because it didn't spark the development push that they had hoped for.
Still, I would think this would limit the number of supercomputer clusters that use PS3's. You'd think the marketing benefits of being a platform in the top 100 supercomputers would be valuable, but perhaps Sony is still willing to work with academic institutions to make this possible still.
top HP Creates First Hybrid Memristor Chip
The reason that memristors were so quick to market is that they were discovered while researching new fabrication processes. This isn't like, say, carbon nanotubes where they're able to make one or two in a lab. For memristors, they had a fabrication process from day one.
top Google Map To Real Piracy
I heard a report on the Diane Rehm show that that's exactly what they're planning to do. The other approach that is forthcoming is convoys where multiple shipping vessels get together with one armed gunboat to protect them. Blackwater has come forth as a potential candidate for the security contract of future convoys.
top Cisco To Buy Jabber
The weird thing about Jabber Inc. is how irrelevant they are to the XMPP scene. There's a huge array of jabber servers and clients out there, and Jabber Inc. doesn't really have anything to do with any of them. Then there's the whole branding shift from calling it XMPP instead of Jabber. I'm not quite sure what Jabber Inc. brings to the table for Cisco to buy them.
mo has no journal entries.