fallenpegasus: amazon (Default)
Brain control headset for gamers

I saw this discussed at OScon. I'm looking forward to seeing the hardware hackers play with this.

On the other hand, someone suggested "Consider a Hulk game where you have to get angry to Hulk-out."

Keyboards and game controllers resulted in all sorts of interesting forms of RSI. What kind of RSI can people get from a neural interface?
fallenpegasus: amazon (Default)
I remember the first time I saw an animated GIF on a web page. I knew what an animated GIF was, because I had read the GIF spec years before that, but you never really encountered them. Now they've moved from a weird exotic to a standard and often tacky part of everyday experence.

I just saw something today that's amazing, brilliant, cool, and obvious in retrospect (it could have been in the net-pbm tools 10+ years ago, if anyone had thought of it), that is currently research demo, but I predict that in a couple of years, it will be an unremarkable and standard part of the user interface, and UIs that don't have it will seem stupid and clunky.

It's called Seam Carving for Content Aware Image Resizing, and it's a way to resize and rescale images in a way that's useful, fast, headslappingly obvious in retrospect.
fallenpegasus: amazon (Default)

In a decade, on-demand virtualized utility computing will be an invisible utility, part of the vital infrastructure of the technological economy.

People will mostly have forgotten what an enormous pain in the ass provisioning computation was today. Today, we don't truly feel that pain, because it seems "normal", everyone has to suffer it together.

The situation right now is, if you have a delivery van, you have to make your own gasoline. And you have to hire and pay for your own mechanics. Seems stupid, doesn't it? It's amazing that there are any delivery vans at all …

Think of the internet itself, what it did to telecoms.

Twenty-five years ago, if you wanted a high speed data connection to a computer in San Francisco, it was a pain. You'd have to come up with a pile of money, and wait a couple of months, at best. Hardware would be dedicated and provisioned, and then finally you would have your connection. To only one place. That you had to pay (a lot for) every month, no matter how many bits you pushed into it. And you had to pay for more than your actual average use, for burst need capacity.

Today, if you want that same sort of connection, you click a hyperlink, or start a VPN. The data stream is virtualized, packetized, and handled by a dozen routers owned by half a dozen companies. It lasts only for the few seconds or minutes that you need it. When it's done, the underlying hardware has forgotten about it, and is carrying some completely different person's completely different data.

Today, do you need a billion teraops of computation? You do the same thing today that you did for datacom then. That is, spend a pile of money, rent or buy hardware and rackspace in a physical datacenter, wait the lead time for purchasing, installation, and provisioning. And then pay for it month to month, year to year. Including paying for the burst capacity you rarely use.

Look forward ten years. Do for computation what the IP router did for datacom. Applications, components, and and dev tools will be architected for dynamic horizontal and vertical scaling as matter of course. When you suddenly need that billion teraops, the utility computing grid will provide it on demand, runs your instances only as you need them, and when demands change, the underlying hardware goes to work running something completely different.

Owning your own big datacenter will be like owning a network of leased lines. You won't do it unless you really really have to, e.g. for legal mandates, or for massive capacity needs, or if you're the actual utility provider.

For everyone else, hardware is going to be a lot smaller, relative to the amount of computation that you do. And even for that hardware you do own, you're still going to run a dynamic virtualizing grid on top of it, just for your own sanity.

Right now, all the existent utility computing outfits are more or less in startup mode. And they are all currently walled gardens, they think. (This is important, but that is a topic for another essay.). But even already, despite being a fledgling industry, they are massively oversubscribed. Massive unmet demand attracts money looking for investment. That money is starting to pour in, and the capacity is going to appear, seemingly overnight. Everything is going to change. Again.

The best known name in the game right now is Amazon's AWS EC2. But there are at least three startups who've talked to just me so far. And the hot rumor is that MSN's datacenters are about to virtualize and sell access. And then there's Google …

Bah, you say. Who suddenly needs a billion teraops with no warning, anyway?

iLike did, only a few weeks ago. Their Facebook application caught the zeitgeist, hit critical mass, and their userbase and attendant demand soared orders of magnitude in days. They started with 2 servers. Suddenly their 40 reserve were consumed, and they had to beg, borrow, buy, and scrounge hundreds more. news article here.

Marc Andreessen recently wrote it up in his blog. But since Marc is a well-funded VC, looking for places where well-funded VCs are needed, he missed the point. His conclusion was that only companies backed with the money and connections of a well-funded VC will be able to survive and afford the sudden demand crunches / instant capacity buildouts that the next generation of tech companies face.

In his words:

unless you already have, or are prepared to quickly procure, a 100-500+ server infrastructure and everything associated with it — networking gear, storage gear, ISP interconnetions, monitoring systems, firewalls, load balancers, provisioning systems, etc. — and a killer operations team, launching a successful Facebook application may well be a self-defeating proposition.

This is a "success kills" scenario — the good news is you're successful, the bad news is you're flat on your back from what amounts to a self-inflicted denial of service attack, unless you have the money and time and knowledge to tackle the resulting scale challenges.

Will every Facebook application go through this?

No, of course not. The ones that nobody uses will not have this problem.

But the successful ones all will.

The implication is, in my view, quite clear — the Facebook Platform is primarily for use by either big companies, or venture-backed startups with the funding and capability to handle the slightly insane scale requirements. Individual developers are going to have a very hard time taking advantage of it in useful ways.

I disagree. (I'll leave it to [livejournal.com profile] krow to say "Bullshit!").

Instead, what is going to happen is that when a company suddenly needs 100x capacity, the application itself is going to ask the grid for more capacity, and get it. And when the CEO and the other staff comes in from their Memorial Day holiday, they will discover that they now work for a company that's 100x as big, which almost no pain on their part.

When that pain goes away and is forgotten, a whole pile of really cool applications are going to go online that we can barely imagine now. Things that we don't think of because the economics and financing don't work. Yet.

Here are a couple of immediately obvious ones:

  • Some little personal webserver contains a page that gets referenced by SlashDot, or BoingBoing, or Instapundit. Today, that server would get smashed. But instead, the grid-aware webservice provisions a http redirector and a dozen httpd instances, holds up under the load, and then a day later shrinks back down.
  • One person builds a networked game, or a SecondLife toy, that suddenly hits a wave of fashion. Fortunately, he build his server to be scalable and grid aware. Instead of a dozen users, he's serving a million users, without being smashed by either "success kills", or having sold his idea into slavery to the VCs. And when the fickle wave of celebrity passes, no hardware or capital funds get wasted.

This is big. This is important. In ten years it's going to have changed everything.

fallenpegasus: amazon (Default)
In the mid-90s, at the very starting edge of the dotcom boom, I was the lead tech guy behind a project, we put a nationwide yellow pages database on the web, ypol.com.

At the time, nobody else had done it, but we were sure that many other people had had the same idea. And since it's rare for someone to care about finding a plumber or locksmith in another state in the middle of the night, of course I had location based searches from the start.

It was, in fact, my first real encounter with using Oracle, Sybase, and using Perl to interface to a DB. (This was before Perl5, so there was no DBI module, so it was all oraperl.)

It looks like someone has actually patented the idea of location based searches on a online yellow pages database, has put together an investor-based lawsuit machine, and has fired the first salvo by suing Verizon.

I'm pretty damn sure my old YPOL project is prior art. I'm even willing to testify on that point...
fallenpegasus: amazon (Default)
Google Calendar has fixed the annoyance in their UI, where if something spans over midnight, it doesn't appear in the hourly scheduling grid.


Thanks, Google.
fallenpegasus: amazon (Default)
When I book travel on an airline or travel website, there should be a feature to download the itinerary as an iCal.

Other related secondary "shoulds" are to download/SMS/SyncML the itenerary to a cellphone / blackberry / smartphone / PDA, and to insert it into Google Calendar.

When my confirmation is emailed to me, it should have an attachment with the iCal object.

Update: I suggested this to AlaskaAirlines website at the same time I wrote this LJ post. I just got the following email back from them:

Dear Mr. Atwood,
Thank you for your suggestion. We have forwarded your email to the web development team. I have heard that they are looking into this feature.

Customer Care for alaskaair.com

fallenpegasus: amazon (Default)
At the Vancouver PHP Conference this past weekend, and then again last night at Ignite Seattle, there were presentations on OpenID.

OpenID is probably the coolest pieces of net "infrastructure" tech I've seen in a year. I'm really excited. I'm looking at places I can help advance the tech, projects I can plug it into.

I've started burbling it to everyone I know who has the geek chops to understand it.

For you non-tech folks, here is what it means. For every site that is OpenID enabled, you don't set up a password there. Ever. You just plug in your OpenID, and go. First thing each morning you will have to login to your OpenID provider.

A neat part of the tech is that some big company can't just be the central "Big Brother", anybody can be your authorizer, and it just works. One of the biggest authorizors right now is LiveJournal. So any OpenID enabled site, I can put "fallenpegasus.livejournal.com" in as my id, and go. Or I can install MyOpenId on a webserver somewhere, and use it. And because I've put a couple of "magic text" in the <head> of my homepage, I can just use "http://mark.atwood.name" as my id. There is a proposal to use something that looks more like a username.

Moz, FF, and IE will magically it even easier in their new versions. They will notice OpenID login boxes, and pop open a widget that will just let you select one, to save typing.

An excellent detailed howto on how to setup your own pages is here, at OpenID for non-SuperUsers. If you have your own HTML "homepage", and/or have access to a PHP enabled "home webserver", I recommend it to you highly.

The only real problem right now with OpenID is that there are actually more sites that will provide an id than will use one. But that will change. Work is underway to plug this into MediaWiki/Wikipedia, into WordPress and the other OSS blog systems, into things like bbPHP and Gallery2, and into Slash.

I'm looking myself at plugging it into Jabber and into PAM.
fallenpegasus: amazon (Default)
I'm running Linux Fedora Core 5, and the latest standard versions of Firefox and Adobe Flash plug-in. Flash stuff does load and run, but text won't display. This is especially problematic for the text in buttons and wigets, because it makes it very hard to use flash bashed applications.

I've removed the plugin and reinstalled it several times.

Any ideas as to the problem, or at least how to better diagnose it?


fallenpegasus: amazon (Default)
Mark Atwood

September 2017

1011121314 1516


RSS Atom

Most Popular Tags

Style Credit

Expand Cut Tags

No cut tags
Page generated Sep. 26th, 2017 03:40 am
Powered by Dreamwidth Studios