One of the most awesome concepts I've been learning about recently is derailing.
I want to write a full post about this someday, but let's just say that when someone in a marginalized group points out bias, and someone in the privileged group responds "Yeah, but everybody...", it's a very familiar pattern to the minority. That's derailing. The majority don't know what bias looks like to the minority — they just need to listen.
John Halstead at HuffPo:
The problem with being “colorblind” — aside from the fact that we’re not really — is that it is really a white privilege to be able to ignore race. White people like me have the luxury of not paying attention to race — white or black. The reason is because whiteness is treated as the default in our society. Whiteness is not a problem for white people, because it blends into the cultural background.
Black people, on the other hand, don’t have the luxury of being “colorblind.” They live in a culture which constantly reminds them of their Black-ness, which tells them in a million large and small ways that they are not as important as white people, that their lives actually do not matter as much as white lives. Which is why saying “Black Lives Matter” is so important.
“All Lives Matter” is a problem because it refocuses the issue away from systemic racism and Black lives. It distracts and diminishes the message that Black lives matter or that they should matter more than they do.
Cherry-picking online arguments and tearing them down is a favorite pastime of people with a soapbox and either no opponents or a weak platform in need of some easy wins.
But hear me out, these are comments from a friend and their friend, from a religious circle close to mine, and a view of non-believers I think many may relate to, so I'm going to do exactly that.
"For many faith [or, life] is not about finding peace but rather trusting that the struggle is meaningful."
And for atheists, life is essentially meaningless.
Life isn't quite meaningless for atheists. They may find meaning in family, hobbies, work, personal achievements, and more, but, those things provide temporary meaning, and so one is often drifting from place to place, thing to thing. Faith provides tangible, lifelong meaning that carries you through all walks of life.
One may make their own meaning in life but overall, there is no purpose. No plan behind the scenes.
This appears to be meant to illustrate a fundamental lack of value of atheism or, I'd rather say, absence of a belief-positive in god.
I am surprised by, and disagree with, both assessments, though the first moreso.
I don't believe there is a fundamental lack of value in the absence of a god, in fact I believe the opposite and here's why.
Some background about me
I don't believe in a specific god, only the possibility of a god. And I believe in the certainty of much more than we know now. In fact, infinitely unknown -- and therefore likely of what we'd call "spiritual" nature.
This is written as if spoken to the original commenters.
First, assuming that an atheist experiences life the way you imagine them doing so is a fallacy -- called the Pyschologist's fallacy (wikipedia). You're assuming your own objectivity for analyzing not just a fact but someone else's subjectivity.
You don't know whether I experience meaning in my own life or of what nature that meaning is. You can only know that you imagine it doesn't feel like what you'd place meaning in.
How do I find meaning?
Second, I'm here to tell you life feels far more meaningful to me now. I don't know how you experience belief in god, but I experienced it as a life void of my own meaning (purpose). All my usefulness in life was defined externally. There was no need to discover self-reward, or intrinsic meaning, because I was told my actions would bring about a literal external (extrinsic) and eternal reward.
As it was, I was also always putting my actions to the judgment of an external set of rules.
When I decided that I didn't care whether I got a reward or not, as long as I knew that as time passed I had been a good person --
-- by that I mean, no matter what actions I take, I will always try to do better --
-- then I realized this meant I believed I was a truly good person, with a good heart. And with innate good inside, I could realize that external rules were actually a crutch keeping me from developing a better conscience. If I kept leaning on those rules, my internal compass never needed to strengthen. The only thing that would increase would be my confidence of my behavior matching the rules.
Of course, this belief doesn't jive with your core beliefs. That's fine. However, you're stating here that believing otherwise makes life meaningless. But my experience, as the person you are talking about, shows differently.
This change of mind allowed me to value myself and my lived experience as the thing most important at the center of my circles, not an external set of rules and externally promised future reward of which we have never observed.
When I saw that value, it follows that every life has equal value at the center of their own circles, which overlap and affect me. So every lived experience is the meaning itself.
We have no knowledge of what will follow -- only beliefs. So a belief in no god (or the likely absence of a god) actually put the entire meaning of one's experience into life itself, based on -- 1. ourselves and 2. every person we interact with, from each and every person's own perspective.
This is the principle "love your neighbor as yourself." We value what we love, therefore life has meaning.
What kind of meaning is it?
Is it temporary, or less inherently valuable?
An atheist only has possibility of meaning of life by the awareness of their own existence. But the same goes for a believer of God. So the potential for meaning is unlocked for both.
So why is the atheist life, just because one may believe they came from dirt and will return to dirt, ascribed to have less meaning?
A believer in god may suppose that the difference in the "meaning" is (from some statements above):
the ever presence of what their purpose in life is
the everlastingness of their life
the reward to come
a greater purpose, or plan behind the scenes -- an external plan which is about all of us.
but contrast this with me:
1. All of this purpose and belief I'm telling you now also go with me everywhere. If I got knocked out, lost all my things, and taken to a foreign country with just my life and my mind, I'd still have all the same meaning.
2. Maybe a longer existence gives more meaning?
Comparing "forever" to a fixed lifetime is a function of time progressing, or in other words, future events.
If an atheist believed in permanence of their influence, an atheist has the same potential for meaning from everlasting purpose.
But even with a belief in a likely dissipation of their life's influence on others over time, an atheist has only the present to apply sure meaning.
Given that the present is what we actually sense, not future events, surely this means the atheist has as much or more faith in the meaning of their present existence?
3. Maybe it's because faith gives knowledge of an immense worth of the reward?
Whether the reward is valuable because it's ongoing or because the experience is so much better than we now have -- you may conclude differently -- I don't believe that some ongoing reward is better than the current experience.
Life itself; the events, experiences, senses, feelings, thoughts, have an intrinsic reward.
Imagination of experiencing this, or any other good feeling I've had, as a reward forever is not an actual experience, so it does not hold intrinsic value to me.
Therefore, my meaning -- as I experience it -- is greater than what I logically gained when I was trying to live life for the future reward.
4. Maybe the meaning comes from knowing we're all working together, a super-plan or the value is for a master plan or all people?
Again, go back to my first statement. Life is the reward, and therefore all lives have a reward. This value is every bit as great to me as if it were part of plan to which we are all bound.
And if it was some master plan, I don't value the feeling of believing I'm doing something according to a grand design any more than I do the inherent experience. That's just me, but again I am the one who observes and therefore values my own meaning.
Family, hobbies, work, personal achievements are merely the actions I may participate in or seek in this already-meaningful system of values. The people and my own experience are the inherent value, so as I said it goes with me no matter what I have or where I go. Time progresses, and as I experience things, meaning therefore exists.
"Drifting from place to place, thing to thing" actually makes it sound like it adds meaning to me, because the variety of my experience increases, instead of repeatedly driving the same path, becoming more numb to the small differences with each passing.
I'm sure god-believers' lives have meaning under their belief systems; but that doesn't mean that they can know whether mine holds meaning for me or not. And while this perspective may not provide meaning for all without a god-centered system, this is hopefully an explanation of why it does for me.
I wish you would step back from that ledge, my friend.
You could, cut ties with all the lies, that you've been living in.
This song is one of the things that first drew me to Laura. We both loved the band. I could play just 8 notes from the hook, but it was enough to fool her into thinking I was musically gifted.
It's also a thing I think most of us can relate to. Living a lie at some point in our lives.
So, have I been living a lie? Not exactly. For years I've been... ok. Not bad. Pretty good. Bringing kids into the world with her was amazing, and they are guaranteed to light up my life every moment I see them. Even outside that, I didn't have any excuses to be down.
There's been quite a bit of controversy lately about the label "transgender". Pride, beliefs, policies, science. Are they other genders? Are they people born with the wrong gender? Are they mentally ill? Are they disturbed or more likely to be criminals? Can they be fixed?
All those labels, controversies and supposed conclusions have basically zero to do with my experience. And that experience is all I can tell you about.
For years I've experienced the world as "Nick". By that I mean, I don't know what labels are me, I just know that I am me. How I see things.
For a long time I assumed that Nick fit into certain boxes -- boy. Mormon. Programmer. Man.
I don't think that tells the whole story for anyone. However, for me I really started to notice some problems. For what I thought represented "mormonism", there were in fact my own core beliefs, yet mormonism wasn't a great match for it. I didn't want to be "just" a programmer or a "just" a designer. I had a hard time convincing people I knew anything about both. I never felt like I became a "man" the way people saw it, or even saw me.
I tried to twist back into the boxes. I'd downplay my design talent and passion so that an interviewer would believe I was a real programmer. I kept trying to fit in at church. I tried all the masculine fashions I could think of to see which one felt like "me".
The problem was, I had no idea how to be things that made me a good "man". I'd look in the mirror and think "I guess other people might think this works?" Nearly everything I did was to fit someone else's model. And I felt a bit bummed every time I'd notice how far my passions, interests, traits, and behaviors were from the ideals and norms.
That's fine, still, that's what everybody experiences, right?
I began to notice that I was very often jealous of women. I knew exactly what kind of woman I would be. That's just what attraction is, right?
Here's where it will get difficult for me to explain.
The thing about experience is, you can never truly understand it unless you have it. That's why expertise takes time. There is no "The Matrix" download. That's why we aren't born knowing ourselves or the world. That's why people matter to us. Why we talk and listen. There is no formula for transferring a learned experience from one person to another. There has to be an effort by the giver to turn it into language, illustration, feeling, and the receiver to think, hear, question, experiment, and process from different perspectives.
I've been to hundreds of hours of therapy and spent thousands of hours thinking, studying. Planning. Unplanning. During that time, I realized that none of those things would teach me exactly where to go.
For almost a year I've been working to figure out what LIVED experiences actually do fit me the best. Those that are the "Nick" box and not just those I was handed when I was born. One that is perfectly "me-shaped". This was the line my therapist used just before I told her about my questions:
I had just described to her a scene that sounded completely perfect to me, and she said "So, it is completely 'you'-shaped?" I thought about the significance of that concept.
Suddenly it was clear. A box could be ME-shaped? Really?
At that moment I felt brave enough to start a journey that has brought more clarity and confusion, highs and lows, a hundred other extremes and changes that are impossible to overstate.
During that year this has involved me changing the way the world sees me, instead of just fitting into the world.
As I expressed myself more authentically, It was like one lightbulb after another turned on. My world has a brand new sense of visibility. So many new feelings, I didn't know I should or could feel.
I had no idea how rewarding the process would be. Despite my absolute terror in doing so, embracing what I want the world to see and just saying "Ok, world. This is me. Anyone there?" -- I finally was able to feel seen. I wasn't just a fly on the wall any more. I was a participant.
I recognized at the beginning of this year, that this would involve changes that would throw on its head how I was seen by everyone in my life. Through an enormous amount of compassion, my wife showed me she could hear me and I opened up to her. I hurt her. She hurt me. It became turbulence and chaos for both of us that we had no idea if or how we would get through. We are still trying to come through this, and we're now apart, but stronger as individuals, and as friends. I believe we are better parents, and though we are not the same family unit we thought we'd go to our graves as before, we are approaching this new phase with courage and integrity.
To be clear: I look different. I dress like a woman. I'd rather be seen as a woman. Yep. (Am I really? Don't get me started on labels again...) This is because part of recognizing what I really was inside me was about showing it to the world. Growth requires light.
I love my name. It means "victorious people". It's part of who I am. Which is why I've gone with Nicole. It was always a beautiful name to me. And it turns out that, inside, I feel this is much more closely shaped to me than a "Nick/him"-shaped box was.
Femininity is a beautiful, wondrous world to me. So much I don't know, and so much I have always known. I have no idea why we have gender on top of our sexes. I don't know how it works. All I know is that when I get to show the world the femininity inside of me, I feel confident, courageous, beautiful, peaceful, happy, excited, generous, loving, kind, passionate, alive.
Laura wrote this beautiful, heartbreakingly perfect poem about this experience for her:
As she says "I know you must have always been around from the start, at least in whatever way you could be."
Sometimes, I was there. In moments, sometimes captured in an authentic smile on film, with my kids. In a moment of sheer hilarity. Those times Laura and I shared that no other two people will ever understand.
But out in the world, I mostly wasn't there. I was too scared. I didn't know myself.
I don't hate "him". There are lots of "him" things that are truly me. The way I showed myself as him to the world was the best I could do at the time.
But now I can smile. Really smile. Really laugh. Cry, dance, and love. Freely.
There is an amazing song by a band I truly love, Hive Riot. It's called Undercover:
There's a prisoner of war.
Knocking at your door.
Open up and you will see
The face is yours.
Step out from behind the shadows.
Shake your shame and all your sorrow.
Take your place and dance out in the sun.
No one should live undercover. No one is above the other. Freedom is a gift for everyone.
I love the beauty of that video I posted a few days ago, and the accompanying quote:
"We are living in a culture entirely hypnotized by the illusion of time, in which the so-called present moment is felt as nothing but an infintesimal hairline between an all-powerfully causative past and an absorbingly important future. We have no present. Our consciousness is almost completely preoccupied with memory and expectation. We do not realize that there never was, is, nor will be any other experience than present experience. We are therefore out of touch with reality. We confuse the world as talked about, described, and measured with the world which actually is."
This is my "present". I am all at once my past, present and future self. I am not locked into what I have been, and I have no guarantees about what I will be, but they are all me -- here I am, world!
Cool front-end developers are always pushing the envelope, jumping out of their seat to use the latest and greatest and shiniest of UI frameworks and libraries. They are often found bridging the gap between native apps and web and so will strive to make the UI look and behave like an app. Which app? you may ask. iPhone? Android? What version? All good questions, alas another topic altogether. However, there is another kind of front-end developer, the boring front-end developer. Here is an ode to the boring front-end developer, BFED if you will.
The trouble with innovation is that truly innovative ideas often look like bad ideas at the time. That’s why they are innovative — until now, nobody ever figured out that they were good ideas.
They focused on what the technology could not do at the time rather than what it could do and might be able to do in the future.
Don’t hate, create.
- Ben Horowitz
Being smart sounds right. So why should I be more stupid? Because the world is full of smart reasons why stuff can't be done.
Stupid, on the other hand, can only try.
I could sit here literally forever and wait for the perfect pitch, the one that is the easiest to knock out of the park, or I could just start swinging.
This concept reminds me of several quotes from the fantastic "Win Like Stupid" article I read a few months ago (hat tip Jason Waters). Some of my favorites:
TOO STUPID FOR CAN’T
STUPID PEOPLE NOT THINK ABOUT CAN’T WIN AT ALL. THEM JUST DO IDEAS UNTIL ONE CAN.
STUPID PEOPLE TOO DUMB FOR ODDS. THEM JUST ASSUME NEXT THING WILL WORK.
CHANCE IF NOT TRY EXACTLY NOTHING.
EVERY SMART PERSON TERRIFIED EVERYONE THINK THEM IDIOT.
STUPID PERSON ALREADY IS ONE, NOT MIND IF PEOPLE KNOW.
WORLD SMART. IT HARD TO OUTSMART WORLD. BE IDIOT. OUTSTUPID WORLD INSTEAD.
First guy: "Is he texting and watching a video at the same time?"
Second guy: "Hey, what are you doing?"
Third guy: "I'm texting and watching a video at the same time."
First guy: "You can't do that."
Second guy: "You can't do that."
Third guy: "And yet I'm doing it."
First and second guys: "Yeah - nice!"
That's right, this ad touts one of Samsung's newest features: texting and watching a video at the same time.
That should say "screen images simulated. Final version may actually line up correctly".
For an OS that seems to be marketed at robots (queue "Droid" voice), maybe this has been a long-time coming. But for humans like me, for whom input and output rarely mix, and who are already working on a screen the size of their palm, I have my doubts about the real-life usefulness of this feature.
My phone already plays audio at the same time as I can text or use other apps. In contrast to Samsung's latest, one feature I would love is for an easier way to pause the audio. When I'm listening to a podcast, I frequently take notes — but I can't do it at the same time, or I end up rewinding anyway.
But it's a feature, and features sell, right?
Is playing video on the same screen as a text composer difficult? Technically speaking, no. It might be a design challenge to make the experience seamless and handle the different modes. But even then, it's achievable.
It's a great read about avoiding featuritis. Ballard's premise:
Users are not really compelled by features; they care only about achieving their goals. The real cost of adding excess features to digital products is unhappy customers.
I understand that Samsung has to innovate (especially given recent events). But let's remember that innovation doesn't always have to be additive.
The Sad User Slope
Hitching your horse to the "features-sell" wagon is actually a very dangerous game.
This is aptly illustrated in Ballard's article with the Happy User Peak, originally published by respected author and creator of the Head First book series, Kathy Sierra.
Unfortunately, so many of these features actually fall on the Sad User Slope. Where do you think texting while watching a video falls?
What puts a product onto the happy user peak, or the sad user slope? Ballard's definition is simple:
What turns a pleasurable solution into a cumbersome tool is the number of
visible features that are irrelevant to the user’s goal.
Why this happens
These concepts aren't revolutionary. Sierra's post about featuritis is from 2005. Don't make me dig up a post about what phones were like then. (Hint: one term for them is feature phones.)
So, let's assume that Samsung isn't filled with idiots. Why is it so hard to resist adding features where the human value is so low — and then create commercials about them?
Knowing exactly what features will make an elegant solution depends on establishing a deep understanding of the users’ goals. When we fail to know what the users’ behaviors, motivations and goals are, we create complex and burdensome tools. This is because instead of first designing digital products with a real understanding of the user, we cast a wide net of features with the hope that we will capture a wide audience. We wrongly feel that the safe way to develop products is to provide as many features as possible. This is self-defeating to our goal of providing real solutions. The secret to success is to get to know your user and be specific with your solution. In short, focus on your customers’ goals and not on features. (emphasis added)
And why wouldn't we think that our users will want these features? After all, they tell us they do:
Users will almost always say yes to more features. “Would you like?” questions will generally be received with a positive response even if the feature will not help users be successful. The idea that “more is more” is so compelling, and users are unable to visualize the experience that will result from applying
Just say no.
Apple is famous for their focus. Recently, John Gruber linked to a fantastic article titled "The Disciplined Pursuit of Less", written by Greg McKeown for Harvard Business Review. In it, McKeown puts forward what he calls the Clarity Paradox:
Why don’t successful people and organizations automatically become very successful? One important explanation is due to what I call “the clarity paradox,” which can be summed up in four predictable phases:
Phase 1: When we really have clarity of purpose, it leads to success.
Phase 2: When we have success, it leads to more options and opportunities.
Phase 3: When we have increased options and opportunities, it leads to diffused efforts.
Phase 4: Diffused efforts undermine the very clarity that led to our success in the first place.
Curiously, and overstating the point in order to make it, success is a catalyst for failure.
In his own post, Gruber suggested that Apple's uncommon focus has allowed it to keep clarity and overcome this pattern.
With the resulting clarity, Apple keeps its products focused on the goal of providing exactly what users need.
Steve Jobs once said:
I'm as proud of the products we have not done as I am of the products we have done.
Replace products with features and you get a similar point. That sentiment is echoed by Jobs himself, as recounted by Ballard:
In a meeting between Steve Jobs and some record label representatives concerning Apple’s iTunes, Jobs kept fielding questions like "Does it do [x]?" or "Do you plan to add [y]?"
Finally Jobs said, "Wait, wait — put your hands down. Listen: I know you have a thousand ideas for all the cool features iTunes could have. So do we. But we don’t want a thousand features. That would be ugly. Innovation is not about saying yes to everything. It’s about saying NO to all but the most crucial features."
We don't yet know if this feature makes an ugly phone; Samsung's screens are simulated, but even that version doesn't give a lot of hope. Neither does the fact that users are having trouble finding how to use it.
Features don't need to be fancy. They don't need to be gimmicky or do new tricks. They just need to get done what the user wants or needs.
Just say no; and allow the features that are done right the spotlight they deserve.
Stripe is a web payments company whose engineering team get web security. They launched a hacking contest. Joseph Tartaro of IOActive has kindly compiled this writeup of the solutions.
It is a must-read for anyone interested in web security. Wait, scratch that — for anyone who even touches web application code.
In February, the engineering team at Stripe (easy, secure web payments) created the first Stripe Capture the Flag, a "security wargame" intended to test your ability to find exploits in vulnerable code. This event was largely based on understanding of Unix systems, C exploits, with one PHP exploit thrown in.
The original event was a huge success, with attention from Hacker News as well as Reddit (of course).
A few days ago the team released Stripe CTF 2.0 which they are calling "Web Edition". They stepped up the support systems for this one, with logins, a leaderboard, and public code on GitHub. But what's even better is the type of exploits that are covered:
The Retina MacBook is the least repairable laptop we’ve ever taken apart: unlike the previous model, the display is fused to the glass—meaning replacing the LCD requires buying an expensive display assembly. The RAM is now soldered to the logic board—making future memory upgrades impossible. And the battery is glued to the case—requiring customers to mail their laptop to Apple every so often for a $200 replacement. ... The design pattern has serious consequences not only for consumers and the environment, but also for the tech industry as a whole.
And he blames us:
We have consistently voted for hardware that’s thinner rather than upgradeable. But we have to draw a line in the sand somewhere. Our purchasing decisions are telling Apple that we’re happy to buy computers and watch them die on schedule. When we choose a short-lived laptop over a more robust model that’s a quarter of an inch thicker, what does that say about our values?
Actually, what does that say about our values?
First of all, "short-lived" is arguable, and I'd argue for "flat-out wrong". I don't take enough laptops through to the end of their life to be a representative sample, but I've purchased two PC laptops and two MacBook Pros. After two years of use, both PCs were essentially falling apart (hinges, power cords, and basically dead batteries) while the MacBooks were running strong.
My 2008 MacBook Pro did get a not-totally-necessary battery replacement after a year, but my 2010 has run strong for two years. I'd expect nothing less from the Airs or new MacBook Pro. So short-lived might be a relative characterization if anything, and only if you consider the need to pay Apple to replace your battery instead of doing it yourself a "death".
Second, and more important, this thought occurred to me: when we look at futuristic computing devices in movies such as Minority Report or Avatar, do we think "Neat, but those definitely don't look upgradeable. No thanks."
Do we imagine that such progress is achieved through the kind of Luddite thinking that leads people to value "hackability" over never-before-achieved levels of precision and portability?
The quote above is the summation of Wiens' argument that "consumer voting" has pushed Apple down this road, and that we need to draw a line in the name of repairability, upgradability and hackability.
I'd argue that Apple's push toward devices that are more about the human interface and less about the components is a form of a categorical imperative, a rule for acting that has no conditions or qualifications — that there is no line, there is only an endless drive towards progress: more portable devices that get the job done with less thinking about the hardware.
That is what drives descriptions like Apple uses in its product announcements: magical, revolutionary — not hacking and upgrading.
Approved this morning, Mac App Store tool Folderwatch — an app that monitors, syncs and mirrors important files automatically — included a small detail in its update notes, stating “Retina graphics” were a new addition to the app.
Rumor is that more than just the Airs will be getting Retina displays:
We reported earlier in the week that Apple may have plans to announce updates to many of its Mac computers at WWDC. The models to be updated include the 15″ MacBook Pro, the 11″ and 13″ MacBook Air models and the iMac.
All of the various portable models being updated are rumored to be getting a Retina display boost.
I've been thinking for a while that the next round of Macbook Airs are perfectly positioned for retina displays, so I think that's pretty much a given at this point. A colleague of mine is thinking that touchscreen displays might also make an appearance -- which would justify the scrolling direction switch in Mac OS X Lion last year.
Yes, nineteen. Zero is another significant number: the number of times I've downloaded the new version of Chrome.
Jeff Atwood wrote about this phenomenon, and the eventual paradise of the infinite version a year ago, stating that the infinite version is a "not if — but when" eventuality.
Chrome is at one (awesome) end of the software upgrade continuum. I've rarely noticed an update to Chrome, and I've rarely checked the version. A couple of months ago I did, when a Facebook game I was working on suddenly started crashing consistently. The problem was tracked to a specific version in Chrome, and within hours, I heard that an update was released. I checked my version again, and Chrome had already been updated with the fix.
This nearly version-free agility has allowed Chrome to innovate a pace not matched by IE, Firefox, or even Safari. I swear the preferences pane has changed every time I open it (not necessarily a good thing).
At the other end you have the enterprise dinosaurs; the inventory, procurement, accounting systems that are just impossible to get rid of — I'm thinking of one local furniture retailer which still uses point-of-sale and inventory systems with black/green DOS-looking screens on the sales floor.
Most other software industries fall somewhere in between, trying to innovate, update, or even just survive while still paying the backward-compatibility price for technical decisions made in years past.
Check out this gem alive and well (erm, alive, at least) in Windows Vista:
Look familiar? That file/directory widget has seen better days; it made its debut in 1992 with Windows 3.1:
Supporting legacy versions is not just a technical problem, though. For some companies and products it's just in their nature to be backward compatible. And sometimes to great success; take Windows in the PC growth generation, or standards like USB, for example. For some opinionated few, backward incompatibility is downright opposed to success.
And then there's Apple.
Apple may not be a shining example of smooth upgrades, but they aren't shy about doing it.
OK, so they aren't entirely unsentimental — Steve Jobs did hold a funeral for Mac OS 9, about a year after OS X replaced it on desktops.
All of this allows Apple to put progress and innovation above all else.
Steve Jobs, in his famous "Thoughts on Flash", made clear his view that Apple doesn't let attachment to anything stand in the way of platform progress:
We know from painful experience that letting a third party layer of software come between the platform and the developer ultimately results in sub-standard apps and hinders the enhancement and progress of the platform. If developers grow dependent on third party development libraries and tools, they can only take advantage of platform enhancements if and when the third party chooses to adopt the new features. We cannot be at the mercy of a third party deciding if and when they will make our enhancements available to our developers.
This attitude has paid off for iOS. Analysis of app downloads shows that uptake for new major versions is rapid. iOS 4.0 just barely got going before it was traded in for iOS 5.0.
The story is not the same for Android. MG Seigler reveals (via DF) that Android 4.0.x "Ice Cream Sandwich" has only seen 7.1% uptake in 7 months. As Seigler notes, Google is expected to announce the next version this month, which would likely occur before ICS even sees 10% adoption.
So at one end of the upgrade-nirvana spectrum, we have: Google. And at the other end: Google.
How could they be so different?
A lot of people want to draw tight comparisons between the mobile OS wars going on now and the desktop OS wars that Apple lost in the 80s and 90s.
Android is also "winning" in market share, with 51 percent of the US smartphone market vs. iOS's almost 31 percent, according to March 2012 Comscore data. Gartner numbers put Android ahead even further, with 56 and 23 percent, respectively (and this has been the story for some time).
Or is it?
Isn't all smartphone market share equal? What's interesting about these numbers is just how different they are from web usage numbers. Shockingly different.
"Smartphone" market share numbers suggest Android is winning, yet certain usage statistics still put Apple in a comfortable lead. Meanwhile, Apple seems quite comfortable blowing away growth estimates again and again while nearly every other device maker is struggling to turn a profit. What does Apple know that the industry doesn't?
Has anyone else noticed a lack of "dumbphones" (non-smartphones) around them lately? Android phones have largely replaced these non-smartphones or feature phones in the "free" section of major cell phone providers' online stores and store shelves alike. This means customers who walk into a store looking for a new phone or a replacement but not knowing (or really caring) what they are shopping for are often ending up with an Android phone — just like they ended up with feature phones in years past.
That's great for market share. But how is it great for business? Generally speaking, these people don't buy apps. They don't promote or evangelize. They aren't repeat customers. Their primary business value is to the service provider; not to the device maker, not to Google.
And they are extremely unlikely to update their operating system.
The good news to Google is that these users who don't upgrade don't represent a large cost to abandon, should they decide to make innovative, backward-compatibility-breaking changes in future versions.
The bad news is that third party developers will be afraid to support new (and possibly risky) versions, or break compatibility with old versions in order to adopt new features. All reasonable version distribution statistics, such as the new Ice Cream Sandwich numbers, show basically no adoption at all. So why take the risk?
I get the feeling that Apple will be happy to continue to innovate and bring new and better features to its customers; and Android will continue to match feature-for-feature on paper, but not where it counts: in the App Store.
The articles on Rands keep getting longer and longer, and as I’m finishing a piece, I worry, “Is it too long?” I worry about this because we live in a lovely world of 140-character quips and status updates, and I fret about whether I’ll be able to hold your attention, which is precisely the wrong thing to worry about. What I should be worried about is, “Have I written something worthy of your attention?”
Houzz.com on "magic mirrors" — computerized touch surfaces on the mirrors and windows in your home:
Magic mirrors and magic windows — in fact, magic glass surfaces all over the house — will soon become commonplace, thanks to breathtaking advancements in computers, computer interfaces and, of all things, glass.
Count me as a skeptic on the word "soon". This technology barely exists, let alone having a good reason to (yet).
Devices should be getting more mobile, not less. To be successful, innovations should also solve a problem. I don't remember having the urge to check my email while leaning over the bathroom sink. It's also fairly counterproductive to OCD-types like me: "Now introducing smudgy screens all over your house, not just in your pocket."
... there are a lot of folks who think gamification means pulling the worst aspects out of games and shoving them into an application. It’s not. Don’t think of gamification as anything other than clever strategies to motivate someone to learn so they can have fun being productive.
The ever-insightful Lukas Mathis took the thoughtful middle ground. After I read his piece, I started a draft of my own. Now that draft will probably never see the light because, although the whole article is worth reading, Lopp has in three sentences summed up my feelings.
Of course, if your purpose in using gamification is anything other than helping the user enjoy learning and to be productive, then you'd do well to hear from the critics how you might be making your users feel about your software.
Creating or expanding business relationships is not about selling – it’s about establishing trust, rapport, and value creation without selling. ...
Engage me, communicate with me, add value to my business, solve my problems, create opportunity for me, educate me, inform me, but don’t try and sell me – it won’t work. An attempt to sell me insults my intelligence and wastes my time. Think about it; do you like to be sold? News flash – nobody does. Now ask yourself this question, do you like to be helped? Most reasonable people do. The difference between the two positions, while subtle, is very meaningful.
Great article, and a lot to think about. A corporate goal like "increase sales by 50%" can be taken two ways.
One way would be to imagine that more people need to be convinced to buy your product.
The other way is to consider how you can add value and find the customers who most need what you are offering.
Create goals that communicate your actual intended action and aren't open to interpretation. Measurements should be customer happiness levels, not dollars of income (that may or may not have been pried from your customers' unwilling fingers).
Focus on the reason that your product exists and help it help people.
Looks like it's downfall-prediction time in the popularity lifecycle for Facebook.
Geoffrey James of Inc. Magazine published an editorial which argues that Facebook, unlike LinkedIn, is vulnerable to being ditched for "something 'cooler'":
LinkedIn is all about business and people's resumes. Because its scope is limited to fundamentally dull information, LinkedIn is simply not vulnerable to something "cooler."
Sure, somebody could launch a site similar to LinkedIn. (And I'm sure plenty of people have.) But why would the customer base bother to change? Nobody on LinkedIn cares about being cool. LinkedIn's beauty is that it's dull but functional–like email and the telephone.
Fair point. Being popular because you're cool does make you vulnerable to the next cool thing. But he doesn't make good arguments — or any at all, really — for his case that people have been using Facebook because it's cool.
Specifically, this argument falls flat for me:
Consumer-oriented social networking sites are like television networks: People will switch when there's something better on another channel.
Actually, consumer-oriented social networking sites are nothing like television networks. Exclusive content provides customers a reason to use more than one network. And, most importantly, there's no cost to switch to something better on another TV channel — in fact, it's common to switch back and forth.
Facebook, on the other hand, with years of posts, photos, and other social interactions (yes, many of them useless), as well as a large current audience, has a huge cost for a user that wants to "switch".
James does address that:
Frankly, I think it's just one online conversion program away from losing its customer base and becoming the next MySpace.
An "online conversion program" might provide a way to minimize the data loss, but Facebook has a much larger asset: market momentum.
ComScore analysis shows that 69% of North American internet users have Facebook accounts, according to a CNET article. People use Facebook because other people use Facebook — their friends, specifically. That's been one of the primary drivers of its virality, and it is now the reason for its ubiquity. In that sense, Facebook is more like email or the telephone than LinkedIn.
No online conversion program is going to move their friends over to the "next" Facebook.
A dead-simple, free, JSON API for retrieving the city and state from a zip code. And, a sample/beta jQuery plugin to show how easy it is to auto-populate those fields when the zip code is typed. Could this be an end to the years of extra-typing madness?