Chapter 12: A Manager’s Guide to the Internet and Telecommunications

12.1 Introduction

There’s all sorts of hidden magic happening whenever you connect to the Internet. But what really makes it possible for you to reach servers halfway around the world in just a fraction of a second? Knowing this is not only flat-out fascinating stuff; it’s also critically important for today’s manager to have at least a working knowledge of how the Internet functions.

That’s because the Internet is a platform of possibilities and a business enabler. Understanding how the Internet and networking works can help you brainstorm new products and services and understand roadblocks that might limit turning your ideas into reality. Marketing professionals who know how the Internet reaches consumers have a better understanding of how technologies can be used to find and target customers. Finance firms that rely on trading speed to move billions in the blink of an eye need to master Internet infrastructure to avoid being swept aside by more nimble market movers. And knowing how the Internet works helps all managers understand where their firms are vulnerable. In most industries today, if your network goes down then you might as well shut your doors and go home; it’s nearly impossible to get anything done if you can’t get online. Managers who know the Net are prepared to take the appropriate steps to secure their firms and keep their organization constantly connected.

12.2 Internet 101: Understanding How the Internet Works

Learning Objectives

After studying this section you should be able to do the following:

  1. Describe how the technologies of the Internet combine to answer these questions: What are you looking for? Where is it? And how do we get there?
  2. Interpret a URL, understand what hosts and domains are, describe how domain registration works, describe cybersquatting, and give examples of conditions that constitute a valid and invalid domain-related trademark dispute.
  3. Describe certain aspects of the Internet infrastructure that are fault-tolerant and support load balancing.
  4. Discuss the role of hosts, domains, IP addresses, and the DNS in making the Internet work.

The Internet is a network of networks—millions of them, actually. If the network at your university, your employer, or in your home has Internet access, it connects to an Internet service provider (ISP)An organization or firm that provides access to the Internet.. Many (but not all) ISPs are big telecommunications companies like Verizon, Comcast, and AT&T. These providers connect to one another, exchanging traffic, and ensuring your messages can get to any other computer that’s online and willing to communicate with you.

The Internet has no center and no one owns it. That’s a good thing. The Internet was designed to be redundant and fault-tolerant—meaning that if one network, connecting wire, or server stops working, everything else should keep on running. Rising from military research and work at educational institutions dating as far back as the 1960s, the Internet really took off in the 1990s, when graphical Web browsing was invented, and much of the Internet’s operating infrastructure was transitioned to be supported by private firms rather than government grants.

Figure 12.1

The Internet is a network of networks, and these networks are connected together. In the diagram above, the “state.edu” campus network is connected to other networks of the Internet via two ISPs: Cogent and Verizon.

Enough history—let’s see how it all works! If you want to communicate with another computer on the Internet then your computer needs to know the answer to three questions: What are you looking for? Where is it? And how do we get there? The computers and software that make up Internet infrastructure can help provide the answers. Let’s look at how it all comes together.

The URL: “What Are You Looking For?”

When you type an address into a Web browser (sometimes called a URLOften used interchangeably with “Web address,” URLs identify resources on the Internet along with the application protocol need to retrieve it. for uniform resource locator), you’re telling your browser what you’re looking for. Figure 12.2 “Anatomy of a Web Address” describes how to read a typical URL.

Figure 12.2 Anatomy of a Web Address

The URL displayed really says, “Use the Web (http://) to find a host server named ‘www’ in the ‘nytimes.com’ network, look in the ‘tech’ directory, and access the ‘index.html’ file.”

The http:// you see at the start of most Web addresses stands for hypertext transfer protocolApplication transfer protocol that allows Web browsers and Web servers to communicate with each other.. A protocolEnables communication by defining the format of data and rules for exchange. is a set of rules for communication—sort of like grammar and vocabulary in a language like English. The http protocol defines how Web browser and Web servers communicate and is designed to be independent from the computer’s hardware and operating system. It doesn’t matter if messages come from a PC, a Mac, a huge mainframe, or a pocket-sized smartphone; if a device speaks to another using a common protocol, then it will be heard and understood.

The Internet supports lots of different applications, and many of these applications use their own application transfer protocol to communicate with each other. The server that holds your e-mail uses something called SMTP, or simple mail transfer protocol, to exchange mail with other e-mail servers throughout the world. FTPApplication transfer protocol that is used to copy files from one computer to another., or file transfer protocol, is used for—you guessed it—file transfer. FTP is how most Web developers upload the Web pages, graphics, and other files for their Web sites. Even the Web uses different protocols. When you surf to an online bank or when you’re ready to enter your payment information at the Web site of an Internet retailer, the http at the beginning of your URL will probably change to https (the “s” is for secure). That means that communications between your browser and server will be encrypted for safe transmission. The beauty of the Internet infrastructure is that any savvy entrepreneur can create a new application that rides on top of the Internet.

Hosts and Domain Names

The next part of the URL in our diagram holds the host and domain name. Think of the domain name as the name of the network you’re trying to connect to, and think of the host as the computer you’re looking for on that network.

Many domains have lots of different hosts. For example, Yahoo!’s main Web site is served from the host named “www” (at the address http://www.yahoo.com), but Yahoo! also runs other hosts including those named “finance” (finance.yahoo.com), “sports” (sports.yahoo.com), and “games” (games.yahoo.com).

Host and Domain Names: A Bit More Complex Than That

While it’s useful to think of a host as a single computer, popular Web sites often have several computers that work together to share the load for incoming requests. Assigning several computers to a host name offers load balancingDistributing a computing or networking workload across multiple systems to avoid congestion and slow performance. and fault toleranceThe ability of a system to continue operation even if a component fails., helping ensure that all visits to a popular site like http://www.google.com won’t overload a single computer, or that Google doesn’t go down if one computer fails.

It’s also possible for a single computer to have several host names. This might be the case if a firm were hosting several Web sites on a single piece of computing hardware.

Some domains are also further broken down into subdomains—many times to represent smaller networks or subgroups within a larger organization. For example, the address http://www.rhsmith.umd.edu is a University of Maryland address with a host “www” located in the subdomain “rhsmith” for the Robert H. Smith School of Business. International URLs might also include a second-level domain classification scheme. British URLs use this scheme, for example, with the BBC carrying the commercial (.co) designation—http://www.bbc.co.uk—and the University of Oxford carrying the academic (.ac) designation—http://www.ox.ac.uk. You can actually go 127 levels deep in assigning subdomains, but that wouldn’t make it easy on those who have to type in a URL that long.

Most Web sites are configured to load a default host, so you can often eliminate the host name if you want to go to the most popular host on a site (the default host is almost always named “www”). Another tip: most browsers will automatically add the “http://” for you, too.

Host and domain names are not case sensitive, so you can use a combination of upper and lower case letters and you’ll still get to your destination.

I Want My Own Domain

You can stake your domain name claim in cyberspace by going through a firm called a domain name registrar. You don’t really buy a domain name; you simply pay a registrar for the right to use that name, with the right renewable over time. While some registrars simply register domain names, others act as Web hosting servicesA firm that provides hardware and services to run the Web sites of others. that are able to run your Web site on their Internet-connected servers for a fee.

Registrars throughout the world are accredited by ICANN (Internet Corporation for Assigning Names and Numbers)Nonprofit organization responsible for managing the Internet’s domain and numbering systems., a nonprofit governance and standards-setting body. Each registrar may be granted the ability to register domain names in one or more of the Net’s generic top-level domains (gTLDs), such as “.com,” “.net,” or “.org.” There are dozens of registrars that can register “.com” domain names, the most popular gTLD.

Some generic top-level domain names, like “.com,” have no restrictions on use, while others limit registration. For example, “.edu” is restricted to U.S.-accredited, postsecondary institutions. ICANN has also announced plans to allow organizations to sponsor their own top-level domains (e.g., “.berlin,” or “.coke”).

There are also separate agencies that handle over 250 different two-character country code top-level domains, or ccTLDs (e.g., “.uk” for the United Kingdom and “.jp” for Japan). Servers or organizations generally don’t need to be housed within a country to use a country code as part of their domain names, leading to a number of creatively named Web sites. The URL-shortening site “bit.ly” uses Libya’s “.ly” top-level domain; many physicians are partial to Moldova’s code (“.md”); and the tiny Pacific island nation of Tuvulu might not have a single broadcast television station, but that doesn’t stop it from licensing its country code to firms that want a “.tv” domain name.K. Maney, “Tuvalu’s Sinking, But Its Domain Is on Solid Ground,” USA Today, April 27, 2004. Recent standards also allow domain names in languages that use non-Latin alphabets such as Arabic and Russian.

Domain name registration is handled on a first-come, first-served basis and all registrars share registration data to ensure that no two firms gain rights to the same name. Start-ups often sport wacky names, partly because so many domains with common words and phrases are already registered to others. While some domain names are held by legitimate businesses, others are registered by investors hoping to resell a name’s rights.

Trade in domain names can be lucrative. For example, the “Insure.com” domain was sold to QuinStreet for $16 million in fall 2009.B. Bosker, “The 11 Most Expensive Domain Names Ever,” The Huffington Post, March 10, 2010. But knowingly registering a domain name to profit from someone else’s firm name or trademark is known as cybersquattingAcquiring a domain name that refers to a firm, individual, product, or trademark, with the goal of exploiting it for financial gain. The practice is illegal in many nations, and ICANN has a dispute resolution mechanism that in some circumstances can strip cybersquatters of registered domains. and that’s illegal. The United States has passed the Anticybersquatting Consumer Protection Act (ACPA), and ICANN has the Domain Name Dispute Resolution Policy that can reach across boarders. Try to extort money by holding a domain name that’s identical to (or in some cases, even similar to) a well-known trademark holder and you could be stripped of your domain name and even fined.

Courts and dispute resolution authorities will sometimes allow a domain that uses the trademark of another organization if it is perceived to have legitimate, nonexploitive reasons for doing so. For example, the now defunct site Verizonreallysucks.com was registered as a protest against the networking giant and was considered fair use since owners didn’t try to extort money from the telecom giant.D. Streitfeld, “Web Site Feuding Enters Constitutional Domain,” The Washington Post, September 11, 2000. However, the courts allowed the owner of the PETA trademark (the organization People for the Ethical Treatment of Animals) to claim the domain name peta.org from original registrant, who had been using that domain to host a site called “People Eating Tasty Animals.”D. McCullagh, “Ethical Treatment of PETA Domain,” Wired, August 25, 2001.

Trying to predict how authorities will rule can be difficult. The musician Sting’s name was thought to be too generic to deserve the rights to Sting.com, but Madonna was able to take back her domain name (for the record, Sting now owns Sting.com).R. Konrad and E. Hansen, “Madonna.com Embroiled in Domain Ownership Spat,” CNET, August 21, 2000. Apple executive Jonathan Ive was denied the right to reclaim domain names incorporating his own name, but that had been registered by another party and without his consent. The publicity-shy design guru wasn’t considered enough of a public figure to warrant protection.D. Morson, “Apple VP Ive Loses Domain Name Bid,” MacWorld, May 12, 2009. And sometimes disputing parties can come to an agreement outside of court or ICANN’s dispute resolution mechanisms. When Canadian teenager Michael Rowe registered a site for his part-time Web design business, a firm south of the border took notice of his domain name—Mikerowesoft.com. The two parties eventually settled in a deal that swapped the domain for an Xbox and a trip to the Microsoft Research Tech Fest.M. Kotadia, “MikeRoweSoft Settles for an Xbox,” CNET, January 26, 2004.

Path Name and File Name

Look to the right of the top-level domain and you might see a slash followed by either a path name, a file name, or both. If a Web address has a path and file name, the path maps to a folder location where the file is stored on the server; the file is the name of the file you’re looking for.

Most Web pages end in “.html,” indicating they are in hypertext markup languageLanguage used to compose Web pages.. While http helps browsers and servers communicate, html is the language used to create and format (render) Web pages. A file, however, doesn’t need to be .html; Web servers can deliver just about any type of file: Acrobat documents (.pdf), PowerPoint documents (.ppt or .pptx), Word docs (.doc or .docx), JPEG graphic images (.jpg), and—as we’ll see in Chapter 13 “Information Security: Barbarians at the Gateway (and Just About Everywhere Else)”—even malware programs that attack your PC. At some Web addresses, the file displays content for every visitor, and at others (like amazon.com), a file will contain programs that run on the Web server to generate custom content just for you.

You don’t always type a path or file name as part of a Web address, but there’s always a file lurking behind the scenes. A Web address without a file name will load content from a default page. For example, when you visit “google.com,” Google automatically pulls up a page called “index.html,” a file that contains the Web page that displays the Google logo, the text entry field, the “Google Search” button, and so on. You might not see it, but it’s there.

Butterfingers, beware! Path and file names are case sensitive—amazon.com/books is considered to be different from amazon.com/BOOKS. Mistype your capital letters after the domain name and you might get a 404 error (the very unfriendly Web server error code that means the document was not found).

IP Addresses and the Domain Name System: “Where Is It? And How Do We Get There?”

The IP Address

If you want to communicate, then you need to have a way for people to find and reach you. Houses and businesses have street addresses, and telephones have phone numbers. Every device connected to the Internet has an identifying address, too—it’s called an IP (Internet protocol) address.

A device gets its IP addressA value used to identify a device that is connected to the Internet. IP addresses are usually expressed as four numbers (from 0 to 255), separated by periods. from whichever organization is currently connecting it to the Internet. Connect using a laptop at your university and your school will assign the laptop’s IP address. Connect at a hotel, and the hotel’s Internet service provider lends your laptop an IP address. Laptops and other end-user machines might get a different IP address each time they connect, but the IP addresses of servers rarely change. It’s OK if you use different IP addresses during different online sessions because services like e-mail and Facebook identify you by your username and password. The IP address simply tells the computers that you’re communicating with where they can find you right now. IP addresses can also be used to identify a user’s physical location, to tailor search results, and to customize advertising. See Chapter 14 “Google: Search, Online Advertising, and Beyond” to learn more.

IP addresses are usually displayed as a string of four numbers between 0 and 255, separated by three periods. Want to know which IP address your smartphone or computer is using? Visit a Web site like ip-adress.com (one “d”), whatismyipaddress.com, or ipchicken.com.

The Internet Is Almost Full

If you do the math, four combinations of 0 to 255 gives you a little over four billion possible IP addresses. Four billion sounds like a lot, but the number of devices connecting to the Internet is exploding! Internet access is now baked into smartphones, tablets, televisions, DVD players, video game consoles, utility meters, thermostats, appliances, picture frames, and more. Another problem is a big chunk of existing addresses weren’t allocated efficiently, and these can’t be easily reclaimed from the corporations, universities, and other organizations that initially received them. All of this means that we’re running out of IP addresses. Experts differ on when ICANN will have no more numbers to dole out, but most believe that time will come by 2012, if not sooner.M. Ward, “Internet Approaches Addressing Limit,” BBC News, May 11, 2010.

There are some schemes to help delay the impact of this IP address drought. For example, a technique known as NAT (network address translation)A technique often used to conserve IP addresses by maps devices on a private network to single Internet-connected device that acts on their behalf. uses a gateway that allows multiple devices to share a single IP address. But NAT slows down Internet access and is complex, cumbersome, and expensive to administer.S. Shankland, “Google Tries to Break IPv6 Logjam by Own Example,” CNET, March 27, 2009.

The only long-term solution is to shift to a new IP scheme. Fortunately, one was developed more than a decade ago. IPv6 increases the possible address space from the 232 (4,294,967,296) addresses used in the current system (called IPv4) to a new theoretical limit of 2128 addresses, which is a really big number—bigger than 34 with 37 zeros after it.

But not all the news is good. Unfortunately, IPv6 isn’t backward compatible with IPv4, and the transition to the new standard has been painfully slow. This gives us the equivalent of many islands of IPv6 in a sea of IPv4, with translation between the two schemes happening when these networks come together. While most modern hardware and operating systems providers now support IPv6, converting a network to IPv6 currently involves a lot of cost with little short-term benefit.S. Shankland, “Google Tries to Break IPv6 Logjam by Own Example,” CNET, March 27, 2009. Upgrading may take years and is likely to result in rollout problems. David Conrad, a general manager at Internet Assigned Numbers Authority (IANA), the agency that grants permission to use IP addresses, has said, “I suspect we are actually beyond a reasonable time frame where there won’t be some disruption. It’s just a question of how much.”B. Arnoldy, “IP Address Shortage to Limit Internet Access,” USA Today, August 3, 2007.

Some organizations have stepped up to try to hasten transition. Google has made most of its services IPv6 accessible, the U.S. government has mandated IPv6 support for most agencies, China has spurred conversion within its borders, and Comcast and Verizon have major IPv6 rollouts under way. While the transition will be slow, when wide scale deployment does arrive, IPv6 will offer other benefits, including potentially improving the speed, reliability, and security of the Internet.

The DNS: The Internet’s Phonebook

You can actually type an IP address of a Web site into a Web browser and that page will show up. But that doesn’t help users much because four sets of numbers are really hard to remember.

This is where the domain name service (DNS)Internet directory service that allows devices and services to be named and discoverable. The DNS, for example, helps your browser locate the appropriate computers when entering an address like http://finance.google.com. comes in. The domain name service is a distributed database that looks up the host and domain names that you enter and returns the actual IP address for the computer that you want to communicate with. It’s like a big, hierarchical set of phone books capable of finding Web servers, e-mail servers, and more. These “phone books” are called nameservers—and when they work together to create the DNS, they can get you anywhere you need to go online.

Figure 12.3

When your computer needs to find the IP address for a host or domain name, it sends a message to a DNS resolver, which looks up the IP address starting at the root nameserver. Once the lookup has taken place, that IP address can be saved in a holding space called a cache, to speed future lookups.

To get a sense of how the DNS works, let’s imagine that you type www.yahoo.com into a Web browser. Your computer doesn’t know where to find that address, but when your computer connected to the network, it learned where to find a service on the network called a DNS resolver. The DNS resolver can look up host/domain name combinations to find the matching IP address using the “phone book” that is the DNS. The resolver doesn’t know everything, but it does know where to start a lookup that will eventually give you the address you’re looking for. If this is the first time anyone on that network has tried to find “www.yahoo.com,” the resolver will contact one of thirteen identical root nameservers. The root acts as a lookup starting place. It doesn’t have one big list, but it can point you to a nameserver for the next level, which would be one of the “.com” nameservers in our example. The “.com” nameserver can then find one of the yahoo.com nameservers. The yahoo.com nameserver can respond to the resolver with the IP address for www.yahoo.com, and the resolver passes that information back to your computer. Once your computer knows Yahoo!’s IP address, it’s then ready to communicate directly with www.yahoo.com. The yahoo.com nameserver includes IP addresses for all Yahoo!’s public sites: www.yahoo.com, games.yahoo.com, sports.yahoo.com, finance.yahoo.com, and so on.

The system also remembers what it’s done so the next time you need the IP address of a host you’ve already looked up, your computer can pull this out of a storage space called a cacheA temporary storage space used to speed computing tasks., avoiding all those nameserver visits. Caches are periodically cleared and refreshed to ensure that data referenced via the DNS stays accurate.

Distributing IP address lookups this way makes sense. It avoids having one huge, hard-to-maintain, and ever-changing list. Firms add and remove hosts on their own networks just by updating entries in their nameserver. And it allows host IP addresses to change easily, too. Moving your Web server off-site to a hosting provider? Just update your nameserver with the new IP address at the hosting provider, and the world will invisibly find that new IP address on the new network by using the same old, familiar host/domain name combination. The DNS is also fault-tolerant—meaning that if one nameserver goes down, the rest of the service can function. There are exact copies at each level, and the system is smart enough to move on to another nameserver if its first choice isn’t responding.

But What If the DNS Gets Hacked?

A hacked DNS would be a disaster! Think about it. If bad guys could change which Web sites load when you type in a host and domain name, they could redirect you to impostor Web sites that look like a bank or e-commerce retailer but are really set up to harvest passwords and credit card data.

This exact scenario played out when the DNS of NET Virtua, a Brazilian Internet service provider, was hacked via a technique called DNS cache poisoning. Cache poisoning exploits a hole in DNS software, redirecting users to sites they didn’t request. The Brazilian DNS hack redirected NET Virtua users wishing to visit the Brazilian bank Bradesco to fraudulent Web sites that attempted to steal passwords and install malware. The hack impacted about 1 percent of the bank’s customers before the attack was discovered.D. Godin, “Cache-Poisoning Attack Snares Top Brazilian Bank,” The Register, April 22, 2009.

The exploit showed the importance of paying attention to security updates. A few months earlier, a group that Wired magazine referred to as “A Secret Geek A-Team”J. Davis, “Secret Geek A-Team Hacks Back, Defends Worldwide Web,” Wired, Nov. 24, 2008. had developed a software update that would have prevented the DNS poisoning exploit used against NET Virtua, but administrators at the Brazilian Internet service provider failed to update their software so the hackers got in. An additional upgrade to a DNS system, known as DNSSEC (domain name service security extensions), promises to further limit the likelihood of cache poisoning, but it may take years for the new standards to be rolled out everywhere.J. Hutchinson, “ICANN, Verisign Place Last Puzzle Pieces in DNSSEC Saga,” NetworkWorld, May 2, 2010.

Key Takeaways

  • The Internet is a network of networks. Internet service providers connect with one another to share traffic, enabling any Internet-connected device to communicate with any other.
  • URLs may list the application protocol, host name, domain name, path name, and file name, in that order. Path and file names are case sensitive.
  • A domain name represents an organization. Hosts are public services offered by that organization. Hosts are often thought of as a single computer, although many computers can operate under a single host name and many hosts can also be run off a single computer.
  • You don’t buy a domain name but can register it, paying for a renewable right to use that domain name. Domains need to be registered within a generic top-level domain such as “.com” or “.org” or within a two-character country code top-level domain such as “.uk,” “.ly,” or “.md.”
  • Registering a domain that uses someone else’s trademark in an attempt to extract financial gain is considered cybersquatting. The United States and other nations have anticybersquatting laws, and ICANN has a dispute resolution system that can overturn domain name claims if a registrant is considered to be cybersquatting.
  • Every device connected to the Internet has an IP address. These addresses are assigned by the organization that connects the user to the Internet. An IP address may be assigned temporarily, for use only during that online session.
  • We’re running out of IP addresses. The current scheme (IPv4) is being replaced by IPv6, a scheme that will give us many more addresses and additional feature benefits but is not backward compatible with the IPv4 standard. Transitioning to IPv6 will be costly and take time.
  • The domain name system is a distributed, fault-tolerant system that uses nameservers to map host/domain name combinations to IP addresses.

Questions and Exercises

  1. Find the Web page for your school’s information systems department. What is the URL that gets you to this page? Label the host name, domain name, path, and file for this URL. Are there additional subdomains? If so, indicate them, as well.
  2. Go to a registrar and see if someone has registered your first or last name as a domain name. If so, what’s hosted at that domain? If not, would you consider registering your name as a domain name? Why or why not?
  3. Investigate cases of domain name disputes. Examine a case that you find especially interesting. Who were the parties involved? How was the issue resolved? Do you agree with the decision?
  4. Describe how the DNS is fault-tolerant and promotes load balancing. Give examples of other types of information systems that might need to be fault-tolerant and offer load balancing. Why?
  5. Research DNS poisoning online. List a case, other than the one mentioned in this chapter, where DNS poisoning took place. Which network was poisoned, who were the victims, and how did hackers exploit the poisoned system? Could this exploit have been stopped? How? Whose responsibility is it to stop these kinds of attacks?
  6. Why is the switch from IPv4 to IPv6 so difficult? What key principles, discussed in prior chapters, are slowing migration to the new standard?

12.3 Getting Where You’re Going

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the layers that make up the Internet—application protocol, transmission control protocol, and Internet protocol—and describe why each is important.
  2. Discuss the benefits of Internet architecture in general and TCP/IP in particular.
  3. Name applications that should use TCP and others that might use UDP.
  4. Understand what a router does and the role these devices play in networking.
  5. Conduct a traceroute and discuss the output, demonstrating how Internet interconnections work in getting messages from point to point.
  6. Understand why mastery of Internet infrastructure is critical to modern finance and be able to discuss the risks in automated trading systems.
  7. Describe VoIP, and contrast circuit versus packet switching, along with organizational benefits and limitations of each.

TCP/IP: The Internet’s Secret Sauce

OK, we know how to read a Web address, we know that every device connected to the Net needs an IP address, and we know that the DNS can look at a Web address and find the IP address of the machine that you want to communicate with. But how does a Web page, an e-mail, or an iTunes download actually get from a remote computer to your desktop?

For our next part of the Internet journey, we’ll learn about two additional protocols: TCP and IP. These protocols are often written as TCP/IP and pronounced by reading all five letters in a row, “T-C-P-I-P” (sometimes they’re also referred to as the Internet protocol suite). TCP and IP are built into any device that a user would use to connect to the Internet—from handhelds to desktops to supercomputers—and together TCP/IP make Internet working happen.

Figure 12.4 TCP/IP in Action

In this example, a server on the left sends a Web page to the user on the right. The application (the Web server) passes the contents of the page to TCP (which is built into the server’s operating system). TCP slices the Web page into packets. Then IP takes over, forwarding packets from router to router across the Internet until it arrives at the user’s PC. Packets sometimes take different routes, and occasionally arrive out of order. TCP running on the receiving system on the right checks that all packets have arrived, requests that damaged or lost packets be resent, puts them in the right order, and sends a perfect, exact copy of the Web page to your browser.

TCP and IP operate below http and the other application transfer protocols mentioned earlier. TCP (transmission control protocol)Works at both ends of most Internet communication to ensure a perfect copy of a message is sent. works its magic at the start and endpoint of the trip—on both your computer and on the destination computer you’re communicating with. Let’s say a Web server wants to send you a large Web page. The Web server application hands the Web page it wants to send to its own version of TCP. TCP then slices up the Web page into smaller chunks of data called packets (or datagrams)A unit of data forwarded by a network. All Internet transmissions—URLs, Web pages, e-mails—are divided into one or more packets.. The packets are like little envelopes containing part of the entire transmission—they’re labeled with a destination address (where it’s going) and a source address (where it came from). Now we’ll leave TCP for a second, because TCP on the Web server then hands those packets off to the second half of our dynamic duo, IP.

It’s the job of IP (Internet protocol)Routing protocol that is in charge of forwarding packets on the Internet. to route the packets to their final destination, and those packets might have to travel over several networks to get to where they’re going. The relay work is done via special computers called routersA computing device that connects networks and exchanges data between them., and these routers speak to each other and to other computers using IP (since routers are connected to the Internet, they have IP addresses, too. Some are even named). Every computer on the Internet is connected to a router, and all routers are connected to at least one (and usually more than one) other router, linking up the networks that make up the Internet.

Routers don’t have perfect, end-to-end information on all points in the Internet, but they do talk to each other all the time, so a router has a pretty good idea of where to send a packet to get it closer to where it needs to end up. This chatter between the routers also keeps the Internet decentralized and fault-tolerant. Even if one path out of a router goes down (a networking cable gets cut, a router breaks, the power to a router goes out), as long as there’s another connection out of that router, then your packet will get forwarded. Networks fail, so good, fault-tolerant network design involves having alternate paths into and out of a network.

Once packets are received by the destination computer (your computer in our example), that machine’s version of TCP kicks in. TCP checks that it has all the packets, makes sure that no packets were damaged or corrupted, requests replacement packets (if needed), and then puts the packets in the correct order, passing a perfect copy of your transmission to the program you’re communicating with (an e-mail server, Web server, etc.).

This progression—application at the source to TCP at the source (slice up the data being sent), to IP (for forwarding among routers), to TCP at the destination (put the transmission back together and make sure it’s perfect), to application at the destination—takes place in both directions, starting at the server for messages coming to you, and starting on your computer when you’re sending messages to another computer.

UDP: TCP’s Faster, Less Reliable Sibling

TCP is a perfectionist and that’s what you want for Web transmissions, e-mail, and application downloads. But sometimes we’re willing to sacrifice perfection for speed. You’d make this sacrifice for streaming media applications like Windows Media Player, Real Player, Internet voice chat, and video conferencing. Having to wait to make sure each packet is perfectly sent would otherwise lead to awkward pauses that interrupt real-time listening. It’d be better to just grab the packets as they come and play them, even if they have minor errors. Packets are small enough that if one packet doesn’t arrive, you can ignore it and move on to the next without too much quality disruption. A protocol called UDP (user datagram protocol)Protocol that operates instead of TCP in applications where delivery speed is important and quality can be sacrificed. does exactly this, working as a TCP stand-in when you’ve got the need for speed, and are willing to sacrifice quality. If you’ve ever watched a Web video or had a Web-based phone call and the quality got sketchy, it’s probably because there were packet problems, but UDP kept on chugging, making the “get it fast” instead of “get it perfect” trade-off.

VoIP: When Phone Calls Are Just Another Internet Application

The increasing speed and reliability of the Internet means that applications such as Internet phone calls (referred to as VoIP, or voice over Internet protocolTransmission technologies that enable voice communications (phone calls) to take place over the Internet as well as private packet-switched networks.) are becoming more reliable. That doesn’t just mean that Skype becomes a more viable alternative for consumer landline and mobile phone calls; it’s also good news for many businesses, governments, and nonprofits.

Many large organizations maintain two networks—one for data and another for POTS (plain old telephone service). Maintaining two networks is expensive, and while conventional phone calls are usually of a higher quality than their Internet counterparts, POTS equipment is also inefficient. Old phone systems use a technology called circuit switching. A “circuit” is a dedicated connection between two entities. When you have a POTS phone call, a circuit is open, dedicating a specific amount of capacity between you and the party on the other end. You’re using that “circuit” regardless of whether you’re talking. Pause between words or put someone on hold, and the circuit is still in use. Anyone who has ever tried to make a phone call at a busy time (say, early morning on Mother’s Day or at midnight on New Year’s Eve) and received an “all circuits are busy” recording has experienced congestion on an inefficient circuit-switched phone network.

But unlike circuit-switched counterparts, Internet networks are packet-switched networks, which can be more efficient. Since we can slice conversations up into packets, we can squeeze them into smaller spaces. If there are pauses in a conversation or someone’s on hold, applications don’t hold up the network. And that creates an opportunity to use the network’s available capacity for other users. The trade-off is one that swaps circuit switching’s quality of service (QoS) with packet switching’s efficiency and cost savings. Try to have a VoIP call when there’s too much traffic on a portion of the network and your call quality will drop. But packet switching quality is getting much better. Networking standards are now offering special features, such as “packet prioritization,” that can allow voice packets to gain delivery priority over packets for applications like e-mail, where a slight delay is OK.

When voice is digitized, “telephone service” simply becomes another application that sits on top of the Internet, like the Web, e-mail, or FTP. VoIP calls between remote offices can save long distance charges. And when the phone system becomes a computer application, you can do a lot more. Well-implemented VoIP systems allow users’ browsers access to their voice mail inbox, one-click video conferencing and call forwarding, point-and-click conference call setup, and other features, but you’ll still have a phone number, just like with POTS.

What Connects the Routers and Computers?

Routers are connected together, either via cables or wirelessly. A cable connecting a computer in a home or office is probably copper (likely what’s usually called an Ethernet cable), with transmissions sent through the copper via electricity. Long-haul cables, those that carry lots of data over long distances, are usually fiber-optic lines—glass lined cables that transmit light (light is faster and travels farther distances than electricity, but fiber-optic networking equipment is more expensive than the copper-electricity kind). Wireless transmission can happen via Wi-Fi (for shorter distances), or cell phone tower or satellite over longer distances. But the beauty of the Internet protocol suite (TCP/IP) is that it doesn’t matter what the actual transmission media are. As long as your routing equipment can connect any two networks, and as long as that equipment “speaks” IP, then you can be part of the Internet.

In reality, your messages likely transfer via lots of different transmission media to get to their final destination. If you use a laptop connected via Wi-Fi, then that wireless connection finds a base station, usually within about three hundred feet. That base station is probably connected to a local area network (LAN) via a copper cable. And your firm or college may connect to fast, long-haul portions of the Internet via fiber-optic cables provided by that firm’s Internet service provider (ISP).

Most big organizations have multiple ISPs for redundancy, providing multiple paths in and out of a network. This is so that if a network connection provided by one firm goes down, say an errant backhoe cuts a cable, other connections can route around the problem (see Figure 12.1).

In the United States (and in most deregulated telecommunications markets), Internet service providers come in all sizes, from smaller regional players to sprawling international firms. When different ISPs connect their networking equipment together to share traffic, it’s called peeringWhen separate ISPs link their networks to swap traffic on the Internet.. Peering usually takes place at neutral sites called Internet exchange points (IXPs), although some firms also have private peering points. Carriers usually don’t charge one another for peering. Instead, “the money is made” in the ISP business by charging the end-points in a network—the customer organizations and end users that an ISP connects to the Internet. Competition among carriers helps keep prices down, quality high, and innovation moving forward.

Finance Has a Need for Speed

When many folks think of Wall Street trading, they think of the open outcry pit at the New York Stock Exchange (NYSE). But human traders are just too slow for many of the most active trading firms. Over half of all U.S. stock trades and a quarter of worldwide currency trades now happen via programs that make trading decisions without any human intervention.H. Timmons, “A London Hedge Fund that Opts for Engineers, Not M.B.A.’s,” New York Times, August 18, 2006. There are many names for this automated, data-driven frontier of finance—algorithmic trading, black-box trading, or high-frequency trading. And while firms specializing in automated, high-frequency trading represent only about 2 percent of the trading firms operating in the United States, they account for about three quarters of all U.S. equity trading volume.R. Iati, “The Real Story of Trading Software Espionage,” Advanced Trading, July 10, 2009.

Programmers lie at the heart of modern finance. “A geek who writes code—those guys are now the valuable guys” says the former head of markets systems at Fidelity Investments, and that rare breed of top programmer can make “tens of millions of dollars” developing these systems.A. Berenson, “Arrest Over Software Illuminates Wall St. Secret,” New York Times, August 23, 2009. Such systems leverage data mining and other model-building techniques to crunch massive volumes of data and discover exploitable market patterns. Models are then run against real-time data and executed the instant a trading opportunity is detected. (For more details on how data is gathered and models are built, see Chapter 11 “The Data Asset: Databases, Business Intelligence, and Competitive Advantage”.)

Winning with these systems means being quick—very quick. Suffer delay (what techies call latency) and you may have missed your opportunity to pounce on a signal or market imperfection. To cut latency, many trading firms are moving their servers out of their own data centers and into colocation facilities. These facilities act as storage places where a firm’s servers get superfast connections as close to the action as possible. And by renting space in a “colo,” a firm gets someone else to manage the electrical and cooling issues, often providing more robust power backup and lower energy costs than a firm might get on its own.

Equinix, a major publicly traded IXP and colocation firm with facilities worldwide, has added a growing number of high-frequency trading firms to a roster of customers that includes e-commerce, Internet, software, and telecom companies. In northern New Jersey alone (the location of many of the servers where “Wall Street” trading takes place), Equinix hosts some eighteen exchanges and trading platforms as well as the NYSE Secure Financial Transaction Infrastructure (SFTI) access node.

Less than a decade ago, eighty milliseconds was acceptably low latency, but now trading firms are pushing below one millisecond into microseconds.I. Schmerken, “High-Frequency Trading Shops Play the Colocation Game,” Advanced Trading, October 5, 2009. So it’s pretty clear that understanding how the Internet works, and how to best exploit it, is of fundamental and strategic importance to those in finance. But also recognize that this kind of automated trading comes with risks. Systems that run on their own can move many billions in the blink of an eye, and the actions of one system may cascade, triggering actions by others.

The spring 2010 “Flash Crash” resulted in a nearly 1,000-point freefall in the Dow Jones Industrial Index, it’s biggest intraday drop ever. Those black boxes can be mysterious—weeks after the May 6th event, experts were still parsing through trading records, trying to unearth how the flash crash happened.E. Daimler and G. Davis, “‘Flash Crash’ Proves Diversity Needed in Market Mechanisms,” Pittsburgh Post-Gazette, May 29, 2010. Regulators and lawmakers recognize they now need to understand technology, telecommunications, and its broader impact on society so that they can create platforms that fuel growth without putting the economy at risk.

Watching the Packet Path via Traceroute

Want to see how packets bounce from router to router as they travel around the Internet? Check out a tool called traceroute. Traceroute repeatedly sends a cluster of three packets starting at the first router connected to a computer, then the next, and so on, building out the path that packets take to their destination.

Traceroute is built into all major desktop operating systems (Windows, Macs, Linux), and several Web sites will run traceroute between locations (traceroute.org and visualroute.visualware.com are great places to explore).

The message below shows a traceroute performed between Irish firm VistaTEC and Boston College. At first, it looks like a bunch of gibberish, but if we look closely, we can decipher what’s going on.

Figure 12.5

The table above shows ten hops, starting at a domain in vistatec.ie and ending in 136.167.9.226 (the table doesn’t say this, but all IP addresses starting with 136.167 are Boston College addresses). The three groups of numbers at the end of three lines shows the time (in milliseconds) of three packets sent out to test that hop of our journey. These numbers might be interesting for network administrators trying to diagnose speed issues, but we’ll ignore them and focus on how packets get from point to point.

At the start of each line is the name of the computer or router that is relaying packets for that leg of the journey. Sometimes routers are named, and sometimes they’re just IP addresses. When routers are named, we can tell what network a packet is on by looking at the domain name. By looking at the router names to the left of each line in the traceroute above, we see that the first two hops are within the vistatec.ie network. Hop 3 shows the first router outside the vistatec.ie network. It’s at a domain named tinet.net, so this must be the name of VistaTEC’s Internet service provider since it’s the first connection outside the vistatec.ie network.

Sometimes routers names suggest their locations (oftentimes they use the same three character abbreviations you’d see in airports). Look closely at the hosts in hops 3 through 7. The subdomains dub20, lon11, lon01, jfk02, and bos01 suggest the packets are going from Dublin, then east to London, then west to New York City (John F. Kennedy International Airport), then north to Boston. That’s a long way to travel in a fraction of a second!

Hop 4 is at tinet.net, but hop 5 is at cogentco.com (look them up online and you’ll find out that cogentco.com, like tinet.net, is also an ISP). That suggests that between those hops peering is taking place and traffic is handed off from carrier to carrier.

Hop 8 is still cogentco.com, but it’s not clear who the unnamed router in hop 9, 38.104.218.10, belongs to. We can use the Internet to sleuth that out, too. Search the Internet for the phrase “IP address lookup” and you’ll find a bunch of tools to track down the organization that “owns” an IP address. Using the tool at whatismyip.com, I found that this number is registered to PSI Net, which is now part of cogentco.com.

Routing paths, ISPs, and peering all revealed via traceroute. You’ve just performed a sort of network “CAT scan” and looked into the veins and arteries that make up a portion of the Internet. Pretty cool!

If you try out traceroute on your own, be aware that not all routers and networks are traceroute friendly. It’s possible that as your trace hits some hops along the way (particularly at the start or end of your journey), three “*” characters will show up at the end of each line instead of the numbers indicating packet speed. This indicates that traceroute has timed out on that hop. Some networks block traceroute because hackers have used the tool to probe a network to figure out how to attack an organization. Most of the time, though, the hops between the source and destination of the traceroute (the steps involving all the ISPs and their routers) are visible.

Traceroute can be a neat way to explore how the Internet works and reinforce the topics we’ve just learned. Search for traceroute tools online or browse the Internet for details on how to use the traceroute command built into your computer.

There’s Another Internet?

If you’re a student at a large research university, there’s a good chance that your school is part of Internet2. Internet2 is a research network created by a consortium of research, academic, industry, and government firms. These organizations have collectively set up a high-performance network running at speeds of up to one hundred gigabits per second to support and experiment with demanding applications. Examples include high-quality video conferencing; high-reliability, high-bandwidth imaging for the medical field; and applications that share huge data sets among researchers.

If your university is an Internet2 member and you’re communicating with another computer that’s part of the Internet2 consortium, then your organization’s routers are smart enough to route traffic through the superfast Internet2 backbone. If that’s the case, you’re likely already using Internet2 without even knowing it!

Key Takeaways

  • TCP/IP, or the Internet protocol suite, helps get perfect copies of Internet transmissions from one location to another. TCP works on the ends of transmission, breaking up transmissions up into manageable packets at the start and putting them back together while checking quality at the end. IP works in the middle, routing packets to their destination.
  • Routers are special computing devices that forward packets from one location to the next. Routers are typically connected with more than one outbound path, so in case one path becomes unavailable, an alternate path can be used.
  • UDP is a replacement for TCP, used when it makes sense to sacrifice packet quality for delivery speed. It’s often used for media streaming.
  • TCP/IP doesn’t care about the transition media. This allows networks of different types—copper, fiber, and wireless—to connect to and participate in the Internet.
  • The ability to swap in new applications, protocols, and media files gives the network tremendous flexibility.
  • Decentralization, fault tolerance, and redundancy help keep the network open and reliable.
  • VoIP allows voice and phone systems to become an application traveling over the Internet. This is allowing many firms to save money on phone calls and through the elimination of old, inefficient circuit-switched networks. As Internet applications, VoIP phone systems can also have additional features that circuit-switched networks lack. The primary limitation of many VoIP systems is quality of service.
  • Many firms in the finance industry have developed automated trading models that analyze data and execute trades without human intervention. Speeds substantially less than one second may be vital to capitalizing on market opportunities, so firms are increasingly moving equipment into collocation facilities that provide high-speed connectivity to other trading systems.

Questions and Exercises

  1. How can the Internet consist of networks of such physically different transmission media—cable, fiber, and wireless?
  2. What is the difference between TCP and UDP? Why would you use one over the other?
  3. Would you recommend a VoIP phone system to your firm or University? Why or why not? What are the advantages? What are the disadvantages? Can you think of possible concerns or benefits not mentioned in this section? Research these concerns online and share your finding with your instructor.
  4. What are the risks in the kinds of automated trading systems described in this section? Conduct research and find an example of where these systems have caused problems for firms and/or the broader market. What can be done to prevent such problems? Whose responsibility is this?
  5. Search the Internet for a traceroute tool, or look online to figure out how to use the traceroute command built into your PC. Run three or more traceroutes to different firms at different locations around the world. List the number of ISPs that show up in the trace. Circle the areas where peering occurs. Do some of the “hops” time out with “*” values returned? If so, why do you think that happened?
  6. Find out if your school or employer is an Internet2 member. If it is, run traceroutes to schools that are and are not members of Internet2. What differences do you see in the results?

12.4 Last Mile: Faster Speed, Broader Access

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the last-mile problem and be able to discuss the pros and cons of various broadband technologies, including DSL, cable, fiber, and various wireless offerings.
  2. Describe 3G and 4G systems, listing major technologies and their backers.
  3. Understand the issue of Net neutrality and put forth arguments supporting or criticizing the concept.

The Internet backboneHigh-speed data lines provided by many firms all across the world that interconnect and collectively form the core of the Internet. is made of fiber-optic lines that carry data traffic over long distances. Those lines are pretty speedy. In fact, several backbone providers, including AT&T and Verizon, are rolling out infrastructure with 100 Gbps transmission speeds (that’s enough to transmit a two-hour high-definition [HD] movie in about eight seconds).T. Spangler, “Cisco Clarifies 100-Gig AT&T Backbone Claim,” Multichannel News, March 9, 2010; Zacks.com, “AT&T Tests 100 Gb Ethernet in Move toward Faster Internet,” SeekingAlpha, March 10, 2010. But when considering overall network speed, remember Amdahl’s LawA system’s speed is determined by its slowest component.: a system’s speed is determined by its slowest component.G. Gilder, Telecosm: How Infinite Bandwidth Will Revolutionize Our World (New York: Free Press, 2000). More often than not, the bottleneck isn’t the backbone but the so-called last mileTechnologies that connect end users to the Internet. The last-mile problem refers to the fact that these connections are usually the slowest part of the network., or the connections that customers use to get online.

High-speed last-mile technologies are often referred to as broadband Internet access (or just broadbandBroadly refers to high-speed Internet connections and is often applied to “last-mile” technologies.). What qualifies as broadband varies. In 2009, the Federal Communications Commission (FCC) redefined broadband as having a minimum speed of 768 Kbps (roughly fourteen times the speed of those old 56 Kbps modems). Other agencies worldwide may have different definitions. But one thing is clear: a new generation of bandwidth-demanding services requires more capacity. As we increasingly consume Internet services like HD streaming, real-time gaming, video conferencing, and music downloads, we are in fact becoming a bunch of voracious, bit-craving gluttons.

With the pivotal role the United States has played in the creation of the Internet, and in pioneering software, hardware, and telecommunications industries, you might expect the United States to lead the world in last-mile broadband access. Not even close. A recent study ranked the United States twenty-sixth in download speeds,S. Lawson, “US Ranks 26th in New Broadband Index,” Computerworld, May 25, 2010. while others have ranked the United States far behind in speed, availability, and price.S. Hansell, “The Broadband Gap: Why Is Theirs Faster?” New York Times, March 10, 2009.

Sounds grim, but help is on the way. A range of technologies and firms are upgrading infrastructure and developing new systems that will increase capacity not just in the United States but also worldwide. Here’s an overview of some of the major technologies that can be used to speed the Internet’s last mile.

Understanding Bandwidth

When folks talk about bandwidthNetwork transmission speeds, typically expressed in some form of bits per second (bps)., they’re referring to data transmission speeds. Bandwidth is often expressed in bits per second, or bps. Prefix letters associated with multiples of bps are the same as the prefixes we mentioned in Chapter 5 “Moore’s Law: Fast, Cheap Computing and What It Means for the Manager” when discussing storage capacity in bytes: Kbps = thousand bits (or kilobits) per second, Mbps = million bits (or megabits) per second, Gbps = billion bits (or gigabits) per second (or terabit), and Tbps = trillion bits (or terabits) per second.

Remember, there are eight bits in a byte, and one byte is a single character. One megabyte is roughly equivalent to one digital book, forty-five seconds of music, or twenty seconds of medium-quality video.R. Farzad, “The Truth about Bandwidth,” BusinessWeek, February 3, 2010. But you can’t just divide the amount of bytes by eight to estimate how many bits you’ll need to transfer. When a file or other transmission is sliced into packets (usually of no more than about 1,500 bytes), there’s some overhead added. Those packets “wrap” data chunks in an envelope surrounded by source and destination addressing and other important information.

Here are some rough demand requirements for streaming media. For streaming audio like Pandora, you’d need at least 150 Kbps for acceptable regular quality, and at least 300 Kbps for high quality.Pandora, “Frequently Asked Questions,” http://blog.pandora.com/faq. For streaming video (via Netflix), at a minimum you’d need 1.5 Mbps, but 3.0 Mbps will ensure decent video and audio. For what Netflix calls HD streaming, you’ll need a minimum of 5 Mbps, but would likely want 8 Mbps or more to ensure the highest quality video and audio.LG Knowledge Base, “Bandwidth Needed for Instant Streaming,” http://lgknowledgebase.com/kb/index.php?View=entry&EntryID=6241.

Cable Broadband

Roughly 90 percent of U.S. homes are serviced by a cable provider, each capable of using a thick copper wire to offer broadband access. That wire (called a coaxial cableInsulated copper cable commonly used by cable television providers. or coax) has shielding that reduces electrical interference, allowing cable signals to travel longer distances without degrading and with less chance of interference than conventional telephone equipment.

One potential weakness of cable technology lies in the fact that most residential providers use a system that requires customers to share bandwidth with neighbors. If the guy next door is a BitTorrent-using bandwidth hog, your traffic could suffer.R. Thompson, “DSL Internet vs. Cable Internet,” High Speed Internet Access Guide, March 23, 2010.

Cable is fast and it’s getting faster. Many cable firms are rolling out a new technology called DOCSIS 3.0 that offers speeds up to and exceeding 50 Mbps (previous high-end speeds were about 16 Mbps and often much less than that). Cable firms are also creating so-called fiber-copper hybrids that run higher-speed fiber-optic lines into neighborhoods, then use lower-cost, but still relatively high-speed, copper infrastructure over short distances to homes.S. Hansell, “The Broadband Gap: Why Is Theirs Faster?” New York Times, March 10, 2009. Those are fast networks, but they are also very expensive to build, since cable firms are laying entirely new lines into neighborhoods instead of leveraging the infrastructure that they’ve already got in place.

DSL: Phone Company Copper

Digital subscriber line (DSL)Broadband technology that uses the wires of a local telephone network. technology uses the copper wire the phone company has already run into most homes. Even as customers worldwide are dropping their landline phone numbers, the wires used to provide this infrastructure can still be used for broadband.

DSL speeds vary depending on the technology deployed. Worldwide speeds may range from 7 Mbps to as much as 100 Mbps (albeit over very short distances).S. Hansell, “The Broadband Gap: Why Is Theirs Faster?” New York Times, March 10, 2009. The Achilles heel of the technology lies in the fact that DSL uses standard copper telephone wiring. These lines lack the shielding used by cable, so signals begin to degrade the further you are from the connecting equipment in telephone company offices. Speeds drop off significantly at less than two miles from a central office or DSL hub. If you go four miles out, the technology becomes unusable. Some DSL providers are also using a hybrid fiber-copper system, but as with cable’s copper hybrids, this is expensive to build.

The superspeedy DSL implementations that are popular in Europe and Asia work because foreign cities are densely populated and so many high-value customers can be accessed over short distances. In South Korea, for example, half the population lives in apartments, and most of those customers live in and around Seoul. This density also impacts costs—since so many people live in apartments, foreign carriers run fewer lines to reach customers, digging up less ground or stringing wires across fewer telephone poles. Their U.S. counterparts by contrast need to reach a customer base sprawled across the suburbs, so U.S. firms have much higher infrastructure costs. S. Hansell, “The Broadband Gap: Why Is Theirs Faster?” New York Times, March 10, 2009.

There’s another company with copper, electricity-carrying cables coming into your home—the electrical utility. BPL, or broadband over power line, technology has been available for years. However, there are few deployments because it is considered to be pricier and less practical than alternatives.R. King, “Telecom Companies Scramble for Funding,” BusinessWeek, August 3, 2009.

Fiber: A Light-Filled Glass Pipe to Your Doorstep

Fiber to the home (FTTH)Broadband service provided via light-transmitting fiber-optic cables. is the fastest last-mile technology around. It also works over long distances. Verizon’s FiOS technology boasts 50 Mbps download speeds but has tested network upgrades that increase speeds by over six times that.S. Higginbotham, “Verizon Tests 10 Gbps to the Home. Yeah, You’ll Have to Share,” GigaOM, December 17, 2009. The problem with fiber is that unlike cable or DSL copper, fiber to the home networks weren’t already in place. That means firms had to build their own fiber networks from scratch.

The cost of this build out can be enormous. Verizon, for example, has spent over $23 billion on its FTTH infrastructure. However, most experts think the upgrade was critical. Verizon has copper into millions of homes, but U.S. DSL is uncompetitive. Verizon’s residential landline business was dying as users switch to mobile phone numbers, and while mobile is growing, Verizon Wireless is a joint venture with the United Kingdom’s Vodaphone, not a wholly owned firm. This means it shares wireless unit profits with its partner. With FiOS, Verizon now offers pay television, competing with cable’s core product. It also offers some of the fastest home broadband services anywhere, and it gets to keep everything it earns.

In 2010, Google also announced plans to bring fiber to the home. Google deems its effort an experiment—it’s more interested in learning how developers and users take advantage of ultrahigh-speed fiber to the home (e.g., what kinds of apps are created and used, how usage and time spent online change), rather than becoming a nationwide ISP itself. Google says it will investigate ways to build and operate networks less expensively and plans to share findings with others. The Google network will be “open,” allowing other service providers to use Google’s infrastructure to resell services to consumers. The firm has pledged to bring speeds of 1 Gbps at competitive prices to at least 50,000 and potentially as many as 500,000 homes. Over 1,100 U.S. communities applied to be part of the Google experimental fiber network.M. Ingersoll and J. Kelly, “Think Big with a Gig: Our Experimental Fiber Network,” The Google Blog, February 2, 2010; L. Rao, “The Final Tally: More Than 1100 Cities Apply for Google’s Fiber Network,” TechCrunch, March 27, 2010.

Wireless

Mobile wireless service from cell phone access providers is delivered via cell towers. While these providers don’t need to build a residential wired infrastructure, they still need to secure space for cell towers, build the towers, connect the towers to a backbone network, and license the wireless spectrumFrequencies used for communication. Most mobile cell phone services have to license spectrum. Some technologies (such as Wi-Fi) use unlicensed public spectrum. (or airwave frequency space) for transmission.

We need more bandwidth for mobile devices, too. AT&T now finds that the top 3 percent of its mobile network users gulp up 40 percent of the network’s capacity (thanks, iPhone users), and network strain will only increase as more people adopt smartphones. These users are streaming Major League Baseball games, exploring the planet with Google Earth, watching YouTube and Netflix, streaming music through Pandora, and more. Get a bunch of iPhone users in a crowded space, like in a college football stadium on game day, and the result is a network-choking data traffic jam. AT&T estimates that it’s not uncommon for 80 percent of game-day iPhone users to take out their phones and surf the Web for stats, snap and upload photos, and more. But cell towers often can’t handle the load.R. Farzad, “AT&T’s iPhone Mess,” BusinessWeek, February 3, 2010. If you’ve ever lost coverage in a crowd, you’ve witnessed mobile network congestion firsthand. Trying to have enough capacity to avoid congestion traffic jams will cost some serious coin. In the midst of customer complaints, AT&T committed to spending $18 billion on network upgrades to address its wireless capacity problem.C. Edwards and O. Kharif, “Sprint’s Bold Play on a 4G Network,” BusinessWeek, March 30, 2010.

Table 12.1 Average Demand Usage by Function

Usage Demand
Voice Calls 4 MB/hr.
iPhone Browsing 40–60 MB/hr.
Net Radio 60 MB/hr.
YouTube 200–400 MB/hr.
Conventional mobile phones use an estimated 100 MB/month, iPhones 560 MB/month, and iPads almost 1 GB/month.

We’re in the midst of transitioning from third generation (3G) to fourth generation (4G) wireless networks. 3G systems offer access speeds usually less than 2 Mbps (often a lot less).K. German, “On Call: Welcome to 4G,” CNET, March 9, 2010. While variants of 3G wireless might employ an alphabet soup of technologies—EV-DO (evolution data optimized), UMTS (universal mobile telecommunications systems), and HSDPA (high-speed downlink packet link access) among them—3G standards can be narrowed down to two camps: those based on the dominant worldwide standard called GSM (global system for mobile communications) and the runner-up standards based on CDMA (code division multiplex access). Most of Europe and a good chunk of the rest of the world use GSM. In the United States, AT&T and T-Mobile use GSM-based 3G. Verizon Wireless and Sprint use the CDMA 3G standard. Typically, handsets designed for one network can’t be used on networks supporting the other standard. CDMA has an additional limitation in not being able to use voice and data at the same time.

But 3G is being replaced by high-bandwidth 4G (fourth-generation) mobile networks. 4G technologies also fall into two standards camps: LTE (Long Term Evolution) and WiMAX (Worldwide Interoperability for Microwave Access).

LTE looks like the global winner. In the United States, every major wireless firm, except for Sprint, is betting on LTE victory. Bandwidth for the service rivals what we’d consider fast cable a few years back. Average speeds range from 5 to 12 Mbps for downloads and 2 to 5 Mbps for upload, although Verizon tests in Boston and Seattle showed download speeds as high as 50 Mbps and upload speeds reaching 25 Mbps.K. German, “On Call: Welcome to 4G,” CNET, March 9, 2010.

Competing with LTE is WiMAX; don’t confuse it with Wi-Fi. As with other 3G and 4G technologies, WiMAX needs cell towers and operators need to have licensed spectrum from their respective governments (often paying multibillion-dollar fees to do so). Average download and upload speeds should start out at 3–6 Mbps and 1 Mbps, respectively, although this may go much higher.N. Lee, “Sprint’s 4G Plans Explained,” CNET, May 19, 2010.

WiMAX looks like a particularly attractive option for cable firms, offering them an opportunity to get into the mobile phone business and offer a “quadruple play” of services: pay television, broadband Internet, home phone, and mobile. Comcast and Time Warner have both partnered with Clearwire (a firm majority-owned by Sprint), to gain access to WiMAX-based 4G mobile.

4G could also rewrite the landscape for home broadband competition. If speeds increase, it may be possible for PCs, laptops, and set-top boxes (STB) to connect to the Internet wirelessly via 4G, cutting into DSL, cable, and fiber markets.

Satellite Wireless

Wireless systems provided by earth-bound base stations like cell phone towers are referred to as terrestrial wireless, but it is possible to provide telecommunications services via satellite. Early services struggled due to a number of problems. For example, the first residential satellite services were only used for downloads, which still needed a modem or some other connection to send any messages from the computer to the Internet. Many early systems also required large antennas and were quite expensive. Finally, some services were based on satellites in geosynchronous earth orbit (GEO). GEO satellites circle the earth in a fixed, or stationary, orbit above a given spot on the globe, but to do so they must be positioned at a distance that is roughly equivalent to the planet’s circumference. That means signals travel the equivalent of an around-the-world trip to reach the satellite and then the same distance to get to the user. The “last mile” became the last 44,000 miles at best. And if you used a service that also provided satellite upload as well as download, double that to about 88,000 miles. All that distance means higher latency (more delay).G. Ou, “Why Satellite Service Is So Slow,” ZDNet, February 23, 2008.

A firm named O3b Networks thinks it might have solved the challenges that plagued early pioneers. O3b has an impressive list of big-name backers that include HSBC bank, cable magnate John Malone, European aerospace firm SES, and Google.

The name O3b stands for the “Other 3 Billion,” of the world’s population who lack broadband Internet access, and the firm hopes to provide “fiber-quality” wireless service to more than 150 countries, specifically targeting underserved portions of the developing world. These “middle earth orbit” satellites will circle closer to the earth to reduce latency (only about 5,000 miles up, less than one-fourth the distance of GEO systems). To maintain the lower orbit, O3b’s satellites orbit faster than the planet spins, but with plans to launch as many as twenty satellites, the system will constantly blanket regions served. If one satellite circles to the other side of the globe, another one will circle around to take its place, ensuring there’s always an O3b “bird” overhead.

Only about 3 percent of the sub-Saharan African population uses the Internet, compared to about 70 percent in the United States. But data rates in the few places served can cost as much as one hundred times the rates of comparable systems in the industrialized world.G. Lamb, “O3b Networks: A Far-Out Plan to Deliver the Web,” Christian Science Monitor, September 24, 2008. O3b hopes to change that equation and significantly lower access rates. O3b customers will be local telecommunication firms, not end users. The plan is for local firms to buy O3b’s services wholesale and then resell it to customers alongside rivals who can do the same thing, collectively providing more consumer access, higher quality, and lower prices through competition. O3b is a big, bold, and admittedly risky plan, but if it works, its impact could be tremendous.

Wi-Fi and Other Hotspots

Many users access the Internet via Wi-FiA term used to brand wireless local-area networking devices. Devices typically connect to an antenna-equipped base station or hotspot, which is then connected to the Internet. Wi-Fi devices use standards known as IEEE 802.11, and various version of this standard (e.g., b, g, n) may operate in different frequency bands and have access ranges. (which stands for wireless fidelity). Computer and mobile devices have Wi-Fi antennas built into their chipsets, but to connect to the Internet, a device needs to be within range of a base station or hotspot. The base station range is usually around three hundred feet (you might get a longer range outdoors and with special equipment; and less range indoors when signals need to pass through solid objects like walls, ceilings, and floors). Wi-Fi base stations used in the home are usually bought by end users, then connected to a cable, DSL, or fiber provider.

And now a sort of mobile phone hotspot is being used to overcome limitations in those services, as well. Mobile providers can also be susceptible to poor coverage indoors. That’s because the spectrum used by most mobile phone firms doesn’t travel well through solid objects. Cell coverage is also often limited in the United States because of a lack of towers, which is a result of the NIMBY problem (not in my backyard). People don’t want an eighty-foot to four-hundred-foot unsightly tower clouding their local landscape, even if it will give their neighborhood better cell phone coverage.G. Dechter and O. Kharif, “How Craig McCaw Built a 4G Network on the Cheap,” BusinessWeek, May 24, 2010. To overcome reception and availability problems, mobile telecom services firms have begun offering fentocells. These devices are usually smaller than a box of cereal and can sell for $150 or less (some are free with specific service contracts). Plug a fentocell into a high-speed Internet connection like an in-home cable or fiber service and you can get “five-bar” coverage in a roughly 5,000-square-foot footprint.C. Mims, “A Personal Cell Phone Tower,” Technology Review, April 7, 2010. That can be a great solution for someone who has an in-home, high-speed Internet connection, but wants to get phone and mobile data service indoors, too.

Net Neutrality: What’s Fair?

Across the world, battle lines are being drawn regarding the topic of Net neutrality. Net neutrality is the principle that all Internet traffic should be treated equally.M. Honan, “Inside Net Neutrality,” MacWorld, February 12, 2008. Sometimes access providers have wanted to offer varying (some say “discriminatory”) coverage, depending on the service used and bandwidth consumed. But where regulation stands is currently in flux. In a pivotal U.S. case, the FCC ordered Comcast to stop throttling (blocking or slowing down) subscriber access to the peer-to-peer file sharing service BitTorrent. BitTorrent users can consume a huge amount of bandwidth—the service is often used to transfer large files, both legitimate (like version of the Linux operating system) and pirated (HD movies). Then in spring 2010, a federal appeals court moved against the FCC’s position, unanimously ruling that the agency did not have the legal authority to dictate terms to Comcast.“What Is Net Neutrality?” The Week, April 7, 2010.

On one side of the debate are Internet service firms, with Google being one of the strongest Net neutrality supporters. In an advocacy paper, Google states, “Just as telephone companies are not permitted to tell consumers who they can call or what they can say, broadband carriers should not be allowed to use their market power to control activity online.”Google, “A Guide to Net Neutrality for Google Users,” 2008, http://www.docstoc.com/docs/1064274/A-Guide-to-Net-Neutrality-for-Google-Users. Many Internet firms also worry that if network providers move away from flat-rate pricing toward usage-based (or metered) schemes, this may limit innovation. Says Google’s Vint Cerf (who is considered one of the “fathers of the Internet” for his work on the original Internet protocol suite) “You are less likely to try things out. No one wants a surprise bill at the end of the month.”M. Jesdanun, “As the Internet Turns 40, Barriers Threaten Growth,” Technology Review, August 31, 2009. Metered billing may limit the use of everything from iTunes to Netflix; after all, if you have to pay for per-bit bandwidth consumption as well as for the download service, then it’s as if you’re paying twice.

The counterargument is that if firms are restricted from charging more for their investment in infrastructure and services, then they’ll have little incentive to continue to make the kinds of multibillion-dollar investments that innovations like 4G and fiber networks require. Telecom industry executives have railed against Google, Microsoft, Yahoo! and others, calling them free riders who earn huge profits by piggybacking off ISP networks, all while funneling no profits back to the firms that provide the infrastructure. One Verizon vice president said, “The network builders are spending a fortune constructing and maintaining the networks that Google intends to ride on with nothing but cheap servers.…It is enjoying a free lunch that should, by any rational account, be the lunch of the facilities providers.”A. Mohammed, “Verizon Executive Calls for End to Google’s ‘Free Lunch,’” Washington Post, February 7, 2006. AT&T’s previous CEO has suggested that Google, Yahoo! and other services firms should pay for “preferred access” to the firm’s customers. The CEO of Spain’s Telefonica has also said the firm is considering charging Google and other Internet service firms for network use.I. Lunden, “Broadband Content Bits: Web Drama Investment, PPL Video Store, Telefonica to Charge?” paidContent:UK, February 11, 2010.

ISPs also lament the relentlessly increasingly bandwidth demands placed on their networks. Back in 2007, YouTube streamed as much data in three months as the world’s radio, cable, and broadcast television channels combined stream in one year,B. Swanson, “The Coming Exaflood,” Wall Street Journal, January 20, 2007. and YouTube has only continued to grow since then. Should ISPs be required to support the strain of this kind of bandwidth hog? And what if this one application clogs network use for other traffic, such as e-mail or Web surfing? Similarly, shouldn’t firms have the right to prioritize some services to better serve customers? Some network providers argue that services like video chat and streaming audio should get priority over, say, e-mail which can afford slight delay without major impact. In that case, there’s a pretty good argument that providers should be able to discriminate against services. But improving efficiency and throttling usage are two different things.

Internet service firms say they create demand for broadband business, broadband firms say Google and allies are ungrateful parasites that aren’t sharing the wealth. The battle lines on the Net neutrality frontier continue to be drawn, and the eventual outcome will impact consumers, investors, and will likely influence the continued expansion and innovation of the Internet.

Summing Up

Hopefully, this chapter helped reveal the mysteries of the Internet. It’s interesting to know how “the cloud” works but it can also be vital. As we’ve seen, the executive office in financial services firms considers mastery of the Internet infrastructure to be critically important to their competitive advantage. Media firms find the Internet both threatening and empowering. The advancement of last-mile technologies and issues of Net neutrality will expose threats and create opportunity. And a manager who knows how the Internet works will be in a better position to make decisions about how to keep the firm and its customers safe and secure, and be better prepared to brainstorm ideas for winning in a world where access is faster and cheaper, and firms, rivals, partners, and customers are more connected.

Key Takeaways

  • The slowest part of the Internet is typically the last mile, not the backbone. While several technologies can offer broadband service over the last mile, the United States continues to rank below many other nations in terms of access speed, availability, and price.
  • Cable firms and phone companies can leverage existing wiring for cable broadband and DSL service, respectively. Cable services are often criticized for shared bandwidth. DSL’s primary limitation is that it only works within a short distance of telephone office equipment.
  • Fiber to the home can be very fast but very expensive to build.
  • An explosion of high-bandwidth mobile applications is straining 3G networks. 4G systems may alleviate congestion by increasing capacities to near-cable speeds. Fentocells are another technology that can improve service by providing a personal mobile phone hotspot that can plug into in-home broadband access.
  • The two major 3G standards (popularly referred to as GSM and CDMA) will be replaced by two unrelated 4G standards (LTE and WiMAX). GSM has been the dominant 3G technology worldwide. LTE looks like it will be the leading 4G technology.
  • Satellite systems show promise in providing high-speed access to underserved parts of the world, but few satellite broadband providers have been successful so far.
  • Net neutrality is the principle that all Internet traffic should be treated equally. Google and other firms say it is vital to maintain the openness of the Internet. Telecommunications firms say they should be able to limit access to services that overtax their networks, and some have suggested charging Google and other Internet firms for providing access to their customers.

Questions and Exercises

  1. Research online for the latest country rankings for broadband service. Where does the United States currently rank? Why?
  2. Which broadband providers can service your home? Which would you choose? Why?
  3. Research the status of Google’s experimental fiber network. Report updated findings to your class. Why do you suppose Google would run this “experiment”? What other Internet access experiments has the firm been involved in?
  4. Show your understanding of the economics and competitive forces of the telecom industry. Discuss why Verizon chose to go with fiber. Do you think this was a wise decision or not? Why? Feel free to do additional research to back up your argument.
  5. Why have other nations enjoyed faster broadband speeds, greater availability, and lower prices?
  6. The iPhone has been called both a blessing and a curse for AT&T. Why do you suppose this is so?
  7. Investigate the status of mobile wireless offerings (3G and 4G). Which firm would you choose? Why? Which factors are most important in your decision?
  8. Name the two dominant 3G standards. What are the differences between the two? Which firms in your nation support each standard?
  9. Name the two dominant 4G standards. Which firms in your nation will support the respective standards?
  10. Have you ever lost communication access—wirelessly or via wired connection? What caused the loss or outage?
  11. What factors shape the profitability of the mobile wireless provider industry? How do these economics compare with the cable and wire line industry? Who are the major players and which would you invest in? Why?
  12. Last-mile providers often advertise very fast speeds, but users rarely see speeds as high as advertised rates. Search online to find a network speed test and try it from your home, office, mobile device, or dorm. How fast is the network? If you’re able to test from home, what bandwidth rates does your ISP advertise? Does this differ from what you experienced? What could account for this discrepancy?
  13. How can 4G technology help cable firms? Why might it hurt them?
  14. What’s the difference between LEO satellite systems and the type of system used by O3b? What are the pros and cons of these efforts? Conduct some additional research. What is the status of O3b and other satellite broadband efforts?
  15. What advantages could broadband offer to underserved areas of the world? Is Internet access important for economic development? Why or why not?
  16. Does your carrier offer a fentocell? Would you use one? Why or why not?
  17. Be prepared to debate the issue of Net neutrality in class. Prepare positions both supporting and opposing Net neutrality. Which do you support and why?
  18. Investigate the status of Net neutrality laws in your nation and report your findings to your instructor. Do you agree with the stance currently taken by your government? Why or why not?

Chapter 11: The Data Asset: Databases, Business Intelligence, and Competitive Advantage

11.1 Introduction

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand how increasingly standardized data, access to third-party data sets, cheap, fast computing and easier-to-use software are collectively enabling a new age of decision making.
  2. Be familiar with some of the enterprises that have benefited from data-driven, fact-based decision making.

The planet is awash in data. Cash registers ring up transactions worldwide. Web browsers leave a trail of cookie crumbs nearly everywhere they go. And with radio frequency identification (RFID), inventory can literally announce its presence so that firms can precisely journal every hop their products make along the value chain: “I’m arriving in the warehouse,” “I’m on the store shelf,” “I’m leaving out the front door.”

A study by Gartner Research claims that the amount of data on corporate hard drives doubles every six months,C. Babcock, “Data, Data, Everywhere”, InformationWeek, January 9, 2006. while IDC states that the collective number of those bits already exceeds the number of stars in the universe.L. Mearian, “Digital Universe and Its Impact Bigger Than We Thought,” Computerworld, March 18, 2008. Wal-Mart alone boasts a data volume well over 125 times as large as the entire print collection of the U.S. Library of Congress.Derived by comparing Wal-Mart’s 2.5 petabytes (E. Lai, “Teradata Creates Elite Club for Petabyte-Plus Data Warehouse Customers,” Computerworld, October 18, 2008) to the Library of Congress estimate of 20 TB (D. Gewirtz, “What If Someone Stole the Library of Congress?” CNN.com/AC360, May 25, 2009). It’s further noted that the Wal-Mart figure is just for data stored on systems provided by the vendor Teradata. Wal-Mart has many systems outside its Teradata-sourced warehouses, too.

And with this flood of data comes a tidal wave of opportunity. Increasingly standardized corporate data, and access to rich, third-party data sets—all leveraged by cheap, fast computing and easier-to-use software—are collectively enabling a new age of data-driven, fact-based decision making. You’re less likely to hear old-school terms like “decision support systems” used to describe what’s going on here. The phrase of the day is business intelligence (BI)A term combining aspects of reporting, data exploration and ad hoc queries, and sophisticated data modeling and analysis., a catchall term combining aspects of reporting, data exploration and ad hoc queries, and sophisticated data modeling and analysis. Alongside business intelligence in the new managerial lexicon is the phrase analyticsA term describing the extensive use of data, statistical and quantitative analysis, explanatory and predictive models, and fact-based management to drive decisions and actions., a term describing the extensive use of data, statistical and quantitative analysis, explanatory and predictive models, and fact-based management to drive decisions and actions.T. Davenport and J. Harris, Competing on Analytics: The New Science of Winning (Boston: Harvard Business School Press, 2007).

The benefits of all this data and number crunching are very real, indeed. Data leverage lies at the center of competitive advantage we’ve studied in the Zara, Netflix, and Google cases. Data mastery has helped vault Wal-Mart to the top of the Fortune 500 list. It helped Harrah’s Casino Hotels grow to be twice as profitable as similarly sized Caesars, and rich enough to acquire this rival. And data helped Capital One find valuable customers that competitors were ignoring, delivering ten-year financial performance a full ten times greater than the S&P 500. Data-driven decision making is even credited with helping the Red Sox win their first World Series in eighty-three years and with helping the New England Patriots win three Super Bowls in four years. To quote from a BusinessWeek cover story on analytics, “Math Will Rock Your World!”S. Baker, “Math Will Rock Your World,” BusinessWeek, January 23, 2006, http://www.businessweek.com/magazine/content/06_04/b3968001.htm.htm.

Sounds great, but it can be a tough slog getting an organization to the point where it has a leveragable data asset. In many organizations data lies dormant, spread across inconsistent formats and incompatible systems, unable to be turned into anything of value. Many firms have been shocked at the amount of work and complexity required to pull together an infrastructure that empowers its managers. But not only can this be done; it must be done. Firms that are basing decisions on hunches aren’t managing; they’re gambling. And the days of uninformed managerial dice rolling are over.

While we’ll study technology in this chapter, our focus isn’t as much on the technology itself as it is on what you can do with that technology. Consumer products giant P&G believes in this distinction so thoroughly that the firm renamed its IT function as “Information and Decision Solutions.”J. Soat, “P&G’s CIO Puts IT at Users’ Service,” InformationWeek, December 15, 2007. Solutions drive technology decisions, not the other way around.

In this chapter we’ll study the data asset, how it’s created, how it’s stored, and how it’s accessed and leveraged. We’ll also study many of the firms mentioned above, and more; providing a context for understanding how managers are leveraging data to create winning models, and how those that have failed to realize the power of data have been left in the dust.

Data, Analytics, and Competitive Advantage

Anyone can acquire technology—but data is oftentimes considered a defensible source of competitive advantage. The data a firm can leverage is a true strategic asset when it’s rare, valuable, imperfectly imitable, and lacking in substitutes (see Chapter 2 “Strategy and Technology: Concepts and Frameworks for Understanding What Separates Winners from Losers”).

If more data brings more accurate modeling, moving early to capture this rare asset can be the difference between a dominating firm and an also-ran. But be forewarned, there’s no monopoly on math. Advantages based on capabilities and data that others can acquire will be short-lived. Those advances leveraged by the Red Sox were originally pioneered by the Oakland A’s and are now used by nearly every team in the major leagues.

This doesn’t mean that firms can ignore the importance data can play in lowering costs, increasing customer service, and other ways that boost performance. But differentiation will be key in distinguishing operationally effective data use from those efforts that can yield true strategic positioning.

Key Takeaways

  • The amount of data on corporate hard drives doubles every six months.
  • In many organizations, available data is not exploited to advantage.
  • Data is oftentimes considered a defensible source of competitive advantage; however, advantages based on capabilities and data that others can acquire will be short-lived.

Questions and Exercises

  1. Name and define the terms that are supplanting discussions of decision support systems in the modern IS lexicon.
  2. Is data a source of competitive advantage? Describe situations in which data might be a source for sustainable competitive advantage. When might data not yield sustainable advantage?
  3. Are advantages based on analytics and modeling potentially sustainable? Why or why not?
  4. What role do technology and timing play in realizing advantages from the data asset?

11.2 Data, Information, and Knowledge

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the difference between data and information.
  2. Know the key terms and technologies associated with data organization and management.

DataRaw facts and figures. refers simply to raw facts and figures. Alone it tells you nothing. The real goal is to turn data into informationData presented in a context so that it can answer a question or support decision making.. Data becomes information when it’s presented in a context so that it can answer a question or support decision making. And it’s when this information can be combined with a manager’s knowledgeInsight derived from experience and expertise.—their insight from experience and expertise—that stronger decisions can be made.

Trusting Your Data

The ability to look critically at data and assess its validity is a vital managerial skill. When decision makers are presented with wrong data, the results can be disastrous. And these problems can get amplified if bad data is fed to automated systems. As an example, look at the series of man-made and computer-triggered events that brought about a billion-dollar collapse in United Airlines stock.

In the wee hours one Sunday morning in September 2008, a single reader browsing back stories on the Orlando Sentinel’s Web site viewed a 2002 article on the bankruptcy of United Airlines (UAL went bankrupt in 2002, but emerged from bankruptcy four years later). That lone Web surfer’s access of this story during such a low-traffic time was enough for the Sentinel’s Web server to briefly list the article as one of the paper’s “most popular.” Google crawled the site and picked up this “popular” news item, feeding it into Google News.

Early that morning, a worker in a Florida investment firm came across the Google-fed story, assumed United had yet again filed for bankruptcy, then posted a summary on Bloomberg. Investors scanning Bloomberg jumped on what looked like a reputable early warning of another United bankruptcy, dumping UAL stock. Blame the computers again—the rapid plunge from these early trades caused automatic sell systems to kick in (event-triggered, computer-automated trading is responsible for about 30 percent of all stock trades). Once the machines took over, UAL dropped like a rock, falling from twelve to three dollars. That drop represented the vanishing of $1 billion in wealth, and all this because no one checked the date on a news story. Welcome to the new world of paying attention!M. Harvey, “Probe into How Google Mix-Up Caused $1 Billion Run on United,” Times Online, September 12, 2008, http://technology.timesonline.co.uk/tol/news/tech_and_web/article4742147.ece.

Understanding How Data Is Organized: Key Terms and Technologies

A databaseA single table or a collection of related tables. is simply a list (or more likely, several related lists) of data. Most organizations have several databases—perhaps even hundreds or thousands. And these various databases might be focused on any combination of functional areas (sales, product returns, inventory, payroll), geographical regions, or business units. Firms often create specialized databases for recording transactions, as well as databases that aggregate data from multiple sources in order to support reporting and analysis.

Databases are created, maintained, and manipulated using programs called database management systems (DBMS)Sometimes called “databade software”; software for creating, maintaining, and manipulating data., sometimes referred to as database software. DBMS products vary widely in scale and capabilities. They include the single-user, desktop versions of Microsoft Access or Filemaker Pro, Web-based offerings like Intuit QuickBase, and industrial strength products from Oracle, IBM (DB2), Sybase, Microsoft (SQL Server), and others. Oracle is the world’s largest database software vendor, and database software has meant big bucks for Oracle cofounder and CEO Larry Ellison. Ellison perennially ranks in the Top 10 of the Forbes 400 list of wealthiest Americans.

The acronym SQL (often pronounced sequel) also shows up a lot when talking about databases. Structured query language (SQL)A language used to create and manipulate databases. is by far the most common language for creating and manipulating databases. You’ll find variants of SQL inhabiting everything from lowly desktop software, to high-powered enterprise products. Microsoft’s high-end database is even called SQL Server. And of course there’s also the open source MySQL (whose stewardship now sits with Oracle as part of the firm’s purchase of Sun Microsystems). Given this popularity, if you’re going to learn one language for database use, SQL’s a pretty good choice. And for a little inspiration, visit Monster.com or another job site and search for jobs mentioning SQL. You’ll find page after page of listings, suggesting that while database systems have been good for Ellison, learning more about them might be pretty good for you, too.

Even if you don’t become a database programmer or database administrator (DBA)Job title focused on directing, performing, or overseeing activities associated with a database or set of databases. These may include (but not necessarily be limited to): database design, creation, implementation, maintenance, backup and recovery, policy setting and enforcement, and security., you’re almost surely going to be called upon to dive in and use a database. You may even be asked to help identify your firm’s data requirements. It’s quite common for nontech employees to work on development teams with technical staff, defining business problems, outlining processes, setting requirements, and determining the kinds of data the firm will need to leverage. Database systems are powerful stuff, and can’t be avoided, so a bit of understanding will serve you well.

Figure 11.1 A Simplified Relational Database for a University Course Registration System

A complete discourse on technical concepts associated with database systems is beyond the scope of our managerial introduction, but here are some key concepts to help get you oriented, and that all managers should know.

  • A table or fileA list of data, arranged in columns (fields) and rows (records). refers to a list of data.
  • A database is either a single table or a collection of related tables. The course registration database above depicts five tables.
  • A column or fieldA column in a database table. Columns represent each category of data contained in a record (e.g., first name, last name, ID number, date of birth). defines the data that a table can hold. The “Students” table above shows columns for STUDENT_ID, FIRST_NAME, LAST_NAME, CAMPU.S._ADDR (the “…” symbols above are meant to indicate that in practice there may be more columns or rows than are shown in this simplified diagram).
  • A row or recordA row in a database table. Records represent a single instance of whatever the table keeps track of (e.g., student, faculty, course title). represents a single instance of whatever the table keeps track of. In the example above, each row of the “Students” table represents a student, each row of the “Enrollment” table represents the enrollment of a student in a particular course, and each row of the “Course List” represents a given section of each course offered by the University.
  • A keyCode that unlocks encryption. is the field used to relate tables in a database. Look at how the STUDENT_ID key is used above. There is one unique STUDENT_ID for each student, but the STUDENT_ID may appear many times in the “Enrollment” table, indicating that each student may be enrolled in many classes. The “1” and “M” in the diagram above indicate the one to many relationships among the keys in these tables.

Databases organized like the one above, where multiple tables are related based on common keys, are referred to as relational databasesThe most common standard for expressing databases, whereby tables (files) are related based on common keys.. There are many other database formats (sporting names like hierarchical, and object-oriented), but relational databases are far and away the most popular. And all SQL databases are relational databases.

We’ve just scratched the surface for a very basic introduction. Expect that a formal class in database systems will offer you far more detail and better design principles than are conveyed in the elementary example above. But you’re already well on your way!

Key Takeaways

  • Data includes raw facts that must be turned into information in order to be useful and valuable.
  • Databases are created, maintained, and manipulated using programs called database management systems (DBMS), sometimes referred to as database software.
  • All data fields in the same database have unique names, several data fields make up a data record, multiple data records make up a table or data file, and one or more tables or data files make up a database.
  • Relational databases are the most common database format.

Questions and Exercises

  1. Define the following terms: table, record, field. Provide another name for each term along with your definition.
  2. Answer the following questions using the course registration database system, diagramed above:

    1. Imagine you also want to keep track of student majors. How would you do this? Would you modify an existing table? Would you add new tables? Why or why not?
    2. Why do you suppose the system needs a “Course Title” table?
    3. This database is simplified for our brief introduction. What additional data would you need to keep track of if this were a real course registration system? What changes would you make in the database above to account for these needs?
  3. Research to find additional examples of organizations that made bad decisions based on bad data. Report your examples to your class. What was the end result of the examples you’re citing (e.g., loss, damage, or other outcome)? What could managers have done to prevent problems in the cases that you cited? What role did technology play in the examples that you cite? What role did people or procedural issues play?
  4. Why is an understanding of database terms and technologies important, even for nontechnical managers and staff? Consider factors associated with both system use and system development. What other skills, beyond technology, may be important when engaged in data-driven decision making?

11.3 Where Does Data Come From?

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand various internal and external sources for enterprise data.
  2. Recognize the function and role of data aggregators, the potential for leveraging third-party data, the strategic implications of relying on externally purchased data, and key issues associated with aggregators and firms that leverage externally sourced data.

Organizations can pull together data from a variety of sources. While the examples that follow aren’t meant to be an encyclopedic listing of possibilities, they will give you a sense of the diversity of options available for data gathering.

Transaction Processing Systems

For most organizations that sell directly to their customers, transaction processing systems (TPS)Systems that record a transaction (some form of business-related exchange), such as a cash register sale, ATM withdrawal, or product return. represent a fountain of potentially insightful data. Every time a consumer uses a point-of-sale system, an ATM, or a service desk, there’s a transactionSome kind of business exchange. (some kind of business exchange) occurring, representing an event that’s likely worth tracking.

The cash register is the data generation workhorse of most physical retailers, and the primary source that feeds data to the TPS. But while TPS can generate a lot of bits, it’s sometimes tough to match this data with a specific customer. For example, if you pay a retailer in cash, you’re likely to remain a mystery to your merchant because your name isn’t attached to your money. Grocers and retailers can tie you to cash transactions if they can convince you to use a loyalty cardSystems that provide rewards and usage incentives, typically in exchange for a method that provides a more detailed tracking and recording of customer activity. In addition to enhancing data collection, loyalty cards can represent a significant switching cost.. Use one of these cards and you’re in effect giving up information about yourself in exchange for some kind of financial incentive. The explosion in retailer cards is directly related to each firm’s desire to learn more about you and to turn you into a more loyal and satisfied customer.

Some cards provide an instant discount (e.g., the CVS Pharmacy ExtraCare card), while others allow you to build up points over time (Best Buy’s Reward Zone). The latter has the additional benefit of acting as a switching cost. A customer may think “I could get the same thing at Target, but at Best Buy, it’ll increase my existing points balance and soon I’ll get a cash back coupon.”

Tesco: Tracked Transactions, Increased Insights, and Surging Sales

UK grocery giant Tesco, the planet’s third-largest retailer, is envied worldwide for what analysts say is the firm’s unrivaled ability to collect vast amounts of retail data and translate this into sales.K. Capell, “Tesco: ‘Wal-Mart’s Worst Nightmare,’” BusinessWeek, December 29, 2008.

Tesco’s data collection relies heavily on its ClubCard loyalty program, an effort pioneered back in 1995. But Tesco isn’t just a physical retailer. As the world’s largest Internet grocer, the firm gains additional data from Web site visits, too. Remove products from your virtual shopping cart? Tesco can track this. Visited a product comparison page? Tesco watches which product you’ve chosen to go with and which you’ve passed over. Done your research online, then traveled to a store to make a purchase? Tesco sees this, too.

Tesco then mines all this data to understand how consumers respond to factors such as product mix, pricing, marketing campaigns, store layout, and Web design. Consumer-level targeting allows the firm to tailor its marketing messages to specific subgroups, promoting the right offer through the right channel at the right time and the right price. To get a sense of Tesco’s laser-focused targeting possibilities, consider that the firm sends out close to ten million different, targeted offers each quarter.T. Davenport and J. Harris, “Competing with Multichannel Marketing Analytics,” Advertising Age, April 2, 2007. Offer redemption rates are the best in the industry, with some coupons scoring an astronomical 90 percent usage!M. Lowenstein, “Tesco: A Retail Customer Divisibility Champion,” CustomerThink, October 20, 2002.

The firm’s data-driven management is clearly delivering results. In April 2009, while operating in the teeth of a global recession, Tesco posted record corporate profits and the highest earnings ever for a British retailer.K. Capell, “Tesco Hits Record Profit, but Lags in U.S.,” BusinessWeek, April 21, 2009.

Enterprise Software (CRM, SCM, and ERP)

Firms increasingly set up systems to gather additional data beyond conventional purchase transactions or Web site monitoring. CRM or customer relationship management systems are often used to empower employees to track and record data at nearly every point of customer contact. Someone calls for a quote? Brings a return back to a store? Writes a complaint e-mail? A well-designed CRM system can capture all these events for subsequent analysis or for triggering follow-up events.

Enterprise software includes not just CRM systems but also categories that touch every aspect of the value chain, including supply chain management (SCM) and enterprise resource planning (ERP) systems. More importantly, enterprise software tends to be more integrated and standardized than the prior era of proprietary systems that many firms developed themselves. This integration helps in combining data across business units and functions, and in getting that data into a form where it can be turned into information (for more on enterprise systems, see Chapter 9 “Understanding Software: A Primer for Managers”).

Surveys

Sometimes firms supplement operational data with additional input from surveys and focus groups. Oftentimes, direct surveys can tell you what your cash register can’t. Zara store managers informally survey customers in order to help shape designs and product mix. Online grocer FreshDirect (see Chapter 2 “Strategy and Technology: Concepts and Frameworks for Understanding What Separates Winners from Losers”) surveys customers weekly and has used this feedback to drive initiatives from reducing packaging size to including star ratings on produce.R. Braddock, “Lessons of Internet Marketing from FreshDirect,” Wall Street Journal, May 11, 2009. Many CRM products also have survey capabilities that allow for additional data gathering at all points of customer contact.

Can Technology “Cure” U.S. Health Care?

The U.S. health care system is broken. It’s costly, inefficient, and problems seem to be getting worse. Estimates suggest that health care spending makes up a whopping 18 percent of U.S. gross domestic product.J. Zhang, “Recession Likely to Boost Government Outlays on Health Care,” Wall Street Journal, February 24, 2009. U.S. automakers spend more on health care than they do on steel.S. Milligan, “Business Warms to Democratic Leaders,” Boston Globe, May 28, 2009. Even more disturbing, it’s believed that medical errors cause as many as ninety-eight thousand unnecessary deaths in the United States each year, more than motor vehicle accidents, breast cancer, or AIDS.R. Appleton, “Less Independent Doctors Could Mean More Medical Mistakes,” InjuryBoard.com, June 14, 2009; and B. Obama, President’s Speech to the American Medical Association, Chicago, IL, June 15, 2009, http://www.whitehouse.gov/the_press_office/Remarks-by-the-President-to-the-Annual-Conference-of-the -American-Medical-Association.

For years it’s been claimed that technology has the potential to reduce errors, improve health care quality, and save costs. Now pioneering hospital networks and technology companies are partnering to help tackle cost and quality issues. For a look at possibilities for leveraging data throughout the doctor-patient value chain, consider the “event-driven medicine” system built by Dr. John Halamka and his team at Boston’s Beth Israel Deaconess Medical Center (part of the Harvard Medical School network).

When docs using Halamka’s system encounter a patient with a chronic disease, they generate a decision support “screening sheet.” Each event in the system: an office visit, a lab results report (think the medical equivalent of transactions and customer interactions), updates the patient database. Combine that electronic medical record information with artificial intelligenceComputer software that seeks to reproduce or mimic (perhaps with improvements) human thought, decision making, or brain functions. on best practice, and the system can offer recommendations for care, such as, “Patient is past due for an eye exam” or, “Patient should receive pneumovax [a vaccine against infection] this season.”J. Halamka, “IT Spending: When Less Is More,” BusinessWeek, March 2, 2009. The systems don’t replace decision making by doctors and nurses, but they do help to ensure that key issues are on a provider’s radar.

More efficiencies and error checks show up when prescribing drugs. Docs are presented with a list of medications covered by that patient’s insurance, allowing them to choose quality options while controlling costs. Safety issues, guidelines, and best practices are also displayed. When correct, safe medication in the right dose is selected, the electronic prescription is routed to the patients’ pharmacy of choice. As Halamka puts it, going from “doctor’s brain to patients vein” without any of that messy physician handwriting, all while squeezing out layers where errors from human interpretation or data entry might occur.

President Obama believes technology initiatives can save health care as much as $120 billion a year, or roughly two thousand five hundred dollars per family.D. McCullagh, “Q&A: Electronic Health Records and You,” CNET/CBSNews.com, May 19, 2009. An aggressive number, to be sure. But with such a large target to aim at, it’s no wonder that nearly every major technology company now has a health solutions group. Microsoft and Google even offer competing systems for electronically storing and managing patient health records. If systems like Halamka’s and others realize their promise, big benefits may be just around the corner.

External Sources

Sometimes it makes sense to combine a firm’s data with bits brought in from the outside. Many firms, for example, don’t sell directly to consumers (this includes most drug companies and packaged goods firms). If your firm has partners that sell products for you, then you’ll likely rely heavily on data collected by others.

Data bought from sources available to all might not yield competitive advantage on its own, but it can provide key operational insight for increased efficiency and cost savings. And when combined with a firm’s unique data assets, it may give firms a high-impact edge.

Consider restaurant chain Brinker, a firm that runs seventeen hundred eateries in twenty-seven countries under the Chili’s, On The Border, and Maggiano’s brands. Brinker (whose ticker symbol is EAT), supplements their own data with external feeds on weather, employment statistics, gas prices, and other factors, and uses this in predictive models that help the firm in everything from determining staffing levels to switching around menu items.R. King, “Intelligence Software for Business,” BusinessWeek podcast, February 27, 2009.

In another example, Carnival Cruise Lines combines its own customer data with third-party information tracking household income and other key measures. This data plays a key role in a recession, since it helps the firm target limited marketing dollars on those past customers that are more likely to be able to afford to go on a cruise. So far it’s been a winning approach. For three years in a row, the firm has experienced double-digit increases in bookings by repeat customers.R. King, “Intelligence Software for Business,” BusinessWeek podcast, February 27, 2009.

Who’s Collecting Data about You?

There’s a thriving industry collecting data about you. Buy from a catalog, fill out a warranty card, or have a baby, and there’s a very good chance that this event will be recorded in a database somewhere, added to a growing digital dossier that’s made available for sale to others. If you’ve ever gotten catalogs, coupons, or special offers from firms you’ve never dealt with before, this was almost certainly a direct result of a behind-the-scenes trafficking in the “digital you.”

Firms that trawl for data and package them up for resale are known as data aggregatorsFirms that collect and resell data.. They include Acxiom, a $1.3 billion a year business that combines public source data on real estate, criminal records, and census reports, with private information from credit card applications, warranty card surveys, and magazine subscriptions. The firm holds data profiling some two hundred million Americans.A. Gefter and T. Simonite, “What the Data Miners Are Digging Up about You,” CNET, December 1, 2008.

Or maybe you’ve heard of Lexis-Nexis. Many large universities subscribe to the firm’s electronic newspaper, journal, and magazine databases. But the firm’s parent, Reed Elsevier, is a data sales giant, with divisions packaging criminal records, housing information, and additional data used to uncover corporate fraud and other risks. In February, 2008, the firm got even more data rich, acquiring Acxiom competitor ChoicePoint for $4.1 billion. With that kind of money involved, it’s clear that data aggregation is very big business.A. Greenberg, “Companies That Profit from Your Data,” Forbes, May 14, 2008.

The Internet also allows for easy access to data that had been public but otherwise difficult to access. For one example, consider home sale prices and home value assessments. While technically in the public record, someone wanting this information previously had to traipse down to their Town Hall and speak to a clerk, who would hand over a printed log book. Not exactly a Google-speed query. Contrast this with a visit to Zillow.com. The free site lets you pull up a map of your town and instantly peek at how much your neighbors paid for their homes. And it lets them see how much you paid for yours, too.

Computerworld’s Robert Mitchell uncovered a more disturbing issue when public record information is made available online. His New Hampshire municipality had digitized and made available some of his old public documents without obscuring that holy grail for identity thieves, his Social Security number.R. Mithchell, “Why You Should Be Worried about Your Privacy on the Web,” Computerworld, May 11, 2009.

Then there are accuracy concerns. A record incorrectly identifying you as a cat lover is one thing, but being incorrectly named to the terrorist watch list is quite another. During a five-week period airline agents tried to block a particularly high profile U.S. citizen from boarding airplanes on five separate occasions because his name resembled an alias used by a suspected terrorist. That citizen? The late Ted Kennedy, who at the time was the senior U.S. senator from Massachusetts.R. Swarns, “Senator? Terrorist? A Watch List Stops Kennedy at Airport,” New York Times, August 20, 2004.

For the data trade to continue, firms will have to treat customer data as the sacred asset it is. Step over that “creep-out” line, and customers will push back, increasingly pressing for tighter privacy laws. Data aggregator Intellius used to track cell phone customers, but backed off in the face of customer outrage and threatened legislation.

Another concern—sometimes data aggregators are just plain sloppy, committing errors that can be costly for the firm and potentially devastating for victimized users. For example, from 2002 to 2003, a hacker stole 1.6 billion records from Acxiom; while in 2005, ChoicePoint accidentally sold records on one hundred and forty five thousand individuals to a cybercrime identity theft ring. The ChoicePoint case resulted in a fifteen-million-dollar fine from the Federal Trade Commission.A. Greenberg, “Companies That Profit from Your Data,” Forbes, May 14, 2008. Just because you can gather data and traffic in bits doesn’t mean that you should. Any data-centric effort should involve input not only from business and technical staff, but from the firm’s legal team, as well (for more, see the box “Privacy Regulation: A Moving Target”).

Privacy Regulation: A Moving Target

New methods for tracking and gathering user information appear daily, testing user comfort levels. For example, the firm Umbria uses software to analyze millions of blog and forum posts every day, using sentence structure, word choice, and quirks in punctuation to determine a blogger’s gender, age, interests, and opinions. In 2009, Apple introduced facial recognition software while integrating iPhoto into Facebook. It’s quite possible that in the future, someone will be able to upload a photo to a service and direct it to find all the accessible photos and video on the Internet that match that person’s features. And while targeting is getting easier, a Carnegie Mellon study showed that it doesn’t take much to find someone with a minimum of data. Simply by knowing gender, birth date, and postal zip code, 87 percent of people in the United States could be pinpointed by name.A. Gefter and T. Simonite, “What the Data Miners Are Digging Up about You,” CNET, December 1, 2008. Another study showed that publicly available data on state and date of birth could be used to predict U.S. Social Security numbers—a potential gateway to identity theft.E. Mills, “Report: Social Security Numbers Can Be Predicted,” CNET, July 6, 2009, http://news.cnet.com/8301-1009_3-10280614-83.html.

Some feel that Moore’s Law, the falling cost of storage, and the increasing reach of the Internet have us on the cusp of a privacy train wreck. And that may inevitably lead to more legislation that restricts data-use possibilities. Noting this, strategists and technologists need to be fully aware of the legal environment their systems face (see Chapter 14 “Google: Search, Online Advertising, and Beyond” for examples and discussion) and consider how such environments may change in the future. Many industries have strict guidelines on what kind of information can be collected and shared.

For example, HIPAA (the U.S. Health Insurance Portability and Accountability Act) includes provisions governing data use and privacy among health care providers, insurers, and employers. The financial industry has strict requirements for recording and sharing communications between firm and client (among many other restrictions). There are laws limiting the kinds of information that can be gathered on younger Web surfers. And there are several laws operating at the state level as well.

International laws also differ from those in the United States. Europe, in particular, has a strict European Privacy Directive. The directive includes governing provisions that limit data collection, require notice and approval of many types of data collection, and require firms to make data available to customers with mechanisms for stopping collection efforts and correcting inaccuracies at customer request. Data-dependent efforts plotted for one region may not fully translate in another effort if the law limits key components of technology use. The constantly changing legal landscape also means that what works today might not be allowed in the future.

Firms beware—the public will almost certainly demand tighter controls if the industry is perceived as behaving recklessly or inappropriately with customer data.

Key Takeaways

  • For organizations that sell directly to their customers, transaction processing systems (TPS) represent a source of potentially useful data.
  • Grocers and retailers can link you to cash transactions if they can convince you to use a loyalty card which, in turn, requires you to give up information about yourself in exchange for some kind of financial incentive such as points or discounts.
  • Enterprise software (CRM, SCM, and ERP) is a source for customer, supply chain, and enterprise data.
  • Survey data can be used to supplement a firm’s operational data.
  • Data obtained from outside sources, when combined with a firm’s internal data assets, can give the firm a competitive edge.
  • Data aggregators are part of a multibillion-dollar industry that provides genuinely helpful data to a wide variety of organizations.
  • Data that can be purchased from aggregators may not in and of itself yield sustainable competitive advantage since others may have access to this data, too. However, when combined with a firm’s proprietary data or integrated with a firm’s proprietary procedures or other assets, third-party data can be a key tool for enhancing organizational performance.
  • Data aggregators can also be quite controversial. Among other things, they represent a big target for identity thieves, are a method for spreading potentially incorrect data, and raise privacy concerns.
  • Firms that mismanage their customer data assets risk lawsuits, brand damage, lower sales, fleeing customers, and can prompt more restrictive legislation.
  • Further raising privacy issues and identity theft concerns, recent studies have shown that in many cases it is possible to pinpoint users through allegedly anonymous data, and to guess Social Security numbers from public data.
  • New methods for tracking and gathering user information are raising privacy issues which possibly will be addressed through legislation that restricts data use.

Questions and Exercises

  1. Why would a firm use a loyalty card? What is the incentive for the firm? What is the incentive for consumers to opt in and use loyalty cards? What kinds of strategic assets can these systems create?
  2. In what ways does Tesco gather data? Can other firms match this effort? What other assets does Tesco leverage that helps the firm remain among top performing retailers worldwide?
  3. Make a list of the kind of data you might give up when using a cash register, a Web site, or a loyalty card, or when calling a firm’s customer support line. How might firms leverage this data to better serve you and improve their performance?
  4. Are you concerned by any of the data-use possibilities that you outlined in prior questions, discussed in this chapter, or that you’ve otherwise read about or encountered? If you are concerned, why? If not, why not? What might firms, governments, and consumers do to better protect consumers?
  5. What are some of the sources data aggregators tap to collect information?
  6. Privacy laws are in a near constant state of flux. Conduct research to identify the current state of privacy law. Has major legislation recently been proposed or approved? What are the implications for firms operating in effected industries? What are the potential benefits to consumers? Do consumers lose anything from this legislation?
  7. Self-regulation is often proposed as an alternative to legislative efforts. What kinds of efforts would provide “teeth” to self-regulation. Are there steps firms could do to make you believe in their ability to self-regulate? Why or why not?
  8. What is HIPPA? What industry does it impact?
  9. How do international privacy laws differ from U.S. privacy laws?

11.4 Data Rich, Information Poor

Learning Objectives

After studying this section you should be able to do the following:

  1. Know and be able to list the reasons why many organizations have data that can’t be converted to actionable information.
  2. Understand why transactional databases can’t always be queried and what needs to be done to facilitate effective data use for analytics and business intelligence.
  3. Recognize key issues surrounding data and privacy legislation.

Despite being awash in data, many organizations are data rich but information poor. A survey by consulting firm Accenture found 57 percent of companies reporting that they didn’t have a beneficial, consistently updated, companywide analytical capability. Among major decisions, only 60 percent were backed by analytics—40 percent were made by intuition and gut instinct.R. King, “Business Intelligence Software’s Time Is Now,” BusinessWeek, March 2, 2009. The big culprit limiting BI initiatives is getting data into a form where it can be used, analyzed, and turned into information. Here’s a look at some factors holding back information advantages.

Incompatible Systems

Just because data is collected doesn’t mean it can be used. This limit is a big problem for large firms that have legacy systemsOlder information systems that are often incompatible with other systems, technologies, and ways of conducting business. Incompatible legacy systems can be a major roadblock to turning data into information, and they can inhibit firm agility, holding back operational and strategic initiatives., outdated information systems that were not designed to share data, aren’t compatible with newer technologies, and aren’t aligned with the firm’s current business needs. The problem can be made worse by mergers and acquisitions, especially if a firm depends on operational systems that are incompatible with its partner. And the elimination of incompatible systems isn’t just a technical issue. Firms might be under extended agreement with different vendors or outsourcers, and breaking a contract or invoking an escape clause may be costly. Folks working in M&A (the area of investment banking focused on valuing and facilitating mergers and acquisitions) beware—it’s critical to uncover these hidden costs of technology integration before deciding if a deal makes financial sense.

Legacy Systems: A Prison for Strategic Assets

The experience of one Fortune 100 firm that your author has worked with illustrates how incompatible information systems can actually hold back strategy. This firm was the largest in its category, and sold identical commodity products sourced from its many plants worldwide. Being the biggest should have given the firm scale advantages. But many of the firm’s manufacturing facilities and international locations developed or purchased separate, incompatible systems. Still more plants were acquired through acquisition, each coming with its own legacy systems.

The plants with different information systems used different part numbers and naming conventions even though they sold identical products. As a result, the firm had no timely information on how much of a particular item was sold to which worldwide customers. The company was essentially operating as a collection of smaller, regional businesses, rather than as the worldwide behemoth that it was.

After the firm developed an information system that standardized data across these plants, it was, for the first time, able to get a single view of worldwide sales. The firm then used this data to approach their biggest customers, negotiating lower prices in exchange for increased commitments in worldwide purchasing. This trade let the firm take share from regional rivals. It also gave the firm the ability to shift manufacturing capacity globally, as currency prices, labor conditions, disaster, and other factors impacted sourcing. The new information system in effect liberated the latent strategic asset of scale, increasing sales by well over a billion and a half dollars in the four years following implementation.

Operational Data Can’t Always Be Queried

Another problem when turning data into information is that most transactional databases aren’t set up to be simultaneously accessed for reporting and analysis. When a customer buys something from a cash register, that action may post a sales record and deduct an item from the firm’s inventory. In most TPS systems, requests made to the database can usually be performed pretty quickly—the system adds or modifies the few records involved and it’s done—in and out in a flash.

But if a manager asks a database to analyze historic sales trends showing the most and least profitable products over time, they may be asking a computer to look at thousands of transaction records, comparing results, and neatly ordering findings. That’s not a quick in-and-out task, and it may very well require significant processing to come up with the request. Do this against the very databases you’re using to record your transactions, and you might grind your computers to a halt.

Getting data into systems that can support analytics is where data warehouses and data marts come in, the topic of our next section.

Key Takeaways

  • A major factor limiting business intelligence initiatives is getting data into a form where it can be used (i.e., analyzed and turned into information).
  • Legacy systems often limit data utilization because they were not designed to share data, aren’t compatible with newer technologies, and aren’t aligned with the firm’s current business needs.
  • Most transactional databases aren’t set up to be simultaneously accessed for reporting and analysis. In order to run analytics the data must first be ported to a data warehouse or data mart.

Questions and Exercises

  1. How might information systems impact mergers and acquisitions? What are the key issues to consider?
  2. Discuss the possible consequences of a company having multiple plants, each with a different information system using different part numbers and naming conventions for identical products.
  3. Why does it take longer, and require more processing power, to analyze sales trends by region and product, as opposed to posting a sales transaction?

11.5 Data Warehouses and Data Marts

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand what data warehouses and data marts are and the purpose they serve.
  2. Know the issues that need to be addressed in order to design, develop, deploy, and maintain data warehouses and data marts.

Since running analytics against transactional data can bog down a system, and since most organizations need to combine and reformat data from multiple sources, firms typically need to create separate data repositories for their reporting and analytics work—a kind of staging area from which to turn that data into information.

Two terms you’ll hear for these kinds of repositories are data warehouseA set of databases designed to support decision making in an organization. and data martA database or databases focused on addressing the concerns of a specific problem (e.g., increasing customer retention, improving product quality) or business unit (e.g., marketing, engineering).. A data warehouse is a set of databases designed to support decision making in an organization. It is structured for fast online queries and exploration. Data warehouses may aggregate enormous amounts of data from many different operational systems.

A data mart is a database focused on addressing the concerns of a specific problem (e.g., increasing customer retention, improving product quality) or business unit (e.g., marketing, engineering).

Marts and warehouses may contain huge volumes of data. For example, a firm may not need to keep large amounts of historical point-of-sale or transaction data in its operational systems, but it might want past data in its data mart so that managers can hunt for patterns and trends that occur over time.

Figure 11.2

Information systems supporting operations (such as TPS) are typically separate, and “feed” information systems used for analytics (such as data warehouses and data marts).

It’s easy for firms to get seduced by a software vendor’s demonstration showing data at your fingertips, presented in pretty graphs. But as mentioned earlier, getting data in a format that can be used for analytics is hard, complex, and challenging work. Large data warehouses can cost millions and take years to build. Every dollar spent on technology may lead to five to seven more dollars on consulting and other services.R. King, “Intelligence Software for Business,” BusinessWeek podcast, February 27, 2009.

Most firms will face a trade-off—do we attempt a large-scale integration of the whole firm, or more targeted efforts with quicker payoffs? Firms in fast-moving industries or with particularly complex businesses may struggle to get sweeping projects completed in enough time to reap benefits before business conditions change. Most consultants now advise smaller projects with narrow scope driven by specific business goals.D. Rigby and D. Ledingham, “CRM Done Right,” Harvard Business Review, November 2004; and R. King, “Intelligence Software for Business,” BusinessWeek podcast, February 27, 2009.

Firms can eventually get to a unified data warehouse but it may take time. Even analytics king Wal-Mart is just getting to that point. In 2007, it was reported that Wal-Mart had seven hundred different data marts and hired Hewlett-Packard for help in bringing the systems together to form a more integrated data warehouse.H. Havenstein, “HP Nabs Wal-Mart as Data Warehousing Customer,” Computerworld, August 1, 2007.

The old saying from the movie Field of Dreams, “If you build it, they will come,” doesn’t hold up well for large-scale data analytics projects. This work should start with a clear vision with business-focused objectives. When senior executives can see objectives illustrated in potential payoff, they’ll be able to champion the effort, and experts agree, having an executive champion is a key success factor. Focusing on business issues will also drive technology choice, with the firm better able to focus on products that best fit its needs.

Once a firm has business goals and hoped-for payoffs clearly defined, it can address the broader issues needed to design, develop, deploy, and maintain its system:Key points adapted from Davenport and J. Harris, Competing on Analytics: The New Science of Winning (Boston: Harvard Business School Press, 2007).

  • Data relevance. What data is needed to compete on analytics and to meet our current and future goals?
  • Data sourcing. Can we even get the data we’ll need? Where can this data be obtained from? Is it available via our internal systems? Via third-party data aggregators? Via suppliers or sales partners? Do we need to set up new systems, surveys, and other collection efforts to acquire the data we need?
  • Data quantity. How much data is needed?
  • Data quality. Can our data be trusted as accurate? Is it clean, complete, and reasonably free of errors? How can the data be made more accurate and valuable for analysis? Will we need to ‘scrub,’ calculate, and consolidate data so that it can be used?
  • Data hosting. Where will the systems be housed? What are the hardware and networking requirements for the effort?
  • Data governance. What rules and processes are needed to manage data from its creation through its retirement? Are there operational issues (backup, disaster recovery)? Legal issues? Privacy issues? How should the firm handle security and access?

For some perspective on how difficult this can be, consider that an executive from one of the largest U.S. banks once lamented at how difficult it was to get his systems to do something as simple as properly distinguishing between men and women. The company’s customer-focused data warehouse drew data from thirty-six separate operational systems—bank teller systems, ATMs, student loan reporting systems, car loan systems, mortgage loan systems, and more. Collectively these legacy systems expressed gender in seventeen different ways: “M” or “F”; “m” or “f”; “Male” or “Female”; “MALE” or “FEMALE”; “1” for man, “0” for woman; “0” for man, “1” for woman and more, plus various codes for “unknown.” The best math in the world is of no help if the values used aren’t any good. There’s a saying in the industry, “garbage in, garbage out.”

E-discovery: Supporting Legal Inquiries

Data archiving isn’t just for analytics. Sometimes the law requires organizations to dive into their electronic records. E-discoveryThe process of identifying and retrieving relevant electronic information to support litigation efforts. refers to identifying and retrieving relevant electronic information to support litigation efforts. E-discovery is something a firm should account for in its archiving and data storage plans. Unlike analytics that promise a boost to the bottom line, there’s no profit in complying with a judge’s order—it’s just a sunk cost. But organizations can be compelled by court order to scavenge their bits, and the cost to uncover difficult to access data can be significant, if not planned for in advance.

In one recent example, the Office of Federal Housing Enterprise Oversight (OFHEO) was subpoenaed for documents in litigation involving mortgage firms Fannie Mae and Freddie Mac. Even though the OFHEO wasn’t a party in the lawsuit, the agency had to comply with the search—an effort that cost $6 million, a full 9 percent of its total yearly budget.A. Conry-Murray, “The Pain of E-discovery,” InformationWeek, June 1, 2009.

Key Takeaways

  • Data warehouses and data marts are repositories for large amounts of transactional data awaiting analytics and reporting.
  • Large data warehouses are complex, can cost millions, and take years to build.

Questions and Exercises

  1. List the issues that need to be addressed in order to design, develop, deploy, and maintain data warehouses and data marts.
  2. What is meant by “data relevance”?
  3. What is meant by “data governance”?
  4. What is the difference between a data mart and a data warehouse?
  5. Why are data marts and data warehouses necessary? Why can’t an organization simply query its transactional database?
  6. How can something as simple as customer gender be difficult to for a large organization to establish in a data warehouse?

11.6 The Business Intelligence Toolkit

Learning Objectives

After studying this section you should be able to do the following:

  1. Know the tools that are available to turn data into information.
  2. Identify the key areas where businesses leverage data mining.
  3. Understand some of the conditions under which analytical models can fail.
  4. Recognize major categories of artificial intelligence and understand how organizations are leveraging this technology.

So far we’ve discussed where data can come from, and how we can get data into a form where we can use it. But how, exactly, do firms turn that data into information? That’s where the various software tools of business intelligence (BI) and analytics come in. Potential products in the business intelligence toolkit range from simple spreadsheets to ultrasophisticated data mining packages leveraged by teams employing “rocket-science” mathematics.

Query and Reporting Tools

The idea behind query and reporting tools is to present users with a subset of requested data, selected, sorted, ordered, calculated, and compared, as needed. Managers use these tools to see and explore what’s happening inside their organizations.

Canned reportsReports that provide regular summaries of information in a predetermined format. provide regular summaries of information in a predetermined format. They’re often developed by information systems staff and formats can be difficult to alter. By contrast, ad hoc reporting toolsTools that put users in control so that they can create custom reports on an as-needed basis by selecting fields, ranges, summary conditions, and other parameters. allow users to dive in and create their own reports, selecting fields, ranges, and other parameters to build their own reports on the fly. DashboardsA heads-up display of critical indicators that allow managers to get a graphical glance at key performance metrics. provide a sort of heads-up display of critical indicators, letting managers get a graphical glance at key performance metrics. Some tools may allow data to be exported into spreadsheets. Yes, even the lowly spreadsheet can be a powerful tool for modeling “what if” scenarios and creating additional reports (of course be careful: if data can be easily exported, then it can potentially leave the firm dangerously exposed, raising privacy, security, legal, and competitive concerns).

Figure 11.3 The Federal IT Dashboard

The Federal IT dashboard offers federal agencies, and the general public, information about the government’s IT investments.

A subcategory of reporting tools is referred to as online analytical processing (OLAP)A method of querying and reporting that takes data from standard relational databases, calculates and summarizes the data, and then stores the data in a special database called a data cube. (pronounced “oh-lap”). Data used in OLAP reporting is usually sourced from standard relational databases, but it’s calculated and summarized in advance, across multiple dimensions, with the data stored in a special database called a data cubeA special database used to store data in OLAP reporting.. This extra setup step makes OLAP fast (sometimes one thousand times faster than performing comparable queries against conventional relational databases). Given this kind of speed boost, it’s not surprising that data cubes for OLAP access are often part of a firm’s data mart and data warehouse efforts.

A manager using an OLAP tool can quickly explore and compare data across multiple factors such as time, geography, product lines, and so on. In fact, OLAP users often talk about how they can “slice and dice” their data, “drilling down” inside the data to uncover new insights. And while conventional reports are usually presented as a summarized list of information, OLAP results look more like a spreadsheet, with the various dimensions of analysis in rows and columns, with summary values at the intersection.

Figure 11.4

This OLAP report compares multiple dimensions. Company is along the vertical axis, and product is along the horizontal access. Many OLAP tools can also present graphs of multidimensional data.

Public Sector Reporting Tools in Action: Fighting Crime and Fighting Waste

Access to ad hoc query and reporting tools can empower all sorts of workers. Consider what analytics tools have done for the police force in Richmond, Virginia. The city provides department investigators with access to data from internal sources such as 911 logs and police reports, and combines this with outside data including neighborhood demographics, payday schedules, weather reports, traffic patterns, sports events, and more.

Experienced officers dive into this data, exploring when and where crimes occur. These insights help the department decide how to allocate its limited policing assets to achieve the biggest impact. While IT staffers put the system together, the tools are actually used by officers with expertise in fighting street crime—the kinds of users with the knowledge to hunt down trends and interpret the causes behind the data. And it seems this data helps make smart cops even smarter—the system is credited with delivering a single-year crime-rate reduction of 20 percent.S. Lohr, “Reaping Results: Data-Mining Goes Mainstream,” New York Times, May 20, 2007.

As it turns out, what works for cops also works for bureaucrats. When administrators for Albuquerque were given access to ad hoc reporting systems, they uncovered all sorts of anomalies, prompting excess spending cuts on everything from cell phone usage to unnecessarily scheduled overtime. And once again, BI performed for the public sector. The Albuquerque system delivered the equivalent of $2 million in savings in just the first three weeks it was used.R. Mulcahy, “ABC: An Introduction to Business Intelligence,” CIO, March 6, 2007.

Data Mining

While reporting tools can help users explore data, modern data sets can be so large that it might be impossible for humans to spot underlying trends. That’s where data mining can help. Data miningThe process of using computers to identify hidden patterns in, and to build models from, large data sets. is the process of using computers to identify hidden patterns and to build models from large data sets.

Some of the key areas where businesses are leveraging data mining include the following:

  • Customer segmentation—figuring out which customers are likely to be the most valuable to a firm.
  • Marketing and promotion targeting—identifying which customers will respond to which offers at which price at what time.
  • Market basket analysis—determining which products customers buy together, and how an organization can use this information to cross-sell more products or services.
  • Collaborative filtering—personalizing an individual customer’s experience based on the trends and preferences identified across similar customers.
  • Customer churn—determining which customers are likely to leave, and what tactics can help the firm avoid unwanted defections.
  • Fraud detection—uncovering patterns consistent with criminal activity.
  • Financial modeling—building trading systems to capitalize on historical trends.
  • Hiring and promotion—identifying characteristics consistent with employee success in the firm’s various roles.

For data mining to work, two critical conditions need to be present: (1) the organization must have clean, consistent data, and (2) the events in that data should reflect current and future trends. The recent financial crisis provides lessons on what can happen when either of these conditions isn’t met.

First lets look at problems with using bad data. A report in the New York Times has suggested that in the period leading up to the 2008 financial crisis, some banking executives deliberately deceived risk management systems in order to skew capital-on-hand requirements. This deception let firms load up on risky debt, while carrying less cash for covering losses.S. Hansell, “How Wall Street Lied to Its Computers,” New York Times, September 18, 2008. Deceive your systems with bad data and your models are worthless. In this case, wrong estimates from bad data left firms grossly overexposed to risk. When debt defaults occurred; several banks failed, and we entered the worst financial crisis since the Great Depression.

Now consider the problem of historical consistency: Computer-driven investment models can be very effective when the market behaves as it has in the past. But models are blind when faced with the equivalent of the “hundred-year flood” (sometimes called black swans); events so extreme and unusual that they never showed up in the data used to build the model.

We saw this in the late 1990s with the collapse of the investment firm Long Term Capital Management. LTCM was started by Nobel Prize–winning economists, but when an unexpected Russian debt crisis caused the markets to move in ways not anticipated by its models, the firm lost 90 percent of its value in less than two months. The problem was so bad that the Fed had to step in to supervise the firm’s multibillion-dollar bailout. Fast forward a decade to the banking collapse of 2008, and we again see computer-driven trading funds plummet in the face of another unexpected event—the burst of the housing bubble.P. Wahba, “Buffeted ‘Quants’ Are Still in Demand,” Reuters, December 22, 2008.

Data mining presents a host of other perils, as well. It’s possible to over-engineerBuild a model with so many variables that the solution arrived at might only work on the subset of data you’ve used to create it. a model, building it with so many variables that the solution arrived at might only work on the subset of data you’ve used to create it. You might also be looking at a random but meaningless statistical fluke. In demonstrating how flukes occur, one quantitative investment manager uncovered a correlation that at first glance appeared statistically to be a particularly strong predictor for historical prices in the S&P 500 stock index. That predictor? Butter production in Bangladesh.P. Coy, “He Who Mines Data May Strike Fool’s Gold,” BusinessWeek, June 16, 1997. Sometimes durable and useful patterns just aren’t in your data.

One way to test to see if you’re looking at a random occurrence in the numbers is to divide your data, building your model with one portion of the data, and using another portion to verify your results. This is the approach Netflix has used to test results achieved by teams in the Netflix Prize, the firm’s million-dollar contest for improving the predictive accuracy of its movie recommendation engine (see Chapter 4 “Netflix: The Making of an E-commerce Giant and the Uncertain Future of Atoms to Bits”).

Finally, sometimes a pattern is uncovered but determining the best choice for a response is less clear. As an example, let’s return to the data-mining wizards at Tesco. An analysis of product sales data showed several money-losing products, including a type of bread known as “milk loaf.” Drop those products, right? Not so fast. Further analysis showed milk loaf was a “destination product” for a loyal group of high-value customers, and that these customers would shop elsewhere if milk loaf disappeared from Tesco shelves. The firm kept the bread as a loss-leader and retained those valuable milk loaf fans.B. Helm, “Getting Inside the Customer’s Mind,” BusinessWeek, September 11, 2008. Data miner, beware—first findings don’t always reveal an optimal course of action.

This last example underscores the importance of recruiting a data mining and business analytics team that possesses three critical skills: information technology (for understanding how to pull together data, and for selecting analysis tools), statistics (for building models and interpreting the strength and validity of results), and business knowledge (for helping set system goals, requirements, and offering deeper insight into what the data really says about the firm’s operating environment). Miss one of these key functions and your team could make some major mistakes.

While we’ve focused on tools in our discussion above, many experts suggest that business intelligence is really an organizational process as much as it is a set of technologies. Having the right team is critical in moving the firm from goal setting through execution and results.

Artificial Intelligence

Data mining has its roots in a branch of computer science known as artificial intelligence (or AI). The goal of AI is create computer programs that are able to mimic or improve upon functions of the human brain. Data mining can leverage neural networksAn AI system that examines data and hunts down and exposes patterns, in order to build models to exploit findings. or other advanced algorithms and statistical techniques to hunt down and expose patterns, and build models to exploit findings.

Expert systemsAI systems that leverages rules or examples to perform a task in a way that mimics applied human expertise. are AI systems that leverage rules or examples to perform a task in a way that mimics applied human expertise. Expert systems are used in tasks ranging from medical diagnoses to product configuration.

Genetic algorithmsModel building techniques where computers examine many potential solutions to a problem, iteratively modifying (mutating) various mathematical models, and comparing the mutated models to search for a best alternative. are model building techniques where computers examine many potential solutions to a problem, iteratively modifying (mutating) various mathematical models, and comparing the mutated models to search for a best alternative. Genetic algorithms have been used to build everything from financial trading models to handling complex airport scheduling, to designing parts for the international space station.Adapted from J. Kahn, “It’s Alive,” Wired, March 2002; O. Port, “Thinking Machines,” BusinessWeek, August 7, 2000; and L. McKay, “Decisions, Decisions,” CRM Magazine, May 1, 2009.

While AI is not a single technology, and not directly related to data creation, various forms of AI can show up as part of analytics products, CRM tools, transaction processing systems, and other information systems.

Key Takeaways

  • Canned and ad hoc reports, digital dashboards, and OLAP are all used to transform data into information.
  • OLAP reporting leverage data cubes, which take data from standard relational databases, calculating and summarizing data for superfast reporting access. OLAP tools can present results through multidimensional graphs, or via spreadsheet-style cross-tab reports.
  • Modern data sets can be so large that it might be impossible for humans to spot underlying trends without the use of data mining tools.
  • Businesses are using data mining to address issues in several key areas including customer segmentation, marketing and promotion targeting, collaborative filtering, and so on.
  • Models influenced by bad data, missing or incomplete historical data, and over-engineering are prone to yield bad results.
  • One way to test to see if you’re looking at a random occurrence in your data is to divide your data, building your model with one portion of the data, and using another portion to verify your results.
  • Analytics may not always provide the total solution for a problem. Sometimes a pattern is uncovered, but determining the best choice for a response is less clear.
  • A competent business analytics team should possess three critical skills: information technology, statistics, and business knowledge.

Questions and Exercises

  1. What are some of the tools used to convert data into information?
  2. What is the difference between a canned reports and an ad hoc reporting?
  3. How do reports created by OLAP differ from most conventional reports?
  4. List the key areas where businesses are leveraging data mining.
  5. What is market basket analysis?
  6. What is customer churn?
  7. For data mining to work, what two critical data-related conditions must be present?
  8. Discus occurrences of model failure caused by missing or incomplete historical data.
  9. Discuss Tesco’s response to their discovery that “milk loaf” was a money-losing product.
  10. List the three critical skills a competent business analytics team should possess.
  11. Do any of the products that you use leverage artificial intelligence? What kinds of AI might be used in Netflix’s movie recommendation system, Apple’s iTunes Genius playlist builder, or Amazon’s Web site personalization? What kind of AI might help a physician make a diagnosis or help an engineer configure a complicated product in the field?

11.7 Data Asset in Action: Technology and the Rise of Wal-Mart

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand how Wal-Mart has leveraged information technology to become the world’s largest retailer.
  2. Be aware of the challenges that face Wal-Mart in the years ahead.

Wal-Mart demonstrates how a physical product retailer can create and leverage a data asset to achieve world-class supply chain efficiencies targeted primarily at driving down costs.

Wal-Mart isn’t just the largest retailer in the world, over the past several years it has popped in and out of the top spot on the Fortune 500 list—meaning that the firm has had revenues greater than any firm in the United States. Wal-Mart is so big that in three months it sells more than a whole year’s worth of sales at number two U.S. retailer, Home Depot.From 2006 through 2009, Wal-Mart has appeared as either number one or number two in the Fortune 100 rankings.

At that size, it’s clear that Wal-Mart’s key source of competitive advantage is scale. But firms don’t turn into giants overnight. Wal-Mart grew in large part by leveraging information systems to an extent never before seen in the retail industry. Technology tightly coordinates the Wal-Mart value chain from tip to tail, while these systems also deliver a mineable data asset that’s unmatched in U.S. retail. To get a sense of the firm’s overall efficiencies, at the end of the prior decade a McKinsey study found that Wal-Mart was responsible for some 12 percent of the productivity gains in the entire U.S. economy.C. Fishman, “The Wal-Mart You Don’t Know,” Fast Company, December 19, 2007. The firm’s capacity as a systems innovator is so respected that many senior Wal-Mart IT executives have been snatched up for top roles at Dell, HP, Amazon, and Microsoft. And lest one think that innovation is the province of only those located in the technology hubs of Silicon Valley, Boston, and Seattle, remember that Wal-Mart is headquartered in Bentonville, Arkansas.

A Data-Driven Value Chain

The Wal-Mart efficiency dance starts with a proprietary system called Retail Link, a system originally developed in 1991 and continually refined ever since. Each time an item is scanned by a Wal-Mart cash register, Retail Link not only records the sale, it also automatically triggers inventory reordering, scheduling, and delivery. This process keeps shelves stocked, while keeping inventories at a minimum. An AMR report ranked Wal-Mart as having the seventh best supply chain in the country (the only other retailer in the top twenty was Tesco, at number fifteen).T. Friscia, K. O’Marah, D. Hofman, and J. Souza, “The AMR Research Supply Chain Top 25 for 2009,” AMR Research, May 28, 2009, http://www.amrresearch.com/Content/View.aspx?compURI=tcm:7-43469. The firm’s annual inventory turnover ratioThe ratio of a company’s annual sales to its inventory. of 8.5 means that Wal-Mart sells the equivalent of its entire inventory roughly every six weeks (by comparison, Target’s turnover ratio is 6.4, Sears’ is 3.4, and the average for U.S. retail is less than 2).Twelve-month figures from midyear 2009, via Forbes and Reuters.

Back-office scanners keep track of inventory as supplier shipments come in. Suppliers are rated based on timeliness of deliveries, and you’ve got to be quick to work with Wal-Mart. In order to avoid a tractor-trailer traffic jam in store parking lots, deliveries are choreographed to arrive at intervals less than ten minutes apart. When Levi’s joined Wal-Mart, the firm had to guarantee it could replenish shelves every two days—no prior retailer had required a shorter than five day window from Levi’s.C. Fishman, “The Wal-Mart You Don’t Know,” Fast Company, December 19, 2007.

Wal-Mart has been a catalyst for technology adoption among its suppliers. The firm is currently leading an adoption effort that requires partners to leverage RFID technology to track and coordinate inventories. While the rollout has been slow, a recent P&G trial showed RFID boosted sales nearly 20 percent by ensuring that inventory was on shelves and located where it should be.D. Joseph, “Supermarket Strategies: What’s New at the Grocer,” BusinessWeek, June 8, 2009.

Data Mining Prowess

Wal-Mart also mines its mother lode of data to get its product mix right under all sorts of varying environmental conditions, protecting the firm from “a retailer’s twin nightmares: too much inventory, or not enough.”C. Hays, “What Wal-Mart Knows about Customer Habits,” New York Times, November 14, 2004. For example, the firm’s data mining efforts informed buyers that customers stock up on certain products in the days leading up to predicted hurricanes. Bumping up prestorm supplies of batteries and bottled water was a no brainer, but the firm also learned that Pop-Tarts sales spike seven fold before storms hit, and that beer is the top prestorm seller. This insight has lead to truckloads full of six packs and toaster pastries streaming into gulf states whenever word of a big storm surfaces.C. Hays, “What Wal-Mart Knows about Customer Habits,” New York Times, November 14, 2004.

Data mining also helps the firm tighten operational forecasts, helping to predict things like how many cashiers are needed at a given store at various times of day throughout the year. Data drives the organization, with mined reports forming the basis of weekly sales meetings, as well as executive strategy sessions.

Sharing Data, Keeping Secrets

While Wal-Mart is demanding of its suppliers, it also shares data with them, too. Data can help firms become more efficient so that Wal-Mart can keep dropping prices, and data can help firms uncover patterns that help suppliers sell more. P&G’s Gillette unit, for example, claims to have mined Wal-Mart data to develop promotions that increased sales as much as 19 percent. More than seventeen thousand suppliers are given access to their products’ Wal-Mart performance across metrics that include daily sales, shipments, returns, purchase orders, invoices, claims and forecasts. And these suppliers collectively interrogate Wal-Mart data warehouses to the tune of twenty-one million queries a year.K. Evans-Correia, “Dillman Replaced as Wal-Mart CIO,” SearchCIO, April 6, 2006.

While Wal-Mart shares sales data with relevant suppliers, the firm otherwise fiercely guards this asset. Many retailers pool their data by sharing it with information brokers like Information Resources and ACNielsen. This sharing allows smaller firms to pool their data to provide more comprehensive insight on market behavior. But Wal-Mart stopped sharing data with these agencies years ago. The firm’s scale is so big, the additional data provided by brokers wasn’t adding much value, and it no longer made sense to allow competitors access to what was happening in its own huge chunk of retail sales.

Other aspects of the firm’s technology remain under wraps, too. Wal-Mart custom builds large portions of its information systems to keep competitors off its trail. As for infrastructure secrets, the Wal-Mart Data Center in McDonald County, Missouri, was considered so off limits that the county assessor was required to sign a nondisclosure statement before being allowed on-site to estimate property value.M. McCoy, “Wal-Mart’s Data Center Remains Mystery,” Joplin Globe, May 28, 2006.

Challenges Abound

But despite success, challenges continue. While Wal-Mart grew dramatically throughout the 1990s, the firm’s U.S. business has largely matured. And as a mature business it faces a problem not unlike the example of Microsoft discussed at the end of Chapter 14 “Google: Search, Online Advertising, and Beyond”; Wal-Mart needs to find huge markets or dramatic cost savings in order to boost profits and continue to move its stock price higher.

The firm’s aggressiveness and sheer size also increasingly make Wal-Mart a target for criticism. Those low prices come at a price, and the firm has faced accusations of subpar wages and remains a magnet for union activists. Others had identified poor labor conditions at some of the firm’s contract manufacturers. Suppliers that compete for Wal-Mart’s business are often faced with a catch-22. If they bypass Wal-Mart they miss out on the largest single chunk of world retail sales. But if they sell to Wal-Mart, the firm may demand prices so aggressively low that suppliers end up cannibalizing their own sales at other retailers. Still more criticism comes from local citizen groups that have accused Wal-Mart of ruining the market for mom-and-pop stores.C. Fishman, “The Wal-Mart You Don’t Know,” Fast Company, December 19, 2007.

While some might see Wal-Mart as invincibly standing at the summit of world retail, it’s important to note that other megaretailers have fallen from grace. In the 1920s and 1930s, the A&P grocery chain once controlled 80 percent of U.S. grocery sales, at its peak operating five times the number of stores that Wal-Mart has today. But market conditions changed, and the government stepped in to draft antipredatory pricing laws when it felt A&Ps parent was too aggressive.

For all of Wal-Mart’s data brilliance, historical data offers little insight on how to adapt to more radical changes in the retail landscape. The firm’s data warehouse wasn’t able to foretell the rise of Target and other up-market discounters. And yet another major battle is brewing, as Tesco methodically attempts to take its globally honed expertise to U.S. shores. Savvy managers recognize that data use is a vital tool, but not the only tool in management’s strategic arsenal.

Key Takeaways

  • Wal-Mart demonstrates how a physical product retailer can create and leverage a data asset to achieve world-class value chain efficiencies.
  • Wal-Mart uses data mining in numerous ways, from demand forecasting to predicting the number of cashiers needed at a store at a particular time.
  • To help suppliers become more efficient, and as a result lower prices, Wal-Mart shares data with them.
  • Despite its success, Wal-Mart is a mature business that needs to find huge markets or dramatic cost savings in order to boost profits and continue to move its stock price higher. The firm’s success also makes it a high impact target for criticism and activism. And the firm’s data assets could not predict impactful industry trends such as the rise of Target and other upscale discounters.

Questions and Exercises

  1. List the functions performed by Retail Link. What is its benefit to Wal-Mart?
  2. Which supplier metrics does Retail Link gather and report? How is this valuable to Wal-Mart and suppliers?
  3. Name the technology does Wal-Mart require partners to use to track and coordinate inventory. Do you know of other uses for this technology?
  4. What steps has Wal-Mart taken to protect its data from competitors?
  5. List the criticisms leveled at Wal-Mart. Do you think these critiques are valid or not? What can Wal-Mart do to counteract this criticism? Should it take these steps? Why or why not?

11.8 Data Asset in Action: Harrah’s Solid Gold CRM for the Service Sector

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand how Harrah’s has used IT to move from an also-ran chain of casinos to become the largest gaming company based on revenue.
  2. Name some of the technology innovations that Harrah’s is using to help it gather more data, and help push service quality and marketing program success.

Harrah’s Entertainment provides an example of exceptional data asset leverage in the service sector, focusing on how this technology enables world-class service through customer relationship management.

Gary Loveman is a sort of management major trifecta. The CEO of Harrah’s Entertainment is a former operations professor who has leveraged information technology to create what may be the most effective marketing organization in the service industry. If you ever needed an incentive to motivate you for cross-disciplinary thinking, Loveman provides it.

Harrah’s has leveraged its data-powered prowess to move from an also-ran chain of casinos to become the largest gaming company by revenue. The firm operates some fifty-three casinos, employing more than eighty-five thousand workers on five continents. Brands include Harrah’s, Caesars Palace, Bally’s, Horseshoe, and Paris Las Vegas. Under Loveman, Harrah’s has aggressively swallowed competitors, the firm’s $9.4 billion buyout of Caesars Entertainment being its largest deal to date.

Collecting Data

Data drives the firm. Harrah’s collects customer data on just about everything you might do at their properties—gamble, eat, grab a drink, attend a show, stay in a room. The data’s then used to track your preferences and to size up whether you’re the kind of customer that’s worth pursuing. Prove your worth, and the firm will surround you with top-tier service and develop a targeted marketing campaign to keep wooing you back.V. Magnini, E. Honeycutt, and S. Hodge, “Data Mining for Hotel Firms: Use and Limitations,” Cornell Hotel and Restaurant Administration Quarterly, April 2003, http://www.entrepreneur.com/tradejournals/article/101938457.html.

The ace in the firm’s data collection hole is its Total Rewards loyalty card system. Launched over a decade ago, the system is constantly being enhanced by an IT staff of seven hundred, with an annual budget in excess of $100 million.P. Swabey, “Nothing Left to Chance,” Information Age, January 18, 2007. Total Rewards is an opt-inProgram (typically a marketing effort) that requires customer consent. This program is contrasted with opt-out programs, which enroll all customers by default. loyalty program, but customers consider the incentives to be so good that the card is used by some 80 percent of Harrah’s patrons, collecting data on over forty-four million customers.M. Wagner, “Harrah’s Places Its Bet On IT,” InformationWeek, September 16, 2008; and L. Haugsted, “Better Take Care of Big Spenders; Harrah’s Chief Offers Advice to Cablers,” Multichannel News, July 30, 2007.

Customers signing up for the card provide Harrah’s with demographic information such as gender, age, and address. Visitors then present the card for various transactions. Slide it into a slot machine, show it to the restaurant hostess, present it to the parking valet, share your account number with a telephone reservation specialist—every contact point is an opportunity to collect data. Between three hundred thousand and one million customers come through Harrah’s doors daily, adding to the firm’s data stash and keeping that asset fresh.N. Hoover, “Chief of the Year: Harrah’s CIO Tim Stanley,” Information Week Research and Reports, 2007.

Who Are the Most Valuable Customers?

All that data is heavily and relentlessly mined. Customer relationship management should include an assessment to determine which customers are worth having a relationship with. And because Harrah’s has so much detailed historical data, the firm can make fairly accurate projections of customer lifetime value (CLV)The present value of the likely future income stream generated by an individual purchaser.. CLV represents the present value of the likely future income stream generated by an individual purchaser.“Which Customers Are Worth Keeping and Which Ones Aren’t? Managerial Uses of CLV,” Knowledge@Wharton, July 30, 2003, http://knowledge.wharton.upenn.edu/article.cfm?articleid=820. Once you know this, you can get a sense of how much you should spend to keep that customer coming back. You can size them up next to their peer group and if they fall below expectations you can develop strategies to improve their spending.

The firm tracks over ninety demographic segments, and each responds differently to different marketing approaches. Identifying segments and figuring out how to deal with each involves an iterative model of mining the data to identify patterns, creating a hypothesis (customers in group X will respond to a free steak dinner; group Y will want ten dollars in casino chips), then testing that hypothesis against a control group, turning again to analytics to statistically verify the outcome.

The firm runs hundreds of these small, controlled experiments each year. Loveman says that when marketers suggest new initiatives, “I ask, did we test it first? And if I find out that we just whole-hogged, went after something without testing it, I’ll kill ’em. No matter how clever they think it is, we test it.”J. Nickell, “Welcome to Harrah’s,” Business 2.0, April 2002. The former ops professor is known to often quote quality guru W. Edwards Deming, saying, “In God we trust; all others must bring data.”

When Harrah’s began diving into the data, they uncovered patterns that defied the conventional wisdom in the gaming industry. Big money didn’t come from European princes, Hong Kong shipping heirs, or the Ocean’s 11 crowd—it came from locals. The less than 30 percent of customers who spent between one hundred and five hundred dollars per visit accounted for over 80 percent of revenues and nearly 100 percent of profits.P. Swabey, “Nothing Left to Chance,” Information Age, January 18, 2007.

The data also showed that the firm’s most important customers weren’t the families that many Vegas competitors were trying to woo with Disneyland-style theme casinos—it was Grandma! Harrah’s focuses on customers forty-five years and older: twenty-somethings have no money, while thirty-somethings have kids and are too busy. To the premiddle-aged crowd, Loveman says, “God bless you, but we don’t need you.”L. Haugsted, “Better Take Care of Big Spenders; Harrah’s Chief Offers Advice to Cablers,” Multichannel News, July 30, 2007.

Data-Driven Service: Get Close (but Not Too Close) to Your Customers

The names for reward levels on the Total Rewards card convey increasing customer value—Gold, Diamond, and Platinum. Spend more money at Harrah’s and you’ll enjoy shorter lines, discounts, free items, and more. And if Harrah’s systems determine you’re a high-value customer, expect white-glove treatment. The firm will lavish you with attention, using technology to try to anticipate your every need. Customers notice the extra treatment that top-tier Total Rewards members receive and actively work to improve their status.

To illustrate this, Loveman points to the obituary of an Ashville, North Carolina, woman who frequented a casino Harrah’s operates on a nearby Cherokee reservation. “Her obituary was published in the Asheville paper and indicated that at the time of her death, she had several grandchildren, she sang in the Baptist choir and she was a holder of the Harrah’s Diamond Total Rewards card.” Quipped Loveman, “When your loyalty card is listed in someone’s obituary, I would maintain you have traction.”G. Loveman, Speech and Comments, Chief Executive Club of Boston College, January 2005; emphasis added.

The degree of customer service pushed through the system is astonishing. Upon check in, a Harrah’s customer who enjoys fine dining may find his or her table is reserved, along with tickets for a show afterward. Others may get suggestions or special offers throughout their stay, pushed via text message to their mobile device.M. Wagner, “Harrah’s Places Its Bet On IT,” InformationWeek, September 16, 2008. The firm even tracks gamblers to see if they’re suffering unusual losses, and Harrah’s will dispatch service people to intervene with a feel-good offer: “Having a bad day? Here’s a free buffet coupon.”T. Davenport and J. Harris, Competing on Analytics: The New Science of Winning (Boston: Harvard Business School Press, 2007).

The firm’s CRM effort monitors any customer behavior changes. If a customer who usually spends a few hundred a month hasn’t shown up in a while, the firm’s systems trigger follow-up contact methods such as sending a letter with a promotion offer, or having a rep make a phone call inviting them back.G. Loveman, Speech and Comments, Chief Executive Club of Boston College, January 2005.

Customers come back to Harrah’s because they feel that those casinos treat them better than the competition. And Harrah’s laser-like focus on service quality and customer satisfaction are embedded into its information systems and operational procedures. Employees are measured on metrics that include speed and friendliness and are compensated based on guest satisfaction ratings. Hourly workers are notoriously difficult to motivate: they tend to be high-turnover, low-wage earners. But at Harrah’s, incentive bonuses depend on an entire location’s ratings. That encourages strong performers to share tips to bring the new guy up to speed. The process effectively changed the corporate culture at Harrah’s from an every-property-for-itself mentality to a collaborative, customer-focused enterprise.V. Magnini, E. Honeycutt, and S. Hodge, “Data Mining for Hotel Firms: Use and Limitations,” Cornell Hotel and Restaurant Administration Quarterly, April 2003, http://www.entrepreneur.com/tradejournals/article/101938457.html.

While Harrah’s is committed to learning how to make your customer experience better, the firm is also keenly sensitive to respecting consumer data. The firm has never sold or given away any of its bits to third parties. And the firm admits that some of its efforts to track customers have misfired, requiring special attention to find the sometimes subtitle line between helpful and “too helpful.” For example, the firm’s CIO has mentioned that customers found it “creepy and Big Brother-ish” when employees tried to greet them by name and talk with them about their past business history at Harrah’s, so the firm backed off.M. Wagner, “Harrah’s Places Its Bet On IT,” InformationWeek, September 16, 2008.

Innovation

Harrah’s is constantly tinkering with new innovations that help it gather more data and help push service quality and marketing program success. When the introduction of gaming in Pennsylvania threatened to divert lucrative New York City gamblers from Harrah’s Atlantic City properties, the firm launched an interactive billboard in New York’s Times Square, allowing passers-by to operate a virtual slot machine using text messages from their cell phones. Players dialing into the video billboard not only control the display, they receive text message offers promoting Harrah’s sites in Atlantic City.“Future Tense: The Global CMO,” Economist Intelligence Unit, September 2008.

At Harrah’s, tech experiments abound. RFID-enabled poker chips and under-table RFID readers allow pit bosses to track and rate game play far better than they could before. The firm is experimenting with using RFID-embedded bracelets for poolside purchases and Total Rewards tracking for when customers aren’t carrying their wallets. The firm has also incorporated drink ordering into gaming machines—why make customers get up to quench their thirst? A break in gambling is a halt in revenue.

The firm was also one of the first to sign on to use Microsoft’s Surface technology—a sort of touch-screen and sensor-equipped tabletop. Customers at these tables can play bowling and group pinball games and even pay for drinks using cards that the tables will automatically identify. Tech even helps Harrah’s fight card counters and crooks, with facial recognition software scanning casino patrons to spot the bad guys.S. Lohr, “Reaping Results: Data-Mining Goes Mainstream,” New York Times, May 20, 2007.

Strategy

A walk around Vegas during Harrah’s ascendency would find rivals with bigger, fancier casinos. Says Loveman, “We had to compete with the kind of place that God would build if he had the money.…The only thing we had was data.”P. Swabey, “Nothing Left to Chance,” Information Age, January 18, 2007.

That data advantage creates intelligence for a high-quality and highly personal customer experience. Data gives the firm a service differentiation edge. The loyalty program also represents a switching cost. And these assets combined to be leveraged across a firm that has gained so much scale that it’s now the largest player in its industry, gaining the ability to cross-sell customers on a variety of properties—Vegas vacations, riverboat gambling, locally focused reservation properties, and more.

Harrah’s chief marketing officer, David Norton, points out that when Total Rewards started, Harrah’s was earning about thirty-six cents on every dollar customers spent gaming—the rest went to competitors. A climb to forty cents would be considered monstrous. By 2005 that number had climbed to forty-five cents, making Harrah’s the biggest monster in the industry.E. Lundquist, “Harrah’s Bets Big on IT,” eWeek, July 20, 2005. Some of the firm’s technology investments have paid back tenfold in just two years—bringing in hundreds of millions of dollars.P. Swabey, “Nothing Left to Chance,” Information Age, January 18, 2007.

The firm’s technology has been pretty tough for others to match, too. Harrah’s holds several patents covering key business methods and technologies used in its systems. After being acquired by Harrah’s, employees of Caesars lamented that they had, for years, unsuccessfully attempted to replicate Harrah’s systems without violating the firm’s intellectual property.N. Hoover, “Chief of the Year: Harrah’s CIO Tim Stanley,” Information Week Research and Reports, 2007.

Challenges

Harrah’s efforts to gather data, extract information, and turn this into real profits is unparalleled, but it’s not a cure-all. Broader events can often derail even the best strategy. Gaming is a discretionary spending item, and when the economy tanks, gambling is one of the first things consumers will cut. Harrah’s has not been immune to the world financial crisis and experienced a loss in 2008.

Also note that if you look up Harrah’s stock symbol you won’t find it. The firm was taken privateThe process by which a publicly held company has its outstanding shares purchased by an individual or by a small group of individuals who wish to obtain complete ownership and control. in January 2008, when buyout firms Apollo Management and TPG Capital paid $30.7 billion for all of the firm’s shares. At that time Loveman signed a five-year deal to remain on as CEO, and he’s spoken positively about the benefits of being private—primarily that with the distraction of quarterly earnings off the table, he’s been able to focus on the long-term viability and health of the business.A. Knightly, “Harrah’s Boss Speaks,” Las Vegas Review-Journal, June 14, 2009.

But the firm also holds $24 billion in debt from expansion projects and the buyout, all at a time when economic conditions have not been favorable to leveraged firms.P. Lattman, “A Buyout-Shop Breather,” Wall Street Journal, May 30, 2009. A brilliantly successful firm that developed best-in-class customer relationship management in now in a position many consider risky due to debt assumed as part of an overly optimistic buyout occurring at precisely the time when the economy went into a terrible funk. Harrah’s awesome risk-reducing, profit-pushing analytics failed to offer any insight on the wisdom (or risk) in the debt and private equity deals.

Key Takeaways

  • Harrah’s Entertainment provides an example of exceptional data asset leverage in the service sector, focusing on how this technology enables world-class service through customer relationship management.
  • Harrah’s uses its Total Rewards loyalty card system to collect customer data on just about everything you might do at their properties—gamble, eat, drink, see a show, stay in a room, and so on.
  • Individual customers signing up for the Total Rewards loyalty card provide Harrah’s with demographic information such as gender, age, and address, which is combined with transactional data as the card is used.
  • Data mining also provides information about ninety-plus customer demographic segments, each of which responds differently to different marketing approaches.
  • If Harrah’s systems determine you’re a high-value customer, you can expect a higher level of perks and service.
  • Harrah’s CRM effort monitors any customer behavior changes.
  • Harrah’s uses its information systems and operating procedures to measure employees based on metrics that include speed and friendliness, and compensates them based on guest satisfaction ratings.

Questions and Exercises

  1. What types of customer data does Harrah’s gather?
  2. How is the data that Harrah’s collects used?
  3. Describe Harrah’s most valuable customers? Approximately what percentage of profits does this broad group deliver to the firm?
  4. List the services a Rewards Card cardholder might expect.
  5. What happens when a good, regular customer stops showing up?
  6. Describe how Harrah’s treats customer data.
  7. List some of the technology innovations that Harrah’s is using to help it gather more data, and help push service quality and marketing program success.
  8. How does Harrah’s Total Rewards loyalty card system represent a switching cost?
  9. What is customer lifetime value? Do you think this is an easier metric to calculate at Harrah’s or Wal-Mart? Why?
  10. How did intellectual property protection benefit Harrah’s?
  11. Discuss the challenges Harrah’s may have to confront in the near future.
  12. Describe the role that testing plays in initiatives? What advantage does testing provide the firm? What’s the CEO’s attitude to testing? Do you agree with this level of commitment? Why or why not?

Chapter 10: Software in Flux: Partly Cloudy and Sometimes Free

10.1 Introduction

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand how low marginal costs, network effects, and switching costs have combined to help create a huge and important industry.
  2. Recognize that the software industry is undergoing significant and broadly impactful change brought about by several increasingly adopted technologies including open source software, cloud computing, and software-as-a-service.

For many, software has been a magnificent business. It is the two-hundred-billion-dollar-per-year juggernautD. Kirkpatrick, “How the Open Source World Plans to Smack Down Microsoft and Oracle, and…,” Fortune, February 23, 2004. that placed Microsoft’s Bill Gates and Oracle’s Larry Ellison among the wealthiest people in the world. Once a successful software product has been written, the economics for a category-leading offering are among the best you’ll find in any industry. Unlike physical products assembled from raw materials, the marginal costThe cost of producing one more unit of a product. to produce an additional copy of a software product is effectively zero. Just duplicate, no additional input required. That quality leads to businesses that can gush cash. Microsoft generates one and a half billion dollars a month from Windows and Office alone.F. Vogelstein, “Rebuilding Microsoft,” Wired, October 2006. Network effects and switching cost can also offer a leading software firm a degree of customer preference and lock in that can establish a firm as a standard, and in many cases creates winner-take-all (or at least winner-take-most) markets.

But as great as the business has been, the fundamental model powering the software industry is under assault. Open source software (OSS)Software that is free and where anyone can look at and potentially modify the code. offerings—free alternatives where anyone can look at and potentially modify a program’s code—pose a direct challenge to the assets and advantages cultivated by market leaders. Giants shudder—“How can we compete with free,” while others wonder, “How can we make money and fuel innovation on free?” And if free software wasn’t enough of a shock, the way firms and users think about software is also changing. A set of services referred to as cloud computingReplacing computing resources—either an organization’s or individual’s hardware or software—with services provided over the Internet. is making it more common for a firm to move software out of its own IS shop so that it is run on someone else’s hardware. In one variant of this approach known as software as a service (SaaS)A form of cloud computing where a firm subscribes to a third-party software and receives a service that is delivered online., users access a vendor’s software over the Internet, usually by simply starting up a Web browser. With SaaS, you don’t need to own the program or install it on your own computer. Hardware clouds can let firms take their software and run it on someone else’s hardware—freeing them from the burden of buying, managing, and maintaining the physical computing that programs need. Another software technology called virtualizationA type of software that allows a single computer (or cluster of connected computers) to function as if it were several different computers, each running its own operating system and software. Virtualization software underpins most cloud computing efforts, and can make computing more efficient, cost-effective, and scalable. can make a single computer behave like many separate machines. This function helps consolidate computing resources and creates additional savings and efficiencies.

These transitions are important. They mean that smaller firms have access to the kinds of burly, sophisticated computing power than only giants had access to in the past. Start-ups can scale quickly and get up and running with less investment capital. Existing firms can leverage these technologies to reduce costs. Got tech firms in your investment portfolio? Understanding what’s at work here can inform decisions you make on which stocks to buy or sell. If you make tech decisions for your firm or make recommendations for others, these trends may point to which firms have strong growth and sustainability ahead, or which may be facing troubled times.

Key Takeaways

  • The software business is attractive due to near-zero marginal costs and an opportunity to establish a standard—creating the competitive advantages of network effects and switching costs.
  • New trends in the software industry, including open source software (OSS), hardware clouds, software as a service (SaaS), and virtualization are creating challenges and opportunity across tech markets. Understanding the impact of these developments can help a manager make better technology choices and investment decisions.

Questions and Exercises

  1. What major trends, outlined in the section above, are reshaping how we think about software? What industries and firms are potentially impacted by these changes? Why do managers, investors, and technology buyers care about these changes?
  2. Which organizations might benefit from these trends? Which might be threatened? Why?
  3. What are marginal costs? Are there other industries that have cost economics similar to the software industry?
  4. Investigate the revenues and net income of major software players: Microsoft, Google, Oracle, Red Hat, and Salesforce.com. Which firms have higher revenues? Net income? Which have better margins? What do the trends in OSS, SaaS, and cloud computing suggest for these and similar firms?
  5. How might the rise of OSS, SaaS, and cloud computing impact hardware sales? How might it impact entrepreneurship and smaller businesses?

10.2 Open Source

Learning Objectives

After studying this section you should be able to do the following:

  1. Define open source software and understand how it differs from conventional offerings.
  2. Provide examples of open source software and how firms might leverage this technology.

Who would have thought a twenty-one-year-old from Finland could start a revolution that continues to threaten the Microsoft Windows empire? But Linus Torvalds did just that. During a marathon six-month coding session, Torvalds created the first version of LinuxD. Diamond, “The Good-Hearted Wizard—Linus Torvalds,” Virtual Finland, January 2008. marshalling open source revolutionaries like no one before him. Instead of selling his operating system, Torvalds gave it away. Now morphed and modified into scores of versions by hundreds of programmers, LinuxAn open source software operating system. can be found just about everywhere, and most folks credit Linux as being the most significant product in the OSS arsenal. Today Linux powers everything from cell phones to stock exchanges, set top boxes to supercomputers. You’ll find the OS on 30 percent of the servers in corporate America,Sarah Lacy, “Open Warfare in Open Source,” BusinessWeek, August 21, 2006. and supporting most Web servers (including those at Google, Amazon, and Facebook). Linux forms the core of the TiVo operating system, it underpins Google’s Android and Chrome OS offerings, and it has even gone interplanetary. Linux has been used to power the Phoenix Lander and to control the Spirit and Opportunity Mars rovers.J. Brockmeier, “NASA Using Linux,” Unix Review, March 2004; and S. Barrett, “Linux on Mars,” Science News, Space News, Technology News, June 6, 2008. Yes, Linux is even on Mars!

How Do You Pronounce Linux?

Most English speakers in the know pronounce Linux in a way that rhymes with “cynics.” You can easily search online to hear video and audio clips of Linus (whose name is actually pronounced “Lean-us” in Finish) pronouncing the name of his OS. In deference to Linux, some geeks prefer something that sounds more like “lean-ooks.”For examples, see http://mostlylinux.ca/pronounce/torvalds-says-linux.wav and http://suseroot.com/about-suse-linux/how-do-you-pronounce-linux.php. Just don’t call it “line-ucks,” or the tech-savvy will think you’re an open source n00bWritten with two zeros, pronounced “newb.” Geek-slang (leet speak) derogatory term for an uninformed or unskilled person.! Oh yeah, and while we’re on the topic of operating system pronunciation, the Macintosh operating system OS X is pronounced “oh es ten.”

Figure 10.1 Tux, the Linux Mascot

Open source software (OSS) is often described as free. While most OSS can be downloaded for free over the Internet, it’s also “free” as in liberated. The source code for OSS products is openly shared. Anyone can look at the source code, change it, and even redistribute it, provided the modified software continues to remain open and free.A list of criteria defining open source software can be found at the Open Source Initiative at http://opensource.org/osr. This openness is in stark contrast to the practice of conventional software firms, who treat their intellectual property as closely guarded secrets, and who almost never provide the source code for their commercial software products. At times, many software industry execs have been downright hostile toward OSS. The former President of SAP once referred to the open source movement as “socialism,” while Microsoft’s Steve Balmer has called Linux a “cancer.”J. Fortt, “Why Larry Loves Linux (and He’s Not Alone),” Fortune, December 19, 2007.

But while execs at some firms see OSS as a threat undermining the lifeblood of their economic model, other big-name technology companies are now solidly behind the open source movement. The old notion of open source being fueled on the contributions of loners tooling away for the glory of contributing to better code is now largely inaccurate. The vast majority of people who work on efforts like Linux are now paid to do so by commercially motivated employers.D. Woods, “The Commercial Bear Hug of Open Source,” Forbes, August 18, 2008. Nearly every major hardware firm has paid staff contributing to open source projects, and most firms also work together to fund foundations that set standards and coordinate the release of product revisions and improvements. Such coordination is critical—helping, for example, to ensure that various versions of Linux work alike. Sun Microsystems claims to have eleven thousand engineers contributing to OSS.C. Preimesberger, “Sun’s ‘Open’-Door Policy,” eWeek, April 21, 2008. Guido van Rossum, the inventor of the open source Python programming language, works for Google where he continues to coordinate development. IBM programmers work on several open source projects, including Linux. The firm has even deeded a commercially developed programming tool (including an IDE) to the Eclipse foundation, where it’s now embraced and supported by dozens of firms.

Turn on the LAMP—It’s Free!

Figure 10.2

Open source is big on the Web. In fact, you’ll often hear Web programmers and open source advocates refer to the LAMP stack. LAMPAn acronym standing for Linux, the Apache Web server software, the MySQL database, and any of several programming languages that start with P (e.g., Perl, Python, or PHP). is an acronym that stands for the Linux operating system, the Apache Web server software, the MySQL database, and any of several programming languages that start with the letter “P”—Perl, Python, and PHP. From Facebook to YouTube, you’ll find LAMP software powering many of the sites you visit each day.

Key Takeaways

  • OSS is not only available for free, but also makes source code available for review and modification (for the Open Source Initiatives list of the criteria that define an open source software product, see http://opensource.org/docs/osd).
  • While open source alternatives are threatening to conventional software firms, some of the largest technology companies now support OSS initiatives and work to coordinate standards, product improvements, and official releases.
  • The flagship OSS product is the Linux operating system, now available on all scales of computing devices from cell phones to supercomputers.
  • The LAMP stack of open source products is used to power many of the Internet’s most popular Web sites. Linux can be found on 30 percent of corporate servers, supports most Web servers, and is integral to TiVo and Android-based cell phones.
  • The majority of persons who work on open source projects are paid by commercially motivated employers.

Questions and Exercises

  1. Who developed Linux?
  2. Who develops it today?
  3. List the components of the LAMP stack. Which commercial products do these components compete with (investigate online, if necessary)?
  4. Why do commercial firms contribute to open source consortia and foundations?
  5. Free doesn’t always win. Why might a firm turn down free software in favor of a commercial alternative?

10.3 Why Open Source?

Learning Objectives

After studying this section you should be able to do the following:

  1. Know the primary reasons firms choose to use OSS.
  2. Understand how OSS can beneficially impact industry and government.

There are many reasons why firms choose open source products over commercial alternatives:

Cost—Free alternatives to costly commercial code can be a tremendous motivator, particularly since conventional software often requires customers to pay for every copy used and to pay more for software that runs on increasingly powerful hardware. Big Lots stores lowered costs by as much as $10 million by finding viable OSSM. Castelluccio, “Enterprise Open Source Adoption,” Strategic Finance, November 2008. to serve their system needs. Online broker E*TRADE estimates that its switch to open source helped save over $13 million a year.R. King, “Cost-Conscious Companies Turn to Open-Source Software,” BusinessWeek, December 1, 2008. And Amazon claimed in SEC filings that the switch to open source was a key contributor to nearly $20 million in tech savings.S. Shankland, M. Kane, and R. Lemos, “How Linux Saved Amazon Millions,” CNET, October 30, 2001. Firms like TiVo, which use OSS in their own products, eliminate a cost spent either developing their own operating system or licensing similar software from a vendor like Microsoft.

Reliability—There’s a saying in the open source community, “Given enough eyeballs, all bugs are shallow.”E. Raymond, The Cathedral and the Bazaar: Musings on Linux and Open Source by an Accidental Revolutionary (Sebastopol, CA: O’Reilly, 1999). What this means is that the more people who look at a program’s code, the greater the likelihood that an error will be caught and corrected. The open source community harnesses the power of legions of geeks who are constantly trawling OSS products, looking to squash bugs and improve product quality. And studies have shown that the quality of popular OSS products outperforms proprietary commercial competitors.J. Ljungberg, “Open Source Movements as a Model for Organizing,” European Journal of Information Systems 9, no. 4 (December 2000): 208–16. In one study, Carnegie Mellon University’s Cylab estimated the quality of Linux code to be less buggy than commercial alternatives by a factor of two hundred!M. Castelluccio, “Enterprise Open Source Adoption,” Strategic Finance, November 2008.

Security—OSS advocates also argue that by allowing “many eyes” to examine the code, the security vulnerabilities of open source products come to light more quickly and can be addressed with greater speed and reliability.D. Wheeler, Secure Programming for Linux and Unix, 2003, http://www.dwheeler.com/secure-programs/Secure-Programs-HOWTO/index.html. High profile hacking contests have frequently demonstrated the strength of OSS products. In one well-publicized 2008 event, laptops running Windows and Macintosh were both hacked (the latter in just two minutes), while a laptop running Linux remained uncompromised.R. McMillan, “Gone in Two Minutes,” InfoWorld, March 27, 2008. Government agencies and the military often appreciate the opportunity to scrutinize open source efforts to verify system integrity (a particularly sensitive issue among foreign governments leery of legislation like the USA PATRIOT Act of 2001).S. Lohr, “Microsoft to Give Governments Access to Code,” New York Times, January 15, 2003. Many OSS vendors offer security focusedAlso known as “hardened.” Term used to describe technology products that contain particularly strong security features. (sometimes called hardened) versions of their products. These can include systems that monitor the integrity of an OSS distribution, checking file size and other indicators to be sure that code has not been modified and redistributed by bad guys who’ve added a back door, malicious routines, or other vulnerabilities.

Scalability—Many major OSS efforts can run on everything from cheap commodity hardware to high-end supercomputing. ScalabilityAbility to either handle increasing workloads or to be easily expanded to manage workload increases. In a software context, systems that aren’t scalable often require significant rewrites or the purchase or development of entirely new systems. allows a firm to scale from start-up to blue chip without having to significantly rewrite their code, potentially saving big on software development costs. Not only can many forms of OSS be migrated to more powerful hardware, packages like Linux have also been optimized to balance a server’s workload among a large number of machines working in tandem. Brokerage firm E*TRADE claims that usage spikes following 2008 U.S. Federal Reserve moves flooded the firm’s systems, creating the highest utilization levels in five years. But E*TRADE credits its scalable open source systems for maintaining performance while competitors’ systems struggled.R. King, “Cost-Conscious Companies Turn to Open-Source Software,” BusinessWeek, December 1, 2008.

Agility and Time to Market—Vendors who use OSS as part of product offerings may be able to skip whole segments of the software development process, allowing new products to reach the market faster than if the entire software system had to be developed from scratch, in-house. Motorola has claimed that customizing products built on OSS has helped speed time-to-market for the firm’s mobile phones, while the team behind the Zimbra e-mail and calendar effort built their first product in just a few months by using some forty blocks of free code.R. Guth, “Virtual Piecework: Trolling the Web for Free Labor, Software Upstarts Are a New Force,” Wall Street Journal, November 13, 2006.

Key Takeaways

  • The most widely cited benefits of using OSS include low cost; increased reliability; improved security and auditing; system scalability; and helping a firm improve its time to market.
  • Free OSS has resulted in cost savings for many large companies in several industries.
  • OSS often has fewer bugs than its commercial counterparts due to the large number of persons who have looked at the code.
  • The huge exposure to scrutiny by developers and other people helps to strengthen the security of OSS.
  • “Hardened” versions of OSS products often include systems that monitor the integrity of an OSS distribution, checking file size and other indicators to be sure that code has not been modified and redistributed by bad guys who have added a back door, malicious routines, or other vulnerabilities.
  • OSS can be easily migrated to more powerful computers as circumstances dictate, and also can balance workload by distributing work over a number of machines.
  • Vendors who use OSS as part of product offerings may be able to skip whole segments of the software development process, allowing new products to reach the market faster.

Questions and Exercises

  1. What advantages does OSS offer TiVo? What alternatives to OSS might the firm consider and why do you suppose the firm decided on OSS?
  2. What’s meant by the phrase, “Given enough eyeballs, all bugs are shallow”? Provide evidence that the insight behind this phrase is an accurate one.
  3. How has OSS benefited E*TRADE? Amazon? Motorola? Zimbra? What benefits were achieved in each of these examples?
  4. Describe how OSS provides a firm with scalability. What does this mean, and why does this appeal to a firm? What issues might a firm face if chosen systems aren’t scalable?
  5. The Web site NetCraft (http://www.netcraft.com) is one of many that provide a tool to see the kind of operating system and Web server software that a given site is running. Visit NetCraft or a similar site and enter the address of some of your favorite Web sites. How many run open source products (e.g., the Linux OS or Apache Web server)? Do some sites show their software as “unknown”? Why might a site be reluctant to broadcast the kind of software that it uses?

10.4 Examples of Open Source Software

Learning Objectives

After studying this section you should be able to do the following:

  1. Recognize that just about every type of commercial product has an open source equivalent.
  2. Be able to list commercial products and their open source competitors.

Just about every type of commercial product has an open source equivalent. SourceForge.net lists over two hundred and thirty thousand such products!See http://sourceforge.net. Many of these products come with the installation tools, support utilities, and full documentation that make them difficult to distinguish from traditional commercial efforts.D. Woods, “The Commercial Bear Hug of Open Source,” Forbes, August 18, 2008. In addition to the LAMP products, some major examples include the following:

  • Firefox—a Web browser that competes with Internet Explorer
  • OpenOffice—a competitor to Microsoft Office
  • Gimp—a graphic tool with features found in Photoshop
  • Alfresco—collaboration software that competes with Microsoft Sharepoint and EMC’s Documentum
  • Marketcetera—an enterprise trading platform for hedge fund managers that competes with FlexTrade and Portware
  • Zimbra—open source e-mail software that competes with Outlook server
  • MySQL, Ingres, and EnterpriseDB—open source database software packages that each go head-to-head with commercial products from Oracle, Microsoft, Sybase, and IBM
  • SugarCRM—customer relationship management software that competes with Salesforce.com and Siebel
  • Asterix—an open source implementation for running a PBX corporate telephony system that competes with offerings from Nortel and Cisco, among others
  • Free BSD and Sun’s OpenSolaris—open source versions of the Unix operating system

Key Takeaways

  • There are thousands of open source products available, covering nearly every software category. Many have a sophistication that rivals commercial software products.
  • Not all open source products are contenders. Less popular open source products are not likely to attract the community of users and contributors necessary to help these products improve over time (again we see network effects are a key to success—this time in determining the quality of an OSS effort).
  • Just about every type of commercial product has an open source equivalent.

Questions and Exercises

  1. Visit http://www.SourceForge.net. Make a brief list of commercial product categories that an individual or enterprise might use. Are there open source alternatives for these categories? Are well-known firms leveraging these OSS offerings? Which commercial firms do they compete with?
  2. Are the OSS efforts you identified above provided by commercial firms, nonprofit organizations, or private individuals? Does this make a difference in your willingness to adopt a particular product? Why or why not? What other factors influence your adoption decision?
  3. Download a popular, end-user version of an OSS tool that competes with a desktop application that you own, or that you’ve used (hint: choose something that’s a smaller file or easy to install). What do you think of the OSS offering compared to the commercial product? Will you continue to use the OSS product? Why or why not?

10.5 Why Give It Away? The Business of Open Source

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the disproportional impact OSS has on the IT market.
  2. Understand how vendors make money on open source.
  3. Know what SQL and MySQL are.

Open source is a sixty-billion-dollar industry,M. Asay, “Open Source Is a $60 Billion Industry,” CNET, May 15, 2008. but it has a disproportionate impact on the trillion-dollar IT market. By lowering the cost of computing, open source efforts make more computing options accessible to smaller firms. More reliable, secure computing also lowers costs for all users. OSS also diverts funds that firms would otherwise spend on fixed costs, like operating systems and databases, so that these funds can be spent on innovation or other more competitive initiatives. Think about Google, a firm that some estimate has over 1.4 million servers. Imagine the costs if it had to license software for each of those boxes!

Commercial interest in OSS has sparked an acquisition binge. Red Hat bought open source application server firm JBoss for $350 million. Novell snapped up SUSE Linux for $210 million. And Sun plunked down over $1 billion for open source database provider MySQL.A. Greenberg, “Sun Snaps Up Database Firm, MySQL,” Forbes, January 16, 2008. And with Oracle’s acquisition of Sun, one of the world’s largest commercial software firms has zeroed in on one of the deepest portfolios of open source products.

But how do vendors make money on open source? One way is by selling support and consulting services. While not exactly Microsoft money, Red Hat, the largest purely OSS firm, reported half a billion dollars in revenue in 2008. The firm had two and a half million paid subscriptions offering access to software updates and support services.A. Greenberg, “Sun Snaps Up Database Firm, MySQL,” Forbes, January 16, 2008. Oracle, a firm that sells commercial ERP and database products, provides Linux for free, selling high-margin Linux support contracts for as much as five hundred thousand dollars.J. Fortt, “Why Larry Loves Linux (and He’s Not Alone),” Fortune, December 19, 2007. The added benefit for Oracle? Weaning customers away from Microsoft—a firm that sells many products that compete head-to-head with Oracle’s offerings. Service also represents the most important part of IBM’s business. The firm now makes more from services than from selling hardware and software.J. Robertson, “IBM Sees Better-Than-Expected 2009 Profit, Earns US$4.4 Billion in Q4,” Associated Press, January 20, 2009, http://humantimes.com/finance/business/sanfrancis/54853. And every dollar saved on buying someone else’s software product means more money IBM customers can spend on IBM computers and services. Sun Microsystems was a leader in OSS, even before the Oracle acquisition bid. The firm has used OSS to drive advanced hardware sales, but the firm also sells proprietary products that augment its open source efforts. These products include special optimization, configuration management, and performance tools that can tweak OSS code to work its best.C. Preimesberger, “Sun’s ‘Open’-Door Policy,” eWeek, April 21, 2008.

Here’s where we also can relate the industry’s evolution to what we’ve learned about standards competition in our earlier chapters. In the pre-Linux days, nearly every major hardware manufacturer made its own, incompatible version of the Unix operating system. These fractured, incompatible markets were each so small that they had difficulty attracting third-party vendors to write application software. Now, much to Microsoft’s dismay, all major hardware firms run Linux. That means there’s a large, unified market that attracts software developers who might otherwise write for Windows.

To keep standards unified, several Linux-supporting hardware and software firms also back the Linux Foundation, the nonprofit effort where Linus Torvalds serves as a fellow, helping to oversee Linux’s evolution. Sharing development expenses in OSS has been likened to going in on a pizza together. Everyone wants a pizza with the same ingredients. The pizza doesn’t make you smarter or better. So why not share the cost of a bigger pie instead of buying by the slice?S. Cohen, “Open Source: The Model Is Broken,” BusinessWeek, December 1, 2008. With OSS, hardware firms spend less money than they would in the brutal, head-to-head competition where each once offered a “me too” operating system that was incompatible with rivals but offered little differentiation. Hardware firms now find their technical talent can be deployed in other value-added services mentioned above: developing commercial software add-ons, offering consulting services, and enhancing hardware offerings.

Linux on the Desktop?

While Linux is a major player in enterprise software, mobile phones, and consumer electronics, the Linux OS can only be found on a tiny fraction of desktop computers. There are several reasons for this. Some suggest Linux simply isn’t as easy to install and use as Windows or the Mac OS. This complexity can raise the total cost of ownership (TCO)All of the costs associated with the design, development, testing, implementation, documentation, training and maintenance of a software system. of Linux desktops, with additional end-user support offsetting any gains from free software. The small number of desktop users also dissuades third party firms from porting popular desktop applications over to Linux. For consumers in most industrialized nations, the added complexity and limited desktop application availability of desktop Linux just it isn’t worth the one to two hundred dollars saved by giving up Windows.

But in developing nations where incomes are lower, the cost of Windows can be daunting. Consider the OLPC, Nicholas Negroponte’s “one-hundred-dollar” laptop. An additional one hundred dollars for Windows would double the target cost for the nonprofit’s machines. It is not surprising that the first OLPC laptops ran Linux. Microsoft recognizes that if a whole generation of first-time computer users grows up without Windows, they may favor open source alternatives years later when starting their own businesses. As a result, Microsoft has begun offering low-cost versions of Windows (in some cases for as little as seven dollars) in nations where populations have much lower incomes. Microsoft has even offered a version of Windows to the backers of the OLPC. While Microsoft won’t make much money on these efforts, the low cost versions will serve to entrench Microsoft products as standards in emerging markets, staving off open source rivals and positioning the firm to raise prices years later when income levels rise.

MySQL: Turning a Ten-Billion-Dollars-a-Year Business into a One-Billion-Dollar One

Finland is not the only Scandinavian country to spawn an open source powerhouse. Uppsala Sweden’s MySQL (pronounced “my sequel”) is the “M” in the LAMP stack, and is used by organizations as diverse as FedEx, Lufthansa, NASA, Sony, UPS, and YouTube.

The “SQL” in name stands for the structured query languageA language for creating and manipulating databases. SQL is by far the most common database standard in use today, and is supported by many commercial and open source products., a standard method for organizing and accessing data. SQL is also employed by commercial database products from Oracle, Microsoft, and Sybase. Even Linux-loving IBM uses SQL in its own lucrative DB2 commercial database product. Since all of these databases are based on the same standard, switching costs are lower, so migrating from a commercial product to MySQL’s open source alternative is relatively easy. And that spells trouble for commercial firms. Granted, the commercial efforts offer some bells and whistles that MySQL doesn’t yet have, but those extras aren’t necessary in a lot of standard database use. Some organizations, impressed with MySQL’s capabilities, are mandating its use on all new development efforts, attempting to cordon off proprietary products in legacy code that is maintained but not expanded.

Savings from using MySQL can be huge. The Web site PriceGrabber pays less than ten thousand dollars in support for MySQL compared to one hundred thousand to two hundred thousand dollars for a comparable Oracle effort. Lycos Europe switched from Oracle to MySQL and slashed costs from one hundred twenty thousand dollars a year to seven thousand dollars. And the travel reservation firm Sabre used open source products such as MySQL to slash ticket purchase processing costs by 80 percent.D. Lyons, “Cheapware,” Forbes, September 6, 2004.

MySQL does make money, just not as much as its commercial rivals. While you can download a version of MySQL over the Net, the flagship product also sells for four hundred ninety-five dollars per server computer compared to a list price for Oracle that can climb as high as one hundred sixty thousand dollars. Of the roughly eleven million copies of MySQL in use, the company only gets paid for about one in a thousand.A. Ricadela, “The Worth of Open Source? Open Question,” BusinessWeek, June 26, 2007. Firms pay for what’s free for one of two reasons: (1) for MySQL service, and (2) for the right to incorporate MySQL’s code into their own products.D. Kirkpatrick, “How the Open Source World Plans to Smack Down Microsoft and Oracle, and…,” Fortune, February 23, 2004. Amazon, Facebook, Gap, NBC, and Sabre pay MySQL for support; Cisco, Ericsson, HP, and Symantec pay for the rights to the code.A. Ricadela, “The Worth of Open Source? Open Question,” BusinessWeek, June 26, 2007. Top-level round-the-clock support for MySQL for up to fifty servers is fifty thousand dollars a year, still a fraction of the cost for commercial alternatives. Founder Marten Mickos has stated an explicit goal of the firm is “turning the $10-billion-a-year database business into a $1 billion one.”D. Kirkpatrick, “How the Open Source World Plans to Smack Down Microsoft and Oracle, and…,” Fortune, February 23, 2004.

When Sun Microsystems spent over $1 billion to buy Mickos’ MySQL in 2008, Sun CEO Jonathan Schwartz called the purchase the “most important acquisition in the company’s history.”S. Shankland, “Google’s Open-Source Android Now Actually Open,” CNET, October 21, 2008, http://news.cnet.com/8301-1001_3-10071093-92.html. Sun hoped the cheap database software could make the firm’s hardware offerings seem more attractive. And it looked like Sun was good for MySQL, with the product’s revenues growing 55 percent in the year after the acquisition.M. Asay, “Open-Source Database Market Shows Muscles,” CNET, February 3, 2009, http://news.cnet.com/8301-13505_3-10156188-16.html.

But here’s where it gets complicated. Sun also had a lucrative business selling hardware to support commercial ERP and database software from Oracle. That put Sun and partner Oracle in a relationship where they were both competitors and collaborators (the “coopetition” or “frenemies” phenomenon mentioned in Chapter 6 “Understanding Network Effects”). Then in spring 2009, Oracle announced it was buying Sun. Oracle CEO Larry Ellison mentioned acquiring the Java language was the crown jewel of the purchase, but industry watchers have raised several questions. Will the firm continue to nurture MySQL and other open source products, even as this software poses a threat to its bread-and-butter database products? Will the development community continue to back MySQL as the de facto standard for open source SQL databases, or will they migrate to an alternative? Or will Oracle find the right mix of free and fee-based products and services that allow MySQL to thrive while Oracle continues to grow? The implications are serious for investors, as well as firms that have made commitments to Sun, Oracle, and MySQL products. The complexity of this environment further demonstrates why technologists need business savvy and market monitoring skills and why business folks need to understand the implications of technology and tech-industry developments.

Legal Risks and Open Source Software: A Hidden and Complex Challenge

Open source software isn’t without its risks. Competing reports cite certain open source products as being difficult to install and maintain (suggesting potentially higher total cost of ownership, or TCO). Adopters of OSS without support contracts may lament having to rely on an uncertain community of volunteers to support their problems and provide innovative upgrades. Another major concern is legal exposure. Firms adopting OSS may be at risk if they distribute code and aren’t aware of the licensing implications. Some commercial software firms have pressed legal action against the users of open source products when there is a perceived violation of software patents or other unauthorized use of their proprietary code.

For example, in 2007 Microsoft suggested that Linux and other open source software efforts violated some two hundred thirty-five of its patents.A. Ricadela, “Microsoft Wants to ‘Kill’ Open Source,” BusinessWeek, May 15, 2007. The firm then began collecting payments and gaining access to the patent portfolios of companies that use the open source Linux operating system in their products, including Fuji, Samsung, and Xerox. Microsoft also cut a deal with Linux vendor Novell in which both firms pledged not to sue each other’s customers for potential patent infringements.

Also complicating issues are the varying open source license agreements (these go by various names, such as GPL and the Apache License), each with slightly different legal provisions—many of which have evolved over time. Keeping legal with so many licensing standards can be a challenge, especially for firms that want to bundle open source code into their own products.Sarah Lacy, “Open Warfare in Open Source,” BusinessWeek, August 21, 2006. An entire industry has sprouted up to help firms navigate the minefield of open source legal licenses. Chief among these are products, such as those offered by the firm Black Duck, which analyze the composition of software source code and report on any areas of concern so that firms can honor any legal obligations associated with their offerings. Keeping legal requires effort and attention, even in an environment where products are allegedly “free.” This also shows that even corporate lawyers had best geek-up if they want to prove they’re capable of navigating a twenty-first-century legal environment.

Key Takeaways

  • Business models for firms in the open source industry are varied, and can include selling services, licensing OSS for incorporation into commercial products, and using OSS to fuel hardware sales.
  • Many firms are trying to use OSS markets to drive a wedge between competitors and their customers.
  • Linux has been very successful on mobile devices and consumer electronics, as well as on high-end server class and above computers. But it has not been as successful on the desktop. The small user base for desktop Linux makes the platform less attractive for desktop software developers. Incompatibility with Windows applications, switching costs, and other network effects-related issues all suggest that Desktop Linux has an uphill climb in more mature markets.
  • MySQL is the dominant open source database software product. Adoption of the SQL standard eases some issues with migrating from commercial products to MySQL.
  • OSS also has several drawbacks and challenges that limit its appeal. These include complexity of some products and a higher total cost of ownership for some products, concern about the ability of a product’s development community to provide support or product improvement, and legal and licensing concerns.

Questions and Exercises

  1. Describe the impact of OSS on the IT market.
  2. Show your understanding of the commercial OSS market. How do Red Hat, Oracle, Oracle’s Sun division, and IBM make money via open source?
  3. Visit Mozilla.org. Which open source products does this organization develop? Investigate how development of these efforts is financed. How does this organization differ from the ones mentioned above?
  4. What is the Linux Foundation? Why is it necessary? Which firms are members, underwriting foundation efforts?
  5. List the reasons why Linux is installed on only a very small fraction of desktop computers. Are there particular categories of products or users who might see Linux as more appealing than conventional operating systems? Do you think Linux’s share of the desktop market will increase? Why or why not?
  6. How is Microsoft combating the threat of open source software and other free tools that compete with its commercial products?
  7. What is the dominant open source database software product? Which firms use this product? Why?
  8. Which firm developed the leading OSS database product? Do you think it’s more or less likely that a firm would switch to an OSS database instead of an OSS office suite or desktop alternative? Why or why not?
  9. How has stewardship of the leading OSS database effort changed in recent years? Who oversees the effort today? What questions does this raise for the product’s future? Although this book is updated regularly, current events continue to change after publication of this chapter. Investigate the current status of this effort—reaction of the developer community, continued reception of the product—and be prepared to share your findings with class.
  10. List some of the risks associated with using OSS. Give examples of firms that might pass on OSS software, and explain why.

10.6 Cloud Computing: Hype or Hope?

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the concept of cloud computing.
  2. Identify the two major categories of cloud computing.

Oracle Chairman Larry Ellison, lamenting the buzzword-chasing character of the tech sector, once complained that the computer industry is more fashion-focused than even the women’s clothing business.D. Farber, “Oracle’s Ellison Nails Cloud Computing,” CNET, September 26, 2008, http://news.cnet.com/8301-13953_3-10052188-80.html?tag=mncol;txt. Ellison has a point: when a technology term becomes fashionable, the industry hype machine shifts into overdrive. The technology attracts press attention, customer interest, and vendor marketing teams scramble to label their products and services as part of that innovation. Recently, few tech trends have been more fashionable than cloud computing.

Like Web 2.0, trying to nail down an exact definition for cloud computing is tough. In fact, it’s been quite a spectacle watching industry execs struggle to clarify the concept. HP’s Chief Strategy Office “politely refused” when asked by BusinessWeek to define the term cloud computing.S. Hamm, “Cloud Computing: Eyes on the Skies,” BusinessWeek, April 24, 2008. Richard Stallman, founder of the Free Software Foundation said about cloud computing, “It’s worse than stupidity. It’s a marketing hype campaign.”L. McKay, “30,000-Foot Views of the Cloud,” Customer Relationship Management, January 2009. And Larry Ellison, always ready with a sound bite, offered up this priceless quip, “Maybe I’m an idiot, but I have no idea what anyone is talking about. What is it? It’s complete gibberish. It’s insane.”D. Lyons, “A Mostly Cloudy Computing Forecast,” Washington Post, November 4, 2008. Insane, maybe, but also big bucks. By year-end 2008, the various businesses that fall under the rubric of cloud computing had already accounted for an estimated thirty-six-billion-dollar market. That represents a whopping 13 percent of global software sales!M. Liedtke, “Cloud Computing: Pie in the Sky Concept or the Next Big Breakthrough on Tech Horizon?” Associated Press Newswires, December 21, 2008.

When folks talk about cloud computing they’re really talking about replacing computing resources—either an organization’s or an individual’s hardware or software—with services provided over the Internet. The name actually comes from the popular industry convention of drawing the Internet or other computer network as a big cloud.

Cloud computing encompasses a bunch of different efforts. We’ll concentrate on describing, providing examples, and analyzing the managerial implications of two separate categories of cloud computing: (1) software as a service (SaaS), where a firm subscribes to a third-party software-replacing service that is delivered online, and (2) models often referred to as utility computingA form of cloud computing where a firm develops its own software, and then runs it over the Internet on a service provider’s computers., platform as a service, or infrastructure as a service. Using these latter techniques, an organization develops its own systems, but runs them over the Internet on someone else’s hardware. A later section on virtualization will discuss how some organizations are developing their own private cloudsPools of computing resources that reside inside an organization and that can be served up for specific tasks as need arrives., pools of computing resources that reside inside an organization and that can be served up for specific tasks as need arrives.

The benefits and risks of SaaS and the utility computing-style efforts are very similar, but understanding the nuances of each effort can help you figure out if and when the cloud makes sense for your organization. The evolution of cloud computing also has huge implications across the industry: from the financial future of hardware and software firms, to cost structure and innovativeness of adopting organizations, to the skill sets likely to be most valued by employers.

Key Takeaways

  • Cloud computing is difficult to define. Managers and techies use the term cloud computing to describe computing services provided over a network, most often commercial services provided over the Internet by a third party that can replace or offload tasks that would otherwise run on a user or organization’s existing hardware or software.
  • Software as a service (SaaS) refers to a third-party software-replacing service that is delivered online.
  • Hardware cloud computing services replace hardware that a firm might otherwise purchase.
  • Estimated to be a thirty-six-billion-dollar industry, cloud computing is reshaping software, hardware, and service markets, and is impacting competitive dynamics across industries.

Questions and Exercises

  1. Identify and contrast the two categories of cloud computing.
  2. Define cloud computing.

10.7 The Software Cloud: Why Buy When You Can Rent?

Learning Objectives

After studying this section you should be able to do the following:

  1. Know how firms using SaaS products can dramatically lower several costs associated with their information systems.
  2. Know how SaaS vendors earn their money.
  3. Be able to list the benefits to users that accrue from using SaaS.
  4. Be able to list the benefits to vendors from deploying SaaS.

If open source isn’t enough of a threat to firms that sell packaged software, a new generation of products, collectively known as SaaS, claims that you can now get the bulk of your computing done through your Web browser. Don’t install software—let someone else run it for you and deliver the results over the Internet.

Software as a service (SaaS) refers to software that is made available by a third party online. You might also see the terms ASP (application service provider) or HSV (hosted software vendor) used to identify this type of offering. SaaS is potentially a very big deal. Firms using SaaS products can dramatically lower several costs associated with the care and feeding of their information systems, including software licenses, server hardware, system maintenance, and IT staff. Most SaaS firms earn money via a usage-based pricing model akin to a monthly subscription. Others offer free services that are supported by advertising, while others promote the sale of upgraded or premium versions for additional fees.

Make no mistake, SaaS is yet another direct assault on traditional software firms. The most iconic SaaS firm is Salesforce.com, an enterprise customer relationship management (CRM) provider. This “un-software” company even sports a logo featuring the word “software” crossed out, Ghostbusters-style.J. Hempel, “Salesforce Hits Its Stride,” Fortune, March 2, 2009.

Figure 10.3

The antisoftware message is evident in the logo of SaaS leader Salesforce.com.

Other enterprise-focused SaaS firms compete directly with the biggest names in software. Some of these upstarts are even backed by leading enterprise software executives. Examples include NetSuite (funded in part by Oracle’s Larry Ellison—the guy’s all over this chapter), which offers a comprehensive SaaS ERP suite; and Workday (launched by founders of Peoplesoft), which has SaaS offerings for managing human resources. Several traditional software firms have countered start-ups by offering SaaS efforts of their own. IBM offers a SaaS version of its Cognos business intelligence products, Oracle offers CRM On Demand, and SAP’s Business ByDesign includes a full suite of enterprise SaaS offerings. Even Microsoft has gone SaaS, with a variety of Web-based services that include CRM, Web meeting tools, collaboration, e-mail, and calendaring.

SaaS is also taking on desktop applications. Intuit has online versions of its QuickBooks, TurboTax, and Quicken finance software. Adobe has an online version of Photoshop. Google and Zoho offer office suites that compete with desktop alternatives, prompting Microsoft’s own introduction of an online version of Office. And if you store photos on Flickr or Picassa instead of your PC’s hard drive, then you’re using SaaS, too.

Figure 10.4

A look at Zoho’s home page shows the diversity of both desktop and enterprise offerings from this SaaS upstart. Note that the firm makes it services available through browsers, phones, and even Facebook.

The Benefits of SaaS

Firms can potentially save big using SaaS. Organizations that adopt SaaS forgo the large upfront costs of buying and installing software packages. For large enterprises, the cost to license, install, and configure products like ERP and CRM systems can easily run into the hundreds of thousands or even millions of dollars. And these costs are rarely a one time fee. Additional costs like annual maintenance contracts have also been rising as rivals fail or get bought up. Less competition among traditional firms recently allowed Oracle and SAP to raise maintenance fees to as much as 20 percent.Sarah Lacy, “On-Demand Computing: A Brutal Slog,” BusinessWeek, July 18, 2008.

Firms that adopt SaaS don’t just save on software and hardware, either. There’s also the added cost for the IT staff needed to run these systems. Forrester Research estimates that SaaS can bring cost savings of 25 to 60 percent if all these costs are factored in.J. Quittner, “How SaaS Helps Cut Small Business Costs,” BusinessWeek, December 5, 2008.

There are also accounting and corporate finance implications for SaaS. Firms that adopt software as a service never actually buy a system’s software and hardware, so these systems become a variable operating expense. This flexibility helps mitigate the financial risks associated with making a large capital investment in information systems. For example, if a firm pays Salesforce.com sixty-five dollars per month per user for its CRM software, it can reduce payments during a slow season with a smaller staff, or pay more during heavy months when a firm might employ temporary workers. At these rates, SaaS not only looks good to large firms, it makes very sophisticated technology available to smaller firms that otherwise wouldn’t be able to afford expensive systems, let alone the IT staff and hardware required to run them.

In addition to cost benefits, SaaS offerings also provide the advantage of being highly scalable. This feature is important because many organizations operate in environments prone to wide variance in usage. Some firms might expect systems to be particularly busy during tax time or the period around quarterly financial reporting deadlines, while others might have their heaviest system loads around a holiday season. A music label might see spikes when an artist drops a new album. Using conventional software, an organization would have to buy enough computing capacity to ensure that it could handle its heaviest anticipated workload. But sometimes these loads are difficult to predict, and if the difference between high workloads and average use is great, a lot of that expensive computer hardware will spend most of its time doing nothing. In SaaS, however, the vendor is responsible for ensuring that systems meet demand fluctuation. Vendors frequently sign a service level agreement (SLA)A negotiated agreement between the customer and the vendor. The SLA may specify the levels of availability, serviceability, performance, operation, or other commitment requirements. with their customers to ensure a guaranteed uptime and define their ability to meet demand spikes.

When looking at the benefits of SaaS, also consider the potential for higher quality and service levels. SaaS firms benefit from economies of scale that not only lower software and hardware costs, but also potentially boost quality. The volume of customers and diversity of their experiences means that an established SaaS vendor is most likely an expert in dealing with all sorts of critical computing issues. SaaS firms handle backups, instantly deploy upgrades and bug fixes, and deal with the continual burden of security maintenance—all costly tasks that must be performed regularly and with care, although each offers little strategic value to firms that perform these functions themselves in-house. The breadth of a SaaS vendor’s customer base typically pushes the firm to evaluate and address new technologies as they emerge, like quickly offering accessibility from mobile platforms like the BlackBerry and iPhone. For all but the savviest of IT shops, an established SaaS vendor can likely leverage its scale and experience to provide better, cheaper, more reliable standard information systems than individual companies typically can.

Software developers who choose to operate as SaaS providers also realize benefits. While a packaged software company like SAP must support multiple versions of its software to accommodate operating systems like Windows, Linux, and various flavors of Unix, an SaaS provider develops, tests, deploys, and supports just one version of the software executing on its own servers.

An argument might also be made that SaaS vendors are more attuned to customer needs. Since SaaS firms run a customer’s systems on their own hardware, they have a tighter feedback loop in understanding how products are used (and why they fail)—potentially accelerating their ability to enhance their offerings. And once made, enhancements or fixes are immediately available to customers the next time they log in.

SaaS applications also impact distribution costs and capacity. As much as 30 percent of the price of traditional desktop software is tied to the cost of distribution—pressing CD-ROMs, packaging them in boxes, and shipping them to retail outlets.M. Drummond, “The End of Software as We Know It,” Fortune, November 19, 2001. Going direct to consumers can cut out the middleman, so vendors can charge less or capture profits that they might otherwise share with a store or other distributor. Going direct also means that SaaS applications are available anywhere someone has an Internet connection, making them truly global applications. This feature has allowed many SaaS firms to address highly specialized markets (sometimes called vertical nichesSometimes referred to as vertical markets. Products and services designed to target a specific industry (e.g., pharmaceutical, legal, apparel retail).). For example, the Internet allows a company writing specialized legal software, for example, or a custom package for the pharmaceutical industry, to have a national deployment footprint from day one. Vendors of desktop applications that go SaaS benefit from this kind of distribution, too.

Finally, SaaS allows a vendor to counter the vexing and costly problem of software piracy. It’s just about impossible to make an executable, illegal copy of a subscription service that runs on a SaaS provider’s hardware.

Gaming in Flux: Is There a Future in Free?

PC game makers are in a particularly tough spot. Development costs are growing as games become more sophisticated. But profits are plummeting as firms face rampant piracy, a growing market for used game sales, and lower sales from rental options from firms like Blockbuster and GameFly. To combat these trends, Electronic Arts (EA) has begun to experiment with a radical alternative to PC game sales—give the base version of the product away for free and make money by selling additional features.

The firm started with the Korean version of its popular FIFA soccer game. Koreans are crazy for the world’s most popular sport; their nation even cohosted the World Cup in 2002. But piracy was killing EA’s sales in Korea. To combat the problem, EA created a free, online version of FIFA that let fans pay for additional features and upgrades, such as new uniforms for their virtual teams, or performance-enhancing add-ons. Each enhancement only costs about one dollar and fifty cents, but the move to a model based on these so-called microtransactionsSmall payments, typically paid for products or services purchased online. Gaming firms, news, music, and other media efforts have all experimented with microtransactions, albeit with varying degrees of success. has brought in big earnings. During the first two years that the microtransaction-based Korean FIFA game was available, EA raked in roughly $1 million a month. The two-year, twenty-four-million-dollar take was twice the sales record for EA’s original FIFA game.

Asian markets have been particularly receptive to microtransactions—this revenue model makes up a whopping 50 percent of the region’s gaming revenues. But whether this model can spread to other parts of the world remains to be seen. The firm’s first free, microtransaction offering outside of Korea leverages EA’s popular Battlefield franchise. Battlefield Heroes sports lower quality, more cartoon-like graphics than EA’s conventional Battlefield offerings, but it will be offered free online. Lest someone think they can rise to the top of player rankings by buying the best military hardware for their virtual armies, EA offers a sophisticated matching engine, pitting players with similar abilities and add-ons against one another.J. Schenker, “EA Leaps into Free Video Games,” BusinessWeek, January 22, 2008.

Players of the first versions of Battlefield Heroes and FIFA Online needed to download software to their PC. But the start-up World Golf Tour shows how increasingly sophisticated games can execute within a browser, SaaS-style. WGT doesn’t have quite the graphics sophistication of the dominant desktop golf game (EA’s Tiger Woods PGA Golf), but the free, ad-supported offering is surprisingly detailed. Buddies can meet up online for a virtual foursome, played on high-resolution representations of the world’s elite courses stitched together from fly-over photographs taken as part of game development. World Golf Tour is ad-supported. The firm hopes that advertisers will covet access to the high-income office workers likely to favor a quick virtual golf game to break up their workday. Zynga’s FarmVille, an app game for Facebook, combines both models. Free online, but offering added features purchased in micropayment-sized chunks, FarmVille made half a million dollars in three days, just by selling five-dollar virtual sweet potatoes.D. MacMillan, P. Burrows, and S. Ante, “Inside the App Economy,” BusinessWeek, October 22, 2009. FIFA Online, Battlefield Heroes, World Golf Tour, and FarmVille all show that the conventional models of gaming software are just as much in flux as those facing business and productivity packages.

Key Takeaways

  • SaaS firms may offer their clients several benefits including the following:

    • lower costs by eliminating or reducing software, hardware, maintenance, and staff expenses
    • financial risk mitigation since start-up costs are so low
    • potentially faster deployment times compared with installed packaged software or systems developed in-house
    • costs that are a variable operating expense rather than a large, fixed capital expense
    • scalable systems that make it easier for firms to ramp up during periods of unexpectedly high system use
    • higher quality and service levels through instantly available upgrades, vendor scale economies, and expertise gained across its entire client base
    • remote access and availability—most SaaS offerings are accessed through any Web browser, and often even by phone or other mobile device
  • Vendors of SaaS products benefit from the following:

    • limiting development to a single platform, instead of having to create versions for different operating systems
    • tighter feedback loop with clients, helping fuel innovation and responsiveness
    • ability to instantly deploy bug fixes and product enhancements to all users
    • lower distribution costs
    • accessibility to anyone with an Internet connection
    • greatly reduced risk of software piracy
  • Microtransactions and ad-supported gaming present alternatives to conventional purchased video games. Firms leveraging these models potentially benefit from a host of SaaS advantages, including direct-to-consumer distribution, instant upgrades, continued revenue streams rather than one-time purchase payments, and a method for combating piracy.

Questions and Exercises

  1. Firms that buy conventional enterprise software spend money buying software and hardware. What additional and ongoing expenses are required as part of the “care and feeding” of enterprise applications?
  2. In what ways can firms using SaaS products dramatically lower costs associated with their information systems?
  3. How do SaaS vendors earn their money?
  4. Give examples of enterprise-focused SaaS vendors and their products. Visit the Web sites of the firms that offer these services. Which firms are listed as clients? Does there appear to be a particular type of firm that uses its services, or are client firms broadly represented?
  5. Give examples of desktop-focused SaaS vendors and their products. If some of these are free, try them out and compare them to desktop alternatives you may have used. Be prepared to share your experiences with your class.
  6. List the cost-related benefits to users that accrue from using SaaS.
  7. List the benefits other than cost-related that accrue to users from using SaaS.
  8. List the benefits realized by vendors that offer SaaS services instead of conventional software.
  9. Microtransactions have been tried in many contexts, but have often failed. Can you think of contexts where microtransactions don’t work well? Are there contexts where you have paid (or would be wiling to pay) for products and services via microtransactions? What do you suppose are the major barriers to the broader acceptance of microtransactions? Do struggles have more to do with technology, consumer attitudes, or both?
  10. Search online to find free and microtransaction-based games. What do you think of these efforts? What kind of gamers do these efforts appeal to? See if you can investigate whether there are examples of particularly successful offerings, or efforts that have failed. What’s the reason behind the success or failure of the efforts that you’ve investigated?

10.8 SaaS: Not without Risks

Learning Objective

After studying this section you should be able to do the following:

  1. Be able to list and appreciate the risks associated with SaaS.

Like any technology, we also recognize there is rarely a silver bullet that solves all problems. A successful manager is able to see through industry hype and weigh the benefits of a technology against its weaknesses and limitations. And there are still several major concerns surrounding SaaS.

The largest concerns involve the tremendous dependence a firm develops with its SaaS vendor. Having all of your eggs in one basket can leave a firm particularly vulnerable. If a traditional software company goes out of business, in most cases its customers can still go on using its products. But if your SaaS vendor goes under, you’re hosed. They’ve got all your data, and even if firms could get their data out, most organizations don’t have the hardware, software, staff, or expertise to quickly absorb an abandoned function.

Beware with whom you partner. Any hot technology is likely to attract a lot of start-ups, and most of these start-ups are unlikely to survive. In just a single year, the leading trade association found the number of SaaS vendors dropped from seven hundred members to four hundred fifty.M. Drummond, “The End of Software as We Know It,” Fortune, November 19, 2001. One of the early efforts to collapse was Pandesic, a joint venture between SAP and Intel—two large firms that might have otherwise instilled confidence among prospective customers. In another example, Danish SaaS firm “IT Factory” was declared “Denmark’s Best IT Company 2008” by Computerworld, only to follow the award one week later with a bankruptcy declaration.R. Wauters, “The Extraordinary Rise and Fall of Denmark’s IT Factory,” TechCrunch, December 2, 2008. Indeed, despite the benefits, the costs of operating as a SaaS vendor can be daunting. NetSuite’s founder claimed it “takes ten years and $100 million to do right”Sarah Lacy, “On-Demand Computing: A Brutal Slog,” BusinessWeek, July 18, 2008.—maybe that’s why the firm still wasn’t profitable, even a year and a half after going public.

Firms that buy and install packaged software usually have the option of sticking with the old stuff as long as it works, but organizations adopting SaaS may find they are forced into adopting new versions. This fact is important because any radical changes in a SaaS system’s user interface or system functionality might result in unforeseen training costs, or increase the chance that a user might make an error.

Keep in mind that SaaS systems are also reliant on a network connection. If a firm’s link to the Internet goes down, its link to its SaaS vendor is also severed. Relying on an Internet connection also means that data is transferred to and from a SaaS firm at Internet speeds, rather the potentially higher speeds of a firm’s internal network. Solutions to many of these issues are evolving as Internet speeds become faster and Internet service providers become more reliable. There are also several programs that allow for offline use of data that is typically stored in SaaS systems, including Google Gears and Adobe AIR. With these products a user can download a subset of data to be offline (say on a plane flight or other inaccessible location), and then sync the data when the connection is restored. Ultimately, though, SaaS users have a much higher level of dependence on their Internet connections.

And although a SaaS firm may have more security expertise than your organization, that doesn’t mean that security issues can be ignored. Any time a firm allows employees to access a corporation’s systems and data assets from a remote location, a firm is potentially vulnerable to abuse and infiltration. Some firms may simply be unacceptably uncomfortable with critical data assets existing outside their own network. There may also be contractual or legal issues preventing data from being housed remotely, especially if a SaaS vendor’s systems are in another country operating under different laws and regulations. “We’re very bound by regulators in terms of client data and country-of-origin issues, so it’s very difficult to use the cloud,” says Rupert Brown, a chief architect at Merrill Lynch.G. Gruman, “Early Experiments in Cloud Computing,” InfoWorld, April 7, 2008.

SaaS systems are often accused of being less flexible than their installed software counterparts—mostly due to the more robust configuration and programming options available in traditional software packages. It is true that many SaaS vendors have improved system customization options and integration with standard software packages. And at times a lack of complexity can be a blessing—fewer choices can mean less training, faster start-up time, and lower costs associated with system use. But firms with unique needs may find SaaS restrictive.

SaaS offerings usually work well when the bulk of computing happens at the server end of a distributed system because the kind of user interface you can create in a browser isn’t as sophisticated as what you can do with a separate, custom-developed desktop program. A comparison of the first few iterations of the Web-based Google office suite, which offers word processing, presentation software, and a spreadsheet, reveals a much more limited feature set than Microsoft’s Office desktop software. The bonus, of course, is that an online office suite is accessible anywhere and makes sharing documents a snap. Again, an understanding of trade-offs is key.

Here’s another challenge for a firm and its IT staff: SaaS means a greater consumerization of technology. Employees, at their own initiative, can go to Socialtext or Google Sites and set up a wiki, WordPress to start blogging, or subscribe to a SaaS offering like Salesforce.com, all without corporate oversight and approval. This work can result in employees operating outside established firm guidelines and procedures, potentially introducing operational inconsistencies or even legal and security concerns.

The consumerization of corporate technology isn’t all bad. Employee creativity can blossom with increased access to new technologies, costs might be lower than home grown solutions, and staff could introduce the firm to new tools that might not otherwise be on the radar of the firm’s IS Department. But all this creates an environment that requires a level of engagement between a firm’s technical staff and the groups that it serves that is deeper than that employed by any prior generation of technology workers. Those working in an organization’s information systems group must be sure to conduct regular meetings with representative groups of employees across the firm to understand their pain points and assess their changing technology needs. Non-IT managers should regularly reach out to IT to ensure that their needs are on the tech staff’s agenda. Organizations with internal IT-staff R&D functions that scan new technologies and critically examine their relevance and potential impact on the firm can help guide an organization through the promise and peril of new technologies. Now more than ever, IT managers must be deeply knowledgeable about business areas, broadly aware of new technologies, and able to bridge the tech and business worlds. Similarly, any manager looking to advance his or her organization has to regularly consider the impact of new technologies.

Key Takeaways

The risks associated with SaaS include the following:

  • dependence on a single vendor.
  • concern about the long-term viability of partner firms.
  • users may be forced to migrate to new versions—possibly incurring unforeseen training costs and shifts in operating procedures.
  • reliance on a network connection—which may be slower, less stable, and less secure.
  • data asset stored off-site—with the potential for security and legal concerns.
  • limited configuration, customization, and system integration options compared to packaged software or alternatives developed in-house.
  • the user interface of Web-based software is often less sophisticated and lacks the richness of most desktop alternatives.
  • ease of adoption may lead to pockets of unauthorized IT being used throughout an organization.

Questions and Exercises

  1. Consider the following two firms: a consulting start-up, and a defense contractor. Leverage what you know about SaaS and advise whether each might consider SaaS efforts for CRM or other enterprise functions? Why or why not?
  2. Think of firms you’ve worked for, or firms you would like to work for. Do SaaS offerings make sense for these firms? Make a case for or against using certain categories of SaaS.
  3. What factors would you consider when evaluating a SaaS vendor? Which firms are more appealing to you and why?
  4. Discuss problems that may arise because SaaS solutions rely on Internet connections. Discuss the advantages of through-the-browser access.
  5. Evaluate trial versions of desktop SaaS offerings (offered by Adobe, Google, Microsoft, Zoho, or others). Do you agree that the interfaces of Web-based versions are not as robust as desktop rivals? Are they good enough for you? For most users?

10.9 The Hardware Cloud: Utility Computing and Its Cousins

Learning Objectives

After studying this section you should be able to do the following:

  1. Distinguish between SaaS and hardware clouds.
  2. Provide examples of firms and uses of hardware clouds.
  3. Understand the concepts of cloud computing, cloudbursting, and black swan events.
  4. Understand the challenges and economics involved in shifting computing hardware to the cloud.

While SaaS provides the software and hardware to replace an internal information system, sometimes a firm develops its own custom software but wants to pay someone else to run it for them. That’s where hardware clouds, utility computing, and related technologies come in. In this model, a firm replaces computing hardware that it might otherwise run on-site with a service provided by a third party online. While the term utility computing was fashionable a few years back (and old timers claim it shares a lineage with terms like hosted computing or even time sharing), now most in the industry have begun referring to this as an aspect of cloud computing, often referred to as hardware cloudsA cloud computing model in which a service provider makes computing resources such as hardware and storage, along with infrastructure management, available to a customer on an as-needed basis. The provider typically charges for specific resource usage rather than a flat rate. In the past, similar efforts have been described as utility computing, hosting, or even time sharing.. Computing hardware used in this scenario exists “in the cloud,” meaning somewhere on the Internet. The costs of systems operated in this manner look more like a utility bill—you only pay for the amount of processing, storage, and telecommunications used. Tech research firm Gartner has estimated that 80 percent of corporate tech spending goes toward data center maintenance.J. Rayport, “Cloud Computing Is No Pipe Dream,” BusinessWeek, December 9, 2008. Hardware-focused cloud computing provides a way for firms to chip away at these costs.

Major players are spending billions building out huge data centers to take all kinds of computing out of the corporate data center and place it in the cloud. Efforts include Sun’s Network.com grid, IBM’s Cloud Labs, Amazon’s EC2 (Elastic Computing Cloud), Google’s App Engine, Microsoft’s Azure, and Salesforce.com’s Force.com. While cloud vendors typically host your software on their systems, many of these vendors also offer additional tools to help in creating and hosting apps in the cloud. Salesforce.com offers Force.com, which includes not only a hardware cloud but also several cloud-supporting tools, including a programming environment (IDE) to write applications specifically tailored for Web-based delivery. Google’s App Engine offers developers a database product called Big Table, while Amazon offers one called Amazon DB. Traditional software firms like Oracle are also making their products available to developers through various cloud initiatives.

Still other cloud computing efforts focus on providing a virtual replacement for operational hardware like storage and backup solutions. These include the cloud-based backup efforts like EMC’s Mozy, and corporate storage services like Amazon’s Simple Storage Solution (S3). Even efforts like Apple’s MobileMe and Microsoft’s Live Mesh that sync user data across devices (phone, multiple desktops) are considered part of the cloud craze. The common theme in all of this is leveraging computing delivered over the Internet to satisfy the computing needs of both users and organizations.

Clouds in Action: A Snapshot of Diverse Efforts

Large, established organizations, small firms and start-ups are all embracing the cloud. The examples below illustrate the wide range of these efforts.

Journalists refer to the New York Times as, “The Old Gray Lady,” but it turns out that the venerable paper is a cloud-pioneering whippersnapper. When the Times decided to make roughly one hundred fifty years of newspaper archives (over fifteen million articles) available over the Internet, it realized that the process of converting scans into searchable PDFs would require more computing power than the firm had available.J. Rayport, “Cloud Computing Is No Pipe Dream,” BusinessWeek, December 9, 2008. To solve the challenge, a Times IT staffer simply broke out a credit card and signed up for Amazon’s EC2 cloud computing and S3 cloud storage services. The Times then started uploading terabytes of information to Amazon, along with a chunk of code to execute the conversion. While anyone can sign up for services online without speaking to a rep, someone from Amazon eventually contacted the Times to check in after noticing the massive volume of data coming into its systems. Using one hundred of Amazon’s Linux servers, the Times job took just twenty-four hours to complete. In fact, a coding error in the initial batch forced the paper to rerun the job. Even the blunder was cheap—just two hundred forty dollars in extra processing costs. Says a member of the Times IT group: “It would have taken a month at our facilities, since we only had a few spare PCs.…It was cheap experimentation, and the learning curve isn’t steep.”G. Gruman, “Early Experiments in Cloud Computing,” InfoWorld, April 7, 2008.

NASDAQ also uses Amazon’s cloud as part of its Market Replay system. The exchange uses Amazon to make terabytes of data available on demand, and uploads an additional thirty to eighty gigabytes every day. Market Reply allows access through an Adobe AIR interface to pull together historical market conditions in the ten-minute period surrounding a trade’s execution. This allows NASDAQ to produce a snapshot of information for regulators or customers who question a trade. Says the exchange’s VP of Product Development, “The fact that we’re able to keep so much data online indefinitely means the brokers can quickly answer a question without having to pull data out of old tapes and CD backups.”P. Grossman, “Cloud Computing Begins to Gain Traction on Wall Street,” Wall Street and Technology, January 6, 2009. NASDAQ isn’t the only major financial organization leveraging someone else’s cloud. Others include Merrill Lynch, which uses IBM’s Blue Cloud servers to build and evaluate risk analysis programs; and Morgan Stanley, which relies on Force.com for recruiting applications.

The Network.com offering from Sun Microsystems is essentially a grid computer in the clouds (see Chapter 5 “Moore’s Law: Fast, Cheap Computing and What It Means for the Manager”). Since grid computers break a task up to spread across multiple processors, the Sun service is best for problems that can be easily divided into smaller mini jobs that can be processed simultaneously by the army of processors in Sun’s grid. The firm’s cloud is particularly useful for performing large-scale image and data tasks. Infosolve, a data management firm, uses the Sun cloud to scrub massive data sets, at times harnessing thousands of processors to comb through client records and correct inconsistent entries.

IBM Cloud Labs, which counts Elizabeth Arden and the U.S. Golf Association among its customers, offers several services, including so-called cloudburstingDescribes the use of cloud computing to provide excess capacity during periods of spiking demand. Cloudbursting is a scalability solution that is usually provided as an overflow sservice, kicking in as needed.. In a cloudbursting scenario a firm’s data center running at maximum capacity can seamlessly shift part of the workload to IBM’s cloud, with any spikes in system use metered, utility style. Cloudbursting is appealing because forecasting demand is difficult and can’t account for the ultrarare, high-impact events, sometimes called black swansUnpredicted, but highly impactful events. Scalable computing resources can help a firm deal with spiking impact from Black Swan events. The phrase entered the managerial lexicon from the 2007 book of the same name by Nassim Taleb.. Planning to account for usage spikes explains why the servers at many conventional corporate IS shops run at only 10 to 20 percent capacity.J. Parkinson, “Green Data Centers Tackle LEED Certification,” SearchDataCenter.com, January 18, 2007. While Cloud Labs cloudbursting service is particularly appealing for firms that already have a heavy reliance on IBM hardware in-house, it is possible to build these systems using the hardware clouds of other vendors, too.

Salesforce.com’s Force.com cloud is especially tuned to help firms create and deploy custom Web applications. The firm makes it possible to piece together projects using premade Web services that provide software building blocks for features like calendaring and scheduling. The integration with the firm’s SaaS CRM effort, and with third-party products like Google Maps allows enterprise mash-ups that can combine services from different vendors into a single application that’s run on Force.com hardware. The platform even includes tools to help deploy Facebook applications. Intuitive Surgical used Force.com to create and host a custom application to gather clinical trial data for the firm’s surgical robots. An IS manager at Intuitive noted, “We could build it using just their tools, so in essence, there was no programming.”G. Gruman, “Early Experiments in Cloud Computing,” InfoWorld, April 7, 2008. Other users include Jobscience, which used Force.com to launch its online recruiting site; and Harrah’s Entertainment, which uses Force.com applications to manage room reservations, air travel programs, and player relations.

These efforts compete with a host of other initiatives, including Google’s App Engine and Microsoft’s Azure Services Platform, hosting firms like Rackspace, and cloud-specific upstarts like GoGrid.

Challenges Remain

Hardware clouds and SaaS share similar benefits and risk, and as our discussion of SaaS showed, cloud efforts aren’t for everyone. Some additional examples illustrate the challenges in shifting computing hardware to the cloud.

For all the hype about cloud computing, it doesn’t work in all situations. From an architectural standpoint, most large organizations run a hodgepodge of systems that include both package applications and custom code written in-house. Installing a complex set of systems on someone else’s hardware can be a brutal challenge and in many cases is just about impossible. For that reason we can expect most cloud computing efforts to focus on new software development projects rather than options for old software. Even for efforts that can be custom-built and cloud-deployed, other roadblocks remain. For example, some firms face stringent regulatory compliance issues. To quote one tech industry executive, “How do you demonstrate what you are doing is in compliance when it is done outside?”G. Gruman, “Early Experiments in Cloud Computing,” InfoWorld, April 7, 2008.

Firms considering cloud computing need to do a thorough financial analysis, comparing the capital and other costs of owning and operating their own systems over time against the variable costs over the same period for moving portions to the cloud. For high-volume, low-maintenance systems, the numbers may show that it makes sense to buy rather than rent. Cloud costs can seem super cheap at first. Sun’s early cloud effort offered a flat fee of one dollar per CPU per hour. Amazon’s cloud storage rates were twenty-five cents per gigabyte per month. But users often also pay for the number of accesses and the number of data transfers.C. Preimesberger, “Sun’s ‘Open’-Door Policy,” eWeek, April 21, 2008. A quarter a gigabyte a month may seem like a small amount, but system maintenance costs often include the need to clean up old files or put them on tape. If unlimited data is stored in the cloud, these costs can add up.

Firms should enter the cloud cautiously, particularly where mission-critical systems are concerned. When one of the three centers supporting Amazon’s cloud briefly went dark in 2008, start-ups relying on the service, including Twitter and SmugMug, reported outages. Apple’s MobileMe cloud-based product for synchronizing data across computers and mobile devices, struggled for months after its introduction when the cloud repeatedly went down. Vendors with multiple data centers that are able to operate with fault-tolerantCapable of continuing operation even if a component fails. provisioning, keeping a firm’s efforts at more than one location to account for any operating interruptions, will appeal to firms with stricter uptime requirements.

Key Takeaways

  • It’s estimated that 80 percent of corporate tech spending goes toward data center maintenance. Hardware-focused cloud computing initiatives from third party firms help tackle this cost by allowing firms to run their own software on the hardware of the provider.
  • Amazon, EMC, Google, IBM, Microsoft, Oracle/Sun, Rackspace, and Salesforce.com are among firms offering platforms to run custom software projects. Some offer additional tools and services, including additional support for cloud-based software development, hosting, application integration, and backup.
  • Users of cloud computing run the gamut of industries, including publishing (the New York Times), finance (NASDAQ), and cosmetics and skin care (Elizabeth Arden).
  • Benefits and risks are similar to those discussed in SaaS efforts. Benefits include the use of the cloud for handling large batch jobs or limited-time tasks, offloading expensive computing tasks, and cloudbursting efforts that handle system overflow when an organization needs more capacity.
  • Most legacy systems can’t be easily migrated to the cloud, meaning most efforts will be new efforts or those launched by younger firms.
  • Cloud (utility) computing doesn’t work in situations where complex legacy systems have to be ported, or where there may be regulatory compliance issues.
  • Some firms may still find TCO and pricing economics favor buying over renting—scale sometimes suggests an organization is better off keeping efforts in-house.

Questions and Exercises

  1. What are hardware clouds? What kinds of services are described by this terms? What are other names for this phenomenon? How does this differ from SaaS?
  2. Which firms are the leading providers of hardware clouds? How are clients using these efforts?
  3. List the circumstances where hardware clouds work best and where it works poorly.
  4. Research cloud-based alternatives for backing up your hard drive. Which are among the best reviewed product or services? Why? Do you or would you use such a service? Why or why not?
  5. Can you think of “black swan” events that have caused computing services to become less reliable? Describe the events and its consequences for computing services. Suggest a method and vendor for helping firms overcome the sorts of events that you encountered.

10.10 Clouds and Tech Industry Impact

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand how cloud computing’s impact across industries is proving to be broad and significant.
  2. Know the effects of cloud computing on high-end server sales and the influence on the trend shifting from hardware sales to service.
  3. Know the effects of cloud computing on innovation and the influence on the changes in the desired skills mix and job outlook for IS workers.
  4. Know that by lowering the cost to access powerful systems and software, cloud computing can decrease barriers to entry.
  5. Understand the importance, size, and metrics of server farms.

Although still a relatively recent phenomenon, cloud computing’s impact across industries is already proving to be broad and significant.

Cloud computing is affecting the competitive dynamics of the hardware, software, and consulting industries. In the past, firms seeking to increase computing capacity invested heavily in expensive, high margin server hardware, creating a huge market for computer manufacturers. But now hardware firms find these markets may be threatened by the cloud. The trend shifting from hardware to services is evident in IBM’s quarterly numbers. The firm recently reported its overall earnings were up 12 percent, even though hardware sales were off by 20 percent.J. Fortt, “Goodbye, PC (and Mac). Hello, Services,” Fortune, February 4, 2009. What made up the difference? The growth of Big Blue’s services business. IBM is particularly well positioned to take advantage of the shift to services because it employs more technology consultants than any other firm in the world, while most of its competitors are forced to partner to offer something comparable. Consulting firm Capgemini’s partnership to offer cloud services through Amazon is one such example.

The shift to cloud computing also alters the margin structure for many in the computing industry. While Moore’s Law has made servers cheap, deploying SaaS and operating a commercial cloud is still very expensive—much more so than simply making additional copies of conventional, packaged software. Microsoft surprised Wall Street when it announced it would need to pour at least $2 billion more than analysts expected into the year’s server farmA massive network of computer servers running software to coordinate their collective use. Server farms provide the infrastructure backbone to SaaS and hardware cloud efforts, as well as many large-scale Internet services. capital spending. The firm’s stock—among the world’s most widely held—sank 11 percent in a day.S. Mehta, “Behold the Server Farm,” Fortune, July 28, 2006. As a result, many portfolio managers started paying closer attention to the business implications of the cloud.

Cloud computing can accelerate innovation and therefore changes the desired skills mix and job outlook for IS workers. If cloud computing customers spend less on expensive infrastructure investments, they potentially have more money to reinvest in strategic efforts and innovation. IT careers may change, too. Demand for nonstrategic skills like hardware operations and maintenance are likely to decrease. Organizations will need more business-focused technologists who intimately understand a firm’s competitive environment, and can create systems that add value and differentiate the firm from its competition.J. Fortt, “Tech Execs Get Sexy,” Fortune, February 12, 2009. While these tech jobs require more business training, they’re also likely to be more durable and less likely to be outsourced to a third party with a limited understanding of the firm.

By lowering the cost to access powerful systems and software, barriers to entry also decrease. Firms need to think about the strategic advantages they can create, even as technology is easily duplicated. This trend means the potential for more new entrants across industries, and since start-ups can do more with less, it’s also influencing entrepreneurship and venture capital. The CTO of SlideShare, a start-up that launched using Amazon’s S3 storage cloud, offers a presentation on his firm’s site labeled “Using S3 to Avoid VC.” Similarly, the CEO of online payments start-up Zuora claims to have saved between half a million and $1 million by using cloud computing: “We have no servers, we run the entire business in the cloud.”E. Ackerman, “Forecast for Computing: Cloudy,” San Jose Mercury News, December 23, 2008. And the sophistication of these tools lowers development time. Enterprise firm Apttus claims it was able to perform the equivalent of six months of development in a couple of weekends by using cloud services. The firm scored its first million-dollar deal in three months, and was break-even in nine months, a ramp-up time that would have been unheard of, had they needed to plan, purchase, and deploy their own data center, and create from scratch the Web services that were provided by its cloud vendor.J. Rayport, “Cloud Computing Is No Pipe Dream,” BusinessWeek, December 9, 2008.

So What’s It Take to Run This Thing?

In the countryside surrounding the Columbia River in the Pacific Northwest, potato farms are yielding to server farms. Turns out the area is tailor made for creating the kinds of massive data installations that form the building blocks of cloud computing. The land is cheap, the region’s hydroelectric power costs a fraction of Silicon Valley rates, and the area is served by ultrafast fiber-optic connections. Even the area’s mild temperatures cut cooling costs.

Most major players in cloud computing have server farms in the region, each with thousands of processors humming away simultaneously. Microsoft’s Quincy, Washington, facility is as big as ten American football fields and has nearly six hundred miles of wiring, 1.5 metric tons of battery backup, and three miles of chiller piping to keep things cool. Storage is big enough to store 6.75 trillion photos. Just a short drive away, Yahoo has two facilities on fifty acres, including one that runs at a zero carbon footprint. Google has a thirty-acre site sprawled across former farmland in The Dalles, Oregon. The Google site includes two massive buildings, with a third on the way. And in Boardman, Oregon, Amazon has a three building petabyte palace that sports its own ten-megawatt electrical substation.R. Katz, “Tech Titans Building Boom,” IEEE Spectrum 46, no. 2 (February 1, 2009): 40–43.

While U.S. activity has been particularly intense in the Pacific Northwest, server farms that support cloud computing are popping up from Shanghai to São Paulo. Not only does a diverse infrastructure offer a degree of fault tolerance and disaster recovery (Oregon down? Shift to North Carolina), the myriad of national laws and industry-specific regulatory environments may require some firms to keep data within a specific country or region. To meet the challenge, cloud vendors are racing to deploy infrastructure worldwide and allowing customers to select regional availability zones for their cloud computing needs.

The build-out race has become so intense that many firms have developed rapid-deployment server farm modules that are preconfigured and packed inside shipping containers. Some of these units contain as many as three thousand servers each. Just drop the containers on-site, link to power, water, and telecom, and presto—you’ve got yourself a data center. More than two hundred containers can be used on a single site. One Microsoft VP claimed the configuration has cut the time to open a data center to just a few days, claiming Microsoft’s San Antonio facility was operational in less time than it took a local western wear firm to deliver her custom-made cowboy boots!P. Burrows, “Microsoft to Google: Get Off of My Cloud,” BusinessWeek, November 21, 2008. Microsoft’s Dublin-based fourth generation data center will be built entirely of containers—no walls or roof—using the outside air for much of the cooling.T. Vanderbilt, “Data Center Overload,” New York Times, June 8, 2009.

Figure 10.5

This Sun server-packed container is designed for rapid data center deployment.

While firms are buying less hardware, cloud vendors have turned out to be the computing industry’s best customers. Amazon has spent well over $2 billion on its cloud infrastructure. Google reportedly has 1.4 million servers operating across three dozen data centers.R. Katz, “Tech Titans Building Boom,” IEEE Spectrum 46, no. 2 (February 1, 2009): 40–43. Demonstrating it won’t be outdone, Microsoft plans to build as many as twenty server farms, at costs of up to $1 billion each.P. Burrows, “Microsoft to Google: Get Off of My Cloud,” BusinessWeek, November 21, 2008. Look for the clouds to pop up in unexpected places. Microsoft has scouted locations in Siberia, while Google has applied to patent a method for floating data centers on an offshore platform powered by wave motions.R. Katz, “Tech Titans Building Boom,” IEEE Spectrum 46, no. 2 (February 1, 2009): 40–43.

Key Takeaways

  • Cloud computing’s impact across industries is proving to be broad and significant.
  • Clouds can lower barriers to entry in an industry, making it easier for start-ups to launch and smaller firms to leverage the backing of powerful technology.
  • Clouds may also lower the amount of capital a firm needs to launch a business, shifting power away from venture firms in those industries that had previously needed more VC money.
  • Clouds can shift resources out of capital spending and into profitability and innovation.
  • Hardware and software sales may drop as cloud use increases, while service revenues will increase.
  • Cloud computing can accelerate innovation and therefore changes the desired skills mix and job outlook for IS workers. Tech skills in data center operations, support, and maintenance may shrink as a smaller number of vendors consolidate these functions.
  • Demand continues to spike for business-savvy technologists. Tech managers will need even stronger business skills and will focus an increasing percentage of their time on strategic efforts. These latter jobs are tougher to outsource, since they involve an intimate knowledge of the firm, its industry, and its operations.
  • The market for expensive, high margin, sever hardware is threatened by companies moving applications to the cloud instead of investing in hardware.
  • Server farms require plenty of cheap land, low cost power, ultrafast fiber-optic connections, and benefit from mild climates.
  • Sun, Microsoft, IBM, and HP have all developed rapid-deployment server farm modules that are pre configured and packed inside shipping containers.

Questions and Exercises

  1. Describe the change in IBM’s revenue stream resulting from the shift to the cloud.
  2. Why is IBM particularly well positioned to take advantage of the shift to services?
  3. Describe the shift in skill sets required for IT workers that is likely to result from the widespread adoption of cloud computing.
  4. Why do certain entry barriers decrease as a result of cloud computing? What is the effect of lower entry barriers on new entrants, entrepreneurship, and venture capital? On existing competitors?
  5. What factors make the Columbia River region of the Pacific Northwest an ideal location for server farms?
  6. What is the estimated number of computers operated by Google?
  7. Why did Microsoft’s shift to cloud computing create an unexpected shock among stock analysts who cover the firm? What does this tell you about the importance of technology understanding among finance and investment professionals?
  8. Why do cloud computing vendors build regional server farms instead of one mega site?
  9. Why would a firm build a container-based data center?

10.11 Virtualization: Software That Makes One Computer Act Like Many

Learning Objectives

After studying this section you should be able to do the following:

  1. Know what virtualization software is and its impact on cloud computing.
  2. Be able to list the benefits to a firm from using virtualization.

The reduced costs and increased power of commodity hardware are not the only contributors to the explosion of cloud computing. The availability of increasingly sophisticated software tools has also had an impact. Perhaps the most important software tool in the cloud computing toolbox is virtualizationA type of software that allows a single computer (or cluster of connected computers) to function as if it were several different computers, each running its own operating system and software. Virtualization software underpins most cloud computing efforts, and can make computing more efficient, cost-effective, and scalable.. Think of virtualization as being a kind of operating system for operating systems. A server running virtualization software can create smaller compartments in memory that each behave as a separate computer with its own operating system and resources. The most sophisticated of these tools also allow firms to combine servers into a huge pool of computing resources that can be allocated as needed.D. Lyons, “A Mostly Cloudy Computing Forecast,” Washington Post, November 4, 2008.

Virtualization can generate huge savings. Some studies have shown that on average, conventional data centers run at 15 percent or less of their maximum capacity. Data centers using virtualization software have increased utilization to 80 percent or more.R. Katz, “Tech Titans Building Boom,” IEEE Spectrum 46, no. 2 (February 1, 2009): 40–43. This increased efficiency means cost savings in hardware, staff, and real estate. Plus it reduces a firm’s IT-based energy consumption, cutting costs, lowering its carbon footprint, and boosting “green cred.”K. Castro, “The Virtues of Virtualization,” BusinessWeek, December 3, 2007. Using virtualization, firms can buy and maintain fewer servers, each running at a greater capacity. It can also power down servers until demand increases require them to come online.

While virtualization is a key software building block that makes public cloud computing happen, it can also be used in-house to reduce an organization’s hardware needs, and even to create a firm’s own private cloud of scalable assets. Bechtel, BT, Merrill Lynch, and Morgan Stanley are among the firms with large private clouds enabled by virtualization.J. Brodkin, “Private Clouds Bring IT Mgmt. Challenges,” NetworkWorld, December 15, 2008. Another kind of virtualization, virtual desktopsWhen a firm runs an instance of a PC’s software on another machine and simply delivers the image of what’s executing to the remote device. Using virtualization, a single server can run dozens of PCs, simplifying backup, upgrade, security, and administration. allow a server to run what amounts to a copy of a PC—OS, applications, and all—and simply deliver an image of what’s executing to a PC or other connected device. This allows firms to scale, back up, secure, and upgrade systems far more easily than if they had to maintain each individual PC. One game start-up hopes to remove the high-powered game console hardware attached to your television and instead put the console in the cloud, delivering games to your TV as they execute remotely on superfast server hardware. Virtualization can even live on your desktop. Anyone who’s ever run Windows in a window on Mac OS X is using virtualization software; these tools inhabit a chunk of your Mac’s memory for running Windows and actually fool this foreign OS into thinking that it’s on a PC.

Interest in virtualization has exploded in recent years. VMware, the virtualization software division of storage firm EMC, was the biggest IPO of 2007. But its niche is getting crowded. Microsoft has entered the market, building virtualization into its server offerings. Dell bought a virtualization software firm for $1.54 billion. And there’s even an open source virtualization product called Xen.K. Castro, “The Virtues of Virtualization,” BusinessWeek, December 3, 2007.

Key Takeaways

  • Virtualization software allows one computing device to function as many. The most sophisticated products also make it easy for organizations to scale computing requirements across several servers.
  • Virtualization software can lower a firm’s hardware needs, save energy, and boost scalability.
  • Data center virtualization software is at the heart of many so-called private clouds and scalable corporate data centers, as well as the sorts of public efforts described earlier.
  • Virtualization also works on the desktop, allowing multiple operating systems (Mac OS X, Linux, Windows) to run simultaneously on the same platform.
  • Virtualization software can increase data center utilization to 80 percent or more.
  • While virtualization is used to make public cloud computing happen, it can also be used in-house to create a firm’s own private cloud.
  • A number of companies, including Microsoft and Dell, have entered the growing virtualization market.

Questions and Exercises

  1. List the benefits to a firm from using virtualization.
  2. What is the average utilization rate for conventional data centers?
  3. List companies that have virtualization-enabled private clouds.
  4. Give an example of desktop virtualization.
  5. Name three companies that are players in the virtualization software industry.

10.12 Make, Buy, or Rent

Learning Objectives

After studying this section you should be able to do the following:

  1. Know the options managers have when determining how to satisfy the software needs of their companies.
  2. Know the factors that must be considered when making the make, buy, or rent decision.

So now you realize managers have a whole host of options when seeking to fulfill the software needs of their firms. An organization can purchase packaged software from a vendor, use open source offerings, leverage SaaS or other type of cloud computing, outsource development or other IT functions to another firm either domestically or abroad, or a firm can develop all or part of the effort themselves. When presented with all of these options, making decisions about technologies and systems can seem pretty daunting.

First, realize that that for most firms, technology decisions are not binary options for the whole organization in all situations. Few businesses will opt for an IT configuration that is 100 percent in-house, packaged, or SaaS. Being aware of the parameters to consider can help a firm make better, more informed decisions. It’s also important to keep in mind that these decisions need to be continuously reevaluated as markets and business needs change. What follows is a summary of some of the key variables to consider.

Competitive AdvantageDo we rely on unique processes, procedures, or technologies that create vital, differentiating competitive advantage? If so, then these functions aren’t a good candidate to outsource or replace with a package software offering. Amazon.com had originally used recommendation software provided by a third party, and Netflix and Dell both considered third-party software to manage inventory fulfillment. But in all three cases, these firms felt that mastery of these functions was too critical to competitive advantage, so each firm developed proprietary systems unique to the circumstances of each firm.

SecurityAre there unacceptable risks associated with using the packaged software, OSS, cloud solution, or an outsourcing vendor? Are we convinced that the prospective solution is sufficiently secure and reliable? Can we trust the prospective vendor with our code, our data, our procedures and our way of doing business? Are there noncompete provisions for vendor staff that may be privy to our secrets? For off-site work, are there sufficient policies in place for on-site auditing? If the answers to any of these questions is no, outsourcing might not be a viable option.

Legal and ComplianceIs our firm prohibited outright from using technologies? Are there specific legal and compliance requirements related to deploying our products or services? Even a technology as innocuous as instant messaging may need to be deployed in such a way that it complies with laws requiring firms to record and reproduce the electronic equivalent of a paper trail. For example, SEC Rule 17a-4 requires broker dealers to retain client communications for a minimum of three years. HIPAA laws governing health care providers state that electronic communications must also be captured and stored.D. Shapiro, “Instant Messaging and Compliance Issues: What You Need to Know,” SearchCIO, May 17, 2004. While tech has gained a seat in the board room, legal also deserves a seat in systems planning meetings.

Skill, Expertise, and Available LaborCan we build it? The firm may have skilled technologists, but they may not be sufficiently experienced with a new technology. Even if they are skilled, managers much consider the costs of allocating staff away from existing projects for this effort.

CostIs this a cost-effective choice for our firm? A host of factors must be considered when evaluating the cost of an IT decision. The costs to build, host, maintain, and support an ongoing effort involve labor (software development, quality assurance, ongoing support, training, and maintenance), consulting, security, operations, licensing, energy, and real estate. Any analysis of costs should consider not only the aggregate spending required over the lifetime of the effort but also whether these factors might vary over time.

TimeDo we have time to build, test, and deploy the system?

Vendor IssuesIs the vendor reputable and in a sound financial position? Can the vendor guarantee the service levels and reliability we need? What provisions are in place in case the vendor fails or is acquired? Is the vendor certified via the Carnegie Mellon Software Institute or other standards organizations in a way that conveys quality, trust, and reliability?

The list above is a starter. It should also be clear that these metrics are sometimes quite tough to estimate. Welcome to the challenges of being a manager! At times an environment in flux can make an executive feel like he or she is working on a surfboard, constantly being buffeted about by unexpected currents and waves. Hopefully the issues outlined in this chapter will give you the surfing skills you need for a safe ride that avoids the organizational equivalent of a wipeout.

Key Takeaways

  • The make, buy, or rent decision may apply on a case-by-case basis that might be evaluated by firm, division, project or project component. Firm and industry dynamics may change in a way that causes firms to reassess earlier decisions, or to alter the direction of new initiatives.
  • Factors that managers should consider when making a make, buy, or rent decision include the following: competitive advantage, security, legal and compliance issues, the organization’s skill and available labor, cost, time, and vendor issues.
  • Factors must be evaluated over the lifetime of a project, not at a single point in time.
  • Managers have numerous options available when determining how to satisfy the software needs of their companies: purchase packaged software from a vendor, use OSS, use SaaS or utility computing, outsourcing development, or developing all or part of the effort themselves.
  • If a company relies on unique processes, procedures, or technologies that create vital, differentiating, competitive advantages, the functions probably aren’t a good candidate to outsource.

Questions and Exercises

  1. What are the options available to managers when seeking to meet the software needs of their companies?
  2. What are the factors that must be considered when making the make, buy, or rent decision?
  3. What are some security-related questions that must be asked when making the make, buy, or rent decision?
  4. What are some vendor-related questions that must be asked when making the make, buy, or rent decision?
  5. What are some of the factors that must be considered when evaluating the cost of an IT decision?
  6. Why must factors be evaluated over the lifetime of a project, not at a single point in time?

Chapter 9: Understanding Software: A Primer for Managers

9.1 Introduction

Learning Objectives

After studying this section you should be able to do the following:

  1. Recognize the importance of software and its implications for the firm and strategic decision making.
  2. Understand that software is everywhere; not just in computers, but also cell phones, cars, cameras, and many other technologies.
  3. Know what software is and be able to differentiate it from hardware.
  4. List the major classifications of software and give examples of each.

We know computing hardwareThe physical components of information technology, which can include the computer itself plus peripherals such as storage devices, input devices like the mouse and keyboard, output devices like monitors and printers, networking equipment, and so on. is getting faster and cheaper, creating all sorts of exciting and disruptive opportunities for the savvy manager. But what’s really going on inside the box? It’s softwareA computer program or a collection of programs. It is a precise set of instructions that tells hardware what to do. that makes the magic of computing happen. Without software, your PC would be a heap of silicon wrapped in wires encased in plastic and metal. But it’s the instructions—the software code—that enable a computer to do something wonderful, driving the limitless possibilities of information technology.

Software is everywhere. An inexpensive cell phone has about one million lines of code, while the average car contains nearly one hundred million.R. Charette, “Why Software Fails,” IEEE Spectrum, September 2005. In this chapter we’ll take a peek inside the chips to understand what software is. A lot of terms are associated with software: operating systems, applications, enterprise software, distributed systems, and more. We’ll define these terms up front, and put them in a managerial context. A follow-up chapter, Chapter 10 “Software in Flux: Partly Cloudy and Sometimes Free”, will focus on changes impacting the software business, including open source software, software as a service (SaaS), and cloud computing. These changes are creating an environment radically different from the software industry that existed in prior decades—confronting managers with a whole new set of opportunities and challenges.

Managers who understand software can better understand the possibilities and impact of technology. They can make better decisions regarding the strategic value of IT and the potential for technology-driven savings. They can appreciate the challenges, costs, security vulnerabilities, legal and compliance issues, and limitations involved in developing and deploying technology solutions. In the next two chapters we will closely examine the software industry and discuss trends, developments and economics—all of which influence decisions managers make about products to select, firms to partner with, and firms to invest in.

What Is Software?

When we refer to computer hardware (sometimes just hardware), we’re talking about the physical components of information technology—the equipment that you can physically touch, including computers, storage devices, networking equipment, and other peripherals.

Software refers to a computer program or collection of programs—sets of instructions that tell the hardware what to do. Software gets your computer to behave like a Web browser or word processor, makes your iPod play music and video, and enables your bank’s ATM to spit out cash.

It’s when we start to talk about the categories of software that most people’s eyes glaze over. To most folks, software is a big, incomprehensible alphabet soup of acronyms and geeky phrases: OS, VB, SAP, SQL, to name just a few.

Don’t be intimidated. The basics are actually pretty easy to understand. But it’s not soup; it’s more of a layer cake. Think about computer hardware as being at the bottom of the layer cake. The next layer is the operating systemThe software that controls the computer hardware and establishes standards for developing and executing applications., the collection of programs that control the hardware. Windows, Mac OS X, and Linux are operating systems. On top of that layer are applicationsIncludes desktop applications, enterprise software, utilities, and other programs that perform specific tasks for users and organizations.—these can range from end-user programs like those in Office, to the complex set of programs that manage a business’s inventory, payroll, and accounting. At the top of the cake are users.

Figure 9.1 The Hardware/Software Layer Cake

The flexibility of these layers gives computers the customization options that managers and businesses demand. Understanding how the layers relate to each other helps you make better decisions on what options are important to your unique business needs, can influence what you buy, and may have implications for everything from competitiveness to cost overruns to security breaches. What follows is a manager’s guide to the main software categories with an emphasis on why each is important.

Key Takeaways

  • Software refers to a computer program or collection of programs. It enables computing devices to perform tasks.
  • You can think of software as being part of a layer cake, with hardware at the bottom; the operating system controlling the hardware and establishing standards, the applications executing one layer up, and the users at the top.
  • How these layers relate to one another has managerial implications in many areas, including the flexibility in meeting business demand, costs, legal issues and security.
  • Software is everywhere—not just in computers, but also in cell phones, cars, cameras, and many other technologies.

Questions and Exercises

  1. Explain the difference between hardware and software.
  2. Why should a manager care about software and how software works? What critical organizational and competitive factors can software influence?
  3. What role has software played in your decision to select certain products? Has this influenced why you favored one product or service over another?
  4. Find the Fortune 500 list online. Which firm is the highest ranked software firm? While the Fortune 500 ranks firms according to revenue, what’s this firm’s profitability rank? What does this discrepancy tell you about the economics of software development? Why is the software business so attractive to entrepreneurs?
  5. Refer to earlier chapters (and particularly to Chapter 2 “Strategy and Technology: Concepts and Frameworks for Understanding What Separates Winners from Losers”): Which resources for competitive advantage might top software firms be able to leverage to ensure their continued dominance? Give examples of firms that have leveraged these assets, and why they are so strong.

9.2 Operating Systems

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand what an operating system is and why computing devices require operating systems.
  2. Appreciate how embedded systems extend Moore’s Law, allowing firms to create “smarter” products and services

Computing hardware needs to be controlled, and that’s the role of the operating system. The operating system (sometimes called the “OS”) provides a common set of controls for managing computer hardware, making it easier for users to interact with computers and for programmers to write application software. Just about every computing device has an operating system—desktops and laptops, enterprise-class server computers, your mobile phone. Even specialty devices like iPods, video game consoles, and television set top boxes run some form of OS.

Some firms, like Apple and Nintendo, develop their own proprietary OS for their own hardware. Microsoft sells operating systems to everyone from Dell to the ATM manufacturer Diebold (listen for the familiar Windows error beep on some cash machines). And there are a host of specialty firms, such as Wind River (purchased by Intel), that help firms develop operating systems for all sorts of devices that don’t necessarily look like a PC, including cars, video editing systems, and fighter jet control panels.

Anyone who has used both a PC and a Mac and has noticed differences across these platforms can get a sense of the breadth of what an operating system does. Even for programs that are otherwise identical for these two systems (like the Firefox browser), subtitle differences are visible. Screen elements like menus, scroll bars, and window borders look different on the Mac than they do in Windows. So do the dialogue boxes that show up when you print or save.

These items look and behave differently because each of these functions touches the hardware, and the team that developed Microsoft Windows created a system distinctly different from their Macintosh counterparts at Apple. Graphical user interface (UI)The mechanism through which users interact with a computing device. The UI includes elements of the graphical user interface (or GUI, pronounced “gooey”), such as windows, scroll bars, buttons, menus, and dialogue boxes; and can also include other forms of interaction, such as touch screens, motion sensing controllers, or tactile devices used by the visually impaired. items like scroll bars and menus are displayed on the hardware of the computer display. Files are saved to the hardware of a hard drive or other storage device. Most operating systems also include control panels, desktop file management, and other support programs to work directly with hardware elements like storage devices, displays, printers, and networking equipment. The Macintosh Finder and the Windows Explorer are examples of components of these operating systems. The consistent look, feel, and functionality that operating systems enforce across various programs help make it easier for users to learn new software, which reduces training costs and operator error. See Figure 9.2 for similarities and differences.

Figure 9.2

Differences between the Windows and Mac operating systems are evident throughout the user interface, particularly when a program interacts with hardware.

Operating systems are also designed to give programmers a common set of commands to consistently interact with the hardware. These commands make a programmer’s job easier by reducing program complexity and making it faster to write software while minimizing the possibility of errors in code. Consider what an OS does for the Wii game developer. Nintendo’s Wii OS provides Wii programmers with a set of common standards to use to access the Wiimote, play sounds, draw graphics, save files, and more. Without this, games would be a lot more difficult to write, they’d likely look differently, be less reliable, would cost more, and there would be fewer titles available.

Similarly, when Apple provided developers with a common set of robust, easy-to-use standards for the iPhone and (via the App Store) an easy way for users to install these applications on top of the iPhone/iPod touch OS, software development boomed, and Apple became hands-down the most versatile mobile computing device available.The iPhone and iPod touch OS is derived from Apple’s Mac OS X operating system. In Apple’s case, some fifty thousand apps became available through the App Store in less than a year. A good OS and software development platform can catalyze network effects (see Chapter 6 “Understanding Network Effects”). While the OS seems geeky, its effective design has very strategic business implications!

Figure 9.3 Operating System Market Share for Desktop, Server, and Mobile Phones

Firmware and Embedded Systems

Most personal computers have an operating system installed on their hard drives. This system allows the OS to be replaced or upgraded easily. But many smaller, special-purpose computing devices have their operating systems installed on nonvolatile memory, often on read-only memory (ROM) chips. Control programs stored on chips are sometimes referred to as firmwareSoftware stored on nonvolatile memory chips (as opposed to being stored on devices such as hard drives or removable discs). Despite the seemingly permanent nature of firmware, many products allow for firmware to be upgraded online or by connecting to another device.. The OS in an iPod, mobile phone, or your TV’s set-top box is most likely stored as firmware. Your PC also has a tiny bit of firmware that allows it to do very basic functions like start-up (boot) and begin loading its operating system from disk.

Another term you might hear is embedded systemsSpecial-purpose software designed and included inside physical products (often on firmware). Embedded systems help make devices “smarter,” sharing usage information, helping diagnose problems, indicating maintenance schedules, providing alerts, or enabling devices to take orders from other systems.. As computing gets cheaper, special-purpose technology is increasingly becoming embedded into all sorts of devices like cars, picture frames, aircraft engines, photocopiers, and heating and air conditioning systems. The software programs that make up embedded systems are often stored as firmware too.

Moore’s Law (see Chapter 5 “Moore’s Law: Fast, Cheap Computing and What It Means for the Manager”) enables embedded systems, and these systems can create real strategic value. The Otis Elevator Company, a division of United Technologies, uses embedded systems in its products to warn its service centers when the firm’s elevators, escalators, and moving walkways need maintenance or repair. This warning provides Otis with several key benefits:

  1. Since products automatically contact Otis when they need attention, these systems generate a lucrative service business for the firm and make it more difficult for third parties to offer a competing business servicing Otis products.
  2. Products contact service technicians to perform maintenance based on exact needs (e.g., lubricant is low, or a part has been used enough to be replaced) rather than guessed schedules, which makes service more cost-effective, products less likely to break down, and customers happier.
  3. Any product failures are immediately detected, with embedded systems typically dispatching technicians before a client’s phone call.
  4. The data is fed back to Otis’s R&D group, providing information on reliability and failure so that engineers can use this info to design better products.

Collectively, software embedded on tiny chips yields very big benefits, for years helping Otis remain at the top of its industry.

Key Takeaways

  • The operating system (OS) controls a computer’s hardware and provides a common set of commands for writing programs.
  • Most computing devices (enterprise-class server computers, PCs, phones, set-top boxes, video games, cars, the Mars Rover) have an operating system.
  • Some products use operating systems provided by commercial firms, while others develop their own operating system. Others may leverage open source alternatives (see Chapter 10 “Software in Flux: Partly Cloudy and Sometimes Free”).
  • Embedded systems are special-purpose computer systems designed to perform one or a few dedicated functions, and are frequently built into conventional products like cars, air conditioners, and elevators.
  • Embedded systems can make products and services more efficient, more reliable, more functional, and can enable entire new businesses and create or reinforce resources for competitive advantage.

Questions and Exercises

  1. What does an operating system do? Why do you need an operating system? How do operating systems make a programmer’s job easier? How do operating systems make life easier for end users?
  2. How has the market for desktop, server, and mobile operating systems changed in recent years? Do certain products seem to be gaining traction? Why do you think this is the case?
  3. What kinds of operating systems are used in the devices that you own? On your personal computer? Your mobile phone? The set-top box on top of your television? Are there other operating systems that you come into contact with? If you can’t tell which operating system is in each of these devices, see if you can search the Internet to find out.
  4. For your list in the prior question (and to the extent that you can), diagram the hardware/software “layer cake” for these devices.
  5. For this same list, do you think each device’s manufacturer wrote all of the software that you use on these devices? Can you add or modify software to all of these devices? Why or why not? What would the implications be for cost, security, complexity, reliability, updates and upgrades, and the appeal of each device?
  6. Some ATM machines use Windows. Why would an ATM manufacturer choose to build its systems owing Windows? Why might it want to avoid this? Are there other non-PC devices you’ve encountered that were running some form of Windows?
  7. What are embedded systems? When might firms want to install software on chips instead of on a hard drive?
  8. It’s important to understand how technology impacts a firm’s strategy and competitive environment. Consider the description of Otis elevator’s use of embedded systems. Which parts of the value chain does this impact? How? Consider the “five forces”: How does the system impact the firm’s competitive environment? Are these systems a source of competitive advantage? If not, explain why not? If they are, what kinds of resources for competitive advantage can these kinds of embedded systems create?
  9. Can you think of other firms that can or do leverage embedded systems? Provide examples and list the kinds of benefits these might offer firms and consumers.
  10. Research the Americans with Disabilities Act of 1990 (or investigate if your nation has a similar law), and the implications of this legislation for software developers and Web site operators. Have firms been successfully sued when their software or Web sites could not be accessed by users with physical challenges? What sorts of issues should developers consider when making their products more accessible? What practices might they avoid?

9.3 Application Software

Learning Objectives

After studying this section you should be able to do the following:

  1. Appreciate the difference between desktop and enterprise software.
  2. List the categories of enterprise software.
  3. Understand what an ERP (enterprise resource planning) software package is.
  4. Recognize the relationship of the DBMS (database system) to the other enterprise software systems.
  5. Recognize both the risks and rewards of installing packaged enterprise systems.

Operating systems are designed to create a platformProducts and services that allow for the development and integration of software products and other complementary goods. Windows, the iPhone, the Wii, and the standards that allow users to create Facebook apps are all platforms. so that programmers can write additional applications, allowing the computer to do even more useful things. While operating systems control the hardware, application software (sometimes referred to as software applications, applications, or even just apps) perform the work that users and firms are directly interested in accomplishing. Think of applications as the place where the users or organization’s real work gets done. As we learned in Chapter 6 “Understanding Network Effects”, the more application software that is available for a platform (the more games for a video game console, the more apps for your phone), the more valuable it potentially becomes.

Desktop softwareApplications installed on a personal computer, typically supporting tasks performed by a single user. refers to applications installed on a personal computer—your browser, your Office suite (e.g., word processor, spreadsheet, presentation software), photo editors, and computer games are all desktop software. Enterprise softwareApplications that address the needs of multiple users throughout an organization or work group. refers to applications that address the needs of multiple, simultaneous users in an organization or work group. Most companies run various forms of enterprise software programs to keep track of their inventory, record sales, manage payments to suppliers, cut employee paychecks, and handle other functions.

Some firms write their own enterprise software from scratch, but this can be time consuming and costly. Since many firms have similar procedures for accounting, finance, inventory management, and human resource functions, it often makes sense to buy a software packageA software product offered commercially by a third party. (a software product offered commercially by a third party) to support some of these functions. So-called enterprise resource planning (ERP)A software package that integrates the many functions (accounting, finance, inventory management, human resources, etc.) of a business. software packages serve precisely this purpose. In the way that Microsoft can sell you a suite of desktop software programs that work together, many companies sell ERP software that coordinates and integrates many of the functions of a business. The leading ERP vendors include the firm’s SAP and Oracle, although there are many firms that sell ERP software. A company doesn’t have to install all of the modules of an ERP suite, but it might add functions over time—for example, to plug in an accounting program that is able to read data from the firm’s previously installed inventory management system. And although a bit more of a challenge to integrate, a firm can also mix and match components, linking software the firm has written with modules purchased from different enterprise software vendors.

Figure 9.4 ERP in ActionAdapted from G. Edmondson, “Silicon Valley on the Rhine,” BusinessWeek International, November 3, 1997.

An ERP system with multiple modules installed can touch many functions of the business:

  • Sales—A sales rep from Vermont-based SnowboardCo. takes an order for five thousand boards from a French sporting goods chain. The system can verify credit history, apply discounts, calculate price (in euros), and print the order in French.
  • Inventory—While the sales rep is on the phone with his French customer, the system immediately checks product availability, signaling that one thousand boards are ready to be shipped from the firm’s Burlington warehouse, the other four thousand need to be manufactured and can be delivered in two weeks from the firm’s manufacturing facility in Guangzhou.
  • Manufacturing—When the customer confirms the order, the system notifies the Guangzhou factory to ramp up production for the model ordered.
  • Human Resources—High demand across this week’s orders triggers a notice to the Guangzhou hiring manager, notifying her that the firm’s products are a hit and that the flood of orders coming in globally mean her factory will have to hire five more workers to keep up.
  • Purchasing—The system keeps track of raw material inventories, too. New orders trigger an automatic order with SnowboardCo’s suppliers, so that raw materials are on hand to meet demand.
  • Order Tracking—The French customer can log in to track her SnowboardCo order. The system shows her other products that are available, using this as an opportunity to cross-sell additional products.
  • Decision Support—Management sees the firm’s European business is booming and plans a marketing blitz for the continent, targeting board models and styles that seem to sell better for the Alps crowd than in the U.S. market.

Other categories of enterprise software that managers are likely to encounter include the following:

  • customer relationship management (CRM)Systems used to support customer-related sales and marketing activities. systems used to support customer-related sales and marketing activities
  • supply chain management (SCM)Systems that can help a firm manage aspects of its value chain, from the flow of raw materials into the firm, through delivery of finished products and services at the point-of-consumption. systems that can help a firm manage aspects of its value chain, from the flow of raw materials into the firm through delivery of finished products and services at the point-of-consumption
  • business intelligence (BI) systemsSystems that use data created by other systems to provide reporting and analysis for organizational decision making., which use data created by other systems to provide reporting and analysis for organizational decision making

Major ERP vendors are now providing products that extend into these and other categories of enterprise application software, as well.

Most enterprise software works in conjunction with a database management system (DBMS)Sometimes referred to as database software; software for creating, maintaining, and manipulating data., sometimes referred to as a “database system.” The database system stores and retrieves the data that an application creates and uses. Think of this as another additional layer in our cake analogy. Although the DBMS is itself considered an application, it’s often useful to think of a firm’s database systems as sitting above the operating system, but under the enterprise applications. Many ERP systems and enterprise software programs are configured to share the same database system so that an organization’s different programs can use a common, shared set of data. This system can be hugely valuable for a company’s efficiency. For example, this could allow a separate set of programs that manage an inventory and point-of-sale system to update a single set of data that tells how many products a firm has to sell and how many it has already sold—information that would also be used by the firm’s accounting and finance systems to create reports showing the firm’s sales and profits.

Firms that don’t have common database systems with consistent formats across their enterprise often struggle to efficiently manage their value chain. Common procedures and data formats created by packaged ERP systems and other categories of enterprise software also make it easier for firms to use software to coordinate programs between organizations. This coordination can lead to even more value chain efficiencies. Sell a product? Deduct it from your inventory. When inventory levels get too low, have your computer systems send a message to your supplier’s systems so that they can automatically build and ship replacement product to your firm. In many cases these messages are sent without any human interaction, reducing time and errors. And common database systems also facilitate the use of BI systems that provide critical operational and competitive knowledge and empower decision making. For more on CRM and BI systems, and the empowering role of data, see Chapter 11 “The Data Asset: Databases, Business Intelligence, and Competitive Advantage”.

Figure 9.5

An organization’s database management system can be set up to work with several applications both within and outside the firm.

The Rewards and Risks of Packaged Enterprise Systems

When set up properly, enterprise systems can save millions of dollars and turbocharge organizations. For example, the CIO of office equipment maker Steelcase credited the firm’s ERP with an eighty-million-dollar reduction in operating expenses saved from eliminating redundant processes and making data more usable. The CIO of Colgate Palmolive also praised their ERP, saying, “The day we turned the switch on, we dropped two days out of our order-to-delivery cycle.”A. Robinson and D. Dilts, “OR and ERP,” ORMS Today, June 1999. Packaged enterprise systems can streamline processes, make data more usable, and ease the linking of systems with software across the firm and with key business partners. Plus, the software that makes up these systems is often debugged, tested, and documented with an industrial rigor that may be difficult to match with proprietary software developed in-house.

But for all the promise of packaged solutions for standard business functions, enterprise software installations have proven difficult. Standardizing business processes in software that others can buy means that those functions are easy for competitors to match, and the vision of a single monolithic system that delivers up wondrous efficiencies has been difficult for many to achieve. The average large company spends roughly $15 million on ERP software, with some installations running into the hundreds of millions of dollars.C. Rettig, “The Trouble with Enterprise Software,” MIT Sloan Management Review 49, no. 1 (2007): 21–27. And many of these efforts have failed disastrously.

FoxMeyer was once a six-billion-dollar drug distributor, but a failed ERP installation led to a series of losses that bankrupted the firm. The collapse was so rapid and so complete that just a year after launching the system, the carcass of what remained of the firm was sold to a rival for less than $80 million. Hershey Foods blamed a $466 million revenue shortfall on glitches in the firm’s ERP rollout. Among the problems, the botched implementation prevented the candy maker from getting product to stores during the critical period before Halloween. Nike’s first SCM and ERP implementation was labeled a “disaster”; their systems were blamed for over $100 million in lost sales.C. Koch, “Nike Rebounds: How (and Why) Nike Recovered from Its Supply Chain Disaster,” CIO, June 15, 2004. Even tech firms aren’t immune to software implementation blunders. HP once blamed a $160 million loss on problems with its ERP systems.R. Charette, “Why Software Fails,” IEEE Spectrum, September 2005. Manager beware—there are no silver bullets. For insight on the causes of massive software failures, and methods to improve the likelihood of success, see Section 9.6 “Total Cost of Ownership (TCO): Tech Costs Go Way beyond the Price Tag”.

Key Takeaways

  • Application software focuses on the work of a user or an organization.
  • Desktop applications are typically designed for a single user. Enterprise software supports multiple users in an organization or work group.
  • Popular categories of enterprise software include ERP (enterprise resource planning), SCM (supply chain management), CRM (customer relationship management), and BI (business intelligence) software, among many others.
  • These systems are used in conjunction with database management systems, programs that help firms organize, store, retrieve, and maintain data.
  • ERP and other packaged enterprise systems can be challenging and costly to implement, but can help firms create a standard set of procedures and data that can ultimately lower costs and streamline operations.
  • The more application software that is available for a platform, the more valuable that platform becomes.
  • The DBMS stores and retrieves the data used by the other enterprise applications. Different enterprise systems can be configured to share the same database system in order share common data.
  • Firms that don’t have common database systems with consistent formats across their enterprise often struggle to efficiently manage their value chain, and often lack the flexibility to introduce new ways of doing business. Firms with common database systems and standards often benefit from increased organizational insight and decision-making capabilities.
  • Enterprise systems can cost millions of dollars in software, hardware, development, and consulting fees, and many firms have failed when attempting large-scale enterprise system integration. Simply buying a system does not guarantee its effective deployment and use.
  • When set up properly, enterprise systems can save millions of dollars and turbocharge organizations by streamlining processes, making data more usable, and easing the linking of systems with software across the firm and with key business partners.

Questions and Exercises

  1. What is the difference between desktop and enterprise software?
  2. Who are the two leading ERP vendors?
  3. List the functions of a business that might be impacted by an ERP.
  4. What do the acronyms ERP, CRM, SCM, and BI stand for? Briefly describe what each of these enterprise systems does.
  5. Where in the “layer cake” analogy does the DBMS lie.
  6. Name two companies that have realized multimillion-dollar benefits as result of installing enterprise systems.
  7. Name two companies that have suffered multimillion-dollar disasters as result of failed enterprise system installations.
  8. How much does the average large company spend annually on ERP software?

9.4 Distributed Computing

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the concept of distributed computing and its benefits.
  2. Understand the client-server model of distributed computing.
  3. Know what Web services are and the benefits that Web services bring to firms.
  4. Appreciate the importance of messaging standards and understand how sending messages between machines can speed processes, cut costs, reduce errors, and enable new ways of doing business.

When computers in different locations can communicate with one another, this is often referred to as distributed computingA form of computing where systems in different locations communicate and collaborate to complete a task.. Distributed computing can yield enormous efficiencies in speed, error reduction, and cost savings and can create entirely new ways of doing business. Designing systems architecture for distributed systems involves many advanced technical topics. Rather than provide an exhaustive decomposition of distributed computing, the examples that follow are meant to help managers understand the bigger ideas behind some of the terms that they are likely to encounter.

Let’s start with the term serverA program that fulfills the requests of a client.. This is a tricky one because it’s frequently used in two ways: (1) in a hardware context a server is a computer that has been configured to support requests from other computers (e.g., Dell sells servers) and (2) in a software context a server is a program that fulfills requests (e.g., the Apache open source Web server). Most of the time, server software resides on server-class hardware, but you can also set up a PC, laptop, or other small computer to run server software, albeit less powerfully. And you can use mainframe or super-computer-class machines as servers, too.

The World Wide Web, like many other distributed computing services, is what geeks call a client-server system. Client-server refers to two pieces of software, a clientA software program that makes requests of a server program. that makes a request, and a server that receives and attempts to fulfill the request. In our WWW scenario, the client is the browser (e.g., Internet Explorer, Firefox, Safari). When you type a Web site’s address into the location field of your browser, you’re telling the client to “go find the Web server software at the address provided, and tell the server to return the Web site requested.”

It is possible to link simple scripting languages to a Web server for performing calculations, accessing databases, or customizing Web sites. But more advanced distributed environments may use a category of software called an application serverSoftware that houses and serves business logic for use (and reuse) by multiple applications.. The application server (or app server) houses business logic for a distributed system. Individual Web servicesSmall pieces of code that are accessed via the application server which permit interoperable machine-to-machine interaction over a network. served up by the app server are programmed to perform different tasks: returning a calculation (“sales tax for your order will be $11.58”), accessing a database program (“here are the results you searched for”), or even making a request to another server in another organization (“Visa, please verify this customer’s credit card number for me”).

Figure 9.6

In this multitiered distributed system, client browsers on various machines (desktop, laptop, mobile) access the system through the Web server. The cash register doesn’t use a Web browser, so instead the cash register logic is programmed to directly access the services it needs from the app server. Web services accessed from the app server may be asked to do a variety of functions, including perform calculations, access corporate databases, or even make requests from servers at other firms (for example, to verify a customer’s credit card).

Those little chunks of code that are accessed via the application server are sometimes referred to as Web services. The World Wide Web consortium defines Web services as software systems designed to support interoperable machine-to-machine interaction over a network.W3C, “Web Services Architecture,” W3C Working Group Note, February 11, 2004. And when computers can talk together (instead of people), this often results in fewer errors, time savings, cost reductions, and can even create whole new ways of doing business! Each Web service defines the standard method for other programs to request it to perform a task and defines the kind of response the calling client can expect back. These standards are referred to as application programming interfaces (APIs)Programming hooks, or guidelines, published by firms that tell other programs how to get a service to perform a task such as send or receive data. For example, Amazon.com provides APIs to let developers write their own applications and Websites that can send the firm orders..

Look at the advantages that Web services bring a firm like Amazon. Using Web services, the firm can allow the same order entry logic to be used by Web browsers, mobile phone applications, or even by third parties who want to access Amazon product information and place orders with the firm (there’s an incentive to funnel sales to Amazon—the firm will give you a cut of any sales that you send Amazon’s way). Organizations that have created a robust set of Web services around their processes and procedures are said to have a service-oriented architecture (SOA)A robust set of Web services built around an organizations processes and procedures.. Organizing systems like this, with separate applications in charge of client presentation, business logic, and database, makes systems more flexible. Code can be reused, and each layer can be separately maintained, upgraded, or migrated to new hardware—all with little impact on the others.

Web services sound geeky, but here’s a concrete example illustrating their power. Southwest Airlines had a Web site where customers could book flights, but many customers also wanted to rent a car or book a hotel, too. To keep customers on Southwest.com, the firm and its hotel and rental car partners created a set of Web services and shared the APIs. Now customers visiting Southwest.com can book a hotel stay and rental car on the same page where they make their flight reservation. This process transforms Southwest.com into a full service travel destination and allows the site to compete head-to-head with the likes of Expedia, Travelocity, and Orbitz.J. McCarthy, “The Standards Body Politic,” InfoWorld, May 17, 2002.

Think about why Web services are important from a strategic perspective. By adding hotel and rental car services, Southwest is now able to eliminate the travel agent, along with any fees they might share with the agent. This shortcut allows the firm to capture more profits or pass on savings to customers, securing its position as the first place customers go for low-cost travel. And perhaps most importantly, Southwest can capture key data from visitor travel searches and bookings (something it likely couldn’t do if customers went to a site like Expedia or Travelocity). Data is a hugely valuable asset, and this kind of customer data can be used by Southwest to send out custom e-mail messages and other marketing campaigns to bring customers back to the airline. As geeky as they might at first seem, Web services can be very strategic!

Figure 9.7

Southwest.com uses Web services to allow car rental and hotel firms to book services through Southwest. This process transforms Southwest.com into a full-service online travel agent.

Messaging Standards

Two additional terms you might hear within the context of distributed computing are EDI and XML. EDI (electronic data interchange)A set of standards for exchanging messages containing formatted data between computer applications. is a set of standards for exchanging information between computer applications. EDI is most often used as a way to send the electronic equivalent of structured documents between different organizations. Using EDI, each element in the electronic document, like a firm name, address, or customer number, is coded so that it can be recognized by the receiving computer program. Eliminating paper documents makes businesses faster and lowers data entry and error costs. One study showed that firms that used EDI decreased their error rates by 82 percent and their cost of producing each document fell by up to 96 percent.“Petroleum Industry Continues to Explore EDI,” National Petroleum News 90, no. 12 (November 1998).

EDI is a very old standard, with roots stretching back to the 1948 Berlin Air Lift. While still in use, a new generation of more-flexible technologies for specifying data standards are taking its place. Chief among the technologies replacing EDI is extensible markup language (XML)A tagging language that can be used to identify data fields made available for use by other applications. Most APIs and Web services send messages where the data exchanged is wrapped in identifying XML tags.. XML has lots of uses, but in the context of distributed systems, it allows software developers to create a set of standards for common data elements that, like EDI messages, can be sent between different kinds of computers, different applications, and different organizations. XML is often thought of as easier to code than EDI, and it’s more robust because it can be extended—organizations can create formats to represent any kind of data (e.g., a common part number, photos, the complaint field collected by customer support personnel). In fact, most messages sent between Web services are coded in XML (the technology is a key enabler in mashups, discussed in Chapter 7 “Peer Production, Social Media, and Web 2.0”). Many computer programs also use XML as a way to export and import data in a common format that can be used regardless of the kind of computer hardware, operating system, or application program used. And if you design Web sites, you might encounter XML as part of the coding behind the cascading style sheets (CSS) that help maintain a consistent look and feel to the various Web pages in a given Web site.

Rearden Commerce: A Business Built on Web Services

Web services, APIs, and open standards not only transform businesses, they can create entire new firms that change how we get things done. For a look at the mashed-up, integrated, hyperautomated possibilities that Web services make possible, check out Rearden Commerce, a Foster City, California, firm that is using this technology to become what AMR’s Chief Research Office referred to as “Travelocity on Steroids.”

Using Rearden, firms can offer their busy employees a sort of Web-based concierge/personal assistant. Rearden offers firms a one-stop shop where employees can not only make the flight, car, and hotel bookings they might do from a travel agent, they can also book dinner reservations, sports and theatre tickets, and arrange for business services like conference calls and package shipping. Rearden doesn’t supply the goods and services it sells. Instead it acts as the middleman between transactions. A set of open APIs to its Web services allows Rearden’s one hundred and sixty thousand suppliers to send product and service data to Rearden, and to receive booking and sales data from the site.

In this ultimate business mashup, a mobile Rearden user could use her phone to book a flight into a client city, see restaurants within a certain distance of her client’s office, have these locations pop up on a Google map, have listings accompanied by Zagat ratings and cuisine type, book restaurant reservations through Open Table, arrange for a car and driver to meet her at her client’s office at a specific time, and sync up these reservations with her firm’s corporate calendaring systems. If something unexpected comes up, like a flight delay, Rearden will be sure she gets the message. The system will keep track of any cancelled reservation credits, and also records travel reward programs, so Rearden can be used to spend those points in the future.

In order to pull off this effort, the Rearden maestros are not only skilled at technical orchestration, but also in coordinating customer and supplier requirements. As TechCrunch’s Erick Schonfeld put it, “The hard part is not only the technology—which is all about integrating an unruly mess of APIs and Web services—[it also involves] signing commercially binding service level agreements with [now over 160,000] merchants across the world.” For its efforts, Rearden gets to keep between 6 percent and 25 percent of every nontravel dollar spent, depending on the service. The firm also makes money from subscriptions, and distribution deals.

The firm’s first customers were large businesses and included ConAgra, GlaxoSmithKline, and Motorola. Rearden’s customers can configure the system around special parameters unique to each firm: to favor a specific airline, benefit from a corporate discount, or to restrict some offerings for approved employees only. Rearden investors include JPMorgan Chase and American Express—both of whom offer Rearden to their employees and customers. Even before the consumer version was available, Rearden had over four thousand corporate customers and two million total users, a user base larger than better-known firms like Salesforce.com.M. Arrington, “Rearden Commerce: Time for the Adults to Come In and Clean House,” TechCrunch, April 5, 2007; E. Schonfeld, “At Rearden Commerce, Addiction Is Job One,” TechCrunch, May 6, 2008; and M. Arrington, “2008: Rearden Commerce Has a Heck of a Year,” TechCrunch, January 13, 2009. For all the pizzazz we recognize that, as a start-up, the future of Rearden Commerce remains uncertain; however, the firm’s effective use of Web services illustrates the business possibilities as technologies allow firms to connect with greater ease and efficiency.

Connectivity has made our systems more productive and enables entire new strategies and business models. But these wonderful benefits come at the price of increased risk. When systems are more interconnected, opportunities for infiltration and abuse also increase. Think of it this way—each “connection” opportunity is like adding another door to a building. The more doors that have to be defended, the more difficult security becomes. It should be no surprise that the rise of the Internet and distributed computing has led to an explosion in security losses by organizations worldwide.

Key Takeaways

  • Client-server computing is a method of distributed computing where one program (a client) makes a request to be fulfilled by another program (a server).
  • Server is a tricky term and is sometimes used to refer to hardware. While server-class hardware refers to more powerful computers designed to support multiple users, just about any PC or notebook can be configured to run server software.
  • Web servers serve up Web sites and can perform some scripting.
  • Most firms serve complex business logic from an application server.
  • Isolating a system’s logic in three or more layers (presentation or user interface, business logic, and database) can allow a firm flexibility in maintenance, reusability, and in handling upgrades.
  • Web services allow different applications to communicate with one another. APIs define the method to call a Web service (e.g., to get it to do something), and the kind of response the calling program can expect back.
  • Web services make it easier to link applications as distributed systems, and can make it easier for firms to link their systems across organizations.
  • Popular messaging standards include EDI (older) and XML. Sending messages between machines instead of physical documents can speed processes, drastically cut the cost of transactions, and reduce errors.
  • Distributed computing can yield enormous efficiencies in speed, error reduction, and cost savings and can create entirely new ways of doing business.
  • When computers can communicate with each other (instead of people), this often results in fewer errors, time savings, cost reductions, and can even create whole new ways of doing business.
  • Web services, APIs, and open standards not only transform businesses, they can create entire new firms that change how we get things done.

Questions and Exercises

  1. Differentiate the term “server” used in a hardware context, from “server” used in a software context.
  2. Describe the “client-server” model of distributed computing. What products that you use would classify as leveraging client-server computing?
  3. List the advantages that Web services have brought to Amazon.
  4. How has Southwest Airlines utilized Web services to its competitive advantage?
  5. What is Rearden Commerce and which technologies does it employ? Describe Rearden Technology’s revenue model. Who were Rearden Technology’s first customers? Who were among their first investors?
  6. What are the security risks associated with connectivity, the Internet, and distributed processing?

9.5 Writing Software

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand, at a managerial level, what programming languages are and how software is developed.
  2. Recognize that an operating system and microprocessor constrain the platform upon which most compiled application software will run.
  3. Understand what Java is and why it is significant.
  4. Know what scripting languages are.

So you’ve got a great idea that you want to express in software—how do you go about creating a program? Programmers write software in a programming languageProvides the standards, syntax, statements, and instructions for writing computer software.. While each language has its strengths and weaknesses, most commercial software is written in C++ (pronounced “see plus plus”) or C# (pronounced “see sharp”). Visual Basic (from Microsoft) and Java (from Sun) are also among the more popular of the dozens of programming languages available. Web developers may favor specialty languages like Ruby and Python, while languages like SQL are used in databases.

Most professional programmers use an integrated development environment (IDE)An application that includes an editor (a sort of programmer’s word processor), debugger, and compiler, among other tools. to write their code. The IDE includes a text editor, a debugger for sleuthing out errors, and other useful programming tools. The most popular IDE for Windows is Visual Studio, while Apple offers the Xcode IDE. Most IDEs can support several different programming languages. The IDE will also compileStep in which program code written in a language that humans can more easily understand, is then converted into a form (expressed in patterns of ones and zeros) that can be understood and executed by a microprocessor. Programmers using conventional programming languages must compile their software before making it available for execution. a programmer’s code, turning the higher-level lines of instructions that are readable by humans into lower-level instructions expressed as the patterns of ones and zeros that are readable by a computer’s microprocessor.

Figure 9.8

Microsoft’s Visual Studio IDE supports desktop, server, mobile, and cloud computing software development.

Look at the side of a box of commercial software and you’re likely to see system requirements that specify the operating system and processor that the software is designed for (e.g., “this software works on computers with Windows 7 and Intel-compatible processors”). Wouldn’t it be great if software could be written once and run everywhere? That’s the idea behind JavaA programming language, initially developed by Sun Microsystems, designed to provide true platform independence (“write once, run anywhere”) for application developers. In most cases, Java apps are developed to be executed by a Java Virtual Machine—an interpreting layer that translates code as it executes, into the format required by the operating system and microprocessor. Without Java, application developers have to write and compile software to execute natively by a specific operating system / microprocessor combination (e.g., Windows/Intel, Linux PowerPC, Mac/Intel, Linux/Intel).—a programming language developed by Sun Microsystems.

Java programmers don’t write code with specific operating system commands (say for Windows, Mac OS X, or Linux), instead they use special Java commands to control their user interface or interact with the display and other hardware. Java programs can run on any computer that has a Java Virtual Machine (JVM), a software layer that interprets Java code so that it can be understood by the operating system and processor of a given computer. Java’s platform independence—the ability for developers to “write once, run everywhere”—is its biggest selling point. Many Web sites execute Java applets to run the animation you might see in advertisements or games. Java has also been deployed on over six billion mobile phones worldwide, and is popular among enterprise programmers who want to be sure their programs can scale from smaller hardware up to high-end supercomputers. As long as the machine receiving the Java code has a JVM, then the Java application should run. However, Java has not been popular for desktop applications. Since Java isn’t optimized to take advantage of interface elements specific to the Mac or Windows, most Java desktop applications look clunky and unnatural. Java code that runs through the JVM interpreter is also slower than code compiled for the native OS and processor that make up a platform.Some offerings have attempted to overcome the speed issues associated with interpreting Java code. Just-in-time compilation stores code in native processor-executable form after each segment is initially interpreted, further helping to speed execution. Other environments allow for Java to be compiled ahead of time so that it can be directly executed by a microprocessor. However, this process eliminates code portability—Java’s key selling point. And developers preparing their code for the JVM actually precompile code into something called Java bytecode, a format that’s less human friendly but more quickly interpreted by JVM software.

Scripting languages are the final category of programming tool that we’ll cover. Scripting languagesProgramming tool that executes within an application. Scripting languages are interpreted within their applications, rather than compiled to run directly by a microprocessor. typically execute within an application. Microsoft offers a scripting language called VB Script (a derivative of Visual Basic) to automate functions in Office. And most browsers and Web servers support JavaScript, a language that helps make the Web more interactive (despite its name, JavaScript is unrelated to Java). Scripting languages are interpretedLanguages where each line of written code is converted (by a software program, called an “interpreter”) for execution at run-time. Most scripting languages are interpreted languages. Many programmers also write Java applications to be interpreted by the Java Virtual Machine. within their applications, rather than compiled to run directly by a microprocessor. This distinction makes them slower than the kinds of development efforts found in most commercial software. But most scripting languages are usually easy to use, and are often used both by professional programmers and power users.

Key Takeaways

  • Programs are often written in a tool called an IDE, an application that includes an editor (a sort of programmer’s word processor), debugger, and compiler, among other tools.
  • Compiling takes code from the high-level language that humans can understand and converts them into the sets of ones and zeros in patterns representing instructions that microprocessors understand.
  • Popular programming languages include C++, C#, Visual Basic, and Java.
  • Most software is written for a platform—a combination of an operating system and microprocessor.
  • Java is designed to be platform independent. Computers running Java have a separate layer called a Java Virtual Machine that translates (interprets) Java code so that it can be executed on an operating system/processor combination. In theory, Java is “write once, run everywhere,” as opposed to conventional applications that are written for an operating system and compiled for an OS/processor combination.
  • Java is popular on mobile phones, enterprise computing, and to make Web sites more interactive. Java has never been a successful replacement for desktop applications, largely because user interface differences among the various operating systems are too great to be easily standardized.
  • Scripting languages are interpreted languages, such as VB Script or Java Script. Many scripting languages execute within an application (like the Office programs, a Web browser, or to support the functions of a Web server). They are usually easier to program, but are less powerful and execute more slowly than compiled languages.

Questions and Exercises

  1. List popular programming languages.
  2. What’s an IDE? Why do programmers use IDEs? Name IDEs popular for Windows and Mac users.
  3. What is the difference between a compiled programming language and an interpreted programming language?
  4. Name one advantage and one disadvantage of scripting languages.
  5. In addition to computers, on what other technology has Java been deployed? Why do you suppose Java is particularly attractive for these kinds of applications?
  6. What’s a JVM? Why do you need it?
  7. What if a programmer wrote perfect Java code, but there was a bug on the JVM installed on a given computer? What might happen?
  8. Why would developers choose to write applications in Java? Why might they skip Java and choose another programming language?
  9. Why isn’t Java popular for desktop applications?
  10. Go to http://www.java.com. Click on “Do I have Java?” Is Java running on your computer? Which version?

9.6 Total Cost of Ownership (TCO): Tech Costs Go Way beyond the Price Tag

Learning Objectives

After studying this section you should be able to do the following:

  1. List the different cost categories that comprise total cost of ownership.
  2. Understand that once a system is implemented, the costs of maintaining and supporting the system continue.
  3. List the reasons that technology development projects fail and the measures that can be taken to increase the probability of success.

Managers should recognize that there are a whole host of costs that are associated with creating and supporting an organization’s information systems. Of course, there are programming costs for custom software as well as purchase, configuration, and licensing costs for packaged software, but there’s much, much more.

There are costs associated with design and documentation (both for programmers and for users). There are also testing costs. New programs should be tested thoroughly across the various types of hardware the firm uses, and in conjunction with existing software and systems, before being deployed throughout the organization. Any errors that aren’t caught can slow down a business or lead to costly mistakes that could ripple throughout an organization and its partners. Studies have shown that errors not caught before deployment could be one hundred times more costly to correct than if they were detected and corrected beforehand.R. Charette, “Why Software Fails,” IEEE Spectrum, September 2005.

Once a system is “turned on,” the work doesn’t end there. Firms need to constantly engage in a host of activities to support the system that may also include the following:

  • providing training and end user support
  • collecting and relaying comments for system improvements
  • auditing systems to ensure complianceEnsuring that an organization’s systems operate within required legal constraints, and industry and organizational obligations (i.e., that the system operates within the firm’s legal constraints and industry obligations)
  • providing regular backup of critical data
  • planning for redundancy and disaster recovery in case of an outage
  • vigilantly managing the moving target of computer security issues

With so much to do, it’s no wonder that firms spend 70 to 80 percent of their information systems (IS) budgets just to keep their systems running.C. Rettig, “The Trouble with Enterprise Software,” MIT Sloan Management Review 49, no. 1 (2007): 21–27. The price tag and complexity of these tasks can push some managers to think of technology as being a cost sink rather than a strategic resource. These tasks are often collectively referred to as the total cost of ownership (TCO)All of the costs associated with the design, development, testing, implementation, documentation, training and maintenance of a software system. of an information system. Understanding TCO is critical when making technology investment decisions. TCO is also a major driving force behind the massive tech industry changes discussed in Chapter 10 “Software in Flux: Partly Cloudy and Sometimes Free”.

Why Do Technology Projects Fail?

Even though information systems represent the largest portion of capital spending at most firms, an astonishing one in three technology development projects fail to be successfully deployed.L. Dignan, “Survey: One in 3 IT Projects Fail; Management OK with It,” ZDNet, December 11, 2007. Imagine if a firm lost its investment in one out of every three land purchases, or when building one in three factories. These statistics are dismal! Writing in IEEE Spectrum, risk consultant Robert Charette provides a sobering assessment of the cost of software failures, stating, “The yearly tab for failed and troubled software conservatively runs somewhere from $60 to $70 billion in the United States alone. For that money, you could launch the space shuttle one hundred times, build and deploy the entire 24-satellite Global Positioning System, and develop the Boeing 777 from scratch—and still have a few billion left over.”R. Charette, “Why Software Fails,” IEEE Spectrum, September 2005.

Why such a bad track record? Sometimes technology itself is to blame, other times it’s a failure to test systems adequately, and sometimes it’s a breakdown of process and procedures used to set specifications and manage projects. In one example, a multimillion-dollar loss on the NASA Mars Observer was traced back to a laughably simple oversight—Lockheed Martin contractors using English measurements, while the folks at NASA used the metric system.R. Lloyd, “Metric Mishap Caused Loss of NASA Orbiter,” CNN, September 20, 1999. Yes, a $125 million taxpayer investment was lost because a bunch of rocket scientists failed to pay attention to third grade math. When it comes to the success or failure of technical projects, the devil really is in the details.

Projects rarely fail for just one reason. Project post-mortems often point to a combination of technical, project management, and business decision blunders. The most common factors include the following:List largely based on R. Charette, “Why Software Fails,” IEEE Spectrum, September 2005.

  • Unrealistic or unclear project goals
  • Poor project leadership and weak executive commitment
  • Inaccurate estimates of needed resources
  • Badly defined system requirements and allowing “feature creep” during development
  • Poor reporting of the project’s status
  • Poor communication among customers, developers, and users
  • Use of immature technology
  • Unmanaged risks
  • Inability to handle the project’s complexity
  • Sloppy development and testing practices
  • Poor project management
  • Stakeholder politics
  • Commercial pressures (e.g., leaving inadequate time or encouraging corner-cutting)

Managers need to understand the complexity involved in their technology investments, and that achieving success rarely lies with the strength of the technology alone.

But there is hope. Information systems organizations can work to implement procedures to improve the overall quality of their development practices. Mechanisms for quality improvement include capability maturity model integration (CMMI)A process-improvement approach (useful for but not limited to software engineering projects) that can assist in assessing the maturity, quality, and development of certain organizational business processes, and suggest steps for their improvement., which gauge an organization’s process maturity and capability in areas critical to developing and deploying technology projects, and provides a carefully chosen set of best practices and guidelines to assist quality and process improvement.R. Kay, “QuickStudy: Capability Maturity Model Integration (CMMI),” Computerworld, January 24, 2005; and Carnegie Mellon Software Engineering Institute, Welcome to CMMI, 2009, http://www.sei.cmu.edu/cmmi.

Firms are also well served to leverage established project planning and software development methodologies that outline critical businesses processes and stages when executing large-scale software development projects. The idea behind these methodologies is straightforward—why reinvent the wheel when there is an opportunity to learn from and follow blueprints used by those who have executed successful efforts. When methodologies are applied to projects that are framed with clear business goals and business metrics, and that engage committed executive leadership, success rates can improve dramatically.A. Shenhar and D. Dvir, Reinventing Project Management: The Diamond Approach to Successful Growth and Innovation (Boston: Harvard Business School Press, 2007).

While software development methodologies are the topic of more advanced technology courses, the savvy manager knows enough to inquire about the development methodologies and quality programs used to support large scale development projects, and can use these investigations as further input when evaluating whether those overseeing large scale efforts have what it takes to get the job done.

Key Takeaways

  • The care and feeding of information systems can be complex and expensive. The total cost of ownership of systems can include software development and documentation, or the purchase price and ongoing license and support fees, plus configuration, testing, deployment, maintenance, support, training, compliance auditing, security, backup, and provisions for disaster recovery. These costs are collectively referred to as TCO, or a system’s total cost of ownership.
  • Information systems development projects fail at a startlingly high rate. Failure reasons can stem from any combination of technical, process, and managerial decisions.
  • IS organizations can leverage software development methodologies to improve their systems development procedures, and firms can strive to improve the overall level of procedures used in the organization through models like CMMI. However, it’s also critical to engage committed executive leadership in projects, and to frame projects using business metrics and outcomes to improve the chance of success.
  • System errors that aren’t caught before deployment can slow down a business or lead to costly mistakes that could ripple throughout an organization. Studies have shown that errors not caught before deployment could be 100 times more costly to correct than if they were detected and corrected beforehand.
  • Firms spend 70 to 80 percent of their IS budgets just to keep their systems running.
  • One in three technology development projects fail to be successfully deployed.
  • IS organizations can employ project planning and software development methodologies to implement procedures to improve the overall quality of their development practices.

Questions and Exercises

  1. List the types of total ownership costs associated with creating and supporting an organization’s information systems.
  2. On average, what percent of firms’ IS budgets is spent to keep their systems running?
  3. What are the possible effects of not detecting and fixing major system errors before deployment?
  4. List some of the reasons for the failure of technology development projects.
  5. What is the estimated yearly cost of failed technology development projects?
  6. What was the reason attributed to the failure of the NASA Mars Observer project?
  7. What is capability maturity model integration (CMMI) and how is it used to improve the overall quality of a firm’s development practices?
  8. Perform an Internet search for “IBM Rational Portfolio Manager.” How might IBM’s Rational Portfolio Manager software help companies realize more benefit from their IT systems development project expenditures? What competing versions of this product offered by other organizations?

Chapter 8: Facebook: Building a Business from the Social Graph

8.1 Introduction

Learning Objectives

After studying this section you should be able to do the following:

  1. Be familiar with Facebook’s origins and rapid rise.
  2. Understand how Facebook’s rapid rise has impacted the firm’s ability to raise venture funding and its founder’s ability to maintain a controlling interest in the firm.

Here’s how much of a Web 2.0 guy Mark Zuckerberg is: during the weeks he spent working on Facebook as a Harvard sophomore, he didn’t have time to study for a course he was taking, “Art in the Time of Augustus,” so he built a Web site containing all of the artwork in class and pinged his classmates to contribute to a communal study guide. Within hours, the wisdom of crowds produced a sort of custom CliffsNotes for the course, and after reviewing the Web-based crib sheet, he aced the test. Turns out he didn’t need to take that exam, anyway. Zuck (that’s what the cool kids call him)For an insider account of Silicon Valley Web 2.0 start-ups, see Sarah Lacy, Once You’re Lucky, Twice You’re Good: The Rebirth of Silicon Valley and the Rise of Web 2.0. (New York: Gotham Books, 2008). dropped out of Harvard later that year.

Zuckerberg is known as both a shy, geeky, introvert who eschews parties, and as a brash Silicon Valley bad boy. After Facebook’s incorporation, Zuckerberg’s job description was listed as “Founder, Master and Commander [and] Enemy of the State.”T. McGinn, “Online Facebooks Duel over Tangled Web of Authorship,” Harvard Crimson, May 28, 2004. An early business card read “I’m CEO…Bitch.”C. Hoffman, “The Battle for Facebook,” Rolling Stone, June 26, 2008, 9. And let’s not forget that Facebook came out of drunken experiments in his dorm room, one of which was a system for comparing classmates to farm animals (Zuckerberg, threatened with expulsion, later apologized). For one meeting with Sequoia Capital, the venerable Menlo Park venture capital firm that backed Google and YouTube, Zuckerberg showed up in his pajamas.C. Hoffman, “The Battle for Facebook,” Rolling Stone, June 26, 2008.

By the age of twenty-three, Mark Zuckerberg had graced the cover of Newsweek, been profiled on 60 Minutes, and was discussed in the tech world with a reverence previously reserved only for Steve Jobs and the Google guys, Sergey Brin and Larry Page. But Mark Zuckerberg’s star rose much faster than any of his predecessors. Just two weeks after Facebook launched, the firm had four thousand users. Ten months later it was up to one million. The growth continued, and the business world took notice. In 2006, Viacom (parent of MTV) saw that its core demographic was spending a ton of time on Facebook and offered to buy the firm for three quarters of a billion dollars. Zuckerberg passed.S. Rosenbush, “Facebook’s on the Block,” BusinessWeek, March 28, 2006. Yahoo! offered up a cool billion (twice). Zuck passed again, both times.

As growth skyrocketed, Facebook built on its stranglehold of the college market (over 85 percent of four-year college students are Facebook members), opening up first to high schoolers, then to everyone. Web hipsters started selling shirts emblazoned with “I Facebooked your Mom!” Even Microsoft wanted some of Facebook’s magic. In 2006, the firm temporarily locked up the right to broker all banner ad sales that run on the U.S. version of Facebook, guaranteeing Zuckerberg’s firm $100 million a year through 2011. In 2007, Microsoft came back, buying 1.6 percent of the firm for $240 million.While Microsoft had cut deals to run banner ads worldwide, Facebook dropped banner ads for poor performance in early 2010; see C. McCarthy, “More Social, Please: Facebook Nixes Banner Ads”, CNET, February 5, 2010.

The investment was a shocker. Do the math and a 1.6 percent stake for $240 million values Facebook at $15 billion (more on that later). That meant that a firm that at the time had only five hundred employees, $150 million in revenues, and was helmed by a twenty-three-year-old college dropout in his first “real job,” was more valuable than General Motors. Rupert Murdoch, whose News Corporation owns rival MySpace, engaged in a little trash talk, referring to Facebook as “the flavor of the month.”B. Morrissey, “Murdoch: Facebook Is ‘Flavor of the Month,’” Media Week, June 20, 2008.

Watch your back, Rupert. Or on second thought, watch Zuckerberg’s. By spring 2009, Facebook had more than twice MySpace’s monthly unique visitors worldwide;E. Schonfeld, “Dear Owen, Good Luck with That,” TechCrunch, April 24, 2009. by June, Facebook surpassed MySpace in the United States;“Facebook Dethrones MySpace in the U.S.,” Los Angeles Times, June 16, 2009, http://articles.latimes.com/2009/jun/16/business/fi-facebook16. by July, Facebook was cash-flow positiveWhen a company’s revenues can cover its operating costs.; and by February 2010 (when Facebook turned six), the firm had over four hundred million users, more than doubling in size in less than a year.D. Gage, “Facebook Claims 250 Million Users,” InformationWeek, July 16, 2009. Murdoch, the media titan who stood atop an empire that includes the Wall Street Journal and Fox, had been outmaneuvered by “the kid.”

Why Study Facebook?

Looking at the “flavor of the month” and trying to distinguish the reality from the hype is a critical managerial skill. In Facebook’s case, there are a lot of folks with a vested interest in figuring out where the firm is headed. If you want to work there, are you signing on to a firm where your stock options and 401k contributions are going to be worth something or worthless? If you’re an investor and Facebook goes publicThe first time a firm sells stock to the public; formally called an initial public stock offering (IPO)., should you shortShort selling is an attempt to profit from a falling stock price. Short sellers sell shares they don’t own with an obligation of later repayment. They do so in the hope that the price of sold shares will fall. They then repay share debt with shares purchased at a lower price and pocket the difference (spread) between initial share price and repayment price. the firm or increase your holdings? Would you invest in or avoid firms that rely on Facebook’s business? Should your firm rush to partner with the firm? Would you extend the firm credit? Offer it better terms to secure its growing business, or worse terms because you think it’s a risky bet? Is this firm the next Google (underestimated at first, and now wildly profitable and influential), the next GeoCities (Yahoo! paid $3 billion for it—no one goes to the site today), or the next Skype (deeply impactful with over half a billion accounts worldwide, but not much of a profit generator)? The jury is still out on all this, but let’s look at the fundamentals with an eye to applying what we’ve learned. No one has a crystal ball, but we do have some key concepts that can guide our analysis. There are a lot of broadly applicable managerial lessons that can be gleaned by examining Facebook’s successes and missteps. Studying the firm provides a context for examining nework effects, platforms, partnerships, issues in the rollout of new technologies, privacy, ad models, and more.

Zuckerberg Rules!

Many entrepreneurs accept start-up capital from venture capitalists (VCs)Investor groups that provide funding in exchange for a stake in the firm, and often, a degree of managerial control (usually in the form of a voting seat or seats on the firm’s board of directors)., investor groups that provide funding in exchange for a stake in the firm, and often, a degree of managerial control (usually in the form of a voting seat or seats on the firm’s board of directorsGroup assigned to govern, advise, and provide oversight for the firm. The board’s many responsibilities typically include hiring and firing the CEO.). Typically, the earlier a firm accepts VC money, the more control these investors can exert (earlier investments are riskier, so VCs can demand more favorable terms). VCs usually have deep entrepreneurial experience and a wealth of contacts, and can often offer important guidance and advice, but strong investor groups can oust a firm’s founder and other executives if they’re dissatisfied with the firm’s performance.

At Facebook, however, Zuckerberg owns an estimated 20 percent to 30 percent of the company, and controls three of five seats on the firm’s board of directors. That means that he’s virtually guaranteed to remain in control of the firm, regardless of what investors say. Maintaining this kind of control is unusual in a start-up, and his influence is a testament to the speed with which Facebook expanded. By the time Zuckerberg reached out to VCs, his firm was so hot that he could call the shots, giving up surprisingly little in exchange for their money.

Key Takeaways

  • Facebook was founded by a nineteen-year-old college sophomore and eventual dropout.
  • It is currently the largest social network in the world, boasting more than four hundred million members and usage rates that would be the envy of most media companies. The firm is now larger than MySpace in both the United States and worldwide.
  • The firm’s rapid rise is the result of network effects and the speed of its adoption placed its founder in a particularly strong position when negotiating with venture firms. As a result, Facebook founder Mark Zuckerberg retains significant influence over the firm.
  • While revenue prospects remain sketchy, some reports have valued the firm at $15 billion, based largely on an extrapolation of a Microsoft stake.

Questions and Exercises

  1. Who started Facebook? How old was he then? Now? How much control does the founding CEO have over his firm? Why?
  2. Which firms have tried to acquire Facebook? Why? What were their motivations and why did Facebook seem attractive? Do you think these bids are justified? Do you think the firm should have accepted any of the buyout offers? Why or why not?
  3. As of late 2007, Facebook boasted an extremely high “valuation.” How much was Facebook allegedly “worth”? What was this calculation based on?
  4. Why study Facebook? Who cares if it succeeds?

8.2 What’s the Big Deal?

Learning Objectives

After studying this section you should be able to do the following:

  1. Recognize that Facebook’s power is allowing it to encroach on and envelop other Internet businesses.
  2. Understand the concept of the “dark Web” and why some feel this may one day give Facebook a source of advantage vis-à-vis Google.
  3. Understand the basics of Facebook’s infrastructure, and the costs required to power the effort.

The prior era’s Internet golden boy, Netscape founder Marc Andreessen, has said that Facebook is “an amazing achievement one of the most significant milestones in the technology industry.”F. Vogelstein, “How Mark Zuckerberg Turned Facebook into the Web’s Hottest Platform,” Wired, September 6, 2007. While still in his twenties, Andreessen founded Netscape, eventually selling it to AOL for over $4 billion. His second firm, Opsware, was sold to HP for $1.6 billion. He joined Facebook’s Board of Directors within months of making this comment. Why is Facebook considered such a big deal?

First there’s the growth: between December 2008 and 2009, Facebook was adding between six hundred thousand and a million users a day. It was as if every twenty-four hours, a group as big or bigger than the entire city of Boston filed into Facebook’s servers to set up new accounts. Roughly half of Facebook users visit the site every single day,D. Gage, “Facebook Claims 250 Million Users,” InformationWeek, July 16, 2009. with the majority spending fifty-five minutes or more getting their daily Facebook fix.“Facebook Facts and Figures (History and Statistics),” Website Monitoring Blog, March 17, 2010. And it seems that Mom really is on Facebook (Dad, too); users thirty-five years and older account for more than half of Facebook’s daily visitors and its fastest growing population.J. Hagel and J. S. Brown, “Life on the Edge: Learning from Facebook,” BusinessWeek, April 2, 2008; and D. Gage, “Facebook Claims 250 Million Users,” InformationWeek, July 16, 2009.

Then there’s what these users are doing on the site: Facebook isn’t just a collection of personal home pages and a place to declare your allegiance to your friends. The integrated set of Facebook services encroaches on a wide swath of established Internet businesses. Facebook has become the first-choice messaging and chat service for this generation. E-mail is for your professors, but Facebook is for friends. In photos, Google, Yahoo! and MySpace all spent millions to acquire photo sharing tools (Picasa, Flickr, and Photobucket, respectively). But Facebook is now the biggest photo-sharing site on the Web, taking in some three billion photos each month.“Facebook Facts and Figures (History and Statistics),” Website Monitoring Blog, March 17, 2010. And watch out, YouTube. Facebookers share eight million videos each month. YouTube will get you famous, but Facebook is a place most go to share clips you only want friends to see.F. Vogelstein, “Mark Zuckerberg: The Wired Interview,” Wired, June 29, 2009.

Facebook is a kingmaker, opinion catalyst, and traffic driver. While in the prior decade news stories would carry a notice saying, “Copyright, do not distribute without permission,” major news outlets today, including the New York Times, display Facebook icons alongside every copyrighted story, encouraging users to “share” the content on their profile pages via Facebook’s “Like” button, scattering it all over the Web. Like digital photos, video, and instant messaging, link sharing is Facebook’s sharp elbow to the competition. Suddenly, Facebook gets space on a page alongside Digg.com and Del.icio.us, even though those guys showed up first.

Facebook Office? Facebook rolled out the document collaboration and sharing service Docs.com in partnership with Microsoft. Facebook is also hard at work on its own e-mail system,H. Blodget, “Facebook’s Plan To Build a Real Email System and Attack Gmail Is Brilliant,” Business Insider, February 5, 2010. music service,J. Kincaid, “What Is This Mysterious Facebook Music App?” TechCrunch, February 2, 2010. and payments mechanism.R. Maher, “Facebook’s New Payment System Off to Great Start, Could Boost Revenue by $250 Million in 2010,” TBI Research, February 1, 2010. Look out, Gmail, Hotmail, Pandora, iTunes, PayPal, and Yahoo!—you may all be in Facebook’s path!

As for search, Facebook’s got designs on that, too. Google and Bing index some Facebook content, but since much of Facebook is private, accessible only among friends, this represents a massive blind spot for Google search. Sites that can’t be indexed by Google and other search engines are referred to as the dark WebInternet content that can’t be indexed by Google and other search engines.. While Facebook’s partnership with Microsoft currently offers Web search results through Bing.com, Facebook has announced its intention to offer its own search engine with real-time access to up-to-the-minute results from status updates, links, and other information made available to you by your friends. If Facebook can tie together standard Internet search with its dark Web content, this just might be enough for some to break the Google habit.

And Facebook is political—in big, regime-threatening ways. The site is considered such a powerful tool in the activist’s toolbox that China, Iran, and Syria are among nations that have, at times, attempted to block Facebook access within their borders. Egyptians have used the site to protest for democracy. Saudi women have used it to lobby for driving privileges. ABC News cosponsored U.S. presidential debates with Facebook. And Facebook cofounder Chris Hughes was even recruited by the Obama campaign to create my.barackobama.com, a social media site considered vital in the 2008 U.S. presidential victory.D. Talbot, “How Obama Really Did It,” Technology Review, September/October 2008; and E. McGirt, “How Chris Hughes Helped Launch Facebook and the Barack Obama Campaign,” Fast Company, March 17, 2009, http://www.fastcompany.com/magazine/134/boy-wonder.html.

So What’s It Take to Run This Thing?

The Facebook cloudA collection of resources available for access over the Internet. (the big group of connected servers that power the site) is scattered across multiple facilities, including server farms in San Francisco, Santa Clara, and northern Virginia.A. Zeichick, “How Facebook Works,” Technology Review, July/August 2008. The innards that make up the bulk of the system aren’t that different from what you’d find on a high-end commodity workstation. Standard hard drives and eight core Intel processors—just a whole lot of them lashed together through networking and software.

Much of what powers the site is open source software (OSS)Software that is free and whose code can be accessed and potentially modified by anyone.. A good portion of the code is in PHP (a scripting language particularly well-suited for Web site development), while the databases are in MySQL (a popular open source database). Facebook also developed Cassandra, a non-SQL database project for large-scale systems that the firm has since turned over to the open source Apache Software Foundation. The object cache that holds Facebook’s frequently accessed objects is in chip-based RAM instead of on slower hard drives and is managed via an open source product called Memcache.

Other code components are written in a variety of languages, including C++, Java, Python, and Ruby, with access between these components managed by a code layer the firm calls Thrift (developed at Facebook, which was also turned over to the Apache Software Foundation). Facebook also developed its own media serving solution, called Haystack. Haystack coughs up photos 50 percent faster than more expensive, proprietary solutions, and since it’s done in-house, it saves Facebook costs that other online outlets spend on third-party content delivery networks (CDN)Systems distributed throughout the Internet (or other network) that help to improve the delivery (and hence loading) speeds of Web pages and other media, typically by spreading access across multiple sites located closer to users. Akamai is the largest CDN, helping firms like CNN and MTV quickly deliver photos, video, and other media worldwide. like Akamai. Facebook receives some fifty million requests per second,S. Gaudin, “Facebook Rolls Out Storage System to Wrangle Massive Photo Stores,” Computerworld, April 1, 2009, http://www.computerworld.com/s/article/9130959/Facebook_rolls_out_storage_system_to_wrangle_massive_photo_stores. yet 95 percent of data queries can be served from a huge, distributed server cache that lives in over fifteen terabytes of RAM (objects like video and photos are stored on hard drives).A. Zeichick, “How Facebook Works,” Technology Review, July/August 2008.

Hot stuff (literally), but it’s not enough. The firm raised several hundred million dollars more in the months following the fall 2007 Microsoft deal, focused largely on expanding the firm’s server network to keep up with the crush of growth. The one hundred million dollars raised in May 2008 was “used entirely for servers.”S. Ante, “Facebook: Friends with Money,” BusinessWeek, May 9, 2008. Facebook will be buying them by the thousands for years to come. And it’ll pay a pretty penny to keep things humming. Estimates suggest the firm spends $1 million a month on electricity, another half million a month on telecommunications bandwidthTransmission rate, typically expressed as the number of bits per second that can be transmitted by a particular telecommunications mechanism., and at least fifteen million dollars a year in office and data center rental payments.A. Arrington, “Facebook Completes Rollout of Haystack to Stem Losses from Massive Photo Uploads,” TechCrunch, April 6, 2009.

Key Takeaways

  • Facebook’s position as the digital center of its members’ online social lives has allowed the firm to envelop related businesses such as photo and video sharing, messaging, bookmarking, and link sharing. Facebook has opportunities to expand into other areas as well.
  • Much of the site’s content is in the dark Web, unable to be indexed by Google or other search engines. Some suggest this may create an opportunity for Facebook to challenge Google in search.
  • Facebook can be a vital tool for organizers—presenting itself as both opportunity and threat to those in power, and an empowering medium for those seeking to bring about change.
  • Facebook’s growth requires a continued and massive infrastructure investment. The site is powered largely on commodity hardware, open source software, and proprietary code tailored to the specific needs of the service.

Questions and Exercises

  1. What is Facebook? How do people use the site? What do they “do” on Facebook?
  2. What markets has Facebook entered? What factors have allowed the firm to gain share in these markets at the expense of established firms? In what ways does it enjoy advantages that a traditional new entrant in such markets would not?
  3. What is the “dark Web” and why is it potentially an asset to Facebook? Why is Google threatened by Facebook’s dark Web? What firms might consider an investment in the firm, if it provided access to this asset? Do you think the dark Web is enough to draw users to a Facebook search product over Google? Why or why not?
  4. As Facebook grows, what kinds of investments continue to be necessary? What are the trends in these costs over time? Do you think Facebook should wait in making these investments? Why or why not?
  5. Investments in servers and other capital expenses typically must be depreciated over time. What does this imply about how the firm’s profitability is calculated?
  6. How have media attitudes toward their copyrighted content changed over the past decade? Why is Facebook a potentially significant partner for firms like the New York Times? What does the Times stand to gain by encouraging “sharing” its content? What do newspapers and others sites really mean when they encourage sites to “share?” What actually is being passed back and forth? Do you think this ultimately helps or undermines the Times and other newspaper and magazine sites? Why?

8.3 The Social Graph

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the concept of feeds, why users rebelled against Facebook feeds, and why users eventually embraced this feature.
  2. Recognize the two strategic resources that are most critical to Facebook’s competitive advantage and why Facebook was able to create these resources while MySpace has fallen short.
  3. Appreciate that while Facebook’s technology can be easily copied, barriers to sustain any new entrant are extraordinarily high, and the likelihood that a firm will win significant share from Facebook by doing the same thing is considerably remote.

At the heart of Facebook’s appeal is a concept Zuckerberg calls the social graphThe global mapping of users and organizations, and how they are connected., which refers to Facebook’s ability to collect, express, and leverage the connections between the site’s users, or as some describe it, “the global mapping of everyone and how they’re related.”A. Iskold, “Social Graph: Concepts and Issues,” ReadWriteWeb, September 12, 2007. Think of all the stuff that’s on Facebook as a node or endpoint that’s connected to other stuff. You’re connected to other users (your friends), photos about you are tagged, comments you’ve posted carry your name, you’re a member of groups, you’re connected to applications you’ve installed—Facebook links them all.A. Zeichick, “How Facebook Works,” Technology Review, July/August 2008.

While MySpace and Facebook are often mentioned in the same sentence, from their founding these sites were conceived differently. It goes beyond the fact that Facebook, with its neat, ordered user profiles, looks like a planned community compared to the garish, Vegas-like free-for-all of MySpace. MySpace was founded by musicians seeking to reach out to unknown users and make them fans. It’s no wonder the firm, with its proximity to Los Angeles and ownership by News Corporation, is viewed as more of a media company. It has cut deals to run network television shows on its site, and has even established a record label. It’s also important to note that from the start anyone could create a MySpace identity, and this open nature meant that you couldn’t always trust what you saw. Rife with bogus profiles, even News Corporation’s Rupert Murdoch has had to contend with the dozens of bogus Ruperts who have popped up on the service!L. Petrecca, “If You See These CEOs on MySpace…,” USA Today, September 25, 2006.

Facebook, however, was established in the relatively safe cocoon of American undergraduate life, and was conceived as a place where you could reinforce contacts among those who, for the most part, you already knew. The site was one of the first social networks where users actually identified themselves using their real names. If you wanted to establish that you worked for a certain firm or were a student of a particular university, you had to verify that you were legitimate via an e-mail address issued by that organization. It was this “realness” that became Facebook’s distinguishing feature—bringing along with it a degree of safety and comfort that enabled Facebook to become a true social utility and build out a solid social graph consisting of verified relationships. Since “friending” (which is a link between nodes in the social graph) required both users to approve the relationship, the network fostered an incredible amount of trust. Today, many Facebook users post their cell phone numbers and their birthdays, offer personal photos, and otherwise share information they’d never do outside their circle of friends. Because of trust, Facebook’s social graph is stronger than MySpace’s.

There is also a strong network effectAlso known as Metcalfe’s Law, or network externalities. When the value of a product or service increases as its number of users expands. to Facebook (see Chapter 6 “Understanding Network Effects”). People are attracted to the service because others they care about are more likely to be there than anywhere else online. Without the network effect Facebook wouldn’t exist. And it’s because of the network effect that another smart kid in a dorm can’t rip off Zuckerberg in any market where Facebook is the biggest fish. Even an exact copy of Facebook would be a virtual ghost town with no social graph (see Note 8.23 “It’s Not the Technology” below).

The switching costsThe cost a consumer incurs when moving from one product to another. It can involve actual money spent (e.g., buying a new product) as well as investments in time, any data loss, and so forth. for Facebook are also extremely powerful. A move to another service means recreating your entire social graph. The more time you spend on the service, the more you’ve invested in your graph and the less likely you are to move to a rival.

It’s Not the Technology

Does your firm have Facebook envy? KickApps, an eighty-person start-up in Manhattan, will give you the technology to power your own social network. All KickApps wants is a cut of the ads placed around your content. In its first two years, the site has provided the infrastructure for twenty thousand “mini Facebooks,” registering three hundred million page views a month.B. Urstadt, “The Business of Social Networks,” Technology Review, July/August 2008. NPR, ABC, AutoByTel, Harley-Davidson, and Kraft all use the service (social networks for Cheez Whiz?).

There’s also Ning, which has enabled users to create over 2.3 million mini networks organized on all sorts of topics as diverse as church groups, radio personalities, vegans, diabetes sufferers networks limited to just family members.

Or how about the offering from Agriya Infoway, based in Chennai, India? The firm will sell you Kootali, a software package that lets developers replicate Facebook’s design and features, complete with friend networks, photos, and mini-feeds. They haven’t stolen any code, but they have copied the company’s look and feel. Those with Zuckerberg ambitions can shell out the four hundred bucks for Kootali. Sites with names like Faceclub.com and Umicity.com have done just that—and gone nowhere.

Mini networks that extend the conversation (NPR) or make it easier to find other rabidly loyal product fans (Harley-Davidson) may hold a niche for some firms. And Ning is a neat way for specialized groups to quickly form in a secure environment that’s all their own (it’s just us, no “creepy friends” from the other networks). While every market has a place for its niches, none of these will grow to compete with the dominant social networks. The value isn’t in the technology; it’s in what the technology has created over time. For Facebook, it’s a huge user base that (for now at least) is not going anywhere else.

Key Takeaways

  • The social graph expresses the connections between individuals and organizations.
  • Trust created through user verification and friend approval requiring both parties to consent encouraged Facebook users to share more and helped the firm establish a stronger social graph than MySpace or other social networking rivals.
  • Facebook’s key resources for competitive advantage are network effects and switching costs. These resources make it extremely difficult for copycat firms to steal market share from Facebook.

Questions and Exercises

  1. Which is bigger, Facebook or MySpace? How are these firms different? Why would a person or organization be attracted to one service over another?
  2. What is the social graph? Why is Facebook’s social graph considered to be stronger than the social graph available to MySpace users?
  3. In terms of features and utility, how are Facebook and MySpace similar? How are they different? Why would a user choose to go to one site instead of another? Are you a member of either of these sites? Both? Why? Do you feel that they are respectively pursuing lucrative markets? Why or why not? If given the opportunity, would you invest in either firm? Why or why not?
  4. If you were a marketer, which firm would you target for an online advertising campaign—Facebook or MySpace? Why?
  5. Does Facebook have to worry about copycat firms from the United States? In overseas markets? Why or why not? If Facebook has a source (or sources) of competitive advantage, explain these. If it has no advantage, discuss why.

8.4 Facebook Feeds—Ebola for Data Flows

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the concept of feeds, why users rebelled, and why users eventually embraced this feature.
  2. Recognize the role of feeds in viral promotions, catalyzing innovation, and supporting rapid organizing.

While the authenticity and trust offered by Facebook was critical, offering News Feeds concentrated and released value from the social graph. With feeds, each time a user performs an activity in Facebook—makes a friend, uploads a picture, joins a group—the feed blasts this information to all of your friends in a reverse chronological list that shows up right when they next log on. An individual user’s activities are also listed within a mini feed that shows up on their profile. Get a new job, move to a new city, read a great article, have a pithy quote—post it to Facebook—the feed picks it up, and the world of your Facebook friends will get an update.

Feeds are perhaps the linchpin of Facebook’s ability to strengthen and deliver user value from the social graph, but for a brief period of time it looked like feeds would kill the company. News Feeds were launched on September 5, 2006, just as many of the nation’s undergrads were arriving on campus. Feeds reflecting any Facebook activity (including changes to the relationship status) became a sort of gossip page splashed right when your friends logged in. To many, feeds were first seen as a viral blast of digital nosiness—a release of information they hadn’t consented to distribute widely.

And in a remarkable irony, user disgust over the News Feed ambush offered a whip-crack demonstration of the power and speed of the feed virus. Protest groups formed, and every student who, for example, joined a group named Students Against Facebook News Feed, had this fact blasted to their friends (along with a quick link where friends, too, could click to join the group). Hundreds of thousands of users mobilized against the firm in just twenty-four hours. It looked like Zuckerberg’s creation had turned on him, Frankenstein style.

The first official Facebook blog post on the controversy came off as a bit condescending (never a good tone to use when your customers feel that you’ve wronged them). “Calm down. Breathe. We hear you,” wrote Zuckerberg on the evening of September 5. The next post, three days after the News Feed launch, was much more contrite (“We really messed this one up,” he wrote). In the 484-word open letter, Zuckerberg apologized for the surprise, explaining how users could opt out of feeds. The tactic worked, and the controversy blew over.F. Vogelstein, “How Mark Zuckerberg Turned Facebook into the Web’s Hottest Platform,” Wired, September 6, 2007. The ability to stop personal information from flowing into the feed stream was just enough to stifle critics, and as it turns out, a lot of people really liked the feeds and found them useful. It soon became clear that if you wanted to use the Web to keep track of your social life and contacts, Facebook was the place to be. Not only did feeds not push users away, by the start of the next semester subscribers had nearly doubled!

Key Takeaways

  • Facebook feeds foster the viral spread of information and activity.
  • Feeds were initially unwanted by many Facebook users. Feeds themselves helped fuel online protests against the feed feature.
  • Today feeds are considered one of the most vital, value-adding features to Facebook and other social networking sites.
  • Users often misperceive technology and have difficulty in recognizing an effort’s value (as well as its risks). They have every right to be concerned and protective of their privacy. It is the responsibility of firms to engage users on new initiatives and to protect user privacy. Failure to do so risks backlash.

Questions and Exercises

  1. What is the “linchpin” of Facebook’s ability to strengthen and deliver user-value from the social graph?
  2. How did users first react to feeds? What could Facebook have done to better manage the launch?
  3. How do you feel about Facebook feeds? Have you ever been disturbed by information about you or someone else that has appeared in the feed? Did this prompt action? Why or why not?
  4. Visit Facebook and experiment with privacy settings. What kinds of control do you have over feeds and data sharing? Is this enough to set your mind at ease? Did you know these settings existed before being prompted to investigate features?
  5. What other Web sites are leveraging features that mimic Facebook feeds? Do you think these efforts are successful or not? Why?

8.5 Facebook as a Platform

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand how Facebook created a platform and the potential value this offers the firm.
  2. Recognize that running a platform also presents a host of challenges to the platform operator.

In May 2007, Facebook followed News Feeds with another initiative that set it head and shoulders above its competition. At the firm’s first f8 (pronounced “fate”) Developers Conference, Mark Zuckerberg stood on stage and announced that he was opening up the screen real estate on Facebook to other application developers. Facebook published a set of application programming interfaces (APIs)Programming hooks, or guidelines, published by firms that tell other programs how to get a service to perform a task such as send or receive data. For example, Amazon.com provides APIs to let developers write their own applications and Websites that can send the firm orders. that specified how programs could be written to run within and interact with Facebook. Now any programmer could write an application that would run inside a user’s profile. Geeks of the world, Facebook’s user base could be yours! Just write something good.

Developers could charge for their wares, offer them for free, and even run ads. And Facebook let developers keep what they made (Facebook does revenue share with app vendors for some services, such as the Facebook Credits payment service, mentioned later). This was a key distinction; MySpace initially restricted developer revenue on the few products designed to run on their site, at times even blocking some applications. The choice was clear, and developers flocked to Facebook.

To promote the new apps, Facebook would run an Applications area on the site where users could browse offerings. Even better, News Feed was a viral injection that spread the word each time an application was installed. Your best friend just put up a slide show app? Maybe you’ll check it out, too. The predictions of $1 billion in social network ad spending were geek catnip, and legions of programmers came calling. Apps could be cobbled together on the quick, feeds made them spread like wildfire, and the early movers offered adoption rates never before seen by small groups of software developers. People began speaking of the Facebook Economy. Facebook was considered a platform. Some compared it to the next Windows, Zuckerberg the next Gates (hey, they both dropped out of Harvard, right?).

And each application potentially added more value and features to the site without Facebook lifting a finger. The initial event launched with sixty-five developer partners and eighty-five applications. There were some missteps along the way. Some applications were accused of spamming friends with invites to install them. There were also security concerns and apps that violated the intellectual property of other firms (see the “Scrabulous” sidebar below), but Facebook worked to quickly remove errant apps, improve the system, and encourage developers. Just one year in, Facebook had marshaled the efforts of some four hundred thousand developers and entrepreneurs, twenty-four thousand applications had been built for the platform, 140 new apps were being added each day, and 95 percent of Facebook members had installed at least one Facebook application. As Sarah Lacy, author of Once You’re Lucky, Twice You’re Good, put it, “with one masterstroke, Zuck had mobilized all of Silicon Valley to innovate for him.”

With feeds to spread the word, Facebook was starting to look like the first place to go to launch an online innovation. Skip the Web, bring it to Zuckerberg’s site first. Consider iLike: within the first three months, the firm saw installs of its Facebook app explode to seven million, more than doubling the number of users the firm was able to attract through the Web site it introduced the previous year. ILike became so cool that by September, platinum rocker KT Tunstall was debuting tracks through the Facebook service. A programmer named Mark Pincus wrote a Texas hold ’em game at his kitchen table.J. Guynn, “A Software Industry @ Facebook,” Los Angeles Times, September 10, 2007. Today his social gaming firm, Zynga, is a powerhouse—a profitable firm with over three dozen apps, over 230 million users,D. MacMillan, “Zynga Enlarges Its War Chest,” BusinessWeek, December 17, 2009. and more than $600 million in annual revenue.M. Learmonth and A. Klaasen, “Facebook Apps Will Make More Money Than Facebook in 2009,” Silicon Alley Insider, May 18, 2009. Some of Zynga’s revenues come from apps that run on MySpace or other networks, too. Also see N. Carolson, “The Profitable, $100 Million-a-Year Startup You’ve Never Heard Of,” Business Insider, July 27, 2009; and N. Carlson and K. Angelova, “Chart of the Day: FarmVille-Maker Zynga’s Revenues Reach $600 Million, Fueled by Social Obligations,” April 26, 2010. Zynga games include MafiaWars, Vampires, and the wildly successful FarmVille, which boasts some twenty times the number of actual farms in the United States. App firm Slide (started by PayPal cofounder Max Levchin) scored investments from Legg Mason, and Fidelity pegged the firm’s value at $500 million.J. Hempel and M. Copeland, “Are These Widgets Worth Half a Billion?” Fortune, March 25, 2008. Playfish, the U.K. social gaming firm behind the Facebook hits Pet Society and Restaurant City, was snapped up by Electronic Arts for $300 million with another $100 million due if the unit hits performance targets. Lee Lorenzen, founder of Altura Ventures, an investment firm exclusively targeting firms creating Facebook apps, said, “Facebook is God’s gift to developers. Never has the path from a good idea to millions of users been shorter.”J. Guynn, “A Software Industry @ Facebook,” Los Angeles Times, September 10, 2007.

I Majored in Facebook

Once Facebook became a platform, Stanford professor BJ Fogg thought it would be a great environment for a programming class. In ten weeks his seventy-three students built a series of applications that collectively received over sixteen million installs. By the final week of class, several applications developed by students, including KissMe, Send Hotness, and Perfect Match, had received millions of users, and class apps collectively generated more than half a million dollars in ad revenue. At least three companies were formed from the course.

But legitimate questions remain. Are Facebook apps really a big deal? Just how important will apps be to adding sustained value within Facebook? And how will firms leverage the Facebook framework to extract their own value? A chart from FlowingData showed the top category, Just for Fun, was larger than the next four categories combined. That suggests that a lot of applications are faddish time wasters. Yes, there is experimentation beyond virtual Zombie Bites. Visa has created a small business network on Facebook (Facebook had some eighty thousand small businesses online at the time of Visa’s launch). Educational software firm Blackboard offered an application that will post data to Facebook pages as soon as there are updates to someone’s Blackboard account (new courses, whether assignments or grades have been posted, etc.). We’re still a long way from Facebook as a Windows rival, but the platform helped push Facebook to number one, and it continues to deliver quirky fun (and then some) supplied by thousands of developers off its payroll.

Scrabulous

Rajat and Jayant Agarwalla, two brothers in Kolkata, India, who run a modest software development company, decided to write a Scrabble clone as a Facebook application. The app, named Scrabulous, was social—users could invite friends to play, or they could search for new players looking for an opponent. Their application was a smash, snagging three million registered users and seven hundred thousand players a day after just a few months. Scrabulous was featured in PC World’s 100 best products of 2008, received coverage in the New York Times, Newsweek, and Wired, and was pulling in about twenty-five thousand dollars a month from online advertising. Way to go, little guys!H. Timmons, “Online Scrabble Craze Leaves Game Sellers at Loss for Words,” New York Times, March 2, 2008.

There is only one problem: the Agarwalla brothers didn’t have the legal rights to Scrabble, and it was apparent to anyone that from the name to the tiles to the scoring—this was a direct rip-off of the well-known board game. Hasbro owns the copyright to Scrabble in the United States and Canada; Mattel owns it everywhere else. Thousands of fans joined Facebook groups with names like “Save Scrabulous” and “Please God, I Have So Little: Don’t Take Scrabulous, Too.” Users in some protest groups pledged never to buy Hasbro games if Scrabulous was stopped. Even if the firms wanted to succumb to pressure and let the Agarwalla brothers continue, they couldn’t. Both Electronic Arts and RealNetworks have contracted with the firms to create online versions of the game.

While the Facebook Scrabulous app is long gone, the tale shows just one of the challenges of creating a platform. In addition to copyright violations, app makers have crafted apps that annoy, raise privacy and security concerns, purvey pornography, or otherwise step over the boundaries of good taste. Firms from Facebook to Apple (through its iTunes Store) have struggled to find the right mix of monitoring, protection, and approval while avoiding cries of censorship.

Key Takeaways

  • Facebook’s platform allows the firm to further leverage the network effect. Developers creating applications create complementary benefits that have the potential to add value to Facebook beyond what the firm itself provides to its users.
  • There is no revenue-sharing mandate among platform partners—whatever an application makes can be kept by its developers (although Facebook does provide some services via revenue sharing, such as Facebook Credits).
  • Most Facebook applications are focused on entertainment. The true, durable, long-term value of Facebook’s platform remains to be seen.
  • Despite this, some estimates claim Facebook platform developers earned more than Facebook itself in 2009.
  • Running a platform can be challenging. Copyright, security, appropriateness, free speech tensions, efforts that tarnish platform operator brands, privacy, and the potential for competition with partners, all can make platform management more complex than simply creating a set of standards and releasing this to the public.

Questions and Exercises

  1. Why did more developers prefer to write apps for Facebook than for MySpace?
  2. What competitive asset does the application platform initiative help Facebook strengthen? For example, how do apps make Facebook stronger when compared to rivals?
  3. What’s Scrabulous? Did the developers make money? What happened to the firm and why?
  4. Have you used Facebook apps? Which are your favorites? What makes them successful?
  5. Leverage your experience or conduct additional research—are there developers who you feel have abused the Facebook app network? Why? What is Facebook’s responsibility (if any) to control such abuse?
  6. How do most app developers make money? Have you ever helped a Facebook app developer earn money? How or why not?
  7. How do Facebook app revenue opportunities differ from those leveraged by a large portion of iTunes Store apps?

8.6 Advertising and Social Networks: A Work in Progress

Learning Objectives

After studying this section you should be able to do the following:

  1. Describe the differences in the Facebook and Google ad models.
  2. Explain the Hunt versus Hike metaphor, contrast the relative success of ad performance on search compared to social networks, and understand the factors behind the latter’s struggles.
  3. Recognize how firms are leveraging social networks for brand and product engagement, be able to provide examples of successful efforts, and give reasons why such engagement is difficult to achieve.

If Facebook is going to continue to give away its services for free, it needs to make money somehow. Right now the bulk of revenue comes from advertising. Fortunately for the firm, online advertising is hot. For years, online advertising has been the only major media category that has seen an increase in spending (see Chapter 14 “Google: Search, Online Advertising, and Beyond”). Firms spend more advertising online than they do on radio and magazine ads, and the Internet will soon beat out spending on cable TV.M. Sweeney, “Internet Ad Spending Will Overtake Television in 2009,” Guardian, May 19, 2008; and T. Wayne, “A Milestone for Internet Ad Revenue,” New York Times, April 25, 2010. But not all Internet advertising is created equal. And there are signs that social networking sites are struggling to find the right ad model.

Google founder Sergey Brin sums up this frustration, saying, “I don’t think we have the killer best way to advertise and monetize social networks yet,” that social networking ad inventory as a whole was proving problematic and that the “monetization work we were doing [in social media] didn’t pan out as well as we had hoped.”“Everywhere and Nowhere,” Economist, March 19, 2008. When Google ad partner Fox Interactive Media (the News Corporation division that contains MySpace) announced that revenue would fall $100 million short of projections, News Corporation’s stock tumbled 5 percent, analysts downgraded the company, and the firm’s chief revenue officer was dismissed.B. Stelter, “MySpace Might Have Friends, but It Wants Ad Money,” New York Times, June 16, 2008.

Why aren’t social networks having the success of Google and other sites? Problems advertising on these sites include content adjacencyConcern that an advertisement will run near offensive material, embarrassing an advertiser and/or degrading their products or brands., and user attention. The content adjacency problem refers to concern over where a firm’s advertisements will run. Consider all of the questionable titles in social networking news groups. Do advertisers really want their ads running alongside conversations that are racy, offensive, illegal, or that may even mock their products? This potential juxtaposition is a major problem with any site offering ads adjacent to free-form social media. Summing up industry wariness, one P&G manager said, “What in heaven’s name made you think you could monetize the real estate in which somebody is breaking up with their girlfriend?”B. Stone, “Facebook Aims to Extends Its Reach across Web,” New York Times, December 1, 2008. An IDC report suggests that it’s because of content adjacency that “brand advertisers largely consider user-generated content as low-quality, brand-unsafe inventory” for running ads.R. Stross, “Advertisers Face Hurdles on Social Networking Sites,” New York Times, December 14, 2008.

Now let’s look at the user attention problem.

Attention Challenges: The Hunt Versus The Hike

In terms of revenue model, Facebook is radically different from Google and the hot-growth category of search advertising. Users of Google and other search sites are on a hunt—a task-oriented expedition to collect information that will drive a specific action. Search users want to learn something, buy something, research a problem, or get a question answered. To the extent that the hunt overlaps with ads, it works. Just searched on a medical term? Google will show you an ad from a drug company. Looking for a toy? You’ll see Google ads from eBay sellers and other online shops. Type in a vacation destination and you get a long list of ads from travel providers aggressively courting your spending. Even better, Google only charges text advertisers when a user clicks through. No clicks? The ad runs at no cost to the firm. From a return on investment perspective, this is extraordinarily efficient. How often do users click on Google ads? Enough for this to be the single most profitable activity among any Internet firm. In 2009, Google revenue totaled nearly $24 billion. Profits exceeded $6.5 billion, almost all of this from pay-per-click ads (see Chapter 14 “Google: Search, Online Advertising, and Beyond” for more details).

While users go to Google to hunt, they go to Facebook as if they were going on a hike—they have a rough idea of what they’ll encounter, but they’re there to explore and look around, enjoy the sights (or site). They’ve usually allocated time for fun and they don’t want to leave the terrain when they’re having conversations, looking at photos or videos, and checking out updates from friends.

These usage patterns are reflected in click-through rates. Google users click on ads around 2 percent of the time (and at a much higher rate when searching for product information). At Facebook, click-throughs are about 0.04 percent.B. Urstadt, “The Business of Social Networks,” Technology Review, July/August 2008. Rates quoted in this piece seem high, but a large discrepancy between site rates holds across reported data.

Most banner ads don’t charge per click but rather CPMCost per thousand impressions (the M representing the roman numeral for one thousand). (cost per thousand) impressionsEach time an ad is served to a user for viewing. (each time an ad appears on someone’s screen). But Facebook banner ads performed so poorly that the firm pulled them in early 2010.C. McCarthy, “More Social, Please: Facebook Nixes Banner Ads,” CNET, February 5, 2010. Lookery, a one-time ad network that bought ad space on Facebook in bulk, had been reselling inventory at a CPM of 7.5 cents (note that Facebook does offer advertisers pay-per-click as well as impression-based, or CPM, options). Even Facebook ads with a bit of targeting weren’t garnering much (Facebook’s Social Ads, which allow advertisers to target users according to location and age, have a floor price of fifteen cents CPM).B. Urstadt, “The Business of Social Networks,” Technology Review, July/August 2008; J. Hempel, “Finding Cracks in Facebook,” Fortune, May 13, 2008; and E. Schonfeld, “Are Facebook Ads Going to Zero? Lookery Lowers Its Guarantee to 7.5-cent CMPs,” TechCrunch, July 22, 2008. Other social networks also suffered. In 2008, MySpace lowered its banner ad rate from $3.25 CPM to less than two dollars. By contrast, information and news-oriented sites do much better, particularly if these sites draw in a valuable and highly targeted audience. The social networking blog Mashable has CPM rates ranging between seven and thirty-three dollars. Technology Review magazine boasts a CPM of seventy dollars. TechTarget, a Web publisher focusing on technology professionals, has been able to command CPM rates of one hundred dollars and above (an ad inventory that valuable helped the firm go public in 2007).

Getting Creative with Promotions: Does It Work?

Facebook and other social networks are still learning what works. Ad inventory displayed on high-traffic home pages have garnered big bucks for firms like Yahoo! With Facebook offering advertisers greater audience reach than most network television programs, there’s little reason to suggest that chunks of this business won’t eventually flow to the social networks. But even more interesting is how Facebook and widget sites have begun to experiment with relatively new forms of advertising. Many feel that Facebook has a unique opportunity to get consumers to engage with their brand, and some initial experiments point where this may be heading.

Many firms have been leveraging so-called engagement adsPromotion technique popular with social media that attempts to get consumers to interact with an ad, then shares that action with friends. by making their products part of the Facebook fun. Using an engagement ad, a firm can set up a promotion where a user can do things such as “Like” or become a fan of a brand, RSVP to an event and invite others, watch and comment on a video and see what your friends have to say, send a “virtual gift” with a personal message, or answer a question in a poll. The viral nature of Facebook allows actions to flow back into the news feed and spread among friends.

COO Sheryl Sandberg discussed Ben & Jerry’s promotion for the ice cream chain’s free cone day event. To promote the upcoming event, Ben & Jerry’s initially contracted to make two hundred and fifty thousand “gift cones” available to Facebook users; they could click on little icons that would gift a cone icon to a friend, and that would show up in their profile. Within a couple of hours, customers had sent all two hundred and fifty thousand virtual cones. Delighted, Ben & Jerry’s bought another two hundred and fifty thousand cones. Within eleven hours, half a million people had sent cones, many making plans with Facebook friends to attend the real free cone day. The day of the Facebook promotion, Ben & Jerry’s Web site registered fifty-three million impressions, as users searched for store locations and wrote about their favorite flavors.Q. Hardy, “Facebook Thinks Outside Boxes,” Forbes, May 28, 2008. The campaign dovetailed with everything Facebook was good at: it was viral, generating enthusiasm for a promotional event and even prompting scheduling.

In other promotions, Honda gave away three quarters of a million hearts during a Valentine’s Day promo,S. Sandberg, “Sheryl Sandberg on Facebook’s Future,” BusinessWeek, April 8, 2009. and the Dr. Pepper Snapple Group offered two hundred and fifty thousand virtual Sunkist sodas, which earned the firm one hundred thirty million brand impressions in twenty-two hours. Says Sunkist’s brand manager, “A Super Bowl ad, if you compare it, would have generated somewhere between six to seven million.”E. Wong, “Ben & Jerry’s, Sunkist, Indy Jones Unwrap Facebook’s ‘Gift of Gab,’” Brandweek, June 1, 2008.

Facebook, Help Get Me a Job!

The papers are filled with stories about employers scouring Facebook for dirt on potential hires. But one creative job seeker turned the tables and used Facebook to make it easier for firms to find him. Recent MBA graduate Eric Barker, a talented former screenwriter with experience in the film and gaming industry, bought ads promoting himself on Facebook, setting them up to run only on the screens of users identified as coming from firms he’d like to work for. In this way, someone Facebook identified as being from Microsoft would see an ad from Eric declaring “I Want to Be at Microsoft” along with an offer to click and learn more. The cost to run the ads was usually less than $5 a day. Said Barker, “I could control my bid price and set a cap on my daily spend. Starbucks put a bigger dent in my wallet than promoting myself online.” The ads got tens of thousands of impressions, hundreds of clicks, and dozens of people called offering assistance. Today, Eric Barker is gainfully employed at a “dream job” in the video game industry.Eric is a former student of mine. His story has been covered by many publications, including J. Zappe, “MBA Grad Seeks Job with Microsoft; Posts Ad on Facebook,” ERE.net, May 27, 2009; G. Sentementes, “‘Hire Me’ Nation: Using the Web & Social Media to Get a Job,” Baltimore Sun, July 15, 2009; and E. Liebert, Facebook Fairytales (New York: Skyhorse, 2010).

Figure 8.1

Eric Barker used Facebook to advertise himself to prospective employers.

Of course, even with this business, Facebook may find that it competes with widget makers. Unlike Apple’s App Store (where much of developer-earned revenue comes from selling apps), the vast majority of Facebook apps are free and supported by ads. That means Facebook and its app providers are both running at a finite pot of advertising dollars. Slide’s Facebook apps have attracted top-tier advertisers, such as Coke and Paramount Pictures—a group Facebook regularly courts as well. By some estimates, in 2009, Facebook app developers took in well over half a billion dollars—exceeding Facebook’s own haul.M. Learmonth and A. Klaasen, “Facebook Apps Will Make More Money Than Facebook in 2009,” Silicon Alley Insider, May 18, 2009. And there’s controversy. Zynga was skewered in the press when some of its partners were accused of scamming users into signing up for subscriptions or installing unwanted software in exchange for game credits (Zynga has since taken steps to screen partners and improve transparency).M. Arrington, “Zynga Takes Steps to Remove Scams from Games,” TechCrunch, November 2, 2009.

While these efforts might be innovative, are they even effective? Some of these programs are considered successes; others, not so much. Jupiter Research surveyed marketers trying to create a viral impact online and found that only about 15 percent of these efforts actually caught on with consumers.M. Cowan, “Marketers Struggle to Get Social,” Reuters, June 19, 2008, http://www.reuters.com/news/video?videoId=84894. While the Ben & Jerry’s gift cones were used up quickly, a visit to Facebook in the weeks after this campaign saw CareerBuilder, Wide Eye Caffeinated Spirits, and Coors Light icons lingering days after their first appearance. Brands seeking to deploy their own applications in Facebook have also struggled. New Media Age reported that applications rolled out by top brands such as MTV, Warner Bros., and Woolworths were found to have as little as five daily users. Congestion may be setting in for all but the most innovative applications, as standing out in a crowd of over 550,000 applications becomes increasingly difficult.Facebook Press Room, Statistics, April 29, 2010, http://www.facebook.com/press/info.php?statistics.

Consumer products giant P&G has been relentlessly experimenting with leveraging social networks for brand engagement, but the results show what a tough slog this can be. The firm did garner fourteen thousand Facebook “fans” for its Crest Whitestrips product, but those fans were earned while giving away free movie tickets and other promos. The New York Times quipped that with those kinds of incentives, “a hemorrhoid cream” could have attracted a similar group of “fans.” When the giveaways stopped, thousands promptly “unfanned” Whitestrips. Results for Procter & Gamble’s “2X Ultra Tide” fan page were also pretty grim. P&G tried offbeat appeals for customer-brand bonding, including asking Facebookers to post “their favorite places to enjoy stain-making moments.” But a check eleven months after launch had garnered just eighteen submissions, two from P&G, two from staffers at spoof news site The Onion, and a bunch of short posts such as “Tidealicious!”R. Stross, “Advertisers Face Hurdles on Social Networking Sites,” New York Times, December 14, 2008.

Efforts around engagement opportunities like events (Ben & Jerry’s) or products consumers are anxious to identify themselves with (a band or a movie) may have more success than trying to promote consumer goods that otherwise offer little allegiance, but efforts are so new that metrics are scarce, impact is tough to gauge, and best practices are still unclear.

Facebook Engagement Ads

http://www.facebook.com/video/video.php?v=629649849493

Key Takeaways

  • Content adjacency and user attention make social networking ads less attractive than search and professionally produced content sites.
  • Google enjoys significantly higher click-through rates than Facebook.
  • Display ads are often charged based on impression. Social networks also offer lower CPM rates than many other, more targeted Web sites.
  • Social networking has been difficult to monetize, as users are online to engage friends, not to hunt for products or be drawn away by clicks.
  • Many firms have begun to experiment with engagement ads. While there have been some successes, engagement campaigns often haven’t yielded significant results.

Questions and Exercises

  1. How are most display ads billed? What acronym is used to describe pricing of most display ads?
  2. How are most text ads billed?
  3. Contrast Facebook and Google click-through rates. Contrast Facebook CPMs with CPMs at professional content sites. Why the discrepancy?
  4. What is the content adjacency problem? Search for examples of firms that have experienced embracement due to content adjacency—describe them, why they occurred, and if site operators could have done something to reduce the likelihood these issues could have occurred.
  5. What kinds of Web sites are most susceptible to content adjacency? Are news sites? Why or why not? What sorts of technical features might act as breeding grounds for content adjacency problems?
  6. If a firm removed user content because it was offensive to an advertiser, what kinds of problems might this create? When (if ever) should a firm remove or take down user content?
  7. How are firms attempting to leverage social networks for brand and product engagement?
  8. What are the challenges that social networking sites face when trying to woo advertisers?
  9. Describe an innovative marketing campaign that has leveraged Facebook or other social networking sites. What factors made this campaign work? Are all firms likely to have this sort of success? Why or why not?
  10. Have advertisers ever targeted you when displaying ads on Facebook? How were you targeted? What did you think of the effort?

8.7 Privacy Peril: Beacon and the TOS Debacle

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the difference between opt-in and opt-out efforts.
  2. Recognize how user issues and procedural implementation can derail even well-intentioned information systems efforts.
  3. Recognize the risks in being a pioneer associated with new media efforts, and understand how missteps led to Facebook and its partners being embarrassed (and in some cases sued) by Beacon’s design and rollout issues.

Conventional advertising may grow into a great business for Facebook, but the firm was clearly sitting on something that was unconventional compared to prior generations of Web services. Could the energy and virulent nature of social networks be harnessed to offer truly useful, consumer information to its users? Word of mouth is considered the most persuasive (and valuable) form of marketing,V. Kumar, J. Andrew Petersen, and Robert Leone, “How Valuable Is Word of Mouth?” Harvard Business Review 85, no. 10 (October 2007): 139–46. and Facebook was a giant word of mouth machine. What if the firm worked with vendors and grabbed consumer activity at the point of purchase to put into the News Feed and post to a user’s profile? If you rented a video, bought a cool product, or dropped something in your wish list, your buddies could get a heads-up and they might ask you about it. The person being asked feels like an expert, the person with the question gets a frank opinion, and the vendor providing the data just might get another sale. It looked like a home run.

This effort, named Beacon, was announced in November 2007. Some forty e-commerce sites signed up, including Blockbuster, Fandango, eBay, Travelocity, Zappos, and the New York Times. Zuckerberg was so confident of the effort that he stood before a group of Madison Avenue ad executives and declared that Beacon would represent a “once-in-a-hundred-years” fundamental change in the way media works.

Like News Feeds, user reaction was swift and brutal. The commercial activity of Facebook users began showing up without their consent. The biggest problem with Beacon was that it was “opt-out” instead of “opt-in.” Facebook (and its partners) assumed users would agree to sharing data in their feeds. A pop-up box did appear briefly on most sites supporting Beacon, but it disappeared after a few seconds.E. Nakashima, “Feeling Betrayed, Facebook Users Force Site to Honor Their Privacy,” Washington Post, November 30, 2007. Many users, blind to these sorts of alerts, either clicked through or ignored the warnings. And well…there are some purchases you might not want to broadcast to the world.

“Facebook Ruins Christmas for Everyone!” screamed one headline from MSNBC.com. Another from U.S. News and World Report read “How Facebook Stole Christmas.” The Washington Post ran the story of Sean Lane, a twenty-eight-year-old tech support worker from Waltham, Massachusetts, who got a message from his wife just two hours after he bought a ring on Overstock.com. “Who is this ring for?” she wanted to know. Facebook had not only posted a feed that her husband had bought the ring, but also that he got it for a 51 percent discount! Overstock quickly announced that it was halting participation in Beacon until Facebook changed its practice to opt in.E. Nakashima, “Feeling Betrayed, Facebook Users Force Site to Honor Their Privacy,” Washington Post, November 30, 2007.

MoveOn.org started a Facebook group and online petition protesting Beacon. The Center for Digital Democracy and the U.S. Public Interest Research Group asked the Federal Trade Commission to investigate Facebook’s advertising programs. And a Dallas woman sued Blockbuster for violating the Video Privacy Protection Act (a 1998 U.S. law prohibiting unauthorized access to video store rental records).

To Facebook’s credit, the firm acted swiftly. Beacon was switched to an opt-in system, where user consent must be given before partner data is sent to the feed. Zuckerberg would later say regarding Beacon: “We’ve made a lot of mistakes building this feature, but we’ve made even more with how we’ve handled them. We simply did a bad job with this release, and I apologize for it.”C. McCarthy, “Facebook’s Zuckerberg: ‘We Simply Did a Bad Job’ Handling Beacon,” CNET, December 5, 2007. Beacon was eventually shut down and $9.5 million was donated to various privacy groups as part of its legal settlement.J. Brodkin, “Facebook Shuts Down Beacon Program, Donates $9.5 Million to Settle Lawsuit,” NetworkWorld, December 8, 2009. Despite the Beacon fiasco, new users continued to flock to the site, and loyal users stuck with Zuck. Perhaps a bigger problem was that many of those forty A-list e-commerce sites that took a gamble with Facebook now had their names associated with a privacy screw-up that made headlines worldwide. A manager so burned isn’t likely to sign up first for the next round of experimentation.

From the Prada example in Chapter 3 “Zara: Fast Fashion from Savvy Systems” we learned that savvy managers look beyond technology and consider complete information systems—not just the hardware and software of technology but also the interactions among the data, people, and procedures that make up (and are impacted by) information systems. Beacon’s failure is a cautionary tale of what can go wrong if users fail to broadly consider the impact and implications of an information system on all those it can touch. Technology’s reach is often farther, wider, and more significantly impactful than we originally expect.

Reputation Damage and Increased Scrutiny—The Facebook TOS Debacle

Facebook also suffered damage to its reputation, brand, and credibility, further reinforcing perceptions that the company acts brazenly, without considering user needs, and is fast and loose on privacy and user notification. Facebook worked through the feeds outrage, eventually convincing users of the benefits of feeds. But Beacon was a fiasco. And now users, the media, and watchdogs were on the alert.

When the firm modified its terms of service (TOS) policy in Spring 2009, the uproar was immediate. As a cover story in New York magazine summed it up, Facebook’s new TOS appeared to state, “We can do anything we want with your content, forever,” even if a user deletes their account and leaves the service.V. Grigoriadis, “Do You Own Facebook? Or Does Facebook Own You?” New York, April 5, 2009. Yet another privacy backlash!

Activists organized, the press crafted juicy, attention-grabbing headlines, and the firm was forced once again to backtrack. But here’s where others can learn from Facebook’s missteps and response. The firm was contrite and reached out to explain and engage users. The old TOS were reinstated, and the firm posted a proposed new version that gave the firm broad latitude in leveraging user content without claiming ownership. And the firm renounced the right to use this content if a user closed their Facebook account. This new TOS was offered in a way that solicited user comments, and it was submitted to a community vote, considered binding if 30 percent of Facebook users participated. Zuckerberg’s move appeared to have turned Facebook into a democracy and helped empower users to determine the firm’s next step.

Despite the uproar, only about 1 percent of Facebook users eventually voted on the measure, but the 74 percent to 26 percent ruling in favor of the change gave Facebook some cover to move forward.J. Smith, “Facebook TOS Voting Concludes, Users Vote for New Revised Documents,” Inside Facebook, April 23, 2009. This event also demonstrates that a tempest can be generated by a relatively small number of passionate users. Firms ignore the vocal and influential at their own peril!

In Facebook’s defense, the broad TOS was probably more a form of legal protection than any nefarious attempt to exploit all user posts ad infinitum. The U.S. legal environment does require that explicit terms be defined and communicated to users, even if these are tough for laypeople to understand. But a “trust us” attitude toward user data doesn’t work, particularly for a firm considered to have committed ham-handed gaffes in the past. Managers must learn from the freewheeling Facebook community. In the era of social media, your actions are now subject to immediate and sustained review. Violate the public trust and expect the equivalent of a high-powered investigative microscope examining your every move, and a very public airing of the findings.

Key Takeaways

  • Word of mouth is the most powerful method for promoting products and services, and Beacon was conceived as a giant word-of-mouth machine with win-win benefits for firms, recommenders, recommendation recipients, and Facebook.
  • Beacon failed because it was an opt-out system that was not thoroughly tested beforehand, and because user behavior, expectations, and system procedures were not completely taken into account.
  • Partners associated with the rapidly rolled out, poorly conceived, and untested effort were embarrassed. Several faced legal action.
  • Facebook also reinforced negative perceptions regarding the firm’s attitudes toward users, notifications, and their privacy. This attitude only served to focus a continued spotlight on the firm’s efforts, and users became even less forgiving.
  • Activists and the media were merciless in criticizing the firm’s Terms of Service changes. Facebook’s democratizing efforts demonstrate lessons other organizations can learn from, regarding user scrutiny, public reaction, and stakeholder engagement.

Questions and Exercises

  1. What is Beacon? Why was it initially thought to be a good idea? What were the benefits to firm partners, recommenders, recommendation recipients, and Facebook? Who were Beacon’s partners and what did they seek to gain through the effort?
  2. Describe “the biggest problem with Beacon?” Would you use Beacon? Why or why not?
  3. How might Facebook and its partners have avoided the problems with Beacon? Could the effort be restructured while still delivering on its initial promise? Why or why not?
  4. Beacon shows the risk in being a pioneer—are there risks in being too cautious and not pioneering with innovative, ground-floor marketing efforts? What kinds of benefits might a firm miss out on? Is there a disadvantage in being late to the party with these efforts, as well? Why or why not?
  5. Why do you think Facebook changed its Terms of Service? Did these changes concern you? Were users right to rebel? What could Facebook have done to avoid the problem? Did Facebook do a good job in follow-up? How would you advise Facebook to apply lessons learned form the TOS controversy?

8.8 Predators and Privacy

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the extent and scope of the predator problem on online social networks.
  2. Recognize the steps firms are taking to proactively engage and limit these problems.

While spoiling Christmas is bad, sexual predators are far worse, and in October 2007, Facebook became an investigation target. Officials from the New York State Attorney General’s office had posed as teenagers on Facebook and received sexual advances. Complaints to the service from investigators posing as parents were also not immediately addressed. These were troubling developments for a firm that prided itself on trust and authenticity.

In a 2008 agreement with forty-nine states, Facebook offered aggressive programs, many of which put it in line with MySpace. MySpace had become known as a lair for predators, and after months of highly publicized tragic incidents, the firm had become very aggressive about protecting minors. To get a sense of the scope of the problem, consider that MySpace claimed that it had found and deleted some twenty-nine thousand accounts from its site after comparing profiles against a database of convicted sex offenders.“Facebook Targets China, World’s Biggest Web Market,” Reuters, June 20, 2008. Following MySpace’s lead, Facebook agreed to respond to complaints about inappropriate content within twenty-four hours and to allow an independent examiner to monitor how it handles complaints. The firm imposed age-locking restrictions on profiles, reviewing any attempt by someone under the age of eighteen to change their date of birth. Profiles of minors were no longer searchable. The site agreed to automatically send a warning message when a child is at risk of revealing personal information to an unknown adult. And links to explicit material, the most offensive Facebook groups, and any material related to cyberbullying were banned.

Busted on Facebook

Chapter 7 “Peer Production, Social Media, and Web 2.0” warned that your digital life will linger forever, and that employers are increasingly plumbing the depths of virtual communities in order to get a sense of job candidates. And it’s not just employers. Sleuths at universities and police departments have begun looking to Facebook for evidence of malfeasance. Oxford University fined graduating students more than £10,000 for their post-exam celebrations, evidence of which was picked up from Facebook. Police throughout the United States have made underage drinking busts and issued graffiti warnings based on Facebook photos, too. Beware—the Web knows!

Key Takeaways

  • Thousands of sex offenders and other miscreants have been discovered on MySpace, Facebook, and other social networks. They are a legitimate risk to the community and they harm otherwise valuable services.
  • A combination of firm policies, computerized and human monitoring, aggressive reporting and follow-up, and engagement with authorities can reduce online predator risks.
  • Firms that fail to fully engage this threat put users and communities at risk and may experience irreparable damage to firms and reputations.

Questions and Exercises

  1. How big was the predator problem on MySpace? What efforts have social networks employed to cut down on the number of predators online?
  2. Investigate the current policies regarding underage users on Facebook. Do you think the firm adequately protects its users? Why or why not?
  3. What age is appropriate for users to begin using social networks? Which services are appropriate at which ages? Are there social networks targeted at very young children? Do you think that these are safe places? Why or why not?

8.9 One Graph to Rule Them All: Facebook Takes Over the Web

Learning Objectives

After studying this section you should be able to do the following:

  1. Describe Facebook’s efforts to integrate its service with other Web sites and the potential strategic benefit for Facebook and its partners.
  2. List and discuss the potential benefits and risks of engaging in the kinds of intersite sharing and collaboration efforts described in this section.

In spring 2010, the world got a sense of the breadth and depth of Mark Zuckerberg’s vision. During the firm’s annual f8 Developers Conference, Facebook launched a series of initiatives that placed the company directly at the center of identity, sharing, and personalization—not just on Facebook but also across the Web.

With just a few lines of HTML code, any developer could add a Facebook “Like” button to their site and take advantage of the social network’s power of viral distribution. A user clicking that page’s “Like” button automatically would then send a link to that page to their news feed, where it has the potential to be seen by all of their friends. No additional sign-in is necessary as long as you logged into Facebook first (reinforcing Facebook’s importance as the first stop in your Internet surfing itinerary). While some sites renamed “Like” to “Recommend” (after all, do you really want to “like” a story about a disaster or tragedy?), the effort was adopted with stunning speed. Facebook’s “Like” button served up more than one billion times across the Web in the first twenty-four hours, and over fifty thousand Web sites signed up to add the “Like” button to their content within the first week.A. Oreskovic, “Facebook Efforts Hint at Growing Ad Clout,” The Guardian, April 30, 2010.

Facebook also offered a system where Web site operators can choose to accept a user’s Facebook credentials for logging in. Users like this because they can access content without the hurdle of creating a new account. Web sites like it because with the burden of signing up out of the way, Facebook becomes an experimentation lubricant: “Oh, I can use my Facebook ID to sign in? Then let me try this out.”

Facebook also lets Web sites embed some Facebook functionality right on their pages. A single line of code added to any page creates a “social toolbar” that shows which of your friends are logged into Facebook, and allows access to Facebook Chat without leaving that site. Site operators who are keen on making it easy for friends to summon friends to their pages can now sprinkle these little bits of Facebook across the Web.

Other efforts allow firms to leverage Facebook data to make their sites more personalized. Firms around the Web can now show if a visitor’s friends have “Liked” items on the site, posted comments, or performed other actions. Using this feature, Facebook users logging into Yelp can see a list of restaurants recommended by trusted friends instead of just the reviews posted by a bunch of strangers. Users of the music-streaming site Pandora can have the service customized based on music tastes pulled from their Facebook profile page. They can share stations with friends and have data flow back to update the music preferences listed in their Facebook profile pages. Visit CNN and the site can pull together a list of stories recommended by friends.J. Valentino-DeVries, “Facebook CEO Zuckerberg on Plans to Dominate the Web,” Wall Street Journal, April 21, 2010. Think about how this strengthens the social graph. While items in the news feed might quickly scroll away and disappear, that data can now be pulled up within a Web site, providing insight from friends when and where you’re likely to want it most.

Taken together, these features enlist Web sites to serve as vassal states in the Facebook empire. Each of these ties makes Facebook membership more valuable by enhancing network effects, strengthening switching costs, and creating larger sets of highly personalized data to leverage.

Facebook: The Bank of the Web?

Those with an eye for business disruption are watching the evolution of Facebook Credits. Credits can be used to pay for items, such as weapons in video games or virtual gifts. Facebook shares credits revenue with application developers but takes 30 percent off the top for acting as banker and transaction clearing house.

There are real bucks to be made from digital make-believe. Analysts estimate that in 2009, virtual goods racked up $1 billion in U.S. transactions and $5 billion worldwide.B. Womack and C. Valerio, “Facebook Says Credits Won’t Pay Off Soon, Adds ‘Like’ Feature,” BusinessWeek, April 22, 2010; and C. Miller and B. Stone, “Virtual Goods Start Bringing Real Paydays,” New York Times, November 6, 2009. Facebook currently isn’t much of a player in virtual goods, but that may change. Many expect Credits use to grow into a thriving standard. Users are far more likely to trust Facebook with their credit card than a little-known app developer. There are also an increasing number of ways to pay for Credits. Facebook’s App2Credits effort lets firms offer Credits in ways that don’t involve a credit card, including getting Credits as part of a card loyalty program, converting unwanted real-world gift cards into Facebook Credits, or earning Credits for shopping or performing other online tasks.J. Kincaid, “A Look at the Future of Facebook Credits,” TechCrunch, April 21, 2010.

Credits were rolled out supporting fifteen international currencies and multiple credit cards. Transaction support is provided through a partnership with PayPal, and a deal with mobile payments start-up Zong allows users to bill credits to their phone.C. McCarthy, “Facebook to Developers: Get Ready for Credits,” CNET, February 25, 2010.

All this banking activity leaves some wondering if Facebook might not have grander ambitions. The Financial Times has referred to Facebook as being on the path to becoming “The Bank of the Web.”C. Nuttall, “Facebook Credits Bank of the Web,” Financial Times, April 23, 2010. Could Facebook morph into an actual real-currency bank? A site that knows how to reach your friends might offer an easy way to, say, settle a dinner tab or hound buddies for their Final Four pool money. This might also be a solid base for even deeper banking links between users and all those firms Facebook has begun to leverage in deeper data-sharing partnerships. This may be something to think about, or perhaps, to bank on!

More Privacy Controversy

The decision to launch these new features as “opt-out” instead of “opt-in” immediately drew the concern of lawmakers. Given the Beacon debacle, the TOS controversy, and Google’s problems with Buzz (see Chapter 14 “Google: Search, Online Advertising, and Beyond”), you’d think Facebook would have known better. But within a week of Beacon’s launch, four U.S. senators contacted the firm, asking why it was so difficult to opt out of the information-sharing platform.F. Lardinois, “Is It Time for Facebook to Make Opt-In the Default?” Read Write Web, April 27, 2010. Amid a crush of negative publicity, the firm was forced to quickly roll out simplified privacy management controls.

Facebook’s struggles show the tension faced by any firm that wants to collect data to improve the user experience (and hopefully make money along the way). Opt-out guarantees the largest possible audience and that’s key to realizing the benefits of network effects, data, and scale. Making efforts opt-in creates the very real risk that not enough users will sign up and that the reach and impact of these kinds of initiatives will be limited.F. Lardinois, “Is It Time for Facebook to Make Opt-In the Default?” Read Write Web, April 27, 2010. Fast Company calls this the paradox of privacy, saying, “We want some semblance of control over our personal data, even if we likely can’t be bothered to manage it.”F. Manjoo, “Does Privacy on Facebook, Google, and Twitter Even Matter?” Fast Company, May 1, 2010. Evidence suggests that most people are accepting some degree of data sharing as long as they know that they can easily turn it off if they want to. For example, when Google rolled out ads that tracked users across the network of Web sites running Google ads, the service also provided a link in each ad where users could visit an “ad preferences manager” to learn how they were being profiled, to change settings, and to opt out (see Chapter 14 “Google: Search, Online Advertising, and Beyond”). It turns out only one in fifteen visitors to the ad preferences manager ended up opting out completely.F. Manjoo, “Does Privacy on Facebook, Google, and Twitter Even Matter?” Fast Company, May 1, 2010. Managers seeking to leverage data should learn from the examples of Facebook and Google and be certain to offer clear controls that empower user choice.

Free Riders and Security Concerns

Facebook also allows third-party developers to create all sorts of apps to access Facebook data. Facebook feeds are now streaming through devices that include Samsung, Vizio, and Sony televisions; Xbox 360 and Wii game consoles; Verizon’s FiOS pay television service; and the Amazon Kindle. While Facebook might never have the time or resources to create apps that put its service on every gadget on the market, they don’t need to. Developers using Facebook’s access tools will gladly pick up the slack.

But there are major challenges with a more open approach, most notably a weakening of strategic assets, revenue sharing, and security. First, let’s discuss weakened assets. Mark Zuckerberg’s geeks have worked hard to make their site the top choice for most of the world’s social networkers and social network application developers. Right now, everyone goes to Facebook because everyone else is on Facebook. But as Facebook opens up access to users and content, it risks supporting efforts that undermine the firm’s two most compelling sources of competitive advantage: network effects and switching costs. Any effort that makes it easier to pack up your “social self” and move it elsewhere risks undermining vital competitive resources advantages (it still remains more difficult to export contacts, e-mails, photos, and video from Facebook than it does from sites supporting OpenSocial, a rival platform backed by Google and supported by many of Facebook’s competitors).F. Vogelstein, “The Great Wall of Facebook,” Wired, July 2009. This situation also puts more pressure on Facebook to behave. Lower those switching costs at a time when users are disgusted with firm behavior, and it’s not inconceivable that a sizable chunk of the population could bolt for a new rival (to Facebook’s credit, the site also reached out to prior critics like MoveOn.org, showing Facebook’s data-sharing features and soliciting input months before their official release).

Along with asset weakening comes the issue of revenue sharing. As mentioned earlier, hosting content (especially photos and rich media) is a very expensive proposition. What incentive does a site have to store data if it will just be sent to a third-party site that will run ads around this content and not share the take? Too much data portability presents a free rider problemWhen others take advantage of a user or service without providing any sort of reciprocal benefit. where firms mooch off Facebook’s infrastructure without offering much in return. Consider services like TweetDeck. The free application allows users to access their Facebook feeds and post status updates—alongside Twitter updates and more—all from one interface. Cool for the user, but bad for Facebook, since each TweetDeck use means Facebook users are “off-site,” not looking at ads, and hence not helping Zuckerberg & Co. earn revenue. It’s as if the site has encouraged the equivalent of an ad blocker, yet Facebook’s openness lets this happen!

Finally, consider security. Allowing data streams that contain potentially private posts and photographs to squirt across the Internet and land where you want them raises all sorts of concerns. What’s to say an errant line of code doesn’t provide a back door to your address book or friends list? To your messaging account? To let others see photos you’d hoped to only share with family? Security breaches can occur on any site, but once the data is allowed to flow freely, every site with access is, for hackers, the equivalent of a potential door to open or a window to crawl through.

Social Networking Goes Global

Facebook will eventually see stellar growth start to slow as the law of large numbers sets in. The shift from growth business to mature one can be painful, and for online firms it can occur relatively quickly. That doesn’t mean these firms will become unprofitable, but to sustain growth (particularly important for keeping up the stock price of a publicly traded company), firms often look to expand abroad.

Facebook’s crowdsourcingThe act of taking a job traditionally performed by a designated agent (usually an employee) and outsourcing it to an undefined generally large group of people in the form of an open call. localizationAdapting products and services for different languages and regional differences. effort, where users were asked to look at Facebook phrases and offer translation suggestions for their local language (see Chapter 7 “Peer Production, Social Media, and Web 2.0”), helped the firm rapidly deploy versions in dozens of markets, blasting the firm past MySpace in global reach. But network effects are both quick and powerful, and late market entry can doom a business reliant on the positive feedback loop of a growing user base.

And global competition is out there. Worldwide, Facebook wannabes include “Studiverzeichnis” (German for “student index”); Vkontakte (“in contact”), Russia’s most popular social networking site; and Renren (formerly Xiaonei), which is said to have registered 90 percent of China’s college students.

China is proving a particularly difficult market for foreign Internet firms. Google, eBay, Yahoo! and MySpace have all struggled there (at one point, Rupert Murdoch even sent his wife, Wendi Deng Murdoch, to head up the MySpace China effort). And don’t be surprised to see some of these well-capitalized overseas innovators making a move on U.S. markets too.

While global growth can seem like a good thing, acquiring global users isn’t the same as making money from them. Free sites with large amounts of users from developing nations face real cost/revenue challenges. As the New York Times points out, there are 1.6 billion Internet users worldwide, but fewer than half of them have disposable incomes high enough to interest major advertisers.B. Stone and M. Helft, “In Developing Countries, Web Grows without Profit,” New York Times, April 27, 2009. Worse still, telecommunications costs in these markets are also often higher, too. Bandwidth costs and dim revenue options caused video site Veoh to block access coming from Africa, Eastern Europe, Latin America, and some parts of Asia. MySpace already offers a stripped-down Lite option as its default in India. And execs at YouTube and Facebook haven’t ruled out lowering the quality of streaming media, file size, or other options, discriminating by region or even by user.

Making money in the face of this so-called “International Paradox” requires an awareness of “fast and cheap” tech trends highlighted in Chapter 5 “Moore’s Law: Fast, Cheap Computing and What It Means for the Manager”, as well as an ability to make accurate predictions regarding regional macroeconomic trends. Ignore a market that’s unprofitable today and a rival could swoop in and establish network effects and other assets that are unbeatable tomorrow. But move too early and losses could drag you down.

Key Takeaways

  • Facebook has extended its reach by allowing other Web sites to leverage the site. Facebook partners can add the “Like” button to encourage viral sharing of content, leverage Facebook user IDs for log-in, and tap a user’s friend and feed data to personalize and customize a user’s experience.
  • These efforts come with risks, including enabling free riders that might exploit the firm’s content without compensation, and the potential for privacy and security risks.
  • Facebook Credits are a currency for use for virtual gifts and games. The service accepts multiple currencies and payment methods; and while virtual goods have the potential to be a big business, some speculate that Facebook may one day be able to develop a payments and banking businesses from this base.
  • Global growth is highly appealing to firms, but expensive bandwidth costs and low prospects for ad revenue create challenges akin to the free rider problem.

Questions and Exercises

  1. Cite effective examples you’ve seen of Facebook features on other Web sites (or if you haven’t seen any, do some background research to uncover such efforts). Why do the efforts you’ve highlighted “work”? How do they benefit various parties? Does everyone benefit? Is anyone at risk? If so, explain the risks.
  2. Should Facebook be as open as it is? In what ways might this benefit the firm? In what ways is it a risk?
  3. How can Facebook limit criticism of its data-sharing features? Do you think it made mistakes during rollout?
  4. What is TweetDeck? Why is a product like this a potential threat to Facebook?
  5. Research OpenSocial online. What is this effort? What challenges does it face in attempting to become a dominant standard?
  6. Facebook has global competitors. What determines the success of a social network within a given country? Why do network effects for social networks often fail to translate across national borders?
  7. How did Facebook localize its site so quickly for various different regions of the world?
  8. What factors encourage firms to grow an international user base as quickly as possible? Why is this a risk? What sorts of firms are at more risk than others?

8.10 Is Facebook Worth It?

Learning Objectives

After studying this section you should be able to do the following:

  1. Question the $15 billion valuation so often cited by the media.
  2. Understand why Microsoft might be willing to offer to invest in Facebook at a higher valuation rate.

It has often been said that the first phase of the Internet was about putting information online and giving people a way to find it. The second phase of the Web is about connecting people with one another. The Web 2.0 movement is big and impactful, but is there much money in it?

While the valuations of private firms are notoriously difficult to pin down due to a lack of financial disclosure, the often-cited $15 billion valuation from the fall of 2007 Microsoft investment was rich, even when made by such a deep-pocketed firm. Using estimates at the time of the deal, if Facebook were a publicly traded company, it would have a price-to-earnings ratio of five hundred; Google’s at the time was fifty-three, and the average for the S&P 500 is historically around fifteen.

But the math behind the deal is a bit more complex than was portrayed in most press reports. The deal was also done in conjunction with an agreement that for a time let Microsoft manage the sale of Facebook’s banner ads worldwide. And Microsoft’s investment was done on the basis of preferred stock, granting the firm benefits beyond common stock, such as preference in terms of asset liquidation.B. Stone, “Facebook Aims to Extends Its Reach across Web,” New York Times, December 1, 2008. Both of these are reasons a firm would be willing to “pay more” to get in on a deal.

Another argument can be made for Microsoft purposely inflating the value of Facebook in order to discourage rival bidders. A fat valuation by Microsoft and a deal locking up ad rights makes the firm seem more expensive, less attractive, and out of reach for all but the richest and most committed suitors. Google may be the only firm that could possibly launch a credible bid, and Zuckerberg is reported to be genuinely uninterested in being absorbed by the search sovereign.F. Vogelstein, “The Great Wall of Facebook,” Wired, July 2009.

Since the fall of 2007, several others have invested private money into Facebook as well, including the Founders Fund and Li Ka-shing, the Hong Kong billionaire behind Hutchison Whampoa. Press reports and court documents suggest that these deals were done at valuations that were lower than what Microsoft accepted. In May 2009 Russian firm Digital Sky paid $200 million for 1.96 percent of the firm, a ten-billion-dollar valuation (also in preferred stock). That’s a one-third haircut off the Microsoft price, albeit without the Redmond-specific strategic benefits of the investment.D. Kirkpatrick, “Why Microsoft Isn’t Buying Facebook,” Fortune, May 9, 2008; and S. Ante, “Facebook: Friends with Money,” BusinessWeek, May 9, 2008. And as the chart in Figure 8.2 “Revenue per User (2009)” shows, Facebook still lags well behind many of its rivals in terms of revenue per user.

Figure 8.2 Revenue per User (2009)

While Facebook’s reach has grown to over half a billion visitors a month, its user base generates far less cash on a per-person basis than many rivals do.H. Blodget, “Whoops—Facebook Is Once Again Overhyped,” Business Insider, April 26, 2010.

So despite the headlines, even at the time of the Microsoft investment, Facebook was almost certainly not valued at a pure $15 billion. This isn’t to say definitively that Facebook won’t be worth $15 billion (or more) someday, but even a valuation at “just” $10 billion is a lot to pay for a then-profitless firm with estimated 2009 revenues of $500 million. Of course, raising more capital enables Zuckerberg to go on the hunt as well. Facebook investor Peter Theil confirmed the firm had already made an offer to buy Twitter (a firm which at the time had zero dollars in revenues and no discernible business model) for a cool half billion dollars.S. Ante, “Facebook’s Thiel Explains Failed Twitter Takeover,” BusinessWeek, March 1, 2009.

Much remains to be demonstrated for any valuation to hold. Facebook is new. Its models are evolving, and it has quite a bit to prove. Consider efforts to try to leverage friend networks. According to Facebook’s own research, “an average Facebook user with 500 friends actively follows the news on only forty of them, communicates with twenty, and keeps in close touch with about ten. Those with smaller networks follow even fewer.”S. Baker, “Learning and Profiting from Online Friendships,” BusinessWeek, May 21, 2009. That might not be enough critical mass to offer real, differentiable impact, and that may have been part of the motivation behind Facebook’s mishandled attempts to encourage more public data sharing. The advantages of leveraging the friend network hinge on increased sharing and trust, a challenge for a firm that has had so many high-profile privacy stumbles. There is promise. Profiling firm Rapleaf found that targeting based on actions within a friend network can increase click-through rates threefold—that’s an advantage advertisers are willing to pay for. But Facebook is still far from proving it can consistently achieve the promise of delivering valuable ad targeting.

Steve Rubel wrote the following on his Micro Persuasion blog: “The Internet amber is littered with fossilized communities that once dominated. These former stalwarts include AOL, Angelfire, theGlobe.com, GeoCities, and Tripod.” Network effects and switching cost advantages can be strong, but not necessarily insurmountable if value is seen elsewhere and if an effort becomes more fad than “must have.” Time will tell if Facebook’s competitive assets and constant innovation are enough to help it avoid the fate of those that have gone before them.

Key Takeaways

  • Not all investments are created equal, and a simple calculation of investment dollars multiplied by the percentage of firm owned does not tell the whole story.
  • Microsoft’s investment entitled the firm to preferred shares; it also came with advertising deal exclusivity.
  • Microsoft may also benefit from offering higher valuations that discourage rivals from making acquisition bids for Facebook.
  • Facebook has continued to invest capital raised in expansion, particularly in hardware and infrastructure. It has also pursued its own acquisitions, including a failed bid to acquire Twitter.
  • The firm’s success will hinge on its ability to create sustainably profitable revenue opportunities. It has yet to prove that data from the friend network will be large enough and can be used in a way that is differentiably attractive to advertisers. However, some experiments in profiling and ad targeting across a friend network have shown very promising results. Firms exploiting these opportunities will need to have a deft hand in offering consumer and firm value while quelling privacy concerns.

Questions and Exercises

  1. Circumstances change over time. Research the current state of Facebook’s financials—how much is the firm “valued at”? How much revenue does it bring in? How profitable is it? Are these figures easy or difficult to find? Why or why not?
  2. Who else might want to acquire Facebook? Is it worth it at current valuation rates?
  3. What motivation does Microsoft have in bidding so much for Facebook?
  4. Do you think Facebook was wise to take funds from Digital Sky? Why or why not?
  5. Do you think Facebook’s friend network is large enough to be leveraged as a source of revenue in ways that are notably different than conventional pay-per-click or CPM-based advertising? Would you be excited about certain possibilities? Creeped out by some? Explain possible scenarios that might work or might fail. Justify your interpretation of these scenarios.
  6. So you’ve had a chance to learn about Facebook, its model, growth, outlook, strategic assets, and competitive environment. How much do you think the firm is worth? Which firms do you think it should compare with in terms of value, influence, and impact? Would you invest in Facebook?
  7. Which firms might make good merger partners with Facebook? Would these deals ever go through? Why or why not?

Chapter 7: Peer Production, Social Media, and Web 2.0

7.1 Introduction

Learning Objectives

After studying this section you should be able to do the following:

  1. Recognize the unexpected rise and impact of social media and peer production systems, and understand how these services differ from prior generation tools.
  2. List the major classifications of social media services.

Over the past few years a fundamentally different class of Internet services has attracted users, made headlines, and increasingly garnered breathtaking market valuations. Often referred to under the umbrella term “Web 2.0A term broadly referring to Internet services that foster collaboration and information sharing; characteristics that distinctly set “Web 2.0” efforts apart from the static, transaction-oriented Web sites of “Web 1.0.” The term is often applied to Web sites and Internet services that foster social media or other sorts of peer production.,” these new services are targeted at harnessing the power of the Internet to empower users to collaborate, create resources, and share information in a distinctly different way from the static Web sites and transaction-focused storefronts that characterized so many failures in the dot-com bubble. Blogs, wikis, social networks, photo and video sharing sites, and tagging systems all fall under the Web 2.0 moniker, as do a host of supporting technologies and related efforts.

The term Web 2.0 is a tricky one because like so many popular technology terms there’s not a precise definition. Coined by publisher and pundit Tim O’Reilly in 2003, techies often joust over the breadth of the Web 2.0 umbrella and over whether Web 2.0 is something new or simply an extension of technologies that have existed since the creation of the Internet. These arguments aren’t really all that important. What is significant is how quickly the Web 2.0 revolution came about, how unexpected it was, and how deeply impactful these efforts have become. Some of the sites and services that have evolved and their Web 1.0 counterparts are listed in Table 7.1 “Web 1.0 versus Web 2.0”.Adapted from T. O’Reilly, “What Is Web 2.0?” O’Reilly, September 30, 2005.

Table 7.1 Web 1.0 versus Web 2.0

Web 1.0 Web 2.0
DoubleClick  →  Google AdSense
Ofoto  →  Flickr
Akamai  →  BitTorrent
mp3.com  →  Napster
Britannica Online  →  Wikipedia
personal Web sites  →  blogging
evite  →  upcoming.org and Eventful
domain name speculation  →  search engine optimization
page views  →  cost per click
screen scraping  →  Web services
publishing  →  participation
content management systems  →  wikis
directories (taxonomy)  →  tagging (“folksonomy”)
stickiness  →  syndication
instant messaging  →  Twitter
Monster.com  →  LinkedIn

To underscore the speed with which Web 2.0 arrived on the scene, and the impact of leading Web 2.0 services, consider the following efforts:

  • According to a spring 2008 report by Morgan Stanley, Web 2.0 services ranked as seven of the world’s top ten most heavily trafficked Internet sites (YouTube, Live.com, MySpace, Facebook, Hi5, Wikipedia, and Orkut); only one of these sites (MySpace) was on the list in 2005.Morgan Stanley, Internet Trends Report, March 2008.
  • With only seven full-time employees and an operating budget of less than $1 million, Wikipedia has become the Internet’s fifth most visited site on the Internet.G. Kane and R. Fichman, “The Shoemaker’s Children: Using Wikis for Information Systems Teaching, Research, and Publication,” MIS Quarterly, March 2009. The site boasts well over fifteen million articles in over two hundred sixty different languages, all of them contributed, edited, and fact-checked by volunteers.
  • Just two years after it was founded, MySpace was bought for $580 million by Rupert Murdoch’s News Corporation (the media giant that owns the Wall Street Journal and the Fox networks, among other properties). By the end of 2007, the site accounted for some 12 percent of Internet minutes and had repeatedly ranked as the most-visited Web site in the United States.D. Chmielewski and J. Guynn, “MySpace Ready to Prove Itself in Faceoff,” Chicago Tribune, June 8, 2008. But rapid rise doesn’t always mean a sustained following, and by the start of 2010, some were beginning to write the service’s obituary as it failed to keep pace with Facebook.O. Malik, “MySpace, R.I.P.,” GigaOM, February 10, 2010.
  • The population of rival Facebook is now so large that it could be considered the third largest “nation” in the world. Half the site’s users log in at least once a day, spending an average of fifty-five minutes a day on the site.“Facebook Facts and Figures (History and Statistics),” Website Monitoring Blog, March 17, 2010. A fall 2007 investment from Microsoft pegged the firm’s overall value at $15 billion, a number that would have made it the fifth most valuable Internet firm, despite annual revenues at the time of only $150 million.M. Arrington, “Perspective: Facebook Is Now Fifth Most Valuable U.S. Internet Company,” TechCrunch, October 25, 2007. Those revenues have been growing, with the privately held firm expected to bring in from $1.2 to $2 billion in 2010.J. Vascellaro, “Facebook CEO in No Rush to ‘Friend’ Wall Street,” Wall Street Journal, March 3, 2010.
  • Just twenty months after its founding, YouTube was purchased by Google for $1.65 billion. While Google struggles to figure out how to make profitable what is currently a money-losing resource hog (over twenty hours of video are uploaded to YouTube each minute)E. Nakashima, “YouTube Ordered to Release User Data,” Washington Post, July 4, 2008. the site has emerged as the Web’s leading destination for video, hosting everything from apologies from JetBlue’s CEO for service gaffes to questions submitted as part of the 2008 U.S. presidential debates. Fifty percent of YouTube’s roughly three hundred million users visit the site at least once a week.Morgan Stanley, Internet Trends Report, March 2008.
  • Twitter has emerged as a major force that can break news and shape public opinion. China and Iran are among the governments so threatened by the power of Twitter-fueled data sharing that each has, at times, blocked Twitter access within their borders. At the first Twitter-focused Chirp conference in April 2010, Twitter boasted a population of over one hundred million users who have collectively posted more than ten billion tweets (Twitter messages). By this time, the service had also spawned an ecosystem of over one hundred thousand registered Twitter-supporting apps. In another nod to the service’s significance, the U.S. Library of Congress announced plans to archive every tweet ever sent.N. Bolton, “Chirp, Twitter’s First Developer Conference, Opens Its Doors,” New York Times, April 14, 2010; and M. Shaer, “Google Launches Archive Search for Twitter,” Christian Science Monitor, April 15, 2010.
  • Services such as Twitter, Yelp, and the highly profitable TripAdvisor have unleashed the voice of the customer so that it is now often captured and broadcast immediately at the point of service. Reviews are now incorporated into search results and maps, making them the first thing many customers see when encountering a brand online. TripAdvisor, with just five hundred employees, contributes over $150 million in profits to parent company Expedia (at roughly 50 percent margins),B. Wash, “Double Duty,” Colby Magazine, Winter 2009. while Yelp has reportedly turned down acquisition offers valuing it at $700 million.P. Burrows, “Hot Tech Companies Like Yelp Are Bypassing IPOs,” BusinessWeek, February 4, 2010.

Table 7.2 Major Social Media Tools

Description Features Technology Providers Use Case Examples
Blogs Short for “Web log”—an online diary that keeps a running chronology of entries. Readers can comment on posts. Can connect to other blogs through blog rolls or trackbacks.

Key uses: Share ideas, obtain feedback, mobilize a community.

  • Ease of use
  • Reverse chronology
  • Comment threads
  • Persistence
  • Searchability
  • Tags
  • Trackbacks
  • Blogger (Google)
  • WordPress
  • Six Apart (TypePad and Movable Type)
  • Tumblr
  • News outlets
  • Google
  • Graco
  • GM
  • Kaiser Permanente
  • Marriott
  • Microsoft
Wikis A Web site that anyone can edit directly from within the browser.

Key uses: Collaborate on common tasks or to create a common knowledge base.

  • All changes are attributed
  • A complete revision history is maintained, with the ability to roll back changes and revert to earlier versions
  • Automatic notification of updates
  • Searchability
  • Tags
  • Monitoring
  • Socialtext
  • PBWorks
  • Google Sites
  • WetPaint
  • Microsoft SharePoint
  • Apple OS X Server
  • Dresdner Kleinwort Wasserstein
  • eBay
  • The FBI, CIA, and other intelligence agencies
  • Intuit
  • Pixar
Electronic Social Network Online community that allows users to establish a personal profile, link to other profiles (i.e., friends), share content, and communicate with members via messaging, posts.

Key Uses: Discover and reinforce affiliations; identify experts; message individuals or groups; virally share media.

  • Detailed personal profiles using multimedia
  • Affiliations with groups
  • Affiliations with individuals
  • Messaging and public discussions
  • Media sharing
  • “Feeds” of recent activity among members
  • Facebook
  • LinkedIn
  • MySpace
  • Ning
  • SelectMinds
  • LiveWorld
  • IBM/Lotus Connections
  • Salesforce.com
  • Socialtext
  • Barack Obama (campaign and government organizing)
  • Currensee (foreign exchange trading)
  • Dell
  • Deloitte Consulting
  • Goldman-Sachs
  • IBM
  • Reuters
  • Starbucks
Microblogging Short, asynchronous messaging system. Users send messages to “followers.”

Key Uses: distribute time-sensitive information, share opinions, virally spread ideas, run contests and promotions, solicit feedback, provide customer support, track commentary on firms/products/issues, organize protests.

  • 140-character messages sent and received from mobile device
  • Ability to respond publicly or privately
  • Can specify tags to classify discussion topics for easy searching and building comment threads
  • Follower lists
  • Twitter
  • Socialtext Signals
  • Yammer
  • Salesforce.com (Chatter)
  • Dell
  • Starbucks
  • Intuit
  • Small businesses
  • Celebrities
  • Zappos

Millions of users, billions of dollars, huge social impact, and these efforts weren’t even on the radar of most business professionals when today’s graduating college seniors first enrolled as freshmen. The trend demonstrates that even some of the world’s preeminent thought leaders and business publications can be sideswiped by the speed of the Internet.

Consider that when management guru Michael Porter wrote a piece titled “Strategy and the Internet” at the end of the dot-com bubble, he lamented the high cost of building brand online, questioned the power of network effects, and cast a skeptical eye on ad-supported revenue models. Well, it turns out Web 2.0 efforts challenged all of these concerns. Among the efforts above, all built brand on the cheap with little conventional advertising, and each owes their hypergrowth and high valuation to their ability to harness the network effect.

While the Web 2.0 moniker is a murky one, we’ll add some precision to our discussion of these efforts by focusing on peer productionWhen users collaboratively work to create content, products, and services. Includes social media sites, open source software, and peer-produced services, such as Skype and BitTorrent, where the participation of users provide the infrastructure and computational resources that enable the service., perhaps Web 2.0’s most powerful feature, where users work, often collaboratively, to create content and provide services online. Web-based efforts that foster peer production are often referred to as social mediaContent that is created, shared, and commented on by a broader community of users. Services that support the production and sharing of social media include blogs, wikis, video sites like YouTube, and social networks. or user-generated content sites. These sites include blogs; wikis; social networks like Facebook and MySpace; communal bookmarking and tagging sites like Del.icio.us; media sharing sites like YouTube and Flickr; and a host of supporting technologies. And it’s not just about media. Peer-produced services like Skype and BitTorrent leverage users’ computers instead of a central IT resource to forward phone calls and video. This ability saves their sponsors the substantial cost of servers, storage, and bandwidth. Peer production is also leveraged to create much of the open source software that supports many of the Web 2.0 efforts described above. Techniques such as crowdsourcing, where initially undefined groups of users band together to solve problems, create code, and develop services, are also a type of peer production. These efforts often seek to leverage the so-called wisdom of crowds, the idea that a large, diverse group often has more collective insight than a single or small group of trained professionals. These efforts will be expanded on below, along with several examples of their use and impact.

Key Takeaways

  • A new generation of Internet applications is enabling consumers to participate in creating content and services online. Examples include Web 2.0 efforts such as social networks, blogs, and wikis, as well as efforts such as Skype and BitTorrent, which leverage the collective hardware of their user communities to provide a service.
  • These efforts have grown rapidly, most with remarkably little investment in promotion. Nearly all of these new efforts leverage network effects to add value and establish their dominance and viral marketing to build awareness and attract users.
  • Experts often argue whether Web 2.0 is something new or merely an extension of existing technologies. The bottom line is the magnitude of the impact of the current generation of services.
  • Peer production and social media fall under the Web 2.0 umbrella. These services often leverage the wisdom of crowds to provide insight or production that can be far more accurate or valuable than that provided by a smaller group of professionals.
  • Network effects play a leading role in enabling Web 2.0 firms. Many of these services also rely on ad-supported revenue models.

Questions and Exercises

  1. What distinguishes Web 2.0 technologies and services from the prior generation of Internet sites?
  2. Several examples of rapidly rising Web 2.0 efforts are listed in this section. Can you think of other dramatic examples? Are there cautionary tales of efforts that may not have lived up to their initial hype or promise? Why do you suppose they failed?
  3. Make your own list of Web 1.0 and Web 2.0 services and technologies. Would you invest in them? Why or why not?
  4. In what ways do Web 2.0 efforts challenge the assumptions that Michael Porter made regarding Strategy and the Internet?

7.2 Blogs

Learning Objectives

After studying this section you should be able to do the following:

  1. Know what blogs are and how corporations, executives, individuals, and the media use them.
  2. Understand the benefits and risks of blogging.
  3. Appreciate the growth in the number of blogs, their influence, and their capacity to generate revenue.

BlogsOnline journal entries, usually made in a reverse chronological order. Blogs typically provide comment mechanisms where users can post feedback for authors and other readers. (short for Web logs) first emerged almost a decade ago as a medium for posting online diaries. (In a perhaps apocryphal story, Wired magazine claimed the term “Web log” was coined by Jorn Barger, a sometimes homeless, yet profoundly prolific, Internet poster). From humble beginnings, the blogging phenomenon has grown to a point where the number of public blogs tracked by Technorati (the popular blog index) has surpassed one hundred million.D. Takahashi, “Technorati Releases Data on State of the Blogosphere: Bloggers of the World Have United,” VentureBeat, September 28, 2008. This number is clearly a long tailIn this context, refers to an extremely large selection of content or products. The long tail is a phenomenon whereby firms can make money by offering a near-limitless selection. phenomenon, loaded with niche content that remains “discoverable” through search engines and blog indexes. TrackbacksLinks in a blog post that refer eaders back to cited sources. Trackbacks allow a blogger to see which and how many other bloggers are referring to their content. A “trackback” field is supported by most blog software and while it’s not required to enter a trackback when citing another post, it’s considered good “netiquette” to do so. (third-party links back to original blog post), and blog rollsA list of a blogger’s favorite blogs. While not all blogs include blog rolls, those that do are often displayed on the right or left column of a blog’s main page. (a list of a blogger’s favorite sites—a sort of shout-out to blogging peers) also help distinguish and reinforce the reputation of widely read blogs.

The most popular blogs offer cutting-edge news and commentary, with postings running the gamut from professional publications to personal diaries. While this cacophony of content was once dismissed, blogging is now a respected and influential medium. Some might say that many of the most popular blogs have grown beyond the term, transforming into robust media enterprises. Consider that the political blog The Huffington Post is now more popular than all but eight newspaper sites and has a valuation higher than many publicly traded papers.E. Alterman, “Out of Print, the Death and Life of the American Newspaper,” New Yorker, March 31, 2008; and M. Learmonth, “Huffington Post More Valuable Than Some Newspaper Cos.,” DigitalNext, December 1, 2008. Keep in mind that this is a site without the sports, local news, weather, and other content offered by most papers. Ratings like this are hard to achieve—most bloggers can’t make a living off their musings. But among the elite ranks, killer subscriber numbers are a magnet for advertisers. Top blogs operating on shoestring budgets can snare several hundred thousand dollars a month in ad revenue.S. Zuckerman, “Yes, Some Blogs Are Profitable—Very Profitable,” San Francisco Chronicle, October 21, 2007. Most start with ad networks like Google AdSense, but the most elite engage advertisers directly for high-value deals and extended sponsorships.

Top blogs have begun to attract well-known journalists away from print media. The Huffington Post hired a former Washington Post editor Lawrence Roberts to head the site’s investigative unit. The popular blog TechCrunch now features posts by Sarah Lacy (a BusinessWeek cover-story writer) and has hired Erick Schonfeld away from Time Warner’s business publishing empire. Schonfeld’s colleague, Om Malik, has gone on to found another highly ranked tech industry blog, GigaOM.

Senior executives from many industries have also begun to weigh in with online ruminations, going directly to the people without a journalist filtering their comments. Hotel chief Bill Marriott, Paul Levy (CEO of health care quality leader Beth Israel Deaconess Medical Center), Toyota’s Akio Toyoda, and Zappos’ CEO Tony Hsieh use their blogs for purposes that include a combination of marketing, sharing ideas, gathering feedback, press response, image shaping, and reaching consumers directly without press filtering. Blogs have the luxury of being more topically focused than traditional media, with no limits on page size, word count, or publication deadline. Some of the best examples engage new developments in topic domains much more quickly and deeply than traditional media. For example, it’s not uncommon for blogs focused on the law or politics to provide a detailed dissection of a Supreme Court opinion within hours of its release—offering analysis well ahead of, and with greater depth, than via what bloggers call the mainstream media (MSM)Refers to newspapers, magazines, television, and radio. The MSM is distinctly different from Internet media such as blogs.. As such, it’s not surprising that most mainstream news outlets have begun supplementing their content with blogs that can offer greater depth, more detail, and deadline-free timeliness.

Blogs

While the feature set of a particular blog depends on the underlying platform and the preferences of the blogger, several key features are common to most blogs:

  • Ease of use. Creating a new post usually involves clicking a single button.
  • Reverse chronology. Posts are listed in reverse order of creation, making it easy to see the most recent content.
  • Comment threads. Readers can offer comments on posts.
  • Persistence. Posts are maintained indefinitely at locations accessible by permanent links.
  • Searchability. Current and archived posts are easily searchable.
  • Tags. Posts are often classified under an organized tagging scheme.
  • Trackbacks. Allows an author to acknowledge the source of an item in their post, which allows bloggers to follow the popularity of their posts among other bloggers.

The voice of the blogosphereA term referring to the collective community of bloggers, as well as those who read and comment on blogs. can wield significant influence. Examples include leading the charge for Dan Rather’s resignation and prompting the design of a new insulin pump. In an example of what can happen when a firm ignores social media, consider the flare-up Ingersoll Rand faced when the online community exposed a design flaw in its Kryptonite bike lock.

Online posts showed the thick metal lock could be broken with a simple ball-point pen. A video showing the hack was posted online. When Ingersoll Rand failed to react quickly, the blogosphere erupted with criticism. Just days after online reports appeared, the mainstream media picked up the story. The New York Times ran a piece titled “The Pen Is Mightier Than the Lock” that included a series of photos demonstrating the ballpoint Kryptonite lock pick. The event tarnished the once-strong brand and eventually resulted in a loss of over $10 million.

Like any Web page, blogs can be public, tucked behind a corporate firewall, or password protected. Most blogs offer a two-way dialogue, allowing users to comment on posts (sort of instant “letters to the editor,” posted online and delivered directly to the author). The running dialogue can read like an electronic bulletin board, and can be an effective way to gather opinion when vetting ideas. Comments help keep a blogger honest. Just as the “wisdom of crowds” keeps Wikipedia accurate, a vigorous community of commenters will quickly expose a blogger’s errors of fact or logic.

Despite this increased popularity, blogging has its downside. Blog comments can be a hothouse for spam and the disgruntled. Ham-handed corporate efforts (such as poor response to public criticism or bogus “praise posts”) have been ridiculed. Employee blogging can be difficult to control and public postings can “live” forever in the bowels of an Internet search engine or as content pasted on other Web sites. Many firms have employee blogging and broader Internet posting policies to guide online conduct that may be linked to the firm (see Section 7.9 “Get SMART: The Social Media Awareness and Response Team”). Bloggers, beware—there are dozens of examples of workers who have been fired for what employers viewed as inappropriate posts.

Blogs can be hosted via third-party services (Google Blogger, WordPress, Tumblr, TypePad, Windows Live Spaces), with most offering a combination of free and premium features. Blogging features have also been incorporated into social networks such as Facebook, MySpace, and Ning, as well as corporate social media platforms such as Socialtext. Blogging software can also be run on third-party servers, allowing the developer more control in areas such as security and formatting. The most popular platform for users choosing to host their own blog server is the open source WordPress system.

In the end, the value of any particular blog derives from a combination of technical and social features. The technical features make it easy for a blogger and his or her community to engage in an ongoing conversation on some topic of shared interest. But the social norms and patterns of use that emerge over time in each blog are what determine whether technology features will be harnessed for good or ill. Some blogs develop norms of fairness, accuracy, proper attribution, quality writing, and good faith argumentation, and attract readers that find these norms attractive. Others mix it up with hotly contested debate, one-sided partisanship, or deliberately provocative posts, attracting a decidedly different type of discourse.

Key Takeaways

  • Blogs provide a rapid way to distribute ideas and information from one writer to many readers.
  • Ranking engines, trackbacks, and comments allow a blogger’s community of readers to spread the word on interesting posts and participate in the conversation, and help distinguish and reinforce the reputations of widely read blogs.
  • Well-known blogs can be powerfully influential, acting as flashpoints on public opinion.
  • Firms ignore influential bloggers at their peril, but organizations should also be cautious about how they use and engage blogs, and avoid flagrantly promotional or biased efforts.
  • Top blogs have gained popularity, valuations, and profits that far exceed those of many leading traditional newspapers, and leading blogs have begun to attract well-known journalists away from print media.
  • Senior executives from several industries use blogs for business purposes, including marketing, sharing ideas, gathering feedback, press response, image shaping, and reaching consumers directly without press filtering.

Questions and Exercises

  1. Visit Technorati and find out which blogs are currently the most popular. Why do you suppose the leaders are so popular?
  2. How are popular blogs discovered? How is their popularity reinforced?
  3. Are blog comment fields useful? If so, to whom or how? What is the risk associated with allowing users to comment on blog posts? How should a blogger deal with comments that they don’t agree with?
  4. Why would a corporation, an executive, a news outlet, or a college student want to blog? What are the benefits? What are the concerns?
  5. Identify firms and executives that are blogging online. Bring examples to class and be prepared to offer your critique of their efforts.
  6. How do bloggers make money? Do all bloggers have to make money? Do you think the profit motive influences their content?
  7. Investigate current U.S. Federal Trade Commission laws (or the laws in your home country) that govern bloggers and other social media use. How do these restrictions impact how firms interact with bloggers? What are the penalties and implications if such rules aren’t followed? Are there unwritten rules of good practice that firms and bloggers should consider as well? What might those be?
  8. According to your reading, how does the blog The Huffington Post compare with the popularity of newspaper Web sites?
  9. What advantage do blogs have over the MSM? What advantage does the MSM have over the most popular blogs?
  10. Start a blog using Blogger.com, WordPress.com, or some other blogging service. Post a comment to another blog. Look for the trackback field when making a post, and be sure to enter the trackback for any content you cite in your blog.

7.3 Wikis

Learning Objectives

After studying this section you should be able to do the following:

  1. Know what wikis are and how they are used by corporations and the public at large.
  2. Understand the technical and social features that drive effective and useful wikis.
  3. Suggest opportunities where wikis would be useful and consider under what circumstances their use may present risks.
  4. Recognize how social media such as wikis and blogs can influence a firm’s customers and brand.

A wikiA Web site that can be modified by anyone, from directly within a Web browser (provided that user is granted edit access). is a Web site anyone can edit directly within a Web browser (provided the site grants the user edit access). Wikis derive their name from the Hawaiian word for “quick.” Ward Cunningham, the “wiki father” christened this new class of software with the moniker in honor of the wiki-wiki shuttle bus at the Honolulu airport. Wikis can indeed be one of the speediest ways to collaboratively create content online. Many popular online wikis serve as a shared knowledge repository in some domain.

The largest and most popular wiki is Wikipedia, but there are hundreds of publicly accessible wikis that anyone can participate in. Each attempts to chronicle a world of knowledge within a particular domain, with examples ranging from Wine Wiki for oenophiles to Wookieepedia, the Star Wars wiki. But wikis can be used for any collaborative effort—from meeting planning to project management. And in addition to the hundreds of public wikis, there are many thousand more that are hidden away behind firewalls, used as proprietary internal tools for organizational collaboration.

Like blogs, the value of a wiki derives from both technical and social features. The technology makes it easy to create, edit, and refine content; learn when content has been changed, how and by whom; and to change content back to a prior state. But it is the social motivations of individuals (to make a contribution, to share knowledge) that allow these features to be harnessed. The larger and more active a wiki community, the more likely it is that content will be up-to-date and that errors will be quickly corrected (again, we see the influence of network effects, where products and services with larger user bases become more valuable). Several studies have shown that large community wiki entries are as or more accurate than professional publication counterparts.S. Robert Lichter, Are Chemicals Killing Us? Statistical Assessment Service, May 21, 2009; J. Kane, R. Fichman, J. Gallaugher, and J. Glaser, “Community Relations 2.0,” Harvard Business Review, November 2009.

Want to add to or edit a wiki entry? On most sites you just click the “Edit” link. Wikis support what you see is what you get (WYSIWYG)A phrase used to describe graphical editing tools, such as those found in a wiki, page layout program, or other design tool. editing that, while not as robust as traditional word processors, is still easy enough for most users to grasp without training or knowledge of arcane code or markup language. Users can make changes to existing content and can easily create new pages or articles and link them to other pages in the wiki. Wikis also provide a version history. Click the “History” link on Wikipedia, for example, and you can see when edits were made and by whom. This feature allows the community to roll backThe ability to revert a wiki page to a prior version. This is useful for restoring earlier work in the event of a posting error, inaccuracy, or vandalism. a wiki to a prior page, in the event that someone accidentally deletes key info, or intentionally defaces a page.

Vandalism is a problem on Wikipedia, but it’s more of a nuisance than a crisis. A Wired article chronicled how Wikipedia’s entry for former U.S. President Jimmy Carter was regularly replaced by a photo of a “scruffy, random unshaven man with his left index finger shoved firmly up his nose.”D. Pink, “The Book Stops Here,” Wired, March 2005. Nasty and inappropriate, to be sure, but the Wikipedia editorial community is now so large and so vigilant that most vandalism is caught and corrected within seconds. Watch-lists for the most active targets (say the Web pages of political figures or controversial topics) tip off the community when changes are made. The accounts of vandals can be suspended, and while mischief-makers can log in under another name, most vandals simply become discouraged and move on. It’s as if an army of do-gooders follows a graffiti tagger and immediately repaints any defacement.

Wikis

As with blogs, a wiki’s features set varies depending on the specific wiki tool chosen, as well as administrator design, but most wikis support the following key features:

  • All changes are attributed, so others can see who made a given edit.
  • A complete revision history is maintained so changes can be compared against prior versions and rolled back as needed.
  • There is automatic notification and monitoring of updates; users subscribe to wiki content and can receive updates via e-mail or RSS feed when pages have been changed or new content has been added.
  • All the pages in a wiki are searchable.
  • Specific wiki pages can be classified under an organized tagging scheme.

Wikis are available both as software (commercial as well as open source varieties) that firms can install on their own computers or as online services (both subscription or ad-supported) where content is hosted off-site by third parties. Since wikis can be started without the oversight or involvement of a firm’s IT department, their appearance in organizations often comes from grassroots user initiative. Many wiki services offer additional tools such as blogs, message boards, or spreadsheets as part of their feature set, making most wikis really more full-featured platforms for social computing.

Jump-starting a wiki can be a challenge, and an underused wiki can be a ghost town of orphan, out-of-date, and inaccurate content. Fortunately, once users see the value of wikis, use and effectiveness often snowballs. The unstructured nature of wikis are also both a strength and weakness. Some organizations employ wikimastersIndividuals often employed by organizations to review community content in order to delete excessive posts, move commentary to the best location, and edit as necessary. to “garden” community content; “prune” excessive posts, “transplant” commentary to the best location, and “weed” as necessary. Wikipatterns.com offers a guide to the stages of wiki adoption and a collection of community-building and content-building strategies.

Examples of Wiki Use

Wikis can be vital tools for collecting and leveraging knowledge that would otherwise be scattered throughout an organization; reducing geographic distance; removing boundaries between functional areas; and flattening preexisting hierarchies. Companies have used wikis in a number of ways:

  • At Pixar, all product meetings have an associated wiki to improve productivity. The online agenda ensures that all attendees can arrive knowing the topics and issues to be covered. Anyone attending the meeting (and even those who can’t make it) can update the agenda, post supporting materials, and make comments to streamline and focus in-person efforts.
  • At European investment bank Dresdner Kleinwort Wasserstein, employees use wikis for everything from setting meeting agendas to building multimedia training for new hires. Six months after launch, wiki use had surpassed activity on the firm’s established intranet. Wikis are also credited with helping to reduce Dresdner e-mail traffic by 75 percent.D. Carlin, “Corporate Wikis Go Viral,” BusinessWeek, March 12, 2007.
  • Sony’s PlayStation team uses wikis to regularly maintain one-page overviews on the status of various projects. In this way, legal, marketing, and finance staff can get quick, up-to-date status reports on relevant projects, including the latest projected deadlines, action items, and benchmark progress. Strong security measures are enforced that limit access to only those who must be in the know, since the overviews often discuss products that have not been released.
  • Employees at investment-advisory firm Manning and Napier use a wiki to collaboratively track news in areas of critical interest. Providing central repositories for employees to share articles and update evolving summaries on topics such as health care legislation, enables the firm to collect and focus what would otherwise be fragmented findings and insight. Now all employees can refer to central pages that each serve as a lightning rod attracting the latest and most relevant findings.
  • Intellipedia is a secure wiki built on Intelink, a U.S. government system connecting sixteen spy agencies, military organizations, and the Department of State. The wiki is a “magnum opus of espionage,” handling some one hundred thousand user accounts and five thousand page edits a day. Access is classified in tiers as “unclassified,” “secret,” and “top secret” (the latter hosting 439,387 pages and 57,248 user accounts). A page on the Mumbai terror attacks was up within minutes of the event, while a set of field instructions relating to the use of chlorine-based terror bombs in Iraq was posted and refined within two days of material identification—with the document edited by twenty-three users at eighteen locations.M. Calabrese, “Wikipedia for Spies: The CIA Discovers Web 2.0,” Time, April 8, 2009.

When brought outside the firewall, corporate wikis can also be a sort of value-generation greenhouse, allowing organizations to leverage input from their customers and partners:

  • Intuit has created a “community wiki” that encourages the sharing of experience and knowledge not just regarding Intuit products, such as QuickBooks, but also across broader topics its customers may be interested in, such as industry-specific issues (e.g., architecture, nonprofit) or small business tips (e.g., hiring and training employees). The TurboTax maker has also sponsored TaxAlmanac.org, a wiki-based tax resource and research community.
  • Microsoft leveraged its customer base to supplement documentation for its Visual Studio software development tool. The firm was able to enter the Brazilian market with Visual Studio in part because users had created product documentation in Portuguese.R. King, “No Rest for the Wiki,” BusinessWeek, March 12, 2007.
  • ABC and CBS have created public wikis for the television programs Lost, The Amazing Race, and CSI, among others, offering an outlet for fans, and a way for new viewers to catch up on character backgrounds and complex plot lines.
  • Executive Travel, owned by American Express Publishing, has created a travel wiki for its more than one hundred and thirty thousand readers with the goal of creating what it refers to as “a digital mosaic that in theory is more authoritative, comprehensive, and useful” than comments on a Web site, and far more up-to-date than any paper-based travel guide.R. King, “No Rest for the Wiki,” BusinessWeek, March 12, 2007. Of course, one challenge in running such a corporate effort is that there may be a competing public effort already in place. Wikitravel.org currently holds the top spot among travel-based wikis, and network effects suggest it will likely grow and remain more current than rival efforts.

Don’t Underestimate the Power of Wikipedia

Not only is the nonprofit Wikipedia, with its enthusiastic army of unpaid experts and editors, replacing the three-hundred-year reference reign of Encyclopedia Britannica, Wikipedia entries can impact nearly all large-sized organizations. Wikipedia is the go-to, first-choice reference site for a generation of “netizens,” and Wikipedia entries are invariably one of the top links, often the first link, to appear in Internet search results.

This position means that anyone from top executives to political candidates to any firm large enough to warrant an entry has to contend with the very public commentary offered up in a Wikipedia entry. In the same way that firms monitor their online reputations in blog posts and Twitter tweets, they’ve also got to keep an eye on wikis.

But firms that overreach and try to influence an entry outside of Wikipedia’s mandated neutral point of view (NPOV)An editorial style that is free of bias and opinion. Wikipedia norms dictate that all articles must be written in NPOV., risk a backlash and public exposure. Version tracking means the wiki sees all. Users on computers at right-leaning Fox News were embarrassingly caught editing the wiki page of the lefty pundit and politician Al Franken (a nemesis of Fox’s Bill O’Reilly);A. Bergman, “Wikipedia Is Only as Anonymous as your I.P.,” O’Reilly Radar, August 14, 2007. Sony staffers were flagged as editing the entry for the Xbox game Halo 3;I. Williams, “Sony Caught Editing Halo 3 Wikipedia Entry,” Vnunet.com, September 5, 2007. and none other than Wikipedia founder Jimmy Wales was criticized for editing his own Wikipedia biographyE. Hansen, “Wikipedia Founder Edits Own Bio,” Wired, December 19, 2005.—acts that some consider bad online form at best, and dishonest at worst.

One last point on using Wikipedia for research. Remember that according to its own stated policies, Wikipedia isn’t an original information source; rather, it’s a clearinghouse for verified information. So citing Wikipedia as a reference usually isn’t considered good form. Instead, seek out original (and verifiable) sources, such as those presented via the links at the bottom of Wikipedia entries.

Key Takeaways

  • Wikis can be powerful tools for many-to-many content collaboration, and can be ideal for creating resources that benefit from the input of many such as encyclopedia entries, meeting agendas, and project status documents.
  • The greater the number of wiki users, the more likely the information contained in the wiki will be accurate and grow in value.
  • Wikis can be public or private.
  • The availability of free or low-cost wiki tools can create a knowledge clearinghouse on topics, firms, products, and even individuals. Organizations can seek to harness the collective intelligence (wisdom of crowds) of online communities. The openness of wikis also acts as a mechanism for promoting organizational transparency and accountability.

Questions and Exercises

  1. Visit a wiki, either an established site like Wikipedia, or a wiki service like Socialtext. Make an edit to a wiki entry or use a wiki service to create a new wiki for your own use (e.g., for a class team to use in managing a group project). Be prepared to share your experience with the class.
  2. What factors determine the value of a wiki? Which key concept, first introduced in Chapter 2 “Strategy and Technology: Concepts and Frameworks for Understanding What Separates Winners from Losers”, drives a wiki’s success?
  3. If anyone can edit a wiki, why aren’t more sites crippled by vandalism or by inaccurate or inappropriate content? Are there technical reasons not to be concerned? Are there “social” reasons that can alleviate concern?
  4. Give examples of corporate wiki use, as well as examples where firms used wikis to engage their customers or partners. What is the potential payoff of these efforts? Are there risks associated with these efforts?
  5. Do you feel that you can trust content in wikis? Do you feel this content is more or less reliable than content in print encyclopedias? Than the content in newspaper articles? Why?
  6. Have you ever run across an error in a wiki entry? Describe the situation.
  7. Is it ethical for a firm or individual to edit their own Wikipedia entry? Under what circumstances would editing a Wikipedia entry seem unethical to you? Why? What are the risks a firm or individual is exposed to when making edits to public wiki entries? How do you suppose individuals and organizations are identified when making wiki edits?
  8. Would you cite Wikipedia as a reference when writing a paper? Why or why not?

7.4 Electronic Social Networks

Learning Objectives

After studying this section you should be able to do the following:

  1. Know what social networks are, be able to list key features, and understand how they are used by individuals, groups, and corporations.
  2. Understand the difference between major social networks MySpace, Facebook, and LinkedIn.
  3. Recognize the benefits and risks of using social networks.
  4. Be aware of trends that may influence the evolution of social networks.

Social networksAn online community that allows users to establish a personal profile and communicate with others. Large public social networks include MySpace, Facebook, LinkedIn, and Google’s Orkut. have garnered increasing attention as established networks grow and innovate, new networks emerge, and value is demonstrated. MySpace signed a billion-dollar deal to carry ads from Google’s AdSense network. Meanwhile, privately held Facebook has blown past the flagging MySpace. Its leadership in privacy management, offering new features, allowing third-party applications on its platform, and providing sophisticated analytics tools to corporations and other on-site sponsors have helped the firm move beyond its college roots. LinkedIn, which rounds out the big three U.S. public social networks, has grown to the point where its influence is threatening recruiting sites like Monster.com and CareerBuilder.M. Boyle, “Recruiting: Enough to Make a Monster Tremble,” BusinessWeek, June 25, 2009. It now offers services for messaging, information sharing, and even integration with the BusinessWeek Web site.

Media reports often mention MySpace, Facebook, and LinkedIn in the same sentence. However, while these networks share some common features, they serve very different purposes. MySpace pages are largely for public consumption. Started by musicians, MySpace casts itself as a media discovery tool bringing together users with similar tastes.B. Johnson, “MySpace Bosses Battle to Oust Facebook from Social Networking Top Spot,” The Guardian, March 15, 2010.

Facebook, by contrast, is more oriented towards reinforcing existing social ties between people who already know each other. This difference leads to varying usage patterns. Since Facebook is perceived by users as relatively secure, with only invited “friends” seeing your profile, over a third of Facebook users post their mobile phone numbers on their profile pages.

LinkedIn was conceived from the start as a social network for business users. The site’s profiles act as a sort of digital Rolodex that users update as they move or change jobs. Users can pose questions to members of their network, engage in group discussions, ask for introductions through mutual contacts, and comment on others’ profiles (e.g., recommending a member). Active members find the site invaluable for maintaining professional contacts, seeking peer advice, networking, and even recruiting. Carmen Hudson, Starbucks manager of enterprise staffing, states LinkedIn is “one of the best things for finding midlevel executives.”R. King, “No Rest for the Wiki,” BusinessWeek, March 12, 2007. Such networks are also putting increasing pressure on firms to work particularly hard to retain top talent. While once HR managers fiercely guarded employee directories for fear that a list of talent may fall into the hands of rivals, today’s social networks make it easy for anyone to gain a list of a firm’s staff, complete with contact information.

While these networks dominate in the United States, the network effect and cultural differences work to create islands where other social networks are favored by a particular culture or region. The first site to gain traction in a given market is usually the winner. Google’s Orkut, Mixi, and Cyworld have small U.S. followings, but are among the largest sites in Brazil, Japan, and South Korea. Research by Ipsos Insight also suggests that users in many global markets, including Brazil, South Korea, and China, are more active social networkers than their U.S. counterparts.Ipsos Insights, Online Video and Social Networking Web Sites Set to Drive the Evolution of Tomorrow’s Digital Lifestyle Globally, July 5, 2007.

Perhaps the most powerful (and controversial) feature of most social networks is the feedAn update on an individual’s activities that are broadcast to a member’s contacts or “friends.” Feeds may include activities such as posting messages, photos, or video, joining groups, or installing applications. (or newsfeed). Pioneered by Facebook but now adopted by most services, feeds provide a timely update on the activities of people or topics that an individual has an association with. Feeds can give you a heads-up when someone makes a friend, joins a group, posts a photo, or installs an application.

Feeds are inherently viralIn this context, information or applications that spread rapidly between users.. By seeing what others are doing on a social network, feeds can rapidly mobilize populations and dramatically spread the adoption of applications. Leveraging feeds, it took just ten days for the Facebook group Support the Monks’ Protest in Burma to amass over one hundred and sixty thousand Facebook members. Feeds also helped music app iLike garner three million Facebook users just two weeks after its launch.S. Lacy, Once You’re Lucky, Twice You’re Good: The Rebirth of Silicon Valley and the Rise of Web 2.0 (New York: Gotham Books, 2008); and K. Nicole, “iLike Sees Exponential Growth with Facebook App,” Mashable, June 11, 2007. Its previous Web-based effort took eight months to reach those numbers.

But feeds are also controversial. Many users react negatively to this sort of public broadcast of their online activity, and feed mismanagement can create public relations snafus, user discontent, and potentially open up a site to legal action. Facebook initially dealt with a massive user outcry at the launch of feeds, and faced a subsequent backlash when its Beacon service broadcast user purchases without first explicitly asking their permission, and during attempts to rework its privacy policy and make Facebook data more public and accessible. (See Chapter 8 “Facebook: Building a Business from the Social Graph” for more details.)

Social Networks

The foundation of a social network is the user profile, but utility goes beyond the sort of listing found in a corporate information directory. Typical features of a social network include support for the following:

  • Detailed personal profiles
  • Affiliations with groups, such as alumni, employers, hobbies, fans, health conditions)
  • Affiliations with individuals (e.g., specific “friends”)
  • Private messaging and public discussions
  • Media sharing (text, photos, video)
  • Discovery-fueling feeds of recent activity among members (e.g., status changes, new postings, photos, applications installed)
  • The ability to install and use third-party applications tailored to the service (games, media viewers, survey tools, etc.), many of which are also social and allow others to interact

Corporate Use of Social Networks

Hundreds of firms have established “fan” pages on Facebook and communites on LinkedIn. These are now legitimate customer- and client-engagement platforms that also support advertising. If a customer has decided to press the “like” button of a firm’s Facebook page and become a “fan,” corporate information will appear in their newsfeed, gaining more user attention than the often-ignored ads that run on the sides of social networks. (For more on social networks and advertising, see Chapter 8 “Facebook: Building a Business from the Social Graph”.)

But social networks have also become organizational productivity tools. Many employees have organized groups using publicly available social networking sites because similar tools are not offered by their firms. Workforce Management reported that MySpace had over forty thousand groups devoted to companies or coworkers, while Facebook had over eight thousand.E. Frauenheim, “Social Revolution,” Workforce Management, October 2007. Assuming a large fraction of these groups are focused on internal projects, this demonstrates a clear pent-up demand for corporate-centric social networks (and creates issues as work dialogue moves outside firm-supported services).

Many firms are choosing to meet this demand by implementing internal social network platforms that are secure and tailored to firm needs. At the most basic level, these networks have supplanted the traditional employee directory. Social network listings are easy to update and expand. Employees are encouraged to add their own photos, interests, and expertise to create a living digital identity.

Firms such as Deloitte, Dow Chemical, and Goldman Sachs have created social networks for “alumni” who have left the firm or retired. These networks can be useful in maintaining contacts for future business leads, rehiring former employees (20 percent of Deloitte’s experienced hires are so-called boomerangs, or returning employees), or recruiting retired staff to serve as contractors when labor is tight.R. King, “Social Networks: Execs Use Them Too,” BusinessWeek, November 11, 2006. Maintaining such networks will be critical in industries like IT and health care that are likely to be plagued by worker shortages for years to come.

Social networking can also be important for organizations like IBM, where some 42 percent of employees regularly work from home or client locations. IBM’s social network makes it easier to locate employee expertise within the firm, organize virtual work groups, and communicate across large distances.W. Bulkley, “Playing Well with Others,” Wall Street Journal, June 18, 2007. As a dialogue catalyst, a social network transforms the public directory into a font of knowledge sharing that promotes organization flattening and value-adding expertise sharing.

While IBM has developed their own social network platforms, firms are increasingly turning to third-party vendors like SelectMinds (adopted by Deloitte, Dow Chemical, and Goldman Sachs) and LiveWorld (adopted by Intuit, eBay, the NBA, and Scientific American). Ning allows anyone to create a social network and currently hosts over 2.3 million separate online communities.K. Swisher, “Ning CEO Gina Bianchini to Step Down—Becomes an EIR at Andreessen Horowitz,” AllThingsD, March 15, 2010.

A Little Too Public?

As with any type of social media, content flows in social networks are difficult to control. Embarrassing disclosures can emerge from public systems or insecure internal networks. Employees embracing a culture of digital sharing may err and release confidential or proprietary information. Networks could serve as a focal point for the disgruntled (imagine the activity on a corporate social network after a painful layoff). Publicly declared affiliations, political or religious views, excessive contact, declined participation, and other factors might lead to awkward or strained employee relationships. Users may not want to add a coworker as a friend on a public network if it means they’ll expose their activities, lives, persona, photos, sense of humor, and friends as they exist outside of work. And many firms fear wasted time as employees surf the musings and photos of their peers.

All are advised to be cautious in their social media sharing. Employers are trawling the Internet, mining Facebook, and scouring YouTube for any tip-off that a would-be hire should be passed over. A word to the wise: those Facebook party pics, YouTube videos of open mic performances, or blog postings from a particularly militant period might not age well and may haunt you forever in a Google search. Think twice before clicking the upload button! As Socialnomics author Erik Qualman puts it, “What happens in Vegas stays on YouTube (and Flickr, Twitter, Facebook…).”

Firms have also created their own online communities to foster brainstorming and customer engagement. Dell’s IdeaStorm.com forum collects user feedback and is credited with prompting line offerings, such as the firm’s introduction of a Linux-based laptop.D. Greenfield, “How Companies Are Using I.T. to Spot Innovative Ideas,” InformationWeek, November 8, 2008. At MyStarbucksIdea.com, the coffee giant has leveraged user input to launch a series of innovations ranging from splash sticks that prevent spills in to-go cups, to new menu items. Both IdeaStorm and MyStarbucksIdea run on a platform offered by Salesforce.com that not only hosts these sites but also provides integration into Facebook and other services. Starbucks (the corporate brand with the most Facebook “fans”) has extensively leveraged the site, using Facebook as a linchpin in the “Free Pastry Day” promotion (credited with generating one million in-store visits in a single day) and promotion of the firm’s AIDS-related (Starbucks) RED campaign, which garnered an astonishing three hundred ninety million “viral impressions” through feeds, wall posts, and other messaging.M. Brandau, “Starbucks Brews Up Spot on the List of Top Social Brands in 2008,” Nation’s Restaurant News, April 6, 2009.

Social Networks and Health Care

Dr. Daniel Palestrant often shows a gruesome slide that provides a powerful anecdote for Sermo, the social network for physicians that he cofounded and where he serves as CEO. The image is of an eight-inch saw blade poking through both sides of the bloodied thumb of a construction worker who’d recently arrived in a hospital emergency room. A photo of the incident was posted to Sermo, along with an inquiry on how to remove the blade without damaging tissue or risking a severed nerve. Within minutes replies started coming back. While many replies advised to get a hand surgeon, one novel approach suggested cutting a straw lengthwise, inserting it under the teeth of the blade, and sliding the protected blade out while minimizing further tissue tears.M. Schulder, “50on50: Saw Blade through Thumb. What Would You Do?” CNN, November 4, 2009. The example illustrates how doctors using tools like Sermo can tap into the wisdom of crowds to save thumbs and a whole lot more.

Sermo is a godsend to remote physicians looking to gain peer opinion on confounding cases or other medical questions. The American Medical Association endorsed the site early on,The AMA and Sermo have since broken ties; see B. Comer, “Sermo and AMA Break Ties,” Medical Marketing and Media, July 9, 2009. and the Nature scientific journals have included a “Discuss on Sermo” button alongside the online versions of their medical articles. Doctors are screened and verified to maintain the integrity of participants. Members leverage the site both to share information with each other and to engage in learning opportunities provided by pharmaceutical companies and other firms. Institutional investors also pay for special access to poll Sermo doctors on key questions, such as opinions on pending FDA drug approval. Sermo posts can send valuable warning signals on issues such as disease outbreaks or unseen drug side effects. And doctors have also used the service to rally against insurance company policy changes.

While Sermo focuses on the provider side of the health care equation, a short walk from the firm’s Cambridge, Massachusetts, headquarters will bring one to PatientsLikeMe (PLM), a social network empowering chronically ill patients across a wide variety of disease states. The firm’s “openness policy” is in contrast to privacy rules posted on many sites and encourages patients to publicly track and post conditions, treatments, and symptom variation over time, using the site’s sophisticated graphing and charting tools. The goal is to help others improve the quality of their own care by harnessing the wisdom of crowds.

Todd Small, a multiple sclerosis sufferer, used the member charts and data on PLM to discover that his physician had been undermedicating him. After sharing site data with his doctor, his physician verified the problem and upped the dose. Small reports that the finding changed his life, helping him walk better than he had in a decade and a half and eliminating a feeling that he described as being trapped in “quicksand.”T. Goetz, “Practicing Patients,” New York Times Magazine, March 23, 2008. In another example of PLM’s people power, the site ran its own clinical trial–like experiment to rapidly investigate promising claims that the drug Lithium could improve conditions for ALS (amyotrophic lateral sclerosis) patients. While community efforts did not support these initial claims, a decision was arrived at in months, whereas previous efforts to marshal researchers and resources to focus on the relatively rare disease would have taken many years, even if funding could be found.J. Kane, R. Fichman, J. Gallaugher, and J. Glaser, “Community Relations 2.0,” Harvard Business Review, November 2009.

Both Sermo and PatientsLikeMe are start-ups that are still exploring the best way to fund their efforts for growth and impact. Regardless of where these firms end up, it should be clear from these examples that social media will remain a powerful force on the health care landscape.

Key Takeaways

  • Electronic social networks help individuals maintain contacts, discover and engage people with common interests, share updates, and organize as groups.
  • Modern social networks are major messaging services, supporting private one-to-one notes, public postings, and broadcast updates or “feeds.”
  • Social networks also raise some of the strongest privacy concerns, as status updates, past messages, photos, and other content linger, even as a user’s online behavior and network of contacts changes.
  • Network effects and cultural differences result in one social network being favored over others in a particular culture or region.
  • Information spreads virally via news feeds. Feeds can rapidly mobilize populations, and dramatically spread the adoption of applications. The flow of content in social networks is also difficult to control and sometimes results in embarrassing public disclosures.
  • Feeds have a downside and there have been instances where feed mismanagement has caused user discontent, public relations problems, and the possibility of legal action.
  • The use of public social networks within private organizations is growing, and many organizations are implementing their own, private, social networks.
  • Firms are also setting up social networks for customer engagement and mining these sites for customer ideas, innovation, and feedback.

Questions and Exercises

  1. Visit the major social networks (MySpace, Facebook, LinkedIn). What distinguishes one from the other? Are you a member of any of these services? Why or why not?
  2. How are organizations like Deloitte, Goldman Sachs, and IBM using social networks? What advantages do they gain from these systems?
  3. What factors might cause an individual, employee, or firm to be cautious in their use of social networks?
  4. How do you feel about the feed feature common in social networks like Facebook? What risks does a firm expose itself to if it leverages feeds? How might a firm mitigate these kinds of risks?
  5. What sorts of restrictions or guidelines should firms place on the use of social networks or the other Web 2.0 tools discussed in this chapter? Are these tools a threat to security? Can they tarnish a firm’s reputation? Can they enhance a firm’s reputation? How so?
  6. Why do information and applications spread so quickly within networks like Facebook? What feature enables this? What key promotional concept (described in Chapter 2 “Strategy and Technology: Concepts and Frameworks for Understanding What Separates Winners from Losers”) does this feature foster?
  7. Why are some social networks more popular in some nations than others?
  8. Investigate social networks on your own. Look for examples of their use for fostering political and social movements; for their use in health care, among doctors, patients, and physicians; and for their use among other professional groups or enthusiasts. Identify how these networks might be used effectively, and also look for any potential risks or downside. How are these efforts supported? Is there a clear revenue model, and do you find these methods appropriate or potentially controversial? Be prepared to share your findings with your class.

7.5 Twitter and the Rise of Microblogging

Learning Objectives

After studying this section you should be able to do the following:

  1. Appreciate the rapid rise of Twitter—its scale, scope, and broad appeal.
  2. Understand how Twitter is being used by individuals, organizations, and political movements.
  3. Contrast Twitter and microblogging with Facebook, conventional blogs, and other Web 2.0 efforts.
  4. Consider the commercial viability of the effort, its competitive environment, and concerns regarding limited revenue.

Spawned in 2006 as a side project at the now-failed podcasting start-up Odeo (an effort backed by Blogger.com founder Evan Williams), Twitter has been on a rocket ride. The site’s user numbers have blasted past both mainstream and new media sites, dwarfing New York Times, LinkedIn, and Digg, among others. Reports surfaced of rebuffed buyout offers as high as $500 million.S. Ante, “Facebook’s Thiel Explains Failed Twitter Takeover,” BusinessWeek, March 1, 2009. By the firm’s first developer conference in April 2010, Twitter and its staff of 175 employees had created a global phenomenon embraced by over one hundred million users worldwide.

Twitter is a microbloggingA type of short-message blogging, often made via mobile device. Microblogs are designed to provide rapid notification to their readership (e.g., a news flash, an update on one’s activities), rather than detailed or in-depth comments. Twitter is the most popular microblogging service. service that allows users to post 140-character messages (tweetsA Twitter post, limited to 140 characters.) via the Web, SMSA text messaging standard used by many mobile phones., or a variety of third-party desktop and smartphone applications. The microblog moniker is a bit of a misnomer. The service actually has more in common with Facebook’s status updates and news feeds than it does with traditional blogs. But unlike Facebook, where most users must approve “friends” before they can see status updates, Twitter’s default setting allows for asymmetrical following (although it is possible to set up private Twitter accounts and to block followers).

Sure, there’s a lot of inane “tweeting” going on—lots of meaningless updates that read, “I’m having a sandwich” or “in line at the airport.” But while not every user may have something worthwhile to tweet, many find that Twitter makes for invaluable reading, offering a sense of what friends, customers, thought leaders, and newsmakers are thinking. Twitter leadership has described the service as communicating “The Pulse of the Planet.”E. Schonfeld, “Twitter’s Internal Strategy Laid Bare: To Be ‘The Pulse of The Planet,’” TechCrunch, July 19, 2009. For many, Twitter is a discovery engine, a taste-making machine, a critical source of market intelligence, a source of breaking news, and an instantaneous way to plug into the moment’s zeitgeist.

Many also find Twitter to be an effective tool for quickly blasting queries to friends, colleagues, or strangers who might offer potentially valuable input. Says futurist Paul Saffo, “Instead of creating the group you want, you send it and the group self-assembles.”C. Miller, “Putting Twitter’s World to Use,” New York Times, April 13, 2009. Users can classify comments on a given topic using hash tagsA method for organizing tweets where keywords are preceeded by the # character. (keywords preceded by the “#” or “hash” symbol), allowing others to quickly find related tweets (e.g., #iranelection, #mumbai, #swineflu, #sxsw). Any user can create a hash tag—just type it into your tweet (you may want to search Twitter first to make sure that the tag is not in use by an unrelated topic and that if it is in use, it appropriately describes how you want your tweet classified).

Twitter users have broken news during disasters, terror attacks, and other major events. Dictators fear the people power Twitter enables, and totalitarian governments worldwide have moved to block citizen access to the service (prompting Twitter to work on censor-evading technology). During the 2009 Iranian election protests, the U.S. State Department even asked Twitter to postpone maintenance to ensure the service would continue to be available to support the voice and activism of Iran’s democracy advocates.C. Ruffini, “State Dept. Asked Twitter to Delay Maintenance,” CBS News, June 16, 2009.

Twitter is also emerging as a legitimate business tool. Consider the following commercial examples:

  • Starbucks uses Twitter in a variety of ways. It has run Twitter-based contests and used the service to spread free samples of new products, such as its VIA instant coffee line. Twitter has also been a way for the company to engage customers in its cause-based marketing efforts, such as (Starbucks) RED, which supports (Product) RED. Starbucks has even recruited staff via Twitter and was one of the first firms to participate in Twitter’s advertising model featuring “promoted tweets.”
  • Dell used Twitter to uncover an early warning sign indicating poor design of the keyboard on its Mini 9 Netbook PC. After a series of tweets from early adopters indicated that the apostrophe and return keys were positioned too closely together, the firm dispatched design change orders quickly enough to correct the problem when the Mini 10 was launched just three months later. By December 2009, Dell also claimed to have netted $6.5 million in outlet store sales referred via the Twitter account @DellOutlet (more than 1.5 million followers)K. Eaton, “Twitter Really Works: Makes $6.5 Million in Sales for Dell,” Fast Company, December 8, 2009. and another $1 million from customers who have bounced from the outlet to the new products site.J. Abel, “Dude—Dell’s Making Money Off Twitter!” Wired News, June 12, 2009.
  • Brooklyn Museum patrons can pay an additional $20 a year for access to the private, members-only “1stFans” Twitter feed that shares information on special events and exclusive access to artist content.
  • Twitter is credited with having raised millions via Text-to-Donate and other fundraising efforts following the Haiti earthquake.
  • Twitter can be a boon for sharing time-sensitive information. The True Massage and Wellness Spa in San Francisco tweets last-minute cancellations to tell customers of an unexpected schedule opening. With Twitter, appointments remain booked solid. Gourmet food trucks, popular in many American cities, are also using Twitter to share location and create hipster buzz. Los Angeles’s Kogi Korean Taco Truck now has over sixty thousand followers and uses Twitter to reveal where it’s parked, ensuring long lines of BBQ-craving foodies. Of the firm’s success, owner Roy Choi says, “I have to give all the credit to Twitter.”A. Romano, “Now 4 Restaurant 2.0,” Newsweek, February 28, 2009.
  • Electronics retailer Best Buy has recruited over 2,300 Blue Shirt and Geek Squad staffers to crowdsource Twitter-driven inquiries via @Twelpforce, the firm’s customer service Twitter account. Best Buy staffers register their personal Twitter accounts on a separate Best Buy–run site. Then any registered employees tweeting using the #twelpforce, will automatically have those posts echoed through @Twelpforce, with the employee’s account credited at the end of the tweet. As of November 2009, Twelpforce had provided answers to over 19,500 customer inquiries.Twitter.com, “Case Study: Best Buy Twelpforce,” Twitter 101, http://business.twitter.com/twitter101/case_bestbuy.

Figure 7.1 A Sampling of Tweets Filtered through Best Buy’s @Twelpforce Twitter Account

Surgeons and residents at Henry Ford Hospital have even tweeted during brain surgery (the teaching hospital sees the service as an educational tool). Some tweets are from those so young they’ve got “negative age.” Twitter.com/kickbee is an experimental fetal monitor band that sends tweets when motion is detected: “I kicked Mommy at 08:52.” And savvy hackers are embedding “tweeting” sensors into all sorts of devices. Botanicalls, for example, offers an electronic flowerpot stick that detects when plants need care and sends Twitter status updates to owners (sample post: “URGENT! Water me!”).

Organizations are well advised to monitor Twitter activity related to the firm, as it can act as a sort of canary-in-a-coal mine uncovering emerging events. Users are increasingly using the service as a way to form flash protest crowds. Amazon.com, for example, was caught off guard over a spring 2009 holiday weekend when thousands used Twitter to rapidly protest the firm’s reclassification of gay and lesbian books (hash tag #amazonfail). Others use the platform for shame and ridicule. BP has endured withering ridicule from the satire account @BPGlobalPR (followed by roughly 200,000 two months after the spill).

For all the excitement, many wonder if Twitter is overhyped. Some reports suggest that many Twitter users are curious experimenters who drop the service shortly after signing up.D. Martin, “Update: Return of the Twitter Quitters,” Nielsen Wire, April 30, 2009. This raises the question of whether Twitter is a durable phenomenon or just a fad.

Pundits also wonder if revenues will ever justify initially high valuations and if rivals could usurp Twitter’s efforts with similar features. Thus far, Twitter has been following a “grow-first-harvest-later” approach.J. Murrell, “Twitter Treads Gently into Advertising Minefield,” San Jose Mercury News, April 13, 2010. The site’s rapid rise has allowed it to attract enough start-up capital to enable it to approach revenue gradually and with caution, in the hopes that it won’t alienate users with too much advertising (an approach not unlike Google’s efforts to nurture YouTube). MIT’s Technology Review reports that data sharing deals with Google and Bing may have brought in enough money to make the service profitable in 2009, but that amount was modest (just $25 million).D. Talbot, “Can Twitter Make Money?” Technology Review, March/April 2010. Twitter’s advertising platform is expected to be far more lucrative. Reflecting Twitter’s “deliberately cautious” approach to revenue development, the ad model featuring sponsored ‘‘promoted tweets” rolled out first as part of the search, with distribution to individual Twitter feeds progressing as the firm experiments and learns what works best for users and advertisers.

Another issue—many Twitter users rarely visit the site. Most active users post and read tweets using one of many—often free—applications provided by third parties, such as Seesmic, TweetDeck, and Twhirl. This happens because Twitter made its data available for free to other developers via API (application programming interface)Programming hooks, or guidelines, published by firms that tell other programs how to get a service to perform a task such as send or receive data. For example, Amazon.com provides APIs to let developers write their own applications and Websites that can send the firm orders.. Exposing data can be a good move as it spawned an ecosystem of over one hundred thousand complementary third-party products and services that enhance Twitter’s reach and usefulness (generating network effects from complementary offerings similar to other “platforms” like Windows, iPhone, and Facebook). There are potential downsides to such openness. If users don’t visit Twitter.com, that lessens the impact of any ads running on the site. This creates what is known as the “free rider problemWhen others take advantage of a user or service without providing any sort of reciprocal benefit.,” where users benefit from a service while offering no value in exchange. Encouraging software and service partners to accept ads for a percentage of the cut could lessen the free rider problem.P. Kafka, “Twitter’s Ad Plan: Copy Google,” AllThingsD, February 25, 2010.

When users don’t visit a service, it makes it difficult to spread awareness of new products and features. It can also create branding challenges and customer frustration. Twitter execs lamented that customers were often confused when they searched for “Twitter” in the iPhone App Store and were presented with scores of offerings but none from Twitter itself.D. Goldman, “Twitter Grows Up: Take a Peek Inside,” CNN, April 16, 2010. Twitter’s purchase of the iPhone app Tweetie (subsequently turned into the free “Twitter for iPhone” app) and the launch of its own URL-shortening service (competing with bit.ly and others) signal that Twitter is willing to move into product and service niches and compete with third parties that are reliant on the Twitter ecosystem.

Microblogging does appear to be here to stay, and the impact of Twitter has been deep, broad, stunningly swift, and at times humbling in the power that it wields. But whether Twitter will be a durable, profit-gushing powerhouse remains to be seen. Speculation on Twitter’s future hasn’t prevented many firms from commercializing new microblogging services, and a host of companies have targeted these tools for internal corporate use. Salesforce.com’s Chatter, Socialtext Signals, and Yammer are all services that have been billed as “Twitter for the Enterprise.” Such efforts allow for Twitter-style microblogging that is restricted for participation and viewing by firm-approved accounts.

Key Takeaways

  • While many public and private microblogging services exist, Twitter remains by far the dominant service.
  • Unlike status updates found on services like Facebook and LinkedIn, Twitter’s default supports asymmetric communication, where someone can follow updates without first getting their approval. This function makes Twitter a good choice for anyone cultivating a following—authors, celebrities, organizations, and brand promoters.
  • You don’t need to tweet to get value. Many Twitter users follow friends, firms, celebrities, and thought leaders, quickly gaining access to trending topics.
  • Twitter hash tags (keywords preceded by the # character) are used to organize “tweets” on a given topic. Users can search on hash tags, and many third-party applications allow for Tweets to be organized and displayed by tag.
  • Firms are leveraging Twitter in a variety of ways, including: promotion, customer response, gathering feedback, and time-sensitive communication.
  • Like other forms of social media, Twitter can serve as a hothouse that attracts opinion and forces organizational transparency and accountability.
  • Activists have leveraged the service worldwide, and it has also served as an early warning mechanism in disasters, terror, and other events.
  • Despite its rapid growth and impact, significant questions remain regarding the firm’s durability, revenue prospects, and enduring appeal to initial users.
  • Twitter makes its data available to third parties via an API (application programming interface). The API has helped a rich ecosystem of over seventy thousand Twitter-supporting products and services emerge. But by making the Twitter stream available to third parties, Twitter may suffer from the free rider problem where others firms benefit from Twitter’s service without providing much benefit back to Twitter itself. New ad models may provide a way to distribute revenue-generating content through these services. Twitter has also begun acquiring firms that compete with other players in its ecosystem.

Questions and Exercises

  1. If you don’t already have one, set up a Twitter account and “follow” several others. Follow a diverse group—corporations, executives, pundits, or other organizations. Do you trust these account holders are who they say they are? Why? Which examples do you think use the service most effectively? Which provide the weaker examples of effective Twitter use? Why? Have you encountered Twitter “spam” or unwanted followers? What can you do to limit such experiences? Be prepared to discuss your experiences with class.
  2. If you haven’t done so, install a popular Twitter application such as TweetDeck, Seesmic, or a Twitter client for your mobile device. Why did you select the product you chose? What advantages does your choice offer over simply using Twitter’s Web page? What challenges do these clients offer Twitter? Does the client you chose have a clear revenue model? Is it backed by a viable business?
  3. Visit search.twitter.com. Which Twitter hash tags are most active at this time? Are there other “trending topics” that aren’t associated with hash tags? What do you think of the activity in these areas? Is there legitimate, productive activity happening? Search Twitter on topics, firms, brand names, and issues of interest to you. What do you think of the quality of the information you’ve uncovered on Twitter? Who might find this to be useful?
  4. Why would someone choose to use Twitter over Facebook’s status update, or other services? Which (if either) do you prefer and why?
  5. What do you think of Twitter’s revenue prospects? Is the firm a viable independent service or simply a feature to be incorporated into other social media activity? Advocate where you think the service will be in two years, five, ten. Would you invest in Twitter? Would you suggest that other firms do so? Why?
  6. Assume the role of a manager for your firm. Advocate how the organization should leverage Twitter and other forms of social media. Provide examples of effective use, and cautionary tales, to back up your recommendation.
  7. Some instructors have mandated Twitter for classroom use. Do you think this is productive? Would your professor advocate tweeting during lectures? What are the pros and cons of such use? Work with your instructor to discuss a set of common guidelines for in-class and course use of social media.
  8. As of this writing, Twitter was just rolling out advertising via “promoted tweets.” Perform some additional research. How have Twitter’s attempts to grow revenues fared? How has user growth been trending? Has the firm’s estimated value increased or decreased from the offer figures cited in this chapter? Why?
  9. What do you think of Twitter’s use of the API? What are the benefits of offering an API? What are the downsides? Would you create a company to take advantage of the Twitter API? Why or why not?
  10. Follow this book’s author at http://twitter.com/gallaugher. Tweet him if you run across interesting examples that you think would be appropriate for the next version of the book.

7.6 Other Key Web 2.0 Terms and Concepts

Learning Objectives

After studying this section you should be able to do the following:

  1. Know key terms related to social media, peer production, and Web 2.0, including RSS, folksonomies, mash-ups, location-based services, virtual worlds, and rich media.
  2. Provide examples of the effective business use of these terms and technologies.

RSS

RSSA method for sending/broadcasting data to users who subscribe to a service’s “RSS feed.” Many Web sites and blogs forward headlines to users who subscribe to their “feed,” making it easy to scan headlines and click to access relevant news and information. (an acronym that stands for both “really simple syndication” and “rich site summary”) enables busy users to scan the headlines of newly available content and click on an item’s title to view items of interest, thus sparing them from having to continually visit sites to find out what’s new. Users begin by subscribing to an RSS feed for a Web site, blog, podcast, or other data source. The title or headline of any new content will then show up in an RSS readerA tool for subscribing to and accessing RSS feeds. Most e-mail programs and Web browsers can also act as RSS readers. There are also many Web sites (including Google Reader) that allow users to subscribe to and read RSS feeds.. Subscribe to the New York Times Technology news feed, for example, and you will regularly receive headlines of tech news from the Times. Viewing an article of interest is as easy as clicking the title you like. Subscribing is often as easy as clicking on the RSS icon appearing on the home page of a Web site of interest.

Many firms use RSS feeds as a way to mange information overload, opting to distribute content via feed rather than e-mail. Some even distribute corporate reports via RSS. RSS readers are offered by third-party Web sites such as Google and Yahoo! and they have been incorporated into all popular browsers and most e-mail programs. Most blogging platforms provide a mechanism for bloggers to automatically publish a feed when each new post becomes available. Google’s FeedBurner is the largest publisher of RSS blog feeds, and offers features to distribute content via e-mail as well.

Figure 7.2

RSS readers like Google Reader can be an easy way to scan blog headlines and click through to follow interesting stories.

Figure 7.3

Web sites that support RSS feeds will have an icon in the address bar. Click it to subscribe.

Folksonomies

FolksonomiesKeyword-based classification systems created by user communities (also known as social tagging). (sometimes referred to as social tagging) are keyword-based classification systems created by user communities as they generate and review content. (The label is meant to refer to a people-powered taxonomy.) Bookmarking site Del.icio.us, photo-sharing site Flickr (both owned by Yahoo!), and Twitter’s hash tags all make heavy use of folksonomies.

With this approach, classification schemes emerge from the people most likely to understand them—the users. By leveraging the collective power of the community to identify and classify content, objects on the Internet become easier to locate, and content carries a degree of recommendation and endorsement.

Flickr cofounder Stewart Butterfield describes the spirit of folksonomies, saying, “The job of tags isn’t to organize all the world’s information into tidy categories, it’s to add value to the giant piles of data that are already out there.”D. Terdiman, “Folksonomies Tap People Power,” Wired, February 1, 2005. The Guggenheim Museum in New York City and the San Francisco Museum of Modern Art, among other museums, are taking a folksonomic approach to their online collections, allowing user-generated categories to supplement the specialized lexicon of curators. Amazon.com has introduced a system that allows readers to classify books, and most blog posts and wiki pages allow for social tagging, oftentimes with hot topics indexed and accessible via a “tag cloud” in the page’s sidebar.

Mash-up

Mash-upsThe combination of two or more technologies or data feeds into a single, integrated tool. are combinations of two or more technologies or data feeds into a single, integrated tool. Some of the best known mash-ups leverage Google’s mapping tools. HousingMaps.com combines Craigslist.org listings with Google Maps for a map-based display for apartment hunters. IBM linked together job feeds and Google Maps to create a job-seeker service for victims of Hurricane Katrina. SimplyHired links job listings with Google Maps, LinkedIn listings, and salary data from PayScale.com. And Salesforce.com has tools that allow data from its customer relationship management (CRM) system to be combined with data feeds and maps from third parties.

Mash-ups are made easy by a tagging system called XMLAbbreviation of Extensible Markup Language. A tagging language that can be used to identify data fields made available for use by other applications. For example, programmers may wrap XML tags around elements in an address data stream (e.g., 〈business name〉, 〈street address〉, 〈city〉, 〈state〉) to allow other programs to recognize and use these data items. (for extensible markup language). Site owners publish the parameters of XML data feeds that a service can accept or offer (e.g., an address, price, product descriptions, images). Other developers are free to leverage these public feeds using application programming interfaces (APIs), published instructions on how to make programs call one another, to share data, or to perform tasks. Using APIs and XML, mash-up authors smoosh together seemingly unrelated data sources and services in new and novel ways. Lightweight, browser-friendly software technologies like Ajax and HTML5 can often make a Web site interface as rich as a desktop application, and rapid deployment frameworks like Ruby on Rails will enable and accelerate mash-up creation and deployment.

Location-Based Services

Computing devices increasingly know where you are—and this is creating all sorts of new opportunities for social media. Twitter, Facebook, and Google Buzz are among the many social services that have added location-based options, allowing you to tweet or post a status update attached with a physical location as determined by your phone’s global positioning system (GPS)A network of satellites and supporting technologies used to identify a device’s physical location., triangulation from nearby cell phone towers, or proximity to neighboring Wi-Fi hotspots. This introduces a whole new way to gather and share information. In a new part of town and curious what folks are saying about the spot? Search for tweets tagged as being posted around that location.

Augmented-realityComputer applications that overlay real-world images with computer-generated imagery and data. apps can overlay real data on top of images from a GPS and compass-equipped smartphone. Swivel your iPhone around with Stella Artois’s Bar Finder app open, and it’ll point you to the nearest Stella-equipped watering hole (it’ll also let you text your friends to join you for a drink and call a cab for a safe ride home). Wikitude overlays images appearing through your phone’s camera lens with geotagged data from Wikipedia. Point your Yelp app down the street and activate the monocle feature to see starred reviews hover over the top of establishments that appear on screen.

Boston-based SCVNGR (pronounced “scavenger”), a gaming app, has allowed over four hundred clients, including Princeton, MetLife, and Boston’s Museum of Fine Arts, to create their own mobile phone-based scavenger hunts. The profitable firm has a 90 percent client return rate and had attracted funding from Google Ventures and Highland Capital Partners all before founder Seth Priebatsch turned twenty-one.J. Kincaid, “SCVNGR Raises $4 Million from Google Ventures,” TechCrunch, December 24, 2009.

Perhaps the best known among the location-based pure plays is Foursquare. The service allows players to “check in” at different locations, allowing players to earn “badges” displayed in the app for completing specific achievements (“gym rat” for exercise buffs, “school night” for weeknight bar hoppers). Check into a location more than anyone else and you become that spot’s “mayor.” Foursquare users can follow public location postings from their friends, discovering when someone’s close by and gaining recommendations on new places to explore. Foursquare grew to over one million users roughly one year after the service debuted at the 2009 South by Southwest conference. Firms are now using Foursquare for promotions and to support loyalty programs—offering “mayor specials” or sending out coupons and other incentives when users are nearby. Starbucks, the Bravo television channel, frozen yogurt chain Tasti D-Lite, and the Milwaukee-based burger chain AJ Bombers are among the diverse clients leveraging the service.

Figure 7.4 A Sampling of Location-Aware Apps

Wikitude shows Wikipedia overlays on top of images appearing through the viewfinder. Stella Artois’s Le Bar will point you to establishments offering the brew, and Foursquare offers vendor promotions as well as revealing nearby tweets.

Of course, all this public location sharing raises privacy concerns. The Web site PleaseRobMe.com was created to draw attention to the potentially dangerous issues around real-time location sharing. After a brief demonstration period, the site stopped its real-time aggregation of publicly accessible user-location data and now serves as an awareness site warning of the “stalkerish” side of location-based apps. In most cases, though, users remain firmly in control—determining if they want to keep a visit private or release their locale to verified “app friends” or to the broader online space.

Virtual Worlds

In virtual worldsA computer-generated environment where users present themselves in the form of an avatar, or animated character., users appear in a computer-generated environment in the form of an avatarAn online identity expressed by an animated or cartoon figure., or animated character. Users can customize the look of their avatar, interact with others by typing or voice chat, and can travel about the virtual world by flying, teleporting, or more conventional means.

The most popular general-purpose virtual world is Second Life by Linden Labs, although many others exist. Most are free, although game-oriented worlds, such as World of Warcarft (with ten million active subscribers), charge a fee. Many corporations and organizations have established virtual outposts by purchasing “land” in the world of Second Life, while still others have contracted with networks to create their own, independent virtual worlds.

Most organizations have struggled to commercialize these Second Life forays, but activity has been wide-ranging in its experimentation. Reuters temporarily “stationed” a reporter in Second Life, presidential candidates have made appearances in the virtual world, organizations ranging from Sun Microsystems to Armani have set up virtual storefronts, and there’s a significant amount of virtual mayhem. Second Life “terrorists” have “bombed” virtual outposts run by several organizations, including ABC News, American Apparel, and Reebok.

Even grade schoolers are heavy virtual world users. Many elementary school students get their first taste of the Web through Webkinz, an online world that allows for an animated accompaniment with each of the firm’s plush toys. Webkinz’s parent company, privately held Ganz, doesn’t release financial figures. But according to Compete.com, by the end of 2008 Webkinz.com had roughly the same number of unique visitors as FoxNews.com. The kiddie set virtual world market is considered so lucrative that Disney acquired ClubPenguin for $350 million with agreements to pay another potential three hundred fifty million if the effort hits growth incentives.B. Barnes, “Disney Acquires Web Site for Children,” New York Times, August 2, 2007.

YouTube, Podcasting, and Rich Media

Blogs, wikis, and social networks not only enable sharing text and photos, they also allow for the creation and distribution of audio and video. PodcastsDigital audio or video files served as a series of programs or a multimedia blog. are digital audio files (some also incorporate video), provided as a series of programs. Podcasts range from a sort of media blog, archives of traditional radio and television programs, and regular offerings of original online content. While the term podcast derives from Apple’s wildly successful iPod, podcasts can be recorded in audio formats such as MP3 that can be played on most portable media players. (In perhaps the ultimate concession to the market leader, even the iPod rival Microsoft Zune refers to serialized audio files as podcasts on its navigation menu).

There are many podcast directories, but Apple’s iTunes is by far the largest. Anyone who wants to make a podcast available on iTunes can do so for free. A podcast publisher simply records an audio file, uploads the file to a blog or other hosting server, then sends the RSS feed to Apple (copyrighted material cannot be used without permission, with violators risking banishment from iTunes). Files are discovered in the search feature of the iTunes music store, and listings seamlessly connect the user with the server hosting the podcast. This path creates the illusion that Apple serves the file even though it resides on a publisher’s servers.

While blogs have made stars of some unknowns, the most popular podcasts are from mainstream media outlets. A recent visit to the podcasting section of iTunes showed that eight of the top ten most popular podcasts were high-quality productions of mainstream media programs, including offerings from CBS, Comedy Central, NPR, and PBS. Podcasts are also revolutionizing education, with scores of universities “open sourcing” their classrooms and offering lectures for public consumption via Apple’s iTunesU.

In contrast to iTunes, YouTube actually hosts video on its own servers, so all you need to do is shoot a video and upload it to the site. YouTube is a bastion of amateur video, with most clips shot and uploaded by nonprofessionals. It’s also become a protest site (e.g., “A Comcast Technician Sleeping on my Couch”). However, YouTube has also become a go-to distribution platform for professional content such as ad clips, customer support guides, music videos, TV shows, movies, and more. Much of this rich mediaContent that is more complex than simple text. The term can refer to images, animation, audio, and video. content can be distributed or streamed within another Web site, blog, or social network profile.

Key Takeaways

  • RSS fosters the rapid sharing and scanning of information, including updates from Web 2.0 services such as blogs, wikis, and social networks. RSS feeds can be received via Web browsers, e-mail, cell phones, and special RSS readers.
  • Folksonomies allow users to collaboratively tag and curate online media, making it easy for others to find useful content. Since folksonomies are created by users themselves, they are often more easily understood and embraced than classification schemes imposed by site owners.
  • Mash-ups promote the useful combination of different Web services, such as maps and other information feeds. Mash-up authors leverage technologies such as APIs and XML to combine seemingly unrelated data sources and services in new and novel ways.
  • Location-based services are increasingly combining geolocated data with social media. Users can now quickly see related social media surrounding an area, even overlaying this data on top of maps and images through a phone’s camera lens. Sites like Foursquare are morphing into loyalty and customer-rewards programs. While users are largely in control of sharing location data, some fear privacy and security issues from oversharing.
  • Virtual worlds allow users to interact with and within a computer-generated alternate reality.
  • Internet media is increasingly becoming “richer,” leveraging audio, video, and animation. Organizations and users are creating and distributing rich media online, with interesting content spreading virally.

Questions and Exercises

  1. What is RSS and an RSS reader? Why would an individual use one? Why would a firm use RSS?
  2. Use an RSS reader like Google Reader, or the features built into your e-mail program or browser, and subscribe to RSS feeds. Discuss your experience with the class. Which feeds did you subscribe to? What did you like or not like about using an RSS reader?
  3. If you have a smartphone, download Foursquare or other location-based app. Is this service popular in your community? Research how firms are leveraging these tools for real business value.
  4. Investigate SCVNGR. Many schools are using the tool for orientation programs. Is your school using this? If so, participate in a SCVNGR game on campus. If not, build a case for considering SCVNGR (or similar service) and share this with your student government or student orientation office.
  5. Are privacy concerns from location-based apps valid? What can users do to be safe even while using location-based apps?
  6. Visit Second Life or another virtual world. Create an avatar and look for examples of corporate/commercial involvement. Be prepared to discuss your experience—both positive and negative.
  7. Investigate some of the many virtual worlds targeted at children, including Webkinz, Club Penguin, and Whyville. What are the revenue models for these efforts? How do these sites ensure they are safe for children? Assume the role of a parent—what are the pros and cons of these sites? Which (if any) would you allow your children to participate in? Why? Would you invest in any of these efforts or advise corporations to enter the kid-focused virtual world space? Why or why not?
  8. If you don’t already own it, download iTunes for free, go to the iTunes music store, and explore the free podcast section. Alternatively, find podcasts from another service. Which podcasts seem to be the most popular? Why? Do you use podcasts or other rich media? Why or why not?
  9. Visit YouTube. Identify examples of corporations using the service. Identify examples of customer use. Be prepared to discuss your findings with the class. Do you think Google is making much money with YouTube? Why or why not?
  10. Which firm do you think spends more on the technology infrastructure that powers its service—Apple with iTunes podcasts or Google with YouTube? Explain your answer.
  11. Why would a firm make its data available via XML to use as a mash-up? What can it gain? Are there any risks involved in providing programming hooks that allow the creation of mash-ups?
  12. Give examples of efforts that take advantage of folksonomies. Why are folksonomies considered to be useful?
  13. Do you spend time in rich media Web sites? Which ones? How much time do you or your friends spend on these sites? How would you describe the quality of rich media content found online?
  14. How might a firm use rich media online? What concerns does a firm or individual face with respect to rich media?
  15. Why do you suppose that the most popular podcasts come from established media firms (e.g., Comedy Central, NPR) rather than amateurs, while the top bloggers emerged outside the professional journalist/writer community?

7.7 Prediction Markets and the Wisdom of Crowds

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the concept of the wisdom of crowds as it applies to social networking.
  2. List the criteria necessary for a crowd to be smart.

Many social software efforts leverage what has come to be known as the wisdom of crowdsThe idea that a group of individuals (the crowd), often consisting of untrained amateurs, will collectively have more insight than a single or small group of trained professionals.. In this concept, a group of individuals (the crowd often consists mostly of untrained amateurs), collectively has more insight than a single or small group of trained professionals. Made popular by author James Surowiecki (whose best-selling book was named after the phenomenon), the idea of crowd wisdom is at the heart of wikis, folksonomy tagging systems, and many other online efforts. An article in the journal Nature positively comparing Wikipedia to Encyclopedia Britannica lent credence to social software’s use in harnessing and distilling crowd wisdom.J. Giles, “Special Report: Internet Encyclopedias Go Head to Head,” Nature 438, no. 15 (December 14, 2005): 900–901.

The crowd isn’t always right, but in many cases where topics are complex, problems are large, and outcomes are uncertain, a large, diverse group may bring collective insight to problem solving that one smart guy or a professional committee lacks. One technique for leveraging the wisdom of crowds is a prediction marketPolling a diverse crowd and aggregating opinions in order to form a forecast of an eventual outcome., where a diverse crowd is polled and opinions aggregated to form a forecast of an eventual outcome. The concept is not new. The stock market is arguably a prediction market, with a stock price representing collective assessment of the discounted value of a firm’s future earnings. But Internet technologies are allowing companies to set up prediction markets for exploring all sorts of problems.

Consider Best Buy, where employees are encouraged to leverage the firm’s TagTrade prediction market to make forecasts, and are offered small gifts as incentives for participation. The idea behind this incentive program is simple: the “blue shirts” (Best Buy employees) are closest to customers. They see traffic patterns and buying cycles, can witness customer reactions first hand, and often have a degree of field insight not available to senior managers at the company’s Minneapolis headquarters. Harness this collective input and you’ve got a group brain where, as wisdom of crowds proponents often put it, “the we is greater than the me.” When Best Buy asked its employees to predict gift card sales, the “crowd’s” collective average answer was 99.5 percent accurate; experts paid to make the prediction were off by 5 percent. Another experiment predicting holiday sales was off by only 1/10 of 1 percent. The experts? Off by 7 percent!P. Dvorak, “Best Buy Taps ‘Prediction Market,’” Wall Street Journal, September 16, 2008; and Renée Dye, “The Promise of Prediction Markets: A Roundtable,” McKinsey Quarterly (2008): 83–93.

In an article in the McKinsey Quarterly, Surowiecki outlined several criteria necessary for a crowd to be “smart.”R. Dye, “The Promise of Prediction Markets: A Roundtable,” McKinsey Quarterly (2008): 83–93. The crowd must

  • be diverse, so that participants are bringing different pieces of information to the table,
  • be decentralized, so that no one at the top is dictating the crowd’s answer,
  • offer a collective verdict that summarizes participant opinions,
  • be independent, so that each focuses on information rather than the opinions of others.

Google, which runs several predictive markets, underscored these principles when it found that predictions were less accurate when users were geographically proximate, meaning folks in the same work group who sat near one another typically thought too much alike.B. Cowgill, J. Wolfers, and E. Zitzewitz, “Using Prediction Markets to Track Information Flows: Evidence from Google,” working paper accessed November 30, 2009, via http://bocowgill.com/GooglePredictionMarketPaper.pdf. Poorer predictive outcomes likely resulted because these relatively homogeneous clusters of users brought the same information to the table (yet another reason why organizations should hire and cultivate diverse teams).

Many firms run predictive markets to aid in key forecasts, and with the potential for real financial payoff. But University of Chicago law professor Todd Henderson warns predictive markets may also hold legal and ethical challenges. The Securities and Exchange Commission may look askance at an employee who gets a heads-up in a predictive market that says a certain drug is going to be approved or fail clinical trials. If she trades on this information is she an insider, subject to prosecution for exploiting proprietary data? Disclosure issues are unclear. Gambling laws are also murky, with Henderson uncertain as to whether certain predictive markets will be viewed as an unregulated form of betting.R. Dye, “The Promise of Prediction Markets: A Roundtable,” McKinsey Quarterly 2 (2008): 83–93.

Publicly accessible prediction markets are diverse in their focus. The Iowa Electronic Market attempts to guess the outcome of political campaigns, with mixed results. Farecast (now part of Microsoft’s Bing knowledge engine) claims a 75 percent accuracy rate for forecasting the future price of airline tickets.“Audit Reveals Farecast Predictive Accuracy at 74.5 percent,” farecast.live.com, May 18, 2007, http://www.prnewswire.com/news-releases/farecast-launches-new-tools-to -help-savvy-travelers-catch-elusive-airfare-price-drops-this-summer-58165652.html. The Hollywood Stock Exchange allows participants to buy and sell prediction shares of movies, actors, directors, and film-related options. The exchange, now owned by investment firm Cantor Fitzgerald, has picked Oscar winners with 90 percent accuracy.J. Surowiecki, “Crowdsourcing the Crystal Ball,” Forbes, October 15, 2007. And at HedgeStreet.com, participants can make microbets, wagering as little as ten dollars on the outcome of economic events, including predictions on the prices of homes, gold, foreign currencies, oil, and even the economic impact of hurricanes and tropical storms. HedgeStreet is considered a market and is subject to oversight by the Commodity Futures Trading Commission.E. Lambert, “Hedging for Dummies,” Forbes, March 13, 2006, 70–72.

Key Takeaways

  • Many Web 2.0 efforts allow firms to tap the wisdom of crowds, identifying collective intelligence.
  • Prediction markets tap crowd opinion with results that are often more accurate than the most accurate expert forecasts and estimates.
  • Prediction markets are most accurate when tapping the wisdom of a diverse and variously skilled and experienced group, and are least accurate when participants are highly similar.

Questions and Exercises

  1. What makes for a “wise” crowd? When might a crowd not be so wise?
  2. Find a prediction market online and participate in the effort. Be prepared to share your experience with your class, including any statistics of predictive accuracy, participant incentives, business model of the effort, and your general assessment of the appeal and usefulness of the effort.
  3. Brainstorm on the kinds of organizations that might deploy prediction markets. Why might you think the efforts you suggest and advocate would be successful?
  4. In what ways are legal issues of concern to prediction market operators?

7.8 Crowdsourcing

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the value of crowdsourcing.
  2. Identify firms that have used crowdsourcing successfully.

The power of Web 2.0 also offers several examples of the democratization of production and innovation. Need a problem solved? Offer it up to the crowd and see if any of their wisdom offers a decent result. This phenomenon, known as crowdsourcingThe act of taking a job traditionally performed by a designated agent (usually an employee) and outsourcing it to an undefined generally large group of people in the form of an open call., has been defined by Jeff Howe, founder of the blog crowdsourcing.com and an associate editor at Wired, as “the act of taking a job traditionally performed by a designated agent (usually an employee) and outsourcing it to an undefined, generally large group of people in the form of an open call.”J. Howe, “The Rise of Crowdsourcing,” Wired, June 2006.

Can the crowd really do better than experts inside a firm? At least one company has literally struck gold using crowdsourcing. As told by Don Tapscott and Anthony Williams in their book Wikinomics, mining firm Goldcorp was struggling to gain a return from its 55,000-acre Canadian property holdings. Executives were convinced there was gold “in them thar hills,” but despite years of efforts, the firm struggled to strike any new pay dirt. CEO Rob McEwen, a former mutual fund manager without geology experience who unexpectedly ended up running Goldcorp after a takeover battle, then made what seemed a Hail Mary pass—he offered up all the firm’s data, on the company’s Web site. Along with the data, McEwen ponied up $575,000 from the firm as prize money for the Goldcorp Challenge to anyone who came up with the best methods and estimates for reaping golden riches. Releasing data was seen as sacrilege in the intensely secretive mining industry, but it brought in ideas the firm had never considered. Taking the challenge was a wildly diverse group of “graduate students, consultants, mathematicians, and military officers.” Eighty percent of the new targets identified by entrants yielded “substantial quantities of gold.” The financial payoff? In just a few years a one-hundred-million-dollar firm grew into a nine-billion-dollar titan. For Goldcorp, the crowd coughed up serious coin.

Netflix followed Goldcorp’s lead, offering anonymous data to any takers, along with a one-million-dollar prize to the first team that could improve the accuracy of movie recommendations by 10 percent. Top performers among the over thirty thousand entrants included research scientists from AT&T Labs, researchers from the University of Toronto, a team of Princeton undergrads, and the proverbial “guy in a garage” (and yes, that was his team name). Frustrated for nearly three years, it took a coalition of four teams from Austria, Canada, Israel, and the United States to finally cross the 10 percent threshold. The winning team represented an astonishing brain trust that Netflix would never have been able to harness on its own.S. Lohr, “And the Winner of the $1 Million Netflix Prize (Probably) Is…” New York Times, June 26, 2009.

Other crowdsourcers include Threadless.com, which produces limited run t-shirts with designs users submit and vote on. Marketocracy runs stock market games and has created a mutual fund based on picks from the 100 top-performing portfolios. Just under seven years into the effort, the firm’s m100 Index reports a 75 percent return versus 35 percent for the S&P 500. The St. Louis Cardinals baseball team is even crowdsourcing. The club’s One for the Birds contest calls for the fans to submit scouting reports on promising players, as the team hopes to broaden its recruiting radar beyond its classic recruiting pool of Division I colleges.

There are several public markets for leveraging crowdsourcing for innovation, or as an alternative to standard means of production. Waltham, Massachusetts–based InnoCentive allows “seekers” to offer cash prizes ranging from ten to one hundred thousand dollars. Over one hundred twenty thousand “solvers” have registered to seek solutions for tasks put forward by seekers that include Dow Chemical, Eli Lilly, and Procter & Gamble. Among the findings offered by the InnoCentive crowd are a biomarker that measures progression of ALS. Amazon.com has even created an online marketplace for crowdsourcing called Mechanical Turk. Anyone with a task to be completed or problem to be solved can put it up for Amazon, setting their price for completion or solution. For its role, Amazon takes a small cut of the transaction. And alpha geeks looking to prove their code chops can turn to TopCoder, a firm that stages coding competitions that deliver real results for commercial clients such as ESPN. By 2009, TopCoder contests had attracted over 175,000 participants from 200 countries.M. Brandel, “Should Your Company ‘Crowdsource’ Its Next Project?” Computerworld, December 6, 2007; M. Brandel, “Crowdsourcing: Are You Ready to Ask the World for Answers?” Computerworld, March 3, 2008; and TopCoder, 2009, http://topcoder.com/home.

Not all crowdsourcers are financially motivated. Some benefit by helping to create a better service. Facebook leveraged crowd wisdom to develop versions of its site localized in various languages. Facebook engineers designated each of the site’s English words or phrases as a separate translatable object. Members were then invited to translate the English into other languages, and rated the translations to determine which was best. Using this form of crowdsourcing, fifteen hundred volunteers cranked out Spanish Facebook in a month. It took two weeks for two thousand German speakers to draft Deutsch Facebook. How does the Facebook concept of “poke” translate around the world? The Spaniards decided on “dar un toque,” Germans settled on “anklopfen,” and the French went with “envoyer un poke.”D. Kirkpatrick, “Help Wanted: Adults on Facebook,” Fortune, March 21, 2008. Vive le crowd!

Key Takeaways

  • Crowdsourcing tackles challenges through an open call to a broader community of potential problem solvers. Examples include Goldcorp’s discovering of optimal mining locations in land it already held, Facebook’s leverage of its users to create translations of the site for various international markets, and Netflix’s solicitation of improvements to its movie recommendation software.
  • Several firms run third-party crowdsourcing forums, among them InnoCentive for scientific R&D, TopCoder for programming tasks, and Amazon’s Mechanical Turk for general work.

Questions and Exercises

  1. What is crowdsourcing? Give examples of organizations that are taking advantage of crowdsourcing and be prepared to describe these efforts.
  2. What ethical issues should firms be aware of when considering crowdsourcing? Are there other concerns firms may have when leveraging this technique?
  3. Assume the role of a manager or consultant. Recommend a firm and a task that would be appropriate for crowdsourcing. Justify your choice, citing factors such as cost, breadth of innovation, time, constrained resources, or other factors. How would you recommend the firm conduct this crowdsourcing effort?

7.9 Get SMART: The Social Media Awareness and Response Team

Learning Objectives

After studying this section you should be able to do the following:

  1. Illustrate several examples of effective and poor social media use.
  2. Recognize the skills and issues involved in creating and staffing an effective social media awareness and response team (SMART).
  3. List and describe key components that should be included in any firm’s social media policy.
  4. Understand the implications of ethical issues in social media such as “sock puppetry” and “astroturfing” and provide examples and outcomes of firms and managers who used social media as a vehicle for dishonesty.
  5. List and describe tools for monitoring social media activity relating to a firm, its brands, and staff.
  6. Understand issues involved in establishing a social media presence, including the embassy approach, openness, and staffing.
  7. Discuss how firms can engage and respond through social media, and how companies should plan for potential issues and crises.

For an example of how outrage can go viral, consider Dave Carroll.The concepts in this section are based on work by J. Kane, R. Fichman, J. Gallaugher, and J. Glasser, many of which are covered in the article “Community Relations 2.0,” Harvard Business Review, November 2009. The Canadian singer-songwriter was traveling with his band Sons of Maxwell on a United Airlines flight from Nova Scotia to Nebraska when, during a layover at Chicago’s O’Hare International Airport, Carroll saw baggage handlers roughly tossing his guitar case. The musician’s $3,500 Taylor guitar was in pieces by the time it arrived in Omaha. In the midst of a busy tour schedule, Carroll didn’t have time to follow up on the incident until after United’s twenty-four-hour period for filing a complaint for restitution had expired. When United refused to compensate him for the damage, Carroll penned the four-minute country ditty “United Breaks Guitars,” performed it in a video, and uploaded the clip to YouTube (sample lyrics: “I should have gone with someone else or gone by car…’cuz United breaks guitars”). Carroll even called out the unyielding United rep by name. Take that, Ms. Irlwig! (Note to customer service reps everywhere: you’re always on.)

The clip went viral, receiving 150,000 views its first day and five million more by the next month. Well into the next year, “United Breaks Guitars” remained the top result on YouTube when searching the term “United.” No other topic mentioning that word—not “United States,” “United Nations,” or “Manchester United”—ranked ahead of this one customer’s outrage.

Video

(click to see video)

Dave Carroll’s ode to his bad airline experience, “United Breaks Guitars,” went viral, garnering millions of views.

Scarring social media posts don’t just come from outside the firm. Earlier that same year employees of Domino’s Pizza outlet in Conover, North Carolina, created what they thought would be a funny gross-out video for their friends. Posted to YouTube, the resulting footage of the firm’s brand alongside vile acts of food prep was seen by over one million viewers before it was removed. Over 4.3 million references to the incident can be found on Google, and many of the leading print and broadcast outlets covered the story. The perpetrators were arrested, the Domino’s storefront where the incident occurred was closed, and the firm’s president made a painful apology (on YouTube, of course).

Not all firms choose to aggressively engage social media. As of this writing some major brands still lack a notable social media presence (Apple comes immediately to mind). But your customers are there and they’re talking about your organization, its products, and its competitors. Your employees are there, too, and without guidance, they can step on a social grenade with your firm left to pick out the shrapnel. Soon, nearly everyone will carry the Internet in their pocket. Phones and MP3 players are armed with video cameras capable of recording every customer outrage, corporate blunder, ethical lapse, and rogue employee. Social media posts can linger forever online, like a graffiti tag attached to your firm’s reputation. Get used to it—that genie isn’t going back in the bottle.

As the “United Breaks Guitars” and “Domino’s Gross Out” incidents show, social media will impact a firm whether it chooses to engage online or not. An awareness of the power of social media can shape customer support engagement and crisis response, and strong corporate policies on social media use might have given the clueless Domino’s pranksters a heads-up that their planned video would get them fired and arrested. Given the power of social media, it’s time for all firms to get SMARTThe social media awareness and response team. A group tasked with creating policies and providing support, training, guidance, and development expertise for and monitoring of a firm’s social media efforts., creating a social media awareness and response team. While one size doesn’t fit all, this section details key issues behind SMART capabilities, including creating the social media team, establishing firmwide policies, monitoring activity inside and outside the firm, establishing the social media presence, and managing social media engagement and response.

Creating the Team

Firms need to treat social media engagement as a key corporate function with clear and recognizable leadership within the organization. Social media is no longer an ad hoc side job or a task delegated to an intern. When McDonald’s named its first social media chief, the company announced that it was important to have someone “dedicated 100% of the time, rather than someone who’s got a day job on top of a day job.”E. York, “McDonald’s Names First Social-Media Chief,” Chicago Business, April 13, 2010. Firms without social media baked into employee job functions often find that their online efforts are started with enthusiasm, only to suffer under a lack of oversight and follow-through. One hotel operator found franchisees were quick to create Facebook pages, but many rarely monitored them. Customers later notified the firm that unmonitored hotel “fan” pages contained offensive messages—a racist rant on one, paternity claims against an employee on another.

Organizations with a clearly established leadership role for social media can help create consistency in firm dialogue; develop and communicate policy; create and share institutional knowledge; provide training, guidance, and suggestions; offer a place to escalate issues in the event of a crisis or opportunity; and catch conflicts that might arise if different divisions engage without coordination.

While firms are building social media responsibility into job descriptions, also recognize that social media is a team sport that requires input from staffers throughout an organization. The social media team needs support from public relations, marketing, customer support, HR, legal, IT, and other groups, all while acknowledging that what’s happening in the social media space is distinct from traditional roles in these disciplines. The team will hone unique skills in technology, analytics, and design, as well as skills for using social media for online conversations, listening, trust building, outreach, engagement, and response. As an example of the interdisciplinary nature of social media practice, consider that the social media team at Starbucks (regarded by some as the best in the business) is organized under the interdisciplinary “vice president of brand, content, and online.”Starbucks was named the best firm for social media engagement in a study by Altimeter Group and WetPaint. See the 2009 ENGAGEMENTdb report at http://engagementdb.com.

Also note that while organizations with SMARTs (social media teams) provide leadership, support, and guidance, they don’t necessarily drive all efforts. GM’s social media team includes representatives from all the major brands. The idea is that employees in the divisions are still the best to engage online once they’ve been trained and given operational guardrails. Says GM’s social media chief, “I can’t go in to Chevrolet and tell them ‘I know your story better than you do, let me tell it on the Web.’”C. Barger, talk at the Social Media Club of Detroit, November 18, 2009. Also available via UStream and DigitalMarketingZen.com. Similarly, the roughly fifty Starbucks “Idea Partners” who participate in MyStarbucksIdea are specialists. Part of their job is to manage the company’s social media. In this way, conversations about the Starbucks Card are handled by card team experts, and merchandise dialogue has a product specialist who knows that business best. Many firms find that the social media team is key for coordination and supervision (e.g., ensuring that different divisions don’t overload consumers with too much or inconsistent contact), but the dynamics of specific engagement still belong with the folks who know products, services, and customers best.

Responsibilities and Policy Setting

In an age where a generation has grown up posting shoot-from-the-hip status updates and YouTube is seen as a fame vehicle for those willing to perform sensational acts, establishing corporate policies and setting employee expectations are imperative for all organizations. The employees who don’t understand the impact of social media on the firm can do serious damage to their employers and their careers (look to Domino’s for an example of what can go wrong).

Many experts suggest that a good social media policy needs to be three things: “short, simple, and clear.”J. Soat, “7 Questions Key to Social Networking Success,” InformationWeek, January 16, 2010. Fortunately, most firms don’t have to reinvent the wheel. Several firms, including Best Buy, IBM, Intel, The American Red Cross, and Australian telecom giant Telstra, have made their social media policies public.

Most guidelines emphasize the “three Rs”: representation, responsibility, and respect.

  • Representation. Employees need clear and explicit guidelines on expectations for social media engagement. Are they empowered to speak on behalf of the firm? If they do, it is critical that employees transparently disclose this to avoid legal action. U.S. Federal Trade Commission rules require disclosure of relationships that may influence online testimonial or endorsement. On top of this, many industries have additional compliance requirements (e.g., governing privacy in the health and insurance fields, retention of correspondence and disclosure for financial services firms). Firms may also want to provide guidelines on initiating and conducting dialogue, when to respond online, and how to escalate issues within the organization.
  • Responsibility. Employees need to take responsibility for their online actions. Firms must set explicit expectations for disclosure, confidentiality and security, and provide examples of engagement done right, as well as what is unacceptable. An effective social voice is based on trust, so accuracy, transparency, and accountability must be emphasized. Consequences for violations should be clear.
  • Respect. Best Buy’s policy for its Twelpforce explicitly states participants must “honor our differences” and “act ethically and responsibly.” Many employees can use the reminder. Sure customer service is a tough task and every rep has a story about an unreasonable client. But there’s a difference between letting off steam around the water cooler and venting online. Virgin Atlantic fired thirteen of the airline’s staffers after they posted passenger insults and inappropriate inside jokes on Facebook.L. Conway, “Virgin Atlantic Sacks 13 Staff for Calling Its Flyers ‘Chavs,’” The Independent, November 1, 2008.

Policies also need to have teeth. Remember, a fourth “R” is at stake—reputation (both the firm’s and the employee’s). Violators should know the consequences of breaking firm rules and policies should be backed by action. Best Buy’s policy simply states, “Just in case you are forgetful or ignore the guidelines above, here’s what could happen. You could get fired (and it’s embarrassing to lose your job for something that’s so easily avoided).”

Despite these concerns, trying to micromanage employee social media use is probably not the answer. At IBM, rules for online behavior are surprisingly open. The firm’s code of conduct reminds employees to remember privacy, respect, and confidentiality in all electronic communications. Anonymity is not permitted on IBM’s systems, making everyone accountable for their actions. As for external postings, the firm insists that employees not disparage competitors or reveal customers’ names without permission and asks that any employee posts from IBM accounts or that mention the firm also include disclosures indicating that opinions and thoughts shared publicly are the individual’s and not Big Blue’s.

Some firms have more complex social media management challenges. Consider hotels and restaurants where outlets are owned and operated by franchisees rather than the firm. McDonald’s social media team provides additional guidance so that regional operations can create, for example, a Twitter handle (e.g., @mcdonalds_cincy) that handle a promotion in Cincinnati that might not run in other regions.E. York, “McDonald’s Names First Social-Media Chief,” Chicago Business, April 13, 2010. A social media team can provide coordination while giving up the necessary control. Without this kind of coordination, customer communication can quickly become a mess.

Training is also a critical part of the SMART mandate. GM offers an intranet-delivered video course introducing newbies to the basics of social media and to firm policies and expectations. GM also trains employees to become “social media proselytizers and teachers.” GM hopes this approach enables experts to interact directly with customers and partners, allowing the firm to offer authentic and knowledgeable voices online.

Training should also cover information security and potential threats. Social media has become a magnet for phishing, virus distribution, and other nefarious online activity. Over one-third of social networking users claim to have been sent malware via social networking sites (see Chapter 13 “Information Security: Barbarians at the Gateway (and Just About Everywhere Else)”). The social media team will need to monitor threats and spread the word on how employees can surf safe and surf smart.

Since social media is so public, it’s easy to amass examples of what works and what doesn’t, adding these to the firm’s training materials. The social media team provides a catch point for institutional knowledge and industry best practice; and the team can update programs over time as new issues, guidelines, technologies, and legislation emerge.

The social media space introduces a tension between allowing expression (among employees and by the broader community) and protecting the brand. Firms will fall closer to one end or the other of this continuum depending on compliance requirements, comfort level, and goals. Expect the organization’s position to move. Firms will be cautious as negative issues erupt, others will jump in as new technologies become hot and early movers generate buzz and demonstrate results. But it’s the SMART responsibility to avoid knee-jerk reaction and to shepherd firm efforts with the professionalism and discipline of other management domains.

Astroturfing and Sock Puppets

Social media can be a cruel space. Sharp-tongued comments can shred a firm’s reputation and staff might be tempted to make anonymous posts defending or promoting the firm. Don’t do it! Not only is it a violation of FTC rules, IP addresses and other online breadcrumbs often leave a trail that exposes deceit.

Whole Foods CEO John Mackey fell victim to this kind of temptation, but his actions were eventually, and quite embarrassingly, uncovered. For years, Mackey used a pseudonym to contribute to online message boards, talking up Whole Foods stock and disparaging competitors. When Mackey was unmasked, years of comments were publicly attributed to him. The New York Times cited one particularly cringe-worthy post where Mackey used the pseudonym to complement his own good looks, writing, “I like Mackey’s haircut. I think he looks cute!”A. Martin, “Whole Foods Executive Used Alias,” New York Times, July 12, 2007.

Fake personas set up to sing your own praises are known as sock puppetsA fake online persona created to promote a particular point of view, often in praise of a firm, product, or individual. Be aware that the use of undisclosed relationships in endorsements is a violation of U.S. Federal Trade Comission rules. among the digerati, and the practice of lining comment and feedback forums with positive feedback is known as astroturfingEngineering the posting of positive comments and reviews of a firm’s product and services (or negative ones of a firm’s competitors). Many ratings sites will penalize firms that offer incentives for positive feedback posts.. Do it and it could cost you. The firm behind the cosmetic procedure known as the Lifestyle Lift was fined $300,000 in civil penalties after the New York Attorney General’s office discovered that the firm’s employees had posed as plastic surgery patients and wrote glowing reviews of the procedure.C. Cain Miller, “Company Settles Case of Reviews It Faked,” July 14, 2009.

Review sites themselves will also take action. TripAdvisor penalizes firms if it’s discovered that customers are offered some sort of incentive for posting positive reviews. The firm also employs a series of sophisticated automated techniques as well as manual staff review to uncover suspicious activity. Violators risk penalties that include being banned from the service.

Your customers will also use social media keep you honest. Several ski resorts have been embarrassed when tweets and other social media posts exposed them as overstating snowfall results. There’s even an iPhone app skiers can use to expose inaccurate claims.L. Rathke, “Report: Ski Resorts Exaggerate Snowfall Totals,” USA Today, January 29, 2010.

So keep that ethical bar high—you never know when technology will get sophisticated enough to reveal wrongdoings.

Monitoring

Concern over managing a firm’s online image has led to the rise of an industry known as online reputation managementThe process of tracking and responding to online mentions of a product, organization, or individual. Services supporting online reputation management range from free Google Alerts to more sophisticated services that blend computer-based and human monitoring of multiple media channels.. Firms specializing in this field will track a client firm’s name, brand, executives’ names, or other keywords, reporting online activity and whether sentiment trends toward the positive or negative.

But social media monitoring is about more than about managing one’s reputation; it also provides critical competitive intelligence, it can surface customer support issues, and it can uncover opportunities for innovation and improvement. Firms that are quick to lament the very public conversations about their brands happening online need to embrace social media as an opportunity to learn more.

Resources for monitoring social media are improving all the time, and a number of tools are available for free. All firms can take advantage of Google Alerts, which flag blog posts, new Web pages, and other publicly accessible content, regularly delivering a summary of new links to your mailbox (for more on using Google for intelligence gathering, see Chapter 14 “Google: Search, Online Advertising, and Beyond”). Twitter search and third-party Twitter clients like TweetDeck can display all mentions of a particular term. Tools like Twitrratr will summarize mentions of a phrase and attempt to classify tweets as “positive,” “neutral,” or “negative.”

Figure 7.5

Tools like Twitrratr attempt to classify the sentiment behind tweets mentioning a key word or phrase. Savvy firms can mine comments for opportunities to provide thoughtful customer service (like the suggestion at the top right to provide toothpaste for those who lose it in U.S. airport security).

Facebook provides a summary of fan page activity to administrators (including stats on visits, new fans, wall posts, etc.), while Facebook’s Insights tool measures user exposure, actions, and response behavior relating to a firm’s Facebook pages and ads.

Bit.ly and many other URL-shortening services allow firms to track Twitter references to a particular page. Since bit.ly applies the same shortened URL to all tweets pointing to a page, it allows firms to follow not only if a campaign has been spread through “retweeting” but also if new tweets were generated outside of a campaign. Graphs plot click-throughs over time, and a list of original tweets can be pulled up to examine what commentary accompanied a particular link.

Location-based services like Foursquare have also rolled out robust tools for monitoring how customers engage with firms in the brick-and-mortar world. Foursquare’s analytics and dashboard present firms with a variety of statistics, such as who has “checked in” and when, a venue’s male-to-female ratio, and which times of day are more active for certain customers. “Business owners will also be able to offer instant promotions to try to engage new customers and keep current ones.”N. Bolton, “Foursquare Introduces New Tools for Businesses,” March 9, 2010. Managers can use the tools to notice if a once-loyal patron has dropped off the map, potentially creating a special promotion to lure her back.

Monitoring should also not be limited to customers and competitors. Firms are leveraging social media both inside their firms and via external services (e.g., corporate groups on Facebook and LinkedIn), and these spaces should also be on the SMART radar. This kind of monitoring can help firms keep pace with employee sentiment and insights, flag discussions that may involve proprietary information or other inappropriate topics, and provide guidance for those who want to leverage social media for the firm’s staff—that is, anything from using online tools to help organize the firm’s softball league to creating a wiki for a project group. Social media are end-user services that are particularly easy to deploy but that can also be used disastrously and inappropriately, so it’s vital for IT experts and other staffers on the social media team to be visible and available, offering support and resources for those who want to take a dip into social media’s waters.

Establishing a Presence

Firms hoping to get in on the online conversation should make it easy for their customers to find them. Many firms take an embassyIn the context of social media, an established online presence where customers can reach and interact with the firm. An effective embassy approach uses a consistent firm name in all its social media properties. approach to social media, establishing presence at various services with a consistent name. Think facebook.com/starbucks, twitter.com/starbucks, youtube.com/starbucks, flickr.com/starbucks, and so on. Corporate e-mail and Web sites can include icons linking to these services in a header or footer. The firm’s social media embassies can also be highlighted in physical space such as in print, on bags and packaging, and on store signage. Firms should try to ensure that all embassies carry consistent design elements, so users see familiar visual cues that underscore they are now at a destination associated with the organization.

As mentioned earlier, some firms establish their own communities for customer engagement. Examples include Dell’s IdeaStorm and MyStarbucksIdea. Not every firm has a customer base that is large and engaged enough to support hosting its own community. But for larger firms, these communities can create a nexus for feedback, customer-driven innovation, and engagement.

Customers expect an open dialogue, so firms engaging online should be prepared to deal with feedback that’s not all positive. Firms are entirely within their right to screen out offensive and inappropriate comments. Noting this, firms might think twice before turning on YouTube comments (described as “the gutter of the Internet” by one leading social media manager).Brad Nelson, presentation at the Social Media Conference NW, Mount Vernon, WA, March 25, 2010. Such comments could expose employees or customers profiled in clips to withering, snarky ridicule. However, firms engaged in curating their forums to present only positive messages should be prepared for the community to rebel and for embarrassing cries of censorship to be disclosed. Firms that believe in the integrity of their work and the substance of their message shouldn’t be afraid. While a big brand like Starbucks is often a target of criticism, social media also provides organizations with an opportunity to respond fairly to that criticism and post video and photos of the firm’s efforts. In Starbucks’ case, the firm shares its work investing in poor coffee-growing communities as well as efforts to support AIDS relief. A social media presence allows a firm to share these works without waiting for conventional public relations (PR) to yield results or for journalists to pick up and interpret the firm’s story. Starbucks executives have described the majority of comments the company receives through social media as “a love letter to the firm.” By contrast, if your firm isn’t prepared to be open or if your products and services are notoriously subpar and your firm is inattentive to customer feedback, then establishing a brand-tarring social media beachhead might not make sense. A word to the self-reflective: Customer conversations will happen online even if you don’t have any social media embassies. Users can form their own groups, hash tags, and forums. A reluctance to participate may signal that the firm is facing deeper issues around its product and service.

While firms can learn a lot from social media consultants and tool providers, it’s considered bad practice to outsource the management of a social media presence to a third-party agency. The voice of the firm should come from the firm. In fact, it should come from employees who can provide authentic expertise. Starbucks’ primary Twitter feed is managed by Brad Nelson, a former barista, while the firm’s director of environmental affairs, Jim Hanna, tweets and engages across social media channels on the firm’s green efforts.

Engage and Respond

Having an effective social media presence offers “four Ms” of engagement: it’s a megaphone allowing for outbound communication; it’s a magnet drawing communities inward for conversation; and it allows for monitoring and mediation of existing conversations.J. Gallaugher and S. Ransbotham, “Social Media and Dialog Management at Starbucks” (presented at the MISQE Social Media Workshop, Phoenix, AZ, December 2009). This dialogue can happen privately (private messaging is supported on most services) or can occur very publicly (with the intention to reach a wide audience). Understanding when, where, and how to engage and respond online requires a deft and experienced hand.

Many firms will selectively and occasionally retweet praise posts, underscoring the firm’s commitment to customer service. Highlighting service heroes also reinforces exemplar behavior to employees who may be following the firm online, too. Users are often delighted when a major brand retweets their comments, posts a comment on their blog, or otherwise acknowledges them online—just be sure to do a quick public profile investigation to make sure your shout-outs are directed at customers you want associated with your firm. Escalation procedures should also include methods to flag noteworthy posts, good ideas, and opportunities that the social media team should be paying attention to. The customer base is often filled with heartwarming stories of positive customer experiences and rich with insight on making good things even better.

Many will also offer an unsolicited apology if the firm’s name or products comes up in a disgruntled post. You may not be able to respond to all online complaints, but selective acknowledgement of the customer’s voice (and attempts to address any emergent trends) is a sign of a firm that’s focused on customer care. Getting the frequency, tone, and cadence for this kind of dialogue is more art than science, and managers are advised to regularly monitor other firms with similar characteristics for examples of what works and what doesn’t.

Many incidents can be responded to immediately and with clear rules of engagement. For example, Starbuck issues corrective replies to the often-tweeted urban legend that the firm does not send coffee to the U.S. military because of a corporate position against the war. A typical response might read, “Not true, get the facts here” with a link to a Web page that sets the record straight.

Reaching out to key influencers can also be extremely valuable. Prominent bloggers and other respected social media participants can provide keen guidance and insight. The goal isn’t to create a mouthpiece, but to solicit input, gain advice, gauge reaction, and be sure your message is properly interpreted. Influencers can also help spread accurate information and demonstrate a firm’s commitment to listening and learning. In the wake of the Domino’s gross-out, executives reached out to the prominent blog The Consumerist.A. Jacques, “Domino’s Delivers during Crisis: The Company’s Step-by-Step Response after a Vulgar Video Goes Viral,” The Public Relations Strategist, October 24, 2009. Facebook has solicited advice and feedback from MoveOn.org months before launching new features.B. Stone, “Facebook Aims to Extend Its Reach across the Web,” New York Times, December 2, 2008. Meanwhile, Kaiser Permanente leveraged advice from well-known health care bloggers in crafting its approach to social media.J. Kane, R. Fichman, J. Gallaugher, and J. Glaser, “Community Relations 2.0,” Harvard Business Review, November 2009.

However, it’s also important to recognize that not every mention is worthy of a response. The Internet is filled with PR seekers, the unsatisfiably disgruntled, axe grinders seeking to trap firms, dishonest competitors, and inappropriate groups of mischief makers commonly referred to as trolls. One such group hijacked Time Magazine’s user poll of the World’s Most Influential People, voting their twenty-one-year-old leader to the top of the list ahead of Barack Obama, Vladimir Putin, and the pope. Prank voting was so finely calibrated among the group that the rankings list was engineered to spell out a vulgar term using the first letter of each nominee’s name.E. Schonfeld, “Time Magazine Throws Up Its Hands As It Gets Pawned by 4Chan,” TechCrunch, April 27, 2009.

To prepare, firms should “war game” possible crises, ensuring that everyone knows their role, and that experts are on call. A firm’s social media policy should also make it clear how employees who spot a crisis might “pull the alarm” and mobilize the crisis response team. Having all employees aware of how to respond gives the firm an expanded institutional radar that can lower the chances of being blindsided. This can be especially important as many conversations take place in the so-called dark Web beyond the reach of conventional search engines and monitoring tools (e.g., within membership communities or sites, such as Facebook, where only “friends” have access).

In the event of an incident, silence can be deadly. Consumers expect a response to major events, even if it’s just “we’re listening, we’re aware, and we intend to fix things.” When director Kevin Smith was asked to leave a Southwest Airline flight because he was too large for a single seat, Smith went ballistic on Twitter, berating Southwest’s service to his thousands of online followers. Southwest responded that same evening via Twitter, posting, “I’ve read the tweets all night from @ThatKevinSmith—He’ll be getting a call at home from our Customer Relations VP tonight.”

In the event of a major crisis, firms can leverage online media outside the social sphere. In the days following the Domino’s incident, the gross-out video consistently appeared near the top of Google searches about the firm. When appropriate, companies can buy ads to run alongside keywords explaining their position and, if appropriate, offering an apology.S. Gregory, “Domino’s YouTube Crisis: 5 Ways to Fight Back,” Time, April 18, 2009. Homeopathic cold remedy Zicam countered blog posts citing inaccurate product information by running Google ads adjacent to these links, containing tag lines such as “Zicam: Get the Facts.”Zicam had regularly been the victim of urban legends claiming negative side effects from use; see Snopes.com, “Zicam Warning,” http://www.snopes.com/medical/drugs/zicam.asp. However, the firm subsequently was cited in an unrelated FDA warning on the usage of its product; see S. Young, “FDA Warns against Using 3 Popular Zicam Cold Meds,” CNN.com, June 16, 2009.

Review sites such as Yelp and TripAdvisor also provide opportunities for firms to respond to negative reviews. This can send a message that a firm recognizes missteps and is making an attempt to address the issue (follow-through is critical, or expect an even harsher backlash). Sometimes a private response is most effective. When a customer of Farmstead Cheeses and Wines in the San Francisco Bay area posted a Yelp complaint that a cashier was rude, the firm’s owner sent a private reply to the poster pointing out that the employee in question was actually hard of hearing. The complaint was subsequently withdrawn and the critic eventually joined the firm’s Wine Club.K. Paterson, “Managing an Online Reputation,” New York Times, July 29, 2009. Private responses may be most appropriate if a firm is reimbursing clients or dealing with issues where public dialogue doesn’t help the situation. One doesn’t want to train members of the community that public griping gets reward. For similar reasons, in some cases store credit rather than reimbursement may be appropriate compensation.

Who Should Speak for Your Firm? The Case of the Cisco Fatty

Using the Twitter handle “TheConnor,” a graduating college student recently offered full-time employment by the highly regarded networking giant Cisco posted this tweet: “Cisco just offered me a job! Now I have to weigh the utility of a fatty paycheck against the daily commute to San Jose and hating the work.” Bad idea. Her tweet was public and a Cisco employee saw the post, responding, “Who is the hiring manager. I’m sure they would love to know that you will hate the work. We here at Cisco are versed in the web.” Snap!

But this is also where the story underscores the subtleties of social media engagement. Cisco employees are right to be stung by this kind of criticism. The firm regularly ranks at the top of Fortune’s list of “Best Firms to Work for in America.” Many Cisco employees take great pride in their work, and all have an interest in maintaining the firm’s rep so that the company can hire the best and brightest and continue to compete at the top of its market. But when an employee went after a college student so publicly, the incident escalated. The media picked up on the post, and it began to look like an old guy picking on a clueless young woman who made a stupid mistake that should have been addressed in private. There was also an online pile-on attacking TheConnor. Someone uncovered the woman’s true identity and posted hurtful and disparaging messages about her. Someone else set up a Web site at CiscoFatty.com. Even Oprah got involved, asking both parties to appear on her show (the offer was declined). A clearer social media policy highlighting the kinds of issues to respond to and offering a reporting hierarchy to catch and escalate such incidents might have headed off the embarrassment and helped both Cisco and TheConnor resolve the issue with a little less public attention.H. Popkin, “Twitter Gets You Fired in 140 Characters or Less,” MSNBC, March 23, 2009.

It’s time to take social media seriously. We’re now deep into a revolution that has rewritten the rules of customer-firm communication. There are emerging technologies and skills to acquire, a shifting landscape of laws and expectations, a minefield of dangers, and a wealth of unexploited opportunities. Organizations that professionalize their approach to social media and other Web 2.0 technologies are ready to exploit the upside—potentially stronger brands, increased sales, sharper customer service, improved innovation, and more. Those that ignore the new landscape risk catastrophe and perhaps even irrelevance.

Key Takeaways

  • Customer conversations are happening and employees are using social media. Even firms that aren’t planning on creating a social media presence need to professionalize the social media function in their firm (consider this a social media awareness and response team, or SMART).
  • Social media is an interdisciplinary practice, and the team should include professionals experienced in technology, marketing, PR, customer service, legal, and human resources.
  • While the social media team provides guidance, training, and oversight, and structures crisis response, it’s important to ensure that authentic experts engage on behalf of the firm. Social media is a conversation, and this isn’t a job for the standard PR-style corporate spokesperson.
  • Social media policies revolve around “three Rs”: representation, responsibility, and respect. Many firms have posted their policies online so it can be easy for a firm to assemble examples of best practice.
  • Firms must train employees and update their knowledge as technologies, effective use, and threats emerge. Security training is a vital component of establishing social media policy. Penalties for violation should be clear and backed by enforcement.
  • While tempting, creating sock puppets to astroturf social media with praise posts violates FTC rules and can result in prosecution. Many users who thought their efforts were anonymous have been embarrassingly exposed and penalized. Customers are also using social media to expose firm dishonesty.
  • Many tools exist for monitoring social media mentions of an organization, brands, competitors, and executives. Google Alerts, Twitter search, TweetDeck, Twitrratr, bit.ly, Facebook, and Foursquare all provide free tools that firms can leverage. For-fee tools and services are available as part of the online reputation management industry (and consultants in this space can also provide advice on improving a firm’s online image and engagement).
  • Social media are easy to adopt and potentially easy to abuse. The social media team can provide monitoring and support for firm-focused efforts inside the company and running on third-party networks, both to improve efforts and prevent unwanted disclosure, compliance, and privacy violations.
  • The embassy approach to social media has firms establish their online presence through consistently named areas within popular services (e.g., facebook.com/starbucks, twitter.com/starbucks, youtube.com/starbucks). Firms can also create their own branded social media sites using tools such as Salesforce.com’s “Ideas” platform.
  • Social media provides “four Ms” of engagement: the megaphone to send out messages from the firm, the magnet to attract inbound communication, and monitoring and mediation—paying attention to what’s happening online and selectively engage conversations when appropriate. Engagement can be public or private.
  • Engagement is often more art than science, and managers can learn a lot by paying attention to the experiences of others. Firms should have clear rules for engagement and escalation when positive or negative issues are worthy of attention.

Questions and Exercises

  1. The “United Breaks Guitars” and “Domino’s Gross Out” incidents are powerful reminders of how customers and employees can embarrass a firm. Find other examples of customer-and-employee social media incidents that reflected negatively on an organization. What happened? What was the result? How might these incidents have been prevented or better dealt with?
  2. Hunt for examples of social media excellence. List an example of an organization that got it right. What happened, and what benefits were received?
  3. Social media critics often lament a lack of ROI (return on investment) for these sorts of efforts. What kind of return should firms expect from social media? Does the return justify the investment? Why or why not?
  4. What kinds of firms should aggressively pursue social media? Which ones might consider avoiding these practices? If a firm is concerned about online conversations, what might this also tell management?
  5. List the skills that are needed by today’s social media professionals. What topics should you study to prepare you for a career in this space?
  6. Search online to find examples of corporate social media policies. Share your findings with your instructor. What points do these policies have in common? Are there aspects of any of these policies that you think are especially strong that other firms might adopt? Are there things in these policies that concern you?
  7. Should firms monitor employee social media use? Should they block external social media sites at work? Why or why not? Why might the answer differ by industry?
  8. Use the monitoring tools mentioned in the reading to search your own name. How would a prospective employer evaluate what they’ve found? How should you curate your online profiles and social media presence to be the most “corporate friendly”?
  9. Investigate incidents where employees were fired for social media use. Prepare to discuss examples in class. Could the employer have avoided these incidents?
  10. Use the monitoring tools mentioned in the reading to search for a favorite firm or brand. What trends do you discover? Is the online dialogue fair? How might the firm use these findings?
  11. Consider the case of the Cisco Fatty. Who was wrong? Advise how a firm might best handle this kind of online commentary.

Chapter 6: Understanding Network Effects

6.1 Introduction

Learning Objectives

After studying this section you should be able to do the following:

  1. Define network effects.
  2. Recognize products and services that are subject to network effects.
  3. Understand the factors that add value to products and services subject to network effects.

Network effectsAlso known as Metcalfe’s Law, or network externalities. When the value of a product or service increases as its number of users expands. are sometimes referred to as “Metcalfe’s Law” or “Network Externalities.” But don’t let the dull names fool you—this concept is rocket fuel for technology firms. Bill Gates leveraged network effects to turn Windows and Office into virtual monopolies and in the process became the wealthiest man in America. Mark Zuckerberg of Facebook, Pierre Omidyar of eBay, Caterina Fake and Stewart Butterfield of Flickr, Kevin Rose of Digg, Evan Williams and Biz Stone of Twitter, Chris DeWolfe and Tom Anderson—the MySpace guys—all of these entrepreneurs have built massive user bases by leveraging the concept. When network effects are present, the value of a product or service increases as the number of users grows. Simply, more users = more value. Of course, most products aren’t subject to network effects—you probably don’t care if someone wears the same socks, uses the same pancake syrup, or buys the same trash bags as you. But when network effects are present they’re among the most important reasons you’ll pick one product or service over another. You may care very much, for example, if others are part of your social network, if your video game console is popular, if the Wikipedia article you’re referencing has had prior readers. And all those folks who bought HD DVD players sure were bummed when the rest of the world declared Blu-ray the winner. In each of these examples, network effects are at work.

Not That Kind of Network

The term “network” sometimes stumps people when first learning about network effects. In this context, a network doesn’t refer to the physical wires or wireless systems that connect pieces of electronics. It just refers to a common user base that is able to communicate and share with one another. So Facebook users make up a network. So do owners of Blu-ray players, traders that buy and sell stock over the NASDAQ, or the sum total of hardware and outlets that support the BS 1363 electrical standard.

Key Takeaway

  • Network effects are among the most powerful strategic resources that can be created by technology-based innovation. Many category-dominating organizations and technologies, including Microsoft, Apple, NASDAQ, eBay, Facebook, and Visa, owe their success to network effects. Network effects are also behind the establishment of most standards, including Blu-ray, Wi-Fi, and Bluetooth.

Questions and Exercises

  1. What are network effects? What are the other names for this concept?
  2. List several products or services subject to network effects. What factors do you believe helped each of these efforts achieve dominance?
  3. Which firm do you suspect has stronger end-user network effects: Google’s online search tool or Microsoft’s Windows operating system? Why?
  4. Network effects are often associated with technology, but tech isn’t a prerequisite for the existence of network effects. Name a product, service, or phenomenon that is not related to information technology that still dominates due to network effects.

6.2 Where’s All That Value Come From?

Learning Objectives

After studying this section you should be able to do the following:

  1. Identify the three primary sources of value for network effects.
  2. Recognize factors that contribute to the staying power and complementary benefits of a product or service subject to network effects.
  3. Understand how firms like Microsoft and Apple each benefit from strong network effects.

The value derived from network effects comes from three sources: exchange, staying power, and complementary benefits.

Exchange

Facebook for one person isn’t much fun, and the first guy in the world with a fax machine didn’t have much more than a paperweight. But as each new Facebook friend or fax user comes online, a network becomes more valuable because its users can potentially communicate with more people. These examples show the importance of exchange in creating value. Every product or service subject to network effects fosters some kind of exchange. For firms leveraging technology, this might include anything you can represent in the ones and zeros of digital storage, such as movies, music, money, video games, and computer programs. And just about any standard that allows things to plug into one another, interconnect, or otherwise communicate will live or die based on its ability to snare network effects.

Exercise: Graph It

Some people refer to network effects by the name Metcalfe’s Law. It got this name when, toward the start of the dot-com boom, Bob Metcalfe (the inventor of the Ethernet networking standard) wrote a column in InfoWorld magazine stating that the value of a network equals its number of users squared. What do you think of this formula? Graph the law with the vertical axis labeled “value” and the horizontal axis labeled “users.” Do you think the graph is an accurate representation of what’s happening in network effects? If so, why? If not, what do you think the graph really looks like?

Staying Power

Users don’t want to buy a product or sign up for a service that’s likely to go away, and a number of factors can halt the availability of an effort: a firm could bankrupt or fail to attract a critical mass of user support, or a rival may successfully invade its market and draw away current customers. Networks with greater numbers of users suggest a stronger staying powerThe long-term viability of a product or service.. The staying power, or long-term viability, of a product or service is particularly important for consumers of technology products. Consider that when someone buys a personal computer and makes a choice of Windows, Mac OS, or Linux, their investment over time usually greatly exceeds the initial price paid for the operating system. One invests in learning how to use a system, buying and installing software, entering preferences or other data, creating files—all of which mean that if a product isn’t supported anymore, much of this investment is lost.

The concept of staying power (and the fear of being stranded in an unsupported product or service) is directly related to switching costsThe cost a consumer incurs when moving from one product to another. It can involve actual money spent (e.g., buying a new product) as well as investments in time, any data loss, and so forth. (the cost a consumer incurs when moving from one product to another) and switching costs can strengthen the value of network effects as a strategic asset. The higher the value of the user’s overall investment, the more they’re likely to consider the staying power of any offering before choosing to adopt it. Similarly, the more a user has invested in a product, the less likely he or she is to leave.

Switching costs also go by other names. You might hear the business press refer to products (particularly Web sites) as being “sticky” or creating “friction.” Others may refer to the concept of “lock-in.” And the elite Boston Consulting Group is really talking about a firm’s switching costs when it refers to how well a company can create customers who are “barnacles” (that are tightly anchored to the firm) and not “butterflies” (that flutter away to rivals). The more friction available to prevent users from migrating to a rival, the greater the switching costs. And in a competitive market where rivals with new innovations show up all the time, that can be a very good thing!

How Important Are Switching Costs to Microsoft?

“It is this switching cost that has given our customers the patience to stick with Windows through all our mistakes, our buggy drivers, our high TCO [total cost of ownership]An economic measure of the full cost of owning a product (typically computing hardware and/or software). TCO includes direct costs such as purchase price, plus indirect costs such as training, support, and maintenance., our lack of a sexy vision at times, and many other difficulties […] Customers constantly evaluate other desktop platforms, [but] it would be so much work to move over that they hope we just improve Windows rather than force them to move. […] In short, without this exclusive franchise [meaning Windows] we would have been dead a long time ago.”

Comments from a Microsoft General Manager in a memo to Bill GatesM. Parsons, “Microsoft: ‘We’d Have Been Dead a Long Time Ago without Windows APIs,” ZDNet UK, April 22, 2004, http://news.zdnet.co.uk/software/0,1000000121 ,39152686,00.htm.

Complementary Benefits

Complementary benefitsProducts or services that add additional value to the primary product or service that makes up a network. are those products or services that add additional value to the network. These products might include “how-to” books, software add-ons, even labor. You’ll find more books on auctioning over eBay, more virtual storefronts in Second Life, and more accountants that know Excel, than on any of their rivals. Why? Book authors, Second Life partners, and accountants invest their time where they’re likely to reach the biggest market and get the greatest benefit. In auctions, virtual worlds, and spreadsheet software, eBay, Second Life, and Excel each dwarf their respective competition.

Products and services that encourage others to offer complementary goods are sometimes called platformsProducts and services that allow for the development and integration of software products and other complementary goods. Windows, the iPhone, the Wii, and the standards that allow users to create Facebook apps are all platforms..T. Eisenmann, G. Parker, and M. Van Alstyne, “Strategies for Two-Sided Markets,” Harvard Business Review, October 2006. Allowing other firms to contribute to your platform can be a brilliant strategy, because those firms will spend their time and money to enhance your offerings. Consider the billion-dollar hardware ecosystem that Apple has cultivated around the iPod. There are over ninety brands selling some 280 models of iPod speaker systems.S. Hansell, “The iPod Economy and C.E.S.,” New York Times, January 7, 2008. Thirty-four auto manufacturers now trumpet their cars as being iPod-ready, many with in-car docking stations and steering wheel iPod navigation systems. Each add-on enhances the value of choosing an iPod over a rival like the Microsoft Zune. And now with the App Store for the iPhone, iPod touch, and iPad, Apple is doing the same thing with software add-ons. Software-based ecosystems can grow very quickly. In less than a year after its introduction, the iTunes App Store boasted over fifty thousand applications, collectively downloaded over one billion times.

The iPod Economy

Products built to work with the iPod range from automobiles to the iCarta toilet paper holder.

Apple offers certification programs, where developers of accessories for the iPod and iPhone that meet certain guidelines can use the depicted logo. Each of these third-party products potentially enhances the value of owning an Apple product, while each logo serves as an additional advertisement for Apple. Apple even receives a royalty from firms that use the “Made for iPod” logo in advertisements and on product packaging.

These three value-adding sources—exchange, staying power, and complementary benefits—often work together to reinforce one another in a way that makes the network effect even stronger. When users exchanging information attract more users, they can also attract firms offering complementary products. When developers of complementary products invest time writing software—and users install, learn, and customize these products—switching costs are created that enhance the staying power of a given network. From a strategist’s perspective this can be great news for dominant firms in markets where network effects exist. The larger your network, the more difficult it becomes for rivals to challenge your leadership position.

Key Takeaways

  • Products and services subject to network effects get their value from exchange, perceived staying power, and complementary products and services. Tech firms and services that gain the lead in these categories often dominate all rivals.
  • Many firms attempt to enhance their network effects by creating a platform for the development of third-party products and services that enhance the primary offering.

Questions and Exercises

  1. What are the factors that contribute to the value created by network effects?
  2. Why is staying power particularly important to many technology products and services?
  3. Think about the kinds of technology products that you own that are subject to network effects. What sorts of exchange do these products leverage (e.g., information, money, software, or other media)?
  4. Think about the kinds of technology projects you own. What sorts of switching costs are inherent in each of these? Are these strong switching costs or weak switching costs? What would it take for you to leave one of these services and use a rival? How might a competitor try to lessen these switching costs to persuade you to adopt their product?
  5. Which other terms are sometimes used to describe the phenomenon of switching costs?
  6. Think about the kinds of technology products that you own that are subject to network effects. What sorts of complementary benefits are available for these products? Are complementary benefits strong or weak (meaning, do people choose the product primarily based on these benefits, or for some other reason)?
  7. Identify firms that you believe have built a strong platform. Can you think of firms that have tried to develop a platform, but have been less successful? Why do you suppose they have struggled?

6.3 One-Sided or Two-Sided Markets?

Learning Objectives

After studying this section you should be able to do the following:

  1. Recognize and distinguish between one-sided and two-sided markets.
  2. Understand same-side and cross-side exchange benefits.

Understanding Network Structure

To understand the key sources of network value, it’s important to recognize the structure of the network. Some networks derive most of their value from a single class of users. An example of this kind of network is instant messaging (IM). While there might be some add-ons for the most popular IM tools, they don’t influence most users’ choice of an IM system. You pretty much choose one IM tool over another based on how many of your contacts you can reach. Economists would call IM a one-sided marketA market that derives most of its value from a single class of users (e.g., instant messaging). (a market that derives most of its value from a single class of users), and the network effects derived from IM users attracting more IM users as being same-side exchange benefitsBenefits derived by interaction among members of a single class of participant (e.g., the exchange value when increasing numbers of IM users gain the ability to message each other). (benefits derived by interaction among members of a single class of participant).

But some markets are comprised of two distinct categories of network participant. Consider video games. People buy a video game console largely based on the number of really great games available for the system. Software developers write games based on their ability to reach the greatest number of paying customers, so they’re most likely to write for the most popular consoles first. Economists would call this kind of network a two-sided marketNetwork markets comprised of two distinct categories of participant, both of which that are needed to deliver value for the network to work (e.g., video game console owners and developers of video games). (network markets comprised of two distinct categories of participant, both of which that are needed to deliver value for the network to work). When an increase in the number of users on one side of the market (console owners, for example) creates a rise in the other side (software developers), that’s called a cross-side exchange benefitWhen an increase in the number of users on one side of the market (console owners, for example) creates a rise in the other side (software developers)..

The Positive Feedback Loop of Network Effects

IM is considered a one-sided market, where the value-creating, positive-feedback loop of network effects comes mostly from same-side benefits from a single group (IM members who attract other IM members who want to communicate with them). Video game consoles, however, are considered a two-sided network, where significant benefits come from two distinct classes of users that add value from cross-side benefits by attracting their opposite group. In the game console market, more users of a console attract more developers who write more software for that console, and that attracts more users. Game availability is the main reason the Sony PlayStation 2 dominated over the original Xbox. And app availability is one of the most significant advantages the iPhone offers over competitive hardware.

It is possible that a network may have both same-side and cross-side benefits. Xbox 360 benefits from cross-side benefits in that more users of that console attract more developers writing more software titles and vice versa. However, the Xbox Live network that allows users to play against each other has same-side benefits. If your buddies use Xbox Live and you want to play against them, you’re more likely to buy an Xbox.

Key Takeaways

  • In one-sided markets, users gain benefits from interacting with a similar category of users (think instant messaging, where everyone can send and receive messages to one another).
  • In two-sided markets, users gain benefits from interacting with a separate, complementary class of users (e.g., in the video game industry console owners are attracted to platforms with the most games, while innovative developers are attracted to platforms that have the most users).

Questions and Exercises

  1. What is the difference between same-side exchange benefits and cross-side exchange benefits?
  2. What is the difference between a one-sided market and a two-sided market?
  3. Give examples of one-sided and two-sided markets.
  4. Identify examples of two-sided markets where both sides pay for a product or service. Identify examples where only one side pays. What factors determine who should pay? Does paying have implications for the establishment and growth of a network effect? What might a firm do to encourage early network growth?
  5. The Apple iPhone Developer Program provides developers access to the App Store where they can distribute their free or commercial applications to millions of iPhone and iPod touch customers. Would the iPhone market be considered a one or two-sided market?

6.4 How Are These Markets Different?

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand how competition in markets where network effects are present differ from competition in traditional markets.
  2. Understand the reasons why it is so difficult for late-moving, incompatible rivals to compete in markets where a dominant, proprietary standard is present.

When network effects play a starring role, competition in an industry can be fundamentally different than in conventional, nonnetwork industries.

First, network markets experience early, fierce competition. The positive-feedback loop inherent in network effects—where the biggest networks become even bigger—causes this. Firms are very aggressive in the early stages of these industries because once a leader becomes clear, bandwagons form, and new adopters begin to overwhelmingly favor the leading product over rivals, tipping the market in favor of one dominant firm or standard. This tipping can be remarkably swift. Once the majority of major studios and retailers began to back Blu-ray over HD DVD, the latter effort folded within weeks.

These markets are also often winner-take-all or winner-take-most, exhibiting monopolistic tendencies where one firm dominates all rivals. Look at all of the examples listed so far—in nearly every case the dominant player has a market share well ahead of all competitors. When, during the U.S. Microsoft antitrust trial, Judge Thomas Penfield Jackson declared Microsoft to be a monopolyA market where there are many buyers but only one dominant seller. (a market where there are many buyers but only one dominant seller), the collective response should have been “of course.” Why? The natural state of a market where network effects are present (and this includes operating systems and Office software) is for there to be one major player. Since bigger networks offer more value, they can charge customers more. Firms with a commanding network effects advantage may also enjoy substantial bargaining power over partners. For example, Apple, which controls over 75 percent of digital music sales, for years was able to dictate song pricing, despite the tremendous protests of the record labels.B. Barnes, “NBC Will Not Renew iTunes Contract,” New York Times, August 31, 2007. In fact, Apple’s stranglehold was so strong that it leveraged bargaining power even though the “Big Four” record labels (Universal, Sony, EMI, and Warner) were themselves an oligopolyA market dominated by a small number of powerful sellers. (a market dominated by a small number of powerful sellers) that together provide over 85 percent of music sold in the United States.

Finally, it’s important to note that the best product or service doesn’t always win. PlayStation 2 dominated the video console market over the original Xbox, despite the fact that nearly every review claimed the Xbox was hands-down a more technically superior machine. Why were users willing to choose an inferior product (PS2) over a superior one (Xbox)? The power of network effects! PS2 had more users, which attracted more developers offering more games.

Figure 6.1

Battling a leader with network effects is tough.Adapted from J. Gallaugher and Y. Wang, “Linux vs. Windows in the Middle Kingdom: A Strategic Valuation Model for Platform Competition” (paper, Proceedings of the 2008 Meeting of Americas Conference on Information Systems, Toronto, CA, August 2008), extending M. Schilling, “Technological Leapfrogging: Lessons from the U.S. Video Game Console Industry,” California Management Review, Spring 2003.

This last note is a critical point to any newcomer wishing to attack an established rival. Winning customers away from a dominant player in a network industry isn’t as easy as offering a product or service that is better. Any product that is incompatible with the dominant network has to exceed the value of the technical features of the leading player, plus (since the newcomer likely starts without any users or third-party product complements) the value of the incumbent’s exchange, switching cost, and complementary product benefit (see Figure 6.1). And the incumbent must not be able to easily copy any of the newcomer’s valuable new innovations; otherwise the dominant firm will quickly match any valuable improvements made by rivals. As such, technological leapfroggingCompeting by offering a new technology that is so superior to existing offerings that the value overcomes the total resistance that older technologies might enjoy via exchange, switching cost, and complementary benefits., or competing by offering a superior generation of technology, can be really tough.M. Schilling, “Technological Leapfrogging: Lessons from the U.S. Video Game Console Industry,” California Management Review, Spring 2003.

Is This Good for Innovation?

Critics of firms that leverage proprietary standards for market dominance often complain that network effects are bad for innovation. But this statement isn’t entirely true. While network effects limit competition against the dominant standard, innovation within a standard may actually blossom. Consider Windows. Microsoft has a huge advantage in the desktop operating system market, so few rivals try to compete with it. Apple’s Mac OS and the open source Linux operating system are the firm’s only credible rivals, and both have tiny market shares. But the dominance of Windows is a magnet for developers to innovate within the standard. Programmers with novel ideas are willing to make the investment in learning to write software for Windows because they’re sure that a Windows version can be used by the overwhelming majority of computer users.

By contrast, look at the mess we initially had in the mobile phone market. With so many different handsets offering different screen sizes, running different software, having different key layouts, and working on different carrier networks, writing a game that’s accessible by the majority of users is nearly impossible. Glu Mobile, a maker of online games, launched fifty-six reengineered builds of Monopoly to satisfy the diverse requirements of just one telecom carrier.N. Hutheesing, “Answer Your Phone, a Videogame Is Calling,” Forbes, August 8, 2006. As a result, entrepreneurs with great software ideas for the mobile market were deterred because writing, marketing, and maintaining multiple product versions is both costly and risky. It wasn’t until Apple’s iPhone arrived, offering developers both a huge market and a consistent set of development standards, that third-party software development for mobile phones really took off.

Key Takeaways

  • Unseating a firm that dominates with network effects can be extremely difficult, especially if the newcomer is not compatible with the established leader. Newcomers will find their technology will need to be so good that it must leapfrog not only the value of the established firm’s tech, but also the perceived stability of the dominant firm, the exchange benefits provided by the existing user base, and the benefits from any product complements. For evidence, just look at how difficult it’s been for rivals to unseat the dominance of Windows.
  • Because of this, network effects might limit the number of rivals that challenge a dominant firm. But the establishment of a dominant standard may actually encourage innovation within the standard, since firms producing complements for the leader have faith the leader will have staying power in the market.

Questions and Exercises

  1. How is competition in markets where network effects are present different from competition in traditional markets?
  2. What are the reasons it is so difficult for late-moving, incompatible rivals to compete in markets where a dominant, proprietary standard is present? What is technological leapfrogging and why is it so difficult to accomplish?
  3. Does it make sense to try to prevent monopolies in markets where network effects exist?
  4. Are network effects good or bad for innovation? Explain.
  5. What is the relationship between network effects and the bargaining power of participants in a network effects “ecosystem”?
  6. Cite examples where the best technology did not dominate a network effects-driven market.

6.5 Competing When Network Effects Matter

Learning Objectives

After studying this section you should be able to do the following:

  1. Plot strategies for competing in markets where network effects are present, both from the perspective of the incumbent firm and the new market entrant.
  2. Give examples of how firms have leveraged these strategies to compete effectively.

Why do you care whether networks are one-sided, two-sided, or some sort of hybrid? Well, when crafting your plan for market dominance, it’s critical to know if network effects exist, how strong they might be, where they come from, and how they might be harnessed to your benefit. Here’s a quick rundown of the tools at your disposal when competing in the presence of network effects.

Strategies for Competing in Markets with Network Effects (Examples in Parentheses)

  • Move early (Yahoo! Auctions in Japan)
  • Subsidize product adoption (PayPal)
  • Leverage viral promotion (Skype; Facebook feeds)
  • Expand by redefining the market to bring in new categories of users (Nintendo Wii) or through convergence (iPhone).
  • Form alliances and partnerships (NYCE vs. Citibank)
  • Establish distribution channels (Java with Netscape; Microsoft bundling Media Player with Windows)
  • Seed the market with complements (Blu-ray; Nintendo)
  • Encourage the development of complementary goods—this can include offering resources, subsidies, reduced fees, market research, development kits, venture capital (Facebook fbFund).
  • Maintain backward compatibility (Apple’s Mac OS X Rosetta translation software for PowerPC to Intel)
  • For rivals, be compatible with larger networks (Apple’s move to Intel; Live Search Maps)
  • For incumbents, constantly innovate to create a moving target and block rival efforts to access your network (Apple’s efforts to block access to its own systems)
  • For large firms with well-known followers, make preannouncements (Microsoft)

Move Early

In the world of network effects, this is a biggie. Being first allows your firm to start the network effects snowball rolling in your direction. In Japan, worldwide auction leader eBay showed up just five months after Yahoo! launched its Japanese auction service. But eBay was never able to mount a credible threat and ended up pulling out of the market. Being just five months late cost eBay billions in lost sales, and the firm eventually retreated, acknowledging it could never unseat Yahoo!’s network effects lead.

Another key lesson from the loss of eBay Japan? Exchange depends on the ability to communicate! EBay’s huge network effects in the United States and elsewhere didn’t translate to Japan because most Japanese aren’t comfortable with English, and most English speakers don’t know Japanese. The language barrier made Japan a “greenfield” market with no dominant player, and Yahoo!’s early move provided the catalyst for victory.

Timing is often critical in the video game console wars, too. Sony’s PlayStation 2 enjoyed an eighteen-month lead over the technically superior Xbox (as well as Nintendo’s GameCube). That time lead helped to create what for years was the single most profitable division at Sony. By contrast, the technically superior PS3 showed up months after Xbox 360 and at roughly the same time as the Nintendo Wii, and has struggled in its early years, racking up multibillion-dollar losses for Sony.C. Null, “Sony’s Losses on PS3: $3 Billion and Counting,” Yahoo! Today in Tech, June 27, 2008, http://tech.yahoo.com/blogs/null/96355.

What If Microsoft Threw a Party and No One Showed Up?

Microsoft launched the Zune media player with features that should be subject to network effects—the ability to share photos and music by wirelessly “squirting” content to other Zune users. The firm even promoted Zune with the tagline “Welcome to the Social.” Problem was the Zune Social was a party no one wanted to attend. The late-arriving Zune garnered a market share of just 3 percent, and users remained hard pressed to find buddies to leverage these neat social features.R. Walker, “AntiPod,” New York Times, August 8, 2008. A cool idea does not make a network effect happen.

Subsidize Adoption

Starting a network effect can be tough—there’s little incentive to join a network if there’s no one in the system to communicate with. In one admittedly risky strategy, firms may offer to subsidize initial adoption in hopes that network effects might kick in shortly after. Subsidies to adopters might include a price reduction, rebate, or other giveaways. PayPal, a service that allows users to pay one another using credit cards, gave users a modest rebate as a sign-up incentive to encourage adoption of its new effort (in one early promotion, users got back fifteen dollars when spending their first thirty dollars). This brief subsidy paid to early adopters paid off handsomely. EBay later tried to enter the market with a rival effort, but as a late mover its effort was never able to overcome PayPal’s momentum. PayPal was eventually purchased by eBay for $1.5 billion, and the business unit is now considered one of eBay’s key drivers of growth and profit.

When Even Free Isn’t Good Enough

Subsidizing adoption after a rival has achieved dominance can be an uphill battle, and sometimes even offering a service for free isn’t enough to combat the dominant firm. When Yahoo! introduced a U.S. auction service to compete with eBay, it initially didn’t charge sellers at all (sellers typically pay eBay a small percentage of each completed auction). The hope was that with the elimination of seller fees, enough sellers would jump from eBay to Yahoo! helping the late-mover catch up in the network effect game.

But eBay sellers were reluctant to leave for two reasons. First, there weren’t enough buyers on Yahoo! to match the high bids they earned on much-larger eBay. Some savvy sellers played an arbitrage game where they’d buy items on Yahoo!’s auction service at lower prices and resell them on eBay, where more users bid prices higher.

Second, any established seller leaving eBay would give up their valuable “seller ratings,” and would need to build their Yahoo! reputation from scratch. Seller ratings represent a critical switching cost, as many users view a high rating as a method for reducing the risk of getting scammed or receiving lower-quality goods.

Auctions work best for differentiated goods. While Amazon has had some success in peeling away eBay sellers who provide commodity products (a real danger as eBay increasingly relies on fixed-price sales), eBay’s dominant share of the online auction market still towers over all rivals.B. Stone, “Amid the Gloom, an E-commerce War,” New York Times, October 12, 2008. While there’s no magic in the servers used to create eBay, the early use of technology allowed the firm to create both network effects and switching costs—a dual strategic advantage that has given it a hammerlock on auctions even as others have attempted to mimic its service and undercut its pricing model.

Leverage Viral Promotion

Since all products and services foster some sort of exchange, it’s often possible to leverage a firm’s customers to promote the product or service. Internet calling service Skype has over five hundred million registered users yet has spent almost nothing on advertising. Most Skype users were recruited by others who shared the word on free and low-cost Internet calls. Within Facebook, feeds help activities to spread virally (see Chapter 8 “Facebook: Building a Business from the Social Graph”). Feeds blast updates on user activities on the site, acting as a catalyst for friends to join groups and load applications that their buddies have already signed up for.

Expand by Redefining the Market

If a big market attracts more users (and in two-sided markets, more complements), why not redefine the space to bring in more users? Nintendo did this when launching the Wii. While Sony and Microsoft focused on the graphics and raw processing power favored by hard-core male gamers, Nintendo chose to develop a machine to appeal to families, women, and age groups that normally shunned alien shoot-’em ups. By going after a bigger, redefined market, Nintendo was able to rack up sales that exceeded the Xbox 360, even though it followed the system by twelve months.M. Sanchanta, “Nintendo’s Wii Takes Console Lead,” Financial Times, September 12, 2007.

Seeking the Blue Ocean

Reggie Fils-Aimé, the President of Nintendo of America, describes the Wii Strategy as a Blue Ocean effort.R. Fils-Aimé (presentation and discussion, Carroll School of Management, Boston College, Chestnut Hill, MA, April 6, 2009). The concept of blue ocean strategyAn approach where firms seek to create and compete in uncontested “blue ocean” market spaces, rather than competing in spaces and ways that have attracted many, similar rivals. was popularized by European Institute of Business Administration (INSEAD) professors W. Chan Kim and Renée Mauborgne (authors of a book with the same title).W. C. Kim and R. Mauborgne, Blue Ocean Strategy: How to Create Uncontested Market Space and Make Competition Irrelevant (Cambridge, MA: Harvard Business Press, 2005). See http://www.blueoceanstrategy.com. The idea—instead of competing in blood-red waters where the sharks of highly competitive firms vie for every available market scrap, firms should seek the blue waters of uncontested, new market spaces.

For Nintendo, the granny gamers, moms, and partygoers who flocked to the Wii represented an undiscovered feast in the Blue Ocean. Talk about new markets! Consider that the best-selling video game at the start of 2009 was Wii Fit—a genre-busting title that comes with a scale so you can weigh yourself each time you play! That’s a far cry from Grand Theft Auto IV, the title ranking fifth in 2008 sales, and trailing four Wii-only exclusives.

Blue ocean strategy often works best when combined with strategic positioning described in Chapter 2 “Strategy and Technology: Concepts and Frameworks for Understanding What Separates Winners from Losers”. If an early mover into a blue ocean can use this lead to create defensible assets for sustainable advantage, late moving rivals may find markets unresponsive to their presence.

Market expansion sometimes puts rivals who previously did not compete on a collision course as markets undergo convergenceWhen two or more markets, once considered distinctly separate, begin to offer features and capabilities. As an example: the markets for mobile phones and media players are converging. (when two or more markets, once considered distinctly separate, begin to offer similar features and capabilities). Consider the market for portable electronic devices. Separate product categories for media players, cameras, gaming devices, phones, and global positioning systems (GPS) are all starting to merge. Rather than cede its dominance as a media player, Apple leveraged a strategy known as envelopmentWhen one market attempts to conquer a new market by making it a subset, component, or feature of its primary offering., where a firm seeks to make an existing market a subset of its product offering. Apple deftly morphed the iPod into the iPhone, a device that captures all of these product categories in one device. But the firm went further; the iPhone is Wi-Fi capable, offers browsing, e-mail, and an application platform based on a scaled-down version of the same OS X operating system used in Macintosh computers. As a “Pocket Mac,” the appeal of the device broadened beyond just the phone or music player markets, and within two quarters of launch, iPhone become the second-leading smartphone in North America—outpacing Palm, Microsoft, Motorola and every other rival, except RIM’s BlackBerry.R. Kim, “iPhone No. 2 Smartphone Platform in North America,” The Tech Chronicles—The San Francisco Chronicle, December 17, 2007.

Alliances and Partnerships

Firms can also use partnerships to grow market share for a network. Sometimes these efforts bring rivals together to take out a leader. In a classic example, consider ATM networks. Citibank was the first major bank in New York City to offer a large ATM network. But the Citi network was initially proprietary, meaning customers of other banks couldn’t take advantage of Citi ATMs. Citi’s innovation was wildly popular and being a pioneer in rolling out cash machines helped the firm grow deposits fourfold in just a few years. Competitors responded with a partnership. Instead of each rival bank offering another incompatible network destined to trail Citi’s lead, competing banks agreed to share their ATM operations through NYCE (New York Cash Exchange). While Citi’s network was initially the biggest, after the NYCE launch a Chase bank customer could use ATMs at a host of other banks that covered a geography far greater than Citi offered alone. Network effects in ATMs shifted to the rival bank alliance, Citi eventually joined NYCE and today, nearly every ATM in the United States carries a NYCE sticker.

Google has often pushed an approach to encourage rivals to cooperate to challenge a leader. Its Open Social standard for social networking (endorsed by MySpace, LinkedIn, Bebo, Yahoo! and others) is targeted at offering a larger alternative to Facebook’s more closed efforts (see Chapter 8 “Facebook: Building a Business from the Social Graph”), while its Android open source mobile phone operating system has gained commitments from many handset makers that collectively compete with Apple’s iPhone.

Share or Stay Proprietary?

Defensive moves like the ones above are often meant to diffuse the threat of a proprietary rival. Sometimes firms decide from the start to band together to create a new, more open standard, realizing that collective support is more likely to jumpstart a network than if one firm tried to act with a closed, proprietary offering. Examples of this include the coalitions of firms that have worked together to advance standards like Bluetooth and Wi-Fi. While no single member firm gains a direct profit from the sale of devices using these standards, the standard’s backers benefit when the market for devices expands as products become more useful because they are more interoperable.

Leverage Distribution Channels

Firms can also think about novel ways to distribute a product or service to consumers. Sun faced a challenge when launching the Java programming language—no computers could run it. In order for Java to work, computers need a little interpreter program called the Java Virtual Machine (JVM). Most users weren’t willing to download the JVM if there were no applications written in Java, and no developers were willing to write in Java if no one could run their code. Sun broke the logjam when it bundled the JVM with Netscape’s browser. When millions of users downloaded Netscape, Sun’s software snuck in, almost instantly creating a platform of millions for would-be Java developers. Today, even though Netscape has failed, Sun’s Java remains one of the world’s most popular programming languages. Indeed, Java was cited as one of the main reasons for Oracle’s 2009 acquisition of Sun, with Oracle’s CEO saying the language represented “the single most important software asset we have ever acquired.”A. Ricadela, “Oracle’s Bold Java Plans,” BusinessWeek, June 2, 2009.

As mentioned in Chapter 2 “Strategy and Technology: Concepts and Frameworks for Understanding What Separates Winners from Losers”, Microsoft is in a particularly strong position to leverage this approach. The firm often bundles its new products into its operating systems, Office suite, Internet Explorer browser, and other offerings. The firm used this tactic to transform once market-leader Real Networks into an also-ran in streaming audio. Within a few years of bundling Windows Media Player (WMP) with its other products, WMP grabbed the majority of the market, while Real’s share had fallen to below 10 percent.BusinessWire, “Media Player Format Share for 2006 Confirms Windows Media Remains Dominant with a 50.8% Share of Video Streams Served, Followed by Flash at 21.9%—‘CDN Growth and Market Share Shifts: 2002–2006,’” December 18, 2006; and T. Eisenmann, G. Parker, and M. Van Alstyne, “Strategies for Two-Sided Markets,” Harvard Business Review, October 2006.

Caution is advised, however. Regional antitrust authorities may consider product bundling by dominant firms to be anticompetitive. European regulators have forced Microsoft to unbundle Windows Media Player from its operating system and to provide a choice of browsers alongside Internet Explorer.

Antitrust: Real Versus Microsoft

From October 2001 to March 2003, Microsoft’s bundling of Windows Media Player in versions of its operating system ensured that the software came preinstalled on nearly all of the estimated 207 million new PCs shipped during that period. By contrast, Real Networks’ digital media player was preinstalled on less than 2 percent of PCs. But here’s the kicker that got to regulators (and Real): Microsoft’s standard contract with PC manufacturers “prevented them not only from removing the Windows Media Player, but even [from] providing a desktop icon for Real Networks.”E. Hansen and D. Becker, “Real Hits Microsoft with $1 Billion Antitrust Suit,” CNET, December 18, 2003, http://news.cnet.com/Real-hits-Microsoft-with-1-billion -antitrust-suit/2100-1025_3-5129316.html; and T. Eisenmann, G. Parker, and M. Van Alstyne, “Strategies for Two-Sided Markets,” Harvard Business Review, October 2006. While network effects create monopolies, governments may balk at allowing a firm to leverage its advantages in ways that are designed to deliberately keep rivals from the market.

Seed the Market

When Sony launched the PS3, it subsidized each console by selling at a price estimated at three hundred dollars below unit cost.C. Null, “Sony’s Losses on PS3: $3 Billion and Counting,” Yahoo! Today in Tech, June 27, 2008, http://tech.yahoo.com/blogs/null/96355. Subsidizing consoles is a common practice in the video game industry—game player manufacturers usually make most of their money through royalties paid by game developers. But Sony’s subsidy had an additional benefit for the firm—it helped sneak a Blu-ray player into every home buying a PS3 (Sony was backing the Blu-ray standard over the rival HD DVD effort). Since Sony is also a movie studio and manufacturer of DVD players and other consumer electronics, it had a particularly strong set of assets to leverage to encourage the adoption of Blu-ray over rival HD DVD.

Giving away products for half of a two-sided market is an extreme example of this kind of behavior, but it’s often used. In two-sided markets, you charge the one who will pay. Adobe gives away the Acrobat reader to build a market for the sale of software that creates Acrobat files. Firms with Yellow Page directories give away countless copies of their products, delivered straight to your home, in order to create a market for selling advertising. And Google does much the same by providing free, ad-supported search.

Encourage the Development of Complementary Goods

There are several ways to motivate others to create complementary goods for your network. These efforts often involve some form of developer subsidy or other free or discounted service. A firm may charge lower royalties or offer a period of royalty-free licensing. It can also offer free software development kits (SDKs), training programs, co-marketing dollars, or even start-up capital to potential suppliers. Microsoft and Apple both allow developers to sell their products online through Xbox LIVE Marketplace and iTunes, respectively. This channel lowers developer expenses by eliminating costs associated with selling physical inventory in brick-and-mortar stores and can provide a free way to reach millions of potential consumers without significant promotional spending.

Venture funds can also prompt firms to create complementary goods. Facebook announced it would spur development for the site in part by administering the fbFund, which initially pledged $10 million in start-up funding (in allotments of up to $250,000 each) to firms writing applications for its platform.

Leverage Backward Compatibility

Those firms that control a standard would also be wise to ensure that new products have backward compatibilityThe ability to take advantage of complementary products developed for a prior generation of technology. with earlier offerings. If not, they reenter a market at installed-base zero and give up a major source of advantage—the switching costs built up by prior customers. For example, when Nintendo introduced its 16-bit Super Nintendo system, it was incompatible with the firm’s highly successful prior generation 8-bit model. Rival Sega, which had entered the 16-bit market two years prior to Nintendo, had already built up a large library of 16-bit games for its system. Nintendo entered with only its debut titles, and no ability to play games owned by customers of its previous system, so there was little incentive for existing Nintendo fans to stick with the firm.M. Schilling, “Technological Leapfrogging: Lessons from the U.S. Video Game Console Industry,” California Management Review, Spring 2003.

Backward compatibility was the centerpiece of Apple’s strategy to revitalize the Macintosh through its move to the Intel microprocessor. Intel chips aren’t compatible with the instruction set used by the PowerPC processor used in earlier Mac models. Think of this as two entirely different languages—Intel speaks French, PowerPC speaks Urdu. To ease the transition, Apple included a free software-based adaptorA product that allows a firm to tap into the complementary products, data, or user base of another product or service., called Rosetta, that automatically emulated the functionality of the old chip on all new Macs (a sort of Urdu to French translator). By doing so, all new Intel Macs could use the base of existing software written for the old chip; owners of PowerPC Macs were able to upgrade while preserving their investment in old software; and software firms could still sell older programs while they rewrote applications for new Intel-based Macs.

Even more significant, since Intel is the same standard used by Windows, Apple developed a free software adaptor called Boot Camp that allowed Windows to be installed on Macs. Boot Camp (and similar solutions by other vendors) dramatically lowered the cost for Windows users to switch to Macs. Within two years of making the switch, Mac sales skyrocketed to record levels. Apple now boasts a commanding lead in notebook sales to the education market,P. Seitz, “An Apple for Teacher, Students: Mac Maker Surges in Education,” Investor’s Business Daily, August 8, 2008. and a survey by Yankee Group found that 87 percent of corporations were using at least some Macintosh computers, up from 48 percent at the end of the PowerPC era two years earlier.P. Burrows, “The Mac in the Gray Flannel Suit,” BusinessWeek, May 1, 2008.

Rivals: Be Compatible with the Leading Network

Companies will want to consider making new products compatible with the leading standard. Microsoft’s Live Maps and Virtual Earth 3D arrived late to the Internet mapping game. Users had already put in countless hours building resources that meshed with Google Maps and Google Earth. But by adopting the same keyhole markup language (KML) standard used by Google, Microsoft could, as TechCrunch put it, “drink from Google’s milkshake.” Any work done by users for Google in KML could be used by Microsoft. Voilà, an instant base of add-on content!

Incumbents: Close Off Rival Access and Constantly Innovate

Oftentimes firms that control dominant networks will make compatibility difficult for rivals who try to connect with their systems. AOL has been reluctant to open up its instant messaging tool to rivals, and Skype for years had been similarly closed to non-Skype clients.

Firms that constantly innovate make it particularly difficult for competitors to become compatible. Again, we can look to Apple as an example of these concepts in action. While Macs run Windows, Windows computers can’t run Mac programs. Apple has embedded key software in Mac hardware, making it tough for rivals to write a software emulator like Boot Camp that would let Windows PCs drink from the Mac milkshake. And if any firm gets close to cloning Mac hardware, Apple sues. The firm also modifies software on other products like the iPhone and iTunes each time wily hackers tap into closed aspects of its systems. And Apple has regularly moved to block third-party hardware, such as Palm’s mobile phones, from plugging into iTunes. Even if firms create adaptors that emulate a standard, a firm that constantly innovates creates a moving target that’s tough for others to keep up with.

Apple has been far more aggressive than Microsoft in introducing new versions of its software. Since the firm never stays still, would-be cloners never get enough time to create a reliable emulator that runs the latest Apple software.

Large, Well-Known Followers: Preannouncements

Large firms that find new markets attractive but don’t yet have products ready for delivery might preannounce efforts in order to cause potential adaptors to sit on the fence, delaying a purchasing decision until the new effort rolls out. Preannouncements only work if a firm is large enough to pose a credible threat to current market participants. Microsoft, for example, can cause potential customers to hold off on selecting a rival because users see that the firm has the resources to beat most players (suggesting staying power). Statements from start-ups, however, often lack credibility to delay user purchases. The tech industry acronym for the impact firms try to impart on markets through preannouncements is FUD for fear, uncertainty, and doubt.

The Osborne Effect

Preannouncers, beware. Announce an effort too early and a firm may fall victim to what’s known as “The Osborne EffectWhen a firm preannounces a forthcoming product or service and experiences a sharp and detrimental drop in sales of current offerings as users wait for the new item..” It’s been suggested that portable computer manufacturer Osborne Computer announced new models too early. Customers opted to wait for the new models, so sales of the firm’s current offerings plummeted. While evidence suggests that Osborne’s decline had more to do with rivals offering better products, the negative impact of preannouncements has hurt a host of other firms.A. Orlowski, “Taking Osborne out of the Osborne Effect,” The Register, June 20, 2005. Among these, Sega, which exited the video game console market entirely after preannouncements of a next-generation system killed enthusiasm for its Saturn console.M. Schilling, “Technological Leapfrogging: Lessons from the U.S. Video Game Console Industry,” California Management Review, Spring 2003.

Too Much of a Good Thing?

When network effects are present, more users attract more users. That’s a good thing as long as a firm can earn money from this virtuous cycle. But sometimes a network effect attracts too many users and a service can be so overwhelmed it becomes unusable. These so-called congestion effectsWhen increasing numbers of users lower the value of a product or service. occur when increasing numbers of users lower the value of a product or service. This most often happens when a key resource becomes increasingly scarce. Users of the game Ultima were disappointed in an early online version that launched without enough monsters to fight or server power to handle the crush of fans. Twitter’s early infrastructure was often unable to handle the demands of a service in hypergrowth (leading to the frequent appearance of a not-in-service graphic known in the Twitter community as the “fail whale”). Facebook users with a large number of friends may also find their attention is a limited resource, as feeds push so much content that it becomes difficult to separate interesting information from the noise of friend actions.

And while network effects can attract positive complementary products, a dominant standard may also be the first place where virus writers and malicious hackers choose to strike.

Figure 6.2 The Twitter Fail Whale

Feel confident! Now you’ve got a solid grounding in network effects, the key resource leveraged by some of the most dominant firms in technology. And these concepts apply beyond the realm of tech, too. Network effects can explain phenomena ranging from why some stock markets are more popular than others to why English is so widely spoken, even among groups of nonnative speakers. On top of that, the strategies explored in the last half of the chapter show how to use these principles to sniff out, create, and protect this key strategic asset. Go forth, tech pioneer—opportunity awaits!

Key Takeaways

  • Moving early matters in network markets—firms that move early can often use that time to establish a lead in users, switching costs, and complementary products that can be difficult for rivals to match.
  • Additional factors that can help a firm establish a network effects lead include subsidizing adoption; leveraging viral marketing, creating alliances to promote a product or to increase a service’s user base; redefining the market to appeal to more users; leveraging unique distribution channels to reach new customers; seeding the market with complements; encouraging the development of complements; and maintaining backward compatibility.
  • Established firms may try to make it difficult for rivals to gain compatibility with their users, standards, or product complements. Large firms may also create uncertainty among those considering adoption of a rival by preannouncing competing products.

Questions and Exercises

  1. Is market entry timing important for network effects markets? Explain and offer an example to back up your point.
  2. How might a firm subsidize adoption? Give an example.
  3. Give an example of a partnership or alliance targeted at increasing network effects.
  4. Is it ever advantageous for firms to give up control of a network and share it with others? Why or why not? Give examples to back up your point.
  5. Do firms that dominate their markets with network effects risk government intervention? Why or why not? Explain through an example.
  6. How did Sony seed the market for Blu-ray players?
  7. What does backward compatibility mean and why is this important? What happens if a firm is not backward compatible?
  8. What tactic did Apple use to increase the acceptability of the Mac platform to a broader population of potential users?
  9. How has Apple kept clones at bay?
  10. What are preannouncements? What is the danger in announcing a product too early? What is the term for negative impacts from premature product announcements?
  11. How did PayPal subsidize adoption?
  12. Name two companies that leveraged viral promotion to compete.
  13. Name a product that is the result of the convergence of media players, cameras, and phones.
  14. What is bundling? What are the upsides and downsides of bundling?
  15. Why does Adobe allow the free download of Acrobat Reader?
  16. What tactic might an established firm employ to make it impossible, or at least difficult, for a competitor to gain access to, or become compatible with, their product or service?
  17. How do Apple, Microsoft, and Facebook encourage the development of complementary products?
  18. What is the “congestion effect”? Give an example.
  19. Do network effects apply in nontech areas? Give examples.

Chapter 5: Moore’s Law: Fast, Cheap Computing and What It Means for the Manager

5.1 Introduction

Learning Objectives

After studying this section you should be able to do the following:

  1. Define Moore’s Law and understand the approximate rate of advancement for other technologies, including magnetic storage (disk drives) and telecommunications (fiber-optic transmission).
  2. Understand how the price elasticity associated with faster and cheaper technologies opens new markets, creates new opportunities for firms and society, and can catalyze industry disruption.
  3. Recognize and define various terms for measuring data capacity.
  4. Consider the managerial implication of faster and cheaper computing on areas such as strategic planning, inventory, and accounting.

Faster and cheaper—those two words have driven the computer industry for decades, and the rest of the economy has been along for the ride. Today it’s tough to imagine a single industry not impacted by more powerful, less expensive computing. Faster and cheaper puts mobile phones in the hands of peasant farmers, puts a free video game in your Happy Meal, and drives the drug discovery that may very well extend your life.

Some Definitions

This phenomenon of “faster, cheaper” computing is often referred to as Moore’s LawChip performance per dollar doubles every eighteen months., after Intel cofounder, Gordon Moore. Moore didn’t show up one day, stance wide, hands on hips, and declare “behold my law,” but he did write a four-page paper for Electronics Magazine in which he described how the process of chip making enabled more powerful chips to be manufactured at cheaper prices.G. Moore, “Cramming More Components onto Integrated Circuits,” Electronics Magazine, April 19, 1965.

Moore’s friend, legendary chip entrepreneur and CalTech professor Carver Mead, later coined the “Moore’s Law” moniker. That name sounded snappy, plus as one of the founders of Intel, Moore had enough geek cred for the name to stick. Moore’s original paper offered language only a chip designer would love, so we’ll rely on the more popular definition: chip performance per dollar doubles every eighteen months (Moore’s original paper assumed two years, but many sources today refer to the eighteen-month figure, so we’ll stick with that).

Moore’s Law applies to chips—broadly speaking, to processors, or the electronics stuff that’s made out of silicon.Although other materials besides silicon are increasingly being used. The microprocessorThe part of the computer that executes the instructions of a computer program. is the brain of a computing device. It’s the part of the computer that executes the instructions of a computer program, allowing it to run a Web browser, word processor, video game, or virus. For processors, Moore’s Law means that next generation chips should be twice as fast in eighteen months, but cost the same as today’s models (or from another perspective, in a year and a half, chips that are same speed as today’s models should be available for half the price).

Random-access memory (RAM)The fast, chip-based volatile storage in a computing device. is chip-based memory. The RAM inside your personal computer is volatile memoryStorage (such as RAM chips) that is wiped clean when power is cut off from a device., meaning that when the power goes out, all is lost that wasn’t saved to nonvolatile memoryStorage that retains data even when powered down (such as flash memory, hard disk, or DVD storage). (i.e., a more permanent storage media like a hard disk or flash memory). Think of RAM as temporary storage that provides fast access for executing computer programs and files. When you “load” or “launch” a program, it usually moves from your hard drive to those RAM chips, where it can be more quickly executed by the processor.

Cameras, MP3 players, USB drives, and mobile phones often use flash memoryNonvolatile, chip-based storage, often used in mobile phones, cameras, and MP3 players. Sometimes called flash RAM, flash memory is slower than conventional RAM, but holds its charge even when the power goes out. (sometimes called flash RAM). It’s not as fast as the RAM used in most traditional PCs, but holds data even when the power is off (so flash memory is also nonvolatile memory). You can think of flash memory as the chip-based equivalent of a hard drive. In fact, flash memory prices are falling so rapidly that several manufactures including Apple and the One Laptop per Child initiative (see the “Tech for the Poor” sidebar later in this section) have begun offering chip-based, nonvolatile memory as an alternative to laptop hard drives. The big advantage? Chips are solid state electronicsSemiconductor-based devices. Solid state components often suffer fewer failures and require less energy than mechanical counterparts because they have no moving parts. RAM, flash memory and microprocessors are solid state devices. Hard drives are not. (meaning no moving parts), so they’re less likely to fail, and they draw less power. The solid state advantage also means that chip-based MP3 players like the iPod nano make better jogging companions than hard drive players, which can skip if jostled. For RAM chips and flash memory, Moore’s Law means that in eighteen months you’ll pay the same price as today for twice as much storage.

Computer chips are sometimes also referred to as semiconductorsA substance such as silicon dioxide used inside most computer chips that is capable of enabling as well as inhibiting the flow of electricity. From a managerial perspective, when someone refers to semiconductors, they are talking about computer chips, and the semiconductor industry is the chip business. (a substance such as silicon dioxide used inside most computer chips that is capable of enabling as well as inhibiting the flow of electricity). So if someone refers to the semiconductor industry, they’re talking about the chip business.Semiconductor materials, like the silicon dioxide used inside most computer chips, are capable of enabling as well as inhibiting the flow of electricity. These properties enable chips to perform math or store data.

Strictly speaking, Moore’s Law does not apply to other technology components. But other computing components are also seeing their price versus performance curves skyrocket exponentially. Data storage doubles every twelve months. Networking speed is on a tear, too. With an equipment change at the ends of the cables, the amount of data that can be squirted over an optical fiber lineA high-speed glass or plastic-lined networking cable used in telecommunications. can double every nine months.Fiber-optic lines are glass or plastic data transmission cables that carry light. These cables offer higher transmission speeds over longer distances than copper cables that transmit electricity. These numbers should be taken as rough approximations and shouldn’t be expected to be strictly precise over time. However, they are useful as rough guides regarding future computing price/performance trends. Despite any fluctuation, it’s clear that the price/performance curve for many technologies is exponential, offering astonishing improvement over time.

Figure 5.1 Advancing Rates of Technology (Silicon, Storage, Telecom)

Get Out Your Crystal Ball

Faster and cheaper makes possible the once impossible. As a manager, your job will be about predicting the future. First, consider how the economics of Moore’s Law opens new markets. When technology gets cheap, price elasticityThe rate at which the demand for a product or service fluctuates with price change. Goods and services that are highly price elastic (e.g., most consumer electronics) see demand spike as prices drop, whereas goods and services that are less price elastic are less responsive to price change (think heart surgery). kicks in. Tech products are highly price elastic, meaning consumers buy more products as they become cheaper.As opposed to goods and services that are price inelastic (like health care and housing), which consumers will try their best to buy even if prices go up. And it’s not just that existing customers load up on more tech; entire new markets open up as firms find new uses for these new chips.

Just look at the five waves of computing we’ve seen over the previous five decades.M. Copeland, “How to Ride the Fifth Wave,” Business 2.0, July 1, 2005. In the first wave in the 1960s, computing was limited to large, room-sized mainframe computers that only governments and big corporations could afford. Moore’s Law kicked in during the 1970s for the second wave, and minicomputers were a hit. These were refrigerator-sized computers that were as speedy as or speedier than the prior generation of mainframes, yet were affordable by work groups, factories, and smaller organizations. The 1980s brought wave three in the form of PCs, and by the end of the decade nearly every white-collar worker in America had a fast and cheap computer on their desk. In the 1990s wave four came in the form of Internet computing—cheap servers and networks made it possible to scatter data around the world, and with more power, personal computers displayed graphical interfaces that replaced complex commands with easy-to-understand menus accessible by a mouse click. At the close of the last century, the majority of the population in many developed countries had home PCs, as did most libraries and schools.

Now we’re in wave five, where computers are so fast and so inexpensive that they have become ubiquitous—woven into products in ways few imagined years before. Silicon is everywhere! It’s in the throwaway radio frequency identification (RFID) tags that track your luggage at the airport. It provides the smarts in the world’s billion-plus mobile phones. It’s the brains inside robot vacuum cleaners, next generation Legos, and the table lamps that change color when the stock market moves up or down. These digital shifts can rearrange entire industries. Consider that today the firm that sells more cameras than any other is Nokia, a firm that offers increasingly sophisticated chip-based digital cameras as a giveaway as part of its primary product, mobile phones. This shift has occurred with such sweeping impact that former photography giants Pentax, Konica, and Minolta have all exited the camera business.

Ambient Devices and the Fifth Wave

Carl Yankowski almost never gets caught in the rain without his umbrella. That’s because Yankowski’s umbrella regularly and wirelessly checks weather reports on its own. If the umbrella gets word it will rain in the next few hours, the handle blinks with increasing urgency, warning its owner with a signal that seems to declare, “You will soon require my services.” Yankowski is former CEO of “fifth wave” firm Ambient Devices, a Massachusetts start-up that’s embedding computing and communications technology into everyday devices in an attempt to make them “smarter” and more useful (the weather-sensing umbrella was developed while he helmed the firm).

Ambient’s ability to pull off this little miracle is evidence of how quickly innovative thinkers are able to take advantage of new opportunities and pioneer new markets enabled by Moore’s Law. The firm’s first product, the Orb, is a lamp that can be set up to change color in real time in reaction to factors such as the performance of your stock portfolio or the intensity of the local pollen count. In just six months, the ten refugees from MIT’s Media Lab that founded Ambient Devices took the idea for the Orb, designed the device and its software, licensed wireless spectrum from a pager firm that had both excess capacity and a footprint to cover over 90 percent of the United States, arranged for manufacturing, and began selling the gizmo through Brookstone and Nieman Marcus.M. Copeland, “How to Ride the Fifth Wave,” Business 2.0, July 1, 2005; and J. Miller, “Goodbye G.U.I? Ambient Orb a Computer ‘Mood Ring,’” Mass High Tech, February 10, 2003.

Ambient has since expanded the product line to several low-cost appliances designed to provide information at a glance. These include the Ambient Umbrella, as well as useful little devices that grab and display data ranging from sports scores to fluctuating energy prices (so you’ll put off running the dishwasher until evening during a daytime price spike). The firm even partnered with LG on a refrigerator that can remind you of an upcoming anniversary as you reach for the milk.

Figure 5.2

Products developed by “fifth wave” firm Ambient Devices include the weather-reading Ambient Umbrella, the Energy Joule, a seven-day forecaster, and the Orb lamp.

Moore’s Law inside Your Medicine Cabinet

Moore’s Law is about to hit your medicine cabinet. The GlowCap from Vitality, Inc., is a “smart” pill bottle that will flash when you’re supposed to take your medicine. It will play a little tune if you’re an hour late for your dose and will also squirt a signal to a night-light that flashes as a reminder (in case you’re out of view of the cap). GlowCaps can also be set to call or send a text if you haven’t responded past a set period of time. And the device will send a report to you, your doc, or whomever else you approve. The GlowCap can even alert your pharmacy when it’s time for refills. Amazon sells the device for $99, but we know how Moore’s Law works—it’ll soon likely be free. The business case for that? The World Health Organization estimates drug adherence at just 50 percent, and analysts estimate that up to $290 billion in increased medical costs are due to patients missing their meds. Vitality CEO David Rose (who incidentally also cofounded Ambient Devices) recently cited a test in which GlowCap users reported a 98 percent medication adherence rate.D. Rose, presentation as part of “From Disruption to Innovation” at the MIT Enterprise Forum, Cambridge, MA, June 23, 2010.

Figure 5.3

The GlowCap from Vitality, Inc., will flash, beep, call, and text you if you’ve skipped your meds. It can also send reports to you, your doctor, and your loved ones and even notify your pharmacy when it’s time for a refill.

And there might also be a chip inside the pills, too! Proteus, a Novartis-backed venture, has developed a sensor made of food and vitamin materials that can be swallowed in medicine. The sensor is activated and powered by the body’s digestive acids (think of your stomach as a battery). Once inside you, the chip sends out a signal with vitals such as heart rate, body angle, temperature, sleep, and more. A waterproof skin patch picks up the signal and can wirelessly relay the pill’s findings when the patient walks within twenty feet of their phone. Proteus will then compile a report from the data and send it to their mobile device or e-mail account. The gizmo’s already in clinical trials for heart disease, hypertension, and tuberculosis and for monitoring psychiatric illnesses.E. Landau, “Tattletale Pills, Bottles Remind You to Take Your Meds,” CNN, February 2, 2010.

One of the most agile surfers of this fifth wave is Apple, Inc.—a firm with a product line that is now so broad that in January 2007, it dropped the word “Computer” from its name. Apple’s breakout resurgence owes a great deal to the iPod. At launch, the original iPod sported a 5 GB hard drive that Steve Jobs declared would “put 1,000 songs in your pocket.” Cost? $399. Less than six years later, Apple’s highest-capacity iPod sold for fifty dollars less than the original, yet held forty times the songs. By that time the firm had sold over one hundred fifty million iPods—an adoption rate faster than the original Sony Walkman. Apple’s high-end models have morphed into Internet browsing devices capable of showing maps, playing videos, and gulping down songs from Starbucks’ Wi-Fi while waiting in line for a latte.

The original iPod has also become the jumping-off point for new business lines including the iPhone, Apple TV, iPad, and iTunes. As an online store, iTunes is always open. ITunes regularly sells tens of millions of songs on Christmas Day alone, a date when virtually all of its offline competition is closed for the holiday. In a short five years after its introduction, iTunes has sold over 4 billion songs and has vaulted past retail giants Wal-Mart, Best Buy, and Target to become the number one music retailer in the world. Today’s iTunes is a digital media powerhouse, selling movies, TV shows, games, and other applications. And with podcasting, Apple’s iTunes University even lets students at participating schools put their professors’ lectures on their gym playlist for free. Surfing the fifth wave has increased the value of Apple stock sixteenfold six years after the iPod’s launch. Ride these waves to riches, but miss the power and promise of Moore’s Law and you risk getting swept away in its riptide. Apple’s rise occurred while Sony, a firm once synonymous with portable music, sat on the sidelines unwilling to get on the surfboard. Sony’s stock stagnated, barely moving in six years. The firm has laid off thousands of workers while ceding leadership in digital music (and video) to Apple.

Table 5.1 Top U.S. Music Retailers

1992 2005 2006 2008
1. Musicland 1. Wal-Mart 1. Wal-Mart 1. iTunes
2. The Handleman 2. Best Buy 2. Best Buy 2. Wal-Mart
3. Tower Records 3. Target 3. Target 3. Best Buy
4. Trans World Music

 

7. iTunes

4. iTunes, Amazon tie

 

4. Amazon, Target tie

 

Moore’s Law restructures industries. The firms that dominated music sales when you were born are now bankrupt, while one that had never sold a physical music CD now sells more than anyone else.

While the change in hard drive prices isn’t directly part of Moore’s Law (hard drives are magnetic storage, not silicon chips), as noted earlier, the faster and cheaper phenomenon applies to storage, too. Look to Amazon as another example of jumping onto a once-impossible opportunity courtesy of the price/performance curve. When Amazon.com was founded in 1995, the largest corporate database was one terabyte, or TB (see Note 5.14 “Bits and Bytes” below) in size. In 2003, the firm offered its “Search Inside the Book” feature, digitizing the images and text from thousands of books in its catalog. “Search Inside the Book” lets customers peer into a book’s contents in a way that’s both faster and more accurate than browsing a physical bookstore. Most importantly for Amazon and its suppliers, titles featured in “Search Inside the Book” enjoyed a 7 percent sales increase over nonsearchable books. When “Search Inside the Book” launched, the database to support this effort was 20 TB in size. In just eight years, the firm found that it made good business sense to launch an effort that was a full twenty times larger than anything used by any firm less than a decade earlier. And of course, all of these capacities seem laughably small by today’s standards. (See Chapter 11 “The Data Asset: Databases, Business Intelligence, and Competitive Advantage”.) For Amazon, the impossible had not just become possible; it became good business. By 2009, digital books weren’t just for search; they were for sale. Amazon’s Kindle reader (a Moore’s Law marvel sporting a microprocessor and flash storage) became the firm’s top-selling product in terms of both unit sales and dollar volume. The real business opportunity for Amazon isn’t Kindle as a consumer electronics device but as an ever-present, never-closing store, which also provides the firm with a migration path from atoms to bits. (For more on that topic, see Chapter 4 “Netflix: The Making of an E-commerce Giant and the Uncertain Future of Atoms to Bits”.) By 2009, Amazon (by then the largest book retailer in North America) reported, “For books that are available on the Kindle, sales are already 35 percent of the same books in print.”E. Schonfeld, “For Books Available on Kindle, Sales Are Now Tracking at 35 Percent of Print Sales,” TechCrunch, May 6, 2009. Apple’s 2010 introduction of the iPad, complete with an iBook store, shows how Moore’s Law rewrites the boundaries of competition—bringing a firm that started as a computer retailer and a firm that started as an online bookstore in direct competition with one another.

Bits and Bytes

Computers express data as bits that are either one or zero. Eight bits form a byte (think of a byte as being a single character you can type from a keyboard). A kilobyte refers to roughly a thousand bytes, or a thousand characters, megabyte = 1 million, gigabyte = 1 billion, terabyte = 1 trillion, petabyte = 1 quadrillion, and exabyte = 1 quintillion bytes.

While storage is most often listed in bytes, telecommunication capacity (bandwidth) is often listed in bits per second (bps). The same prefixes apply (Kbps = kilobits, or one thousand bits, per second, Mbps = megabits per second, Gbps = gigabits per second, and Tbps = terabits per second).

These are managerial definitions, but technically, a kilobyte is 210 or 1,024 bytes, mega = 220, giga = 230, tera = 240, peta = 250, and exa = 260. To get a sense for how much data we’re talking about, see the table below.E. Schuman, “At Wal-Mart, World’s Largest Retail Data Warehouse Gets Even Larger,” eWeek, October 13, 2004; and J. Huggins, “How Much Data Is That?” Refrigerator Door, August 19, 2008.

Table 5.2 Bytes Defined

Managerial Definition Exact Amount To Put It in Perspective
1 Byte One keyboard character 8 bits 1 letter or number = 1 byte
1 Kilobyte (KB) One thousand bytes 210 bytes 1 typewritten page = 2 KB
1 digital book (Kindle) = approx. 500–800 KB
1 Megabyte (MB) One million bytes 220 bytes 1 digital photo (7 megapixels) = 1.3 MB
1 MP3 song = approx. 3 MB
1 CD = approx. 700 MB
1 Gigabyte (GB) One billion bytes 230 bytes 1 DVD movie = approx. 4.7 GB
1 Blu-ray movie = approx. 25 GB
1 Terabyte (TB) One trillion bytes 240 bytes Printed collection of the Library of Congress = 20 TB
1 Petabyte (PB) One quadrillion bytes 250 bytes Wal-Mart data warehouse (2008) = 2.5 PB
1 Exabyte (EB) One quintillion bytes 260 bytes
1 Zettabyte (ZB) One sextillion bytes 270 bytes Amount of data consumed by U.S. households in 2008 = 3.6 ZB

Here’s another key implication—if you are producing products with a significant chip-based component, the chips inside that product rapidly fall in value. That’s great when it makes your product cheaper and opens up new markets for your firm, but it can be deadly if you overproduce and have excess inventory sitting on shelves for long periods of time. Dell claims its inventory depreciates as much as a single percentage point in value each week.B. Breen, “Living in Dell Time,” Fast Company, November 24, 2004. That’s a big incentive to carry as little inventory as possible, and to unload it, fast!

While the strategic side of tech may be the most glamorous, Moore’s Law impacts mundane management tasks, as well. From an accounting and budgeting perspective, as a manager you’ll need to consider a number of questions: How long will your computing equipment remain useful? If you keep upgrading computing and software, what does this mean for your capital expense budget? Your training budget? Your ability to make well-reasoned predictions regarding tech’s direction will be key to answering these questions.

Tech for the Poor

Nicholas Negroponte, the former head of MIT’s Media Lab, is on a mission. His OLPC (One Laptop per Child) project aims to deliver education to children in the world’s poorest communities via ultralow-cost computing devices that the firm has developed. The first offering, the XO laptop, costs roughly $175, although a sub-$100 tablet is in the works. The XO sports a rubberized keyboard and entirely solid-state design (flash RAM rather than hard drive) that helps make the machine durable. The XO’s ultrabright screen is readable in daylight and can be flipped to convert into an e-book reader. And a host of open source software and wiki tools for courseware development all aim to keep the costs low. Mesh networking allows laptops within a hundred feet or so to communicate with each other, relaying a single Internet connection for use by all. And since the XO is targeted at the world’s poorest kids in communities where power generation is unreliable or nonexistent, several battery-charging power generation schemes are offered, including a hand crank and foldout flexible solar panels. The OLPC Foundation delivered over 1.6 million laptops to children in twenty-four countries.C. Lawton, “The X.O. Laptop Two Years Later,” Wired, June 19, 2009. The XO is a product made possible by the rapidly falling price of computing.

Figure 5.4 The XO PC

While the success of the OLPC effort will reveal itself over time, another tech product containing a microprocessor is already transforming the lives of some of the world’s most desperate poor—the cell phone. There are three billion people worldwide that don’t yet have a phone, but they will, soon. In the ultimate play of Moore’s Law opening up new markets, mobiles from Vodafone and Indian telecom provider Spice sell for $25 or less. While it took roughly twenty years to sell a billion mobile phones worldwide, the second billion sold in four years, and the third billion took just two years. Today, some 80 percent of the world’s population lives within cellular network range (double the 2000 level), and the vast majority of mobile subscriptions are in developing countries.S. Corbett, “Can the Cellphone Help End Global Poverty?” New York Times Magazine, April 13, 2008.

Why such demand? Mobiles change lives for the better. According to Columbia economist Jeffrey Sachs, “The cell phone is the single most transformative technology for world economic development.”J. Ewing, “Upwardly Mobile in Africa,” BusinessWeek, September 24, 2007, 64–71. Think about the farmer who can verify prices and locate buyers before harvesting and transporting perishable crops to market; the laborer who was mostly unemployed but with a mobile is now reachable by those who have day-to-day work; the mother who can find out if a doctor is in and has medicine before taking off work to make the costly trek to a remote clinic with her sick child; or the immigrant laborer serving as a housekeeper who was “more or less an indentured servant until she got a cell phone” enabling new customers to call and book her services.S. Corbett, “Can the Cellphone Help End Global Poverty?” New York Times Magazine, April 13, 2008.

As an example of impact, look to poor fishermen in the Indian state of Kerala. By using mobile phones to find the best local marketplace prices for sardines, these fishermen were able to increase their profits by an average of 8 percent even though consumer prices for fish dropped 4 percent. The trends benefiting both buyer and seller occurred because the fishermen no longer had to throw away unsold catch previously lost by sailing into a port after all the buyers had left. The phone-equipped fleet now see more consistent pricing, spreading their catch more evenly whereas previous fisherman often inefficiently clustered in one market, overserving one population while underserving another. A London Business School study found that for every ten mobile phones per one hundred people, a country’s GDP bumps up 0.5 percent.J. Ewing, “Upwardly Mobile in Africa,” BusinessWeek, September 24, 2007, 64–71.

Bangladeshi economist Mohammed Yunus won the Nobel Peace Prize based on his work in the microfinance movement, an effort that provides very small loans to the world’s poorest entrepreneurs. Microfinance loans grew the market for Grameen Phone Ltd., a firm that has empowered over two hundred and fifty thousand Bangladeshi “phone ladies” to start businesses that helped their communities become more productive. Phone ladies buy a phone and special antenna on microcredit for about $150 each. These special long-life battery phones allow them to become a sort of village operator, charging a small commission for sending and receiving calls. Through phone ladies, the power of the mobile reaches even those too poor to afford buying one outright. Grameen Phone now has annual revenues of over $1 billion and is Bangladesh’s largest telecom provider.

In another ingenious scheme, phone minutes become a proxy for currency. The New York Times reports that a person “working in Kampala, for instance, who wishes to send the equivalent of five dollars back to his mother in a village will buy a five-dollar prepaid airtime card, but rather than entering the code into his own phone, he will call the village phone operator and read the code to her. [The operator] then uses the airtime for her phone and completes the transaction by giving the man’s mother the money, minus a small commission.”S. Corbett, “Can the Cellphone Help End Global Poverty?” New York Times Magazine, April 13, 2008.

South Africa’s WIZZIT and GCASH in the Philippines allow customers to use mobile phones to store cash credits sent from another phone or purchased through a post office or kiosk operator. When phones can be used as currency for purchases or payments, who needs Visa? Vodafone’s Kenyan-based M-PESA mobile banking program landed 200,000 new customers in a month—they’d expected it would take a year to hit that mark. With 1.6 million customers by that time, the service is spreading throughout Africa. The “mobile phone as bank” may bring banking to a billion unserved customers in a few years.

Key Takeaways

  • Moore’s Law applies to the semiconductor industry. The widely accepted managerial interpretation of Moore’s Law states that for the same money, roughly eighteen months from now you should be able to purchase computer chips that are twice as fast or store twice as much information. Or over that same time period, chips with the speed or storage of today’s chips should cost half as much as they do now.
  • Nonchip-based technology also advances rapidly. Disk drive storage doubles roughly every twelve months, while equipment to speed transmissions over fiber-optic lines has doubled every nine months. While these numbers are rough approximations, the price/performance curve of these technologies continues to advance exponentially.
  • These trends influence inventory value, depreciation accounting, employee training, and other managerial functions. They also help improve productivity and keep interest rates low.
  • From a strategic perspective, these trends suggest that what is impossible from a cost or performance perspective today may be possible in the future. This fact provides an opportunity to those who recognize and can capitalize on the capabilities of new technology. As technology advances, new industries, business models, and products are created, while established firms and ways of doing business can be destroyed.
  • Managers must regularly study trends and trajectory in technology to recognize opportunity and avoid disruption.

Questions and Exercises

  1. What is Moore’s Law? What does it apply to?
  2. Are other aspects of computing advancing as well? At what rates?
  3. What is a microprocessor? What devices do you or your family own that contain microprocessors (and hence are impacted by Moore’s Law)?
  4. What is a semiconductor? What is the substance from which most semiconductors are made?
  5. How does flash memory differ from the memory in a PC? Are both solid state?
  6. Which of the following are solid state devices: an iPod shuffle, a TiVo DVR, a typical laptop PC?
  7. Why is Moore’s Law important for managers? How does it influence managerial thinking?
  8. What is price elasticity? How does Moore’s Law relate to this concept? What’s special about falling chip prices compared to price drops for products like clothing or food?
  9. Give examples of firms that have effectively leveraged the advancement of processing, storage, and networking technology.
  10. What are the five waves of computing? Give examples of firms and industries impacted by the fifth wave.
  11. As Moore’s Law advances, technology becomes increasingly accessible to the poor. Give examples of how tech has benefited those who likely would not have been able to afford the technology of a prior generation.
  12. How have cheaper, faster chips impacted the camera industry? Give an example of the leadership shifts that have occurred in this industry.
  13. What has been the impact of “faster, cheaper” on Apple’s business lines?
  14. How did Amazon utilize the steep decline in magnetic storage costs to its advantage?
  15. How does Moore’s Law impact production and inventory decisions?

5.2 The Death of Moore’s Law?

Learning Objectives

After studying this section you should be able to do the following:

  1. Describe why Moore’s Law continues to advance and discuss the physical limitations of this advancement.
  2. Name and describe various technologies that may extend the life of Moore’s Law.
  3. Discuss the limitations of each of these approaches.

Moore simply observed that we’re getting better over time at squeezing more stuff into tinier spaces. Moore’s Law is possible because the distance between the pathways inside silicon chips gets smaller with each successive generation. While chip plants (semiconductor fabrication facilities, or fabsSemiconductor fabrication facilities; the multibillion dollar plants used to manufacture semiconductors.) are incredibly expensive to build, each new generation of fabs can crank out more chips per silicon waferA thin, circular slice of material used to create semiconductor devices. Hundreds of chips may be etched on a single wafer, where they are eventually cut out for individual packaging.. And since the pathways are closer together, electrons travel shorter distances. If electronics now travel half the distance to make a calculation, that means the chip is twice as fast.

But the shrinking can’t go on forever, and we’re already starting to see three interrelated forces—size, heat, and power—threatening to slow down the Moore’s Law gravy train. When you make processors smaller, the more tightly packed electrons will heat up a chip—so much so that unless today’s most powerful chips are cooled down, they will melt inside their packaging. To keep the fastest computers cool, most PCs, laptops, and video game consoles need fans, and most corporate data centers have elaborate and expensive air conditioning and venting systems to prevent a meltdown. A trip through the Facebook data center during its recent rise would show that the firm was a “hot” start-up in more ways than one. The firm’s servers ran so hot that the Plexiglas sides of the firm’s server racks were warped and melting!E. McGirt, “Hacker, Dropout, C.E.O.,” Fast Company, May 2007. The need to cool modern data centers draws a lot of power and that costs a lot of money.

The chief eco officer at Sun Microsystems has claimed that computers draw 4 to 5 percent of the world’s power. Google’s chief technology officer has said that the firm spends more to power its servers than the cost of the servers themselves.D. Kirkpatrick, “The Greenest Computer Company under the Sun,” April 13, 2007. Microsoft, Yahoo! and Google have all built massive data centers in the Pacific Northwest, away from their corporate headquarters, specifically choosing these locations for access to cheap hydroelectric power. Google’s location in The Dalles, Oregon, is charged a cost per kilowatt hour of two cents by the local power provider, less than one-fifth of the eleven-cent rate the firm pays in Silicon Valley.S. Mehta, “Behold the Server Farm,” Fortune, August 1, 2006. Also see Chapter 10 “Software in Flux: Partly Cloudy and Sometimes Free” in this book. This difference means big savings for a firm that runs more than a million servers.

And while these powerful shrinking chips are getting hotter and more costly to cool, it’s also important to realize that chips can’t get smaller forever. At some point Moore’s Law will run into the unyielding laws of nature. While we’re not certain where these limits are, chip pathways certainly can’t be shorter than a single molecule, and the actual physical limit is likely larger than that. Get too small and a phenomenon known as quantum tunneling kicks in, and electrons start to slide off their paths. Yikes!

Buying Time

One way to overcome this problem is with multicore microprocessorsMicroprocessors with two or more (typically lower power) calculating processor cores on the same piece of silicon., made by putting two or more lower power processor cores (think of a core as the calculating part of a microprocessor) on a single chip. Philip Emma, IBM’s Manager of Systems Technology and Microarchitecture, offers an analogy. Think of the traditional fast, hot, single-core processors as a three hundred-pound lineman, and a dual-core processor as two 160-pound guys. Says Emma, “A 300-pound lineman can generate a lot of power, but two 160-pound guys can do the same work with less overall effort.”A. Ashton, “More Life for Moore’s Law,” BusinessWeek, June 20, 2005. For many applications, the multicore chips will outperform a single speedy chip, while running cooler and drawing less power. Multicore processors are now mainstream.

Today, most PCs and laptops sold have at least a two-core (dual-core) processor. The Microsoft Xbox 360 has three cores. The PlayStation 3 includes the so-called cell processor developed by Sony, IBM, and Toshiba that runs nine cores. By 2010, Intel began shipping PC processors with eight cores, while AMD introduced a twelve-core chip. Intel has even demonstrated chips with upwards of fifty cores.

Multicore processors can run older software written for single-brain chips. But they usually do this by using only one core at a time. To reuse the metaphor above, this is like having one of our 160-pound workers lift away, while the other one stands around watching. Multicore operating systems can help achieve some performance gains. Versions of Windows or the Mac OS that are aware of multicore processors can assign one program to run on one core, while a second application is assigned to the next core. But in order to take full advantage of multicore chips, applications need to be rewritten to split up tasks so that smaller portions of a problem are executed simultaneously inside each core.

Writing code for this “divide and conquer” approach is not trivial. In fact, developing software for multicore systems is described by Shahrokh Daijavad, software lead for next-generation computing systems at IBM, as “one of the hardest things you learn in computer science.”A. Ashton, “More Life for Moore’s Law,” BusinessWeek, June 20, 2005. Microsoft’s chief research and strategy officer has called coding for these chips “the most conceptually different [change] in the history of modern computing.”M. Copeland, “A Chip Too Far?” Fortune, September 1, 2008. Despite this challenge, some of the most aggressive adaptors of multicore chips have been video game console manufacturers. Video game applications are particularly well-suited for multiple cores since, for example, one core might be used to render the background, another to draw objects, another for the “physics engine” that moves the objects around, and yet another to handle Internet communications for multiplayer games.

Another approach to breathing life into Moore’s Law is referred to as stacked or three-dimensional semiconductorsSemiconductors that are manufactured as a stack of multiple, interconnected layers instead of in one flat plane.. In this approach, engineers slice a flat chip into pieces, then reconnect the pieces vertically, making a sort of “silicon sandwich.” The chips are both faster and cooler since electrons travel shorter distances. What was once an end-to-end trip on a conventional chip might just be a tiny movement up or down on a stacked chip. But stacked chips present their own challenges. In the same way that a skyscraper is more difficult and costly to design and build than a ranch house, 3-D semiconductors are tougher to design and manufacture. IBM has developed stacked chips for mobile phones, claiming the technique improves power efficiency by up to 40 percent. HP Labs is using a technology called memristors, or memory resistors, to improve on conventional transistors and speed the transition to 3-D chips, yielding significant improvement over 2-D offerings.J. Markoff, “HP Sees a Revolution in Memory Chip,” New York Times, April 7, 2010.

Quantum Leaps, Chicken Feathers, and the Indium Gallium Arsenide Valley?

Think about it—the triple threat of size, heat, and power means that Moore’s Law, perhaps the greatest economic gravy train in history, will likely come to a grinding halt in your lifetime. Multicore and 3-D semiconductors are here today, but what else is happening to help stave off the death of Moore’s Law?

Every once in a while a material breakthrough comes along that improves chip performance. A few years back researchers discovered that replacing a chip’s aluminum components with copper could increase speeds up to 30 percent. Now scientists are concentrating on improving the very semiconductor material that chips are made of. While the silicon used in chips is wonderfully abundant (it has pretty much the same chemistry found in sand), researchers are investigating other materials that might allow for chips with even tighter component densities. Researchers have demonstrated that chips made with supergeeky-sounding semiconductor materials such as indium gallium arsenide, germanium, and bismuth telluride can run faster and require less wattage than their silicon counterparts.Y. L. Chen, J. G. Analytis, J.-H. Chu, Z. K. Liu, S.-K. Mo, X. L. Qi, H. J. Zhang, et al., “Experimental Realization of a Three-Dimensional Topological Insulator, Bi2Te3,” Science 325, no. 5937 (July 10, 2009): 178–81; K. Greene, “Intel Looks Beyond Silicon,” Technology Review, December 11, 2007; and A. Cane, “A Valley By Any Other Name…” Financial Times, December 11, 2006. Perhaps even more exotic (and downright bizarre), researchers at the University of Delaware have experimented with a faster-than-silicon material derived from chicken feathers! Hyperefficient chips of the future may also be made out of carbon nanotubes, once the technology to assemble the tiny structures becomes commercially viable.

Other designs move away from electricity over silicon. Optical computing, where signals are sent via light rather than electricity, promises to be faster than conventional chips, if lasers can be mass produced in miniature (silicon laser experiments show promise). Others are experimenting by crafting computing components using biological material (think a DNA-based storage device).

One yet-to-be-proven technology that could blow the lid off what’s possible today is quantum computing. Conventional computing stores data as a combination of bits, where a bit is either a one or a zero. Quantum computers, leveraging principles of quantum physics, employ qubits that can be both one and zero at the same time. Add a bit to a conventional computer’s memory and you double its capacity. Add a bit to a quantum computer and its capacity increases exponentially. For comparison, consider that a computer model of serotonin, a molecule vital to regulating the human central nervous system, would require 1094 bytes of information. Unfortunately there’s not enough matter in the universe to build a computer that big. But modeling a serotonin molecule using quantum computing would take just 424 qubits.P. Kaihla, “Quantum Leap,” Business 2.0, August 1, 2004.

Some speculate that quantum computers could one day allow pharmaceutical companies to create hyperdetailed representations of the human body that reveal drug side effects before they’re even tested on humans. Quantum computing might also accurately predict the weather months in advance or offer unbreakable computer security. Ever have trouble placing a name with a face? A quantum computer linked to a camera (in your sunglasses, for example) could recognize the faces of anyone you’ve met and give you a heads-up to their name and background.P. Schwartz, C. Taylor, and R. Koselka, “The Future of Computing: Quantum Leap,” Fortune, August 2, 2006. Opportunities abound. Of course, before quantum computing can be commercialized, researchers need to harness the freaky properties of quantum physics wherein your answer may reside in another universe, or could disappear if observed (Einstein himself referred to certain behaviors in quantum physics as “spooky action at a distance”).

Pioneers in quantum computing include IBM, HP, NEC, and a Canadian start-up named D-Wave. If or when quantum computing becomes a reality is still unknown, but the promise exists that while Moore’s Law may run into limits imposed by Mother Nature, a new way of computing may blow past anything we can do with silicon, continuing to make possible the once impossible.

Key Takeaways

  • As chips get smaller and more powerful, they get hotter and present power-management challenges. And at some, point Moore’s Law will stop because we will no longer be able to shrink the spaces between components on a chip.
  • Multicore chips use two or more low-power calculating “cores” to work together in unison, but to take optimal advantage of multicore chips, software must be rewritten to “divide” a task among multiple cores.
  • 3-D or stackable semiconductors can make chips faster and run cooler by shortening distances between components, but these chips are harder to design and manufacture.
  • New materials may extend the life of Moore’s Law, allowing chips to get smaller, still. Entirely new methods for calculating, such as quantum computing, may also dramatically increase computing capabilities far beyond what is available today.

Questions and Exercises

  1. What three interrelated forces threaten to slow the advancement of Moore’s Law?
  2. Which commercial solutions, described in the section above, are currently being used to counteract the forces mentioned above? How do these solutions work? What are the limitations of each?
  3. Will multicore chips run software designed for single-core processors?
  4. As chips grow smaller they generate increasing amounts of heat that needs to be dissipated. Why is keeping systems cool such a challenge? What are the implications for a firm like Yahoo! or Google? For a firm like Apple or Dell?
  5. What are some of the materials that may replace the silicon that current chips are made of?
  6. What kinds of problems might be solved if the promise of quantum computing is achieved? How might individuals and organizations leverage quantum computing? What sorts of challenges could arise from the widespread availability of such powerful computing technology?

5.3 Bringing Brains Together: Supercomputing and Grid Computing

Learning Objectives

After studying this section you should be able to do the following:

  1. Give examples of the business use of supercomputing and grid computing.
  2. Describe grid computing and discuss how grids transform the economics of supercomputing.
  3. Understand the characteristics of problems that are and are not well suited for supercomputing and grid computing.

As Moore’s Law makes possible the once impossible, businesses have begun to demand access to the world’s most powerful computing technology. SupercomputersComputers that are among the fastest of any in the world at the time of their introduction. are computers that are among the fastest of any in the world at the time of their introduction.A list of the current supercomputer performance champs can be found at http://www.top500.org. Supercomputing was once the domain of governments and high-end research labs, performing tasks such as simulating the explosion of nuclear devices, or analyzing large-scale weather and climate phenomena. But it turns out with a bit of tweaking, the algorithms used in this work are profoundly useful to business. Consider perhaps the world’s most well-known supercomputer, IBM’s Deep Blue, the machine that rather controversially beat chess champion Garry Kasparov. While there is not a burning need for chess-playing computers in the world’s corporations, it turns out that the computing algorithms to choose the best among multiple chess moves are similar to the math behind choosing the best combination of airline flights.

One of the first customers of Deep Blue technologies was United Airlines, which gained an ability to examine three hundred and fifty thousand flight path combinations for its scheduling systems—a figure well ahead of the previous limit of three thousand. Estimated savings through better yield management? Over $50 million! Finance found uses, too. An early adopter was CIBC (the Canadian Imperial Bank of Commerce), one of the largest banks in North America. Each morning CIBC uses a supercomputer to run its portfolio through Monte Carlo simulations that aren’t all that different from the math used to simulate nuclear explosions. An early adopter of the technology, at the time of deployment, CIBC was the only bank that international regulators allowed to calculate its own capital needs rather than use boilerplate ratios. That cut capital on hand by hundreds of millions of dollars, a substantial percentage of the bank’s capital, saving millions a year in funding costs. Also noteworthy: the supercomputer-enabled, risk-savvy CIBC was relatively unscathed by the subprime crisis.

Modern supercomputing is typically done via a technique called massively parallelComputers designed with many microprocessors that work together, simultaneously, to solve problems. processing (computers designed with many microprocessors that work together, simultaneously, to solve problems). The fastest of these supercomputers are built using hundreds of microprocessors, all programmed to work in unison as one big brain. While supercomputers use special electronics and software to handle the massive load, the processors themselves are often of the off-the-shelf variety that you’d find in a typical PC. Virginia Tech created what at the time was the world’s third-fastest supercomputer by using chips from 1,100 Macintosh computers lashed together with off-the-shelf networking components. The total cost of the system was just $5.2 million, far less than the typical cost for such burly hardware. The Air Force recently issued a request-for-proposal to purchase 2,200 PlayStation 3 systems in hopes of crafting a supercheap, superpowerful machine using off-the-shelf parts.

Another technology, known as grid computingA type of computing that uses special software to enable several computers to work together on a common problem as if they were a massively parallel supercomputer., is further transforming the economics of supercomputing. With grid computing, firms place special software on its existing PCs or servers that enables these computers to work together on a common problem. Large organizations may have thousands of PCs, but they’re not necessarily being used all the time, or at full capacity. With grid software installed on them, these idle devices can be marshaled to attack portions of a complex task as if they collectively were one massively parallel supercomputer. This technique radically changes the economics of high-performance computing. BusinessWeek reports that while a middle-of-the-road supercomputer could run as much as $30 million, grid computing software and services to perform comparable tasks can cost as little as twenty-five thousand dollars, assuming an organization already has PCs and servers in place.

An early pioneer in grid computing is the biotech firm Monsanto. Monsanto enlists computers to explore ways to manipulate genes to create crop strains that are resistant to cold, drought, bugs, pesticides, or that are more nutritious. Previously with even the largest computer Monsanto had in-house, gene analysis was taking six weeks and the firm was able to analyze only ten to fifty genes a year. But by leveraging grid computing, Monsanto has reduced gene analysis to less than a day. The fiftyfold time savings now lets the firm consider thousands of genetic combinations in a year.P. Schwartz, C. Taylor, and R. Koselka, “The Future of Computing: Quantum Leap,” Fortune, August 2, 2006. Lower R&D time means faster time to market—critical to both the firm and its customers.

Grids are now everywhere. Movie studios use them to create special effects and animated films. Proctor & Gamble has used grids to redesign the manufacturing process for Pringles potato chips. GM and Ford use grids to simulate crash tests, saving millions in junked cars and speeding time to market. Pratt and Whitney test aircraft engine designs on a grid. And biotech firms including Aventis, GlaxoSmithKline, and Pfizer push their research through a quicker pipeline by harnessing grid power. JP Morgan Chase even launched a grid effort that mimics CIBC’s supercomputer, but at a fraction of the latter’s cost. By the second year of operation, the JPMorgan Chase grid was saving the firm $5 million per year.

You can join a grid, too. SETI@Home turns your computer screen saver into a method to help “search for extraterrestrial intelligence,” analyzing data from the Arecibo radio telescope system in Puerto Rico (no E.T. spotted yet). FightAids@Home will enlist your PC to explore AIDS treatments. And Folding@Home is an effort by Stanford researchers to understanding the science of protein-folding within diseases such as Alzheimer’s, cancer, and cystic fibrosis. A version of Folding@Home software for the PlayStation 3 had enlisted over half a million consoles within months of release. Having access to these free resources is an enormous advantage for researchers. Says the director of Folding@Home, “Even if we were given all of the NSF supercomputing centers combined for a couple of months, that is still fewer resources than we have now.”G. Johnson, “Supercomputing ‘@Home’ Is Paying Off,” New York Times, April 23, 2002.

Multicore, massively parallel, and grid computing are all related in that each attempts to lash together multiple computing devices so that they can work together to solve problems. Think of multicore chips as having several processors in a single chip. Think of massively parallel supercomputers as having several chips in one computer, and think of grid computing as using existing computers to work together on a single task (essentially a computer made up of multiple computers). While these technologies offer great promise, they’re all subject to the same limitation: software must be written to divide existing problems into smaller pieces that can be handled by each core, processor, or computer, respectively. Some problems, such as simulations, are easy to split up, but for problems that are linear (where, for example, step two can’t be started until the results from step one are known), the multiple-brain approach doesn’t offer much help.

Massive clusters of computers running software that allows them to operate as a unified service also enable new service-based computing models, such as software as a service (SaaS)A form of cloud computing where a firm subscribes to a third-party software and receives a service that is delivered online. and cloud computingReplacing computing resources—either an organization’s or individual’s hardware or software—with services provided over the Internet.. In these models, organizations replace traditional software and hardware that they would run in-house with services that are delivered online. Google, Microsoft, Salesforce.com, and Amazon are among the firms that have sunk billions into these Moore’s Law–enabled server farmsA massive network of computer servers running software to coordinate their collective use. Server farms provide the infrastructure backbone to SaaS and hardware cloud efforts, as well as many large-scale Internet services., creating entirely new businesses that promise to radically redraw the software and hardware landscape while bringing gargantuan computing power to the little guy. (See Chapter 10 “Software in Flux: Partly Cloudy and Sometimes Free”.)

Moore’s Law will likely hit its physical limit in your lifetime, but no one really knows if this “Moore’s Wall” is a decade away or more. What lies ahead is anyone’s guess. Some technologies, such as still-experimental quantum computing, could make computers that are more powerful than all the world’s conventional computers combined. Think strategically—new waves of innovation might soon be shouting “surf’s up!”

Key Takeaways

  • Most modern supercomputers use massive sets of microprocessors working in parallel.
  • The microprocessors used in most modern supercomputers are often the same commodity chips that can be found in conventional PCs and servers.
  • Moore’s Law means that businesses as diverse as financial services firms, industrial manufacturers, consumer goods firms, and film studios can now afford access to supercomputers.
  • Grid computing software uses existing computer hardware to work together and mimic a massively parallel supercomputer. Using existing hardware for a grid can save a firm the millions of dollars it might otherwise cost to buy a conventional supercomputer, further bringing massive computing capabilities to organizations that would otherwise never benefit from this kind of power.
  • Massively parallel computing also enables the vast server farms that power online businesses like Google and Facebook, and which create new computing models, like software as a service (SaaS) and cloud computing.
  • The characteristics of problems best suited for solving via multicore systems, parallel supercomputers, or grid computers are those that can be divided up so that multiple calculating components can simultaneously work on a portion of the problem. Problems that are linear—where one part must be solved before moving to the next and the next—may have difficulty benefiting from these kinds of “divide and conquer” computing. Fortunately many problems such as financial risk modeling, animation, manufacturing simulation, and gene analysis are all suited for parallel systems.

Questions and Exercises

  1. What is the difference between supercomputing and grid computing? How is each phenomenon empowered by Moore’s Law?
  2. How does grid computing change the economics of supercomputing?
  3. Which businesses are using supercomputing and grid computing? Describe these uses and the advantages they offer their adopting firms. Are they a source of competitive advantage? Why or why not?
  4. What are the characteristics of problems that are most easily solved using the types of parallel computing found in grids and modern day supercomputers? What are the characteristics of the sorts of problems not well suited for this type of computing?
  5. Visit the SETI@Home Web site (http://seti.ssl.berkeley.edu). What is the purpose of the SETI@Home project? How do you participate? Is there any possible danger to your computer if you choose to participate? (Read their rules and policies.)
  6. Search online to identify the five fastest supercomputers currently in operation. Who sponsors these machines? What are they used for? How many processors do they have?
  7. What is “Moore’s Wall”?
  8. What is the advantage of using grid computing to simulate an automobile crash test as opposed to actually staging a crash?

5.4 E-waste: The Dark Side of Moore’s Law

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the magnitude of the environmental issues caused by rapidly obsolete, faster and cheaper computing.
  2. Explain the limitations of approaches attempting to tackle e-waste.
  3. Understand the risks firms are exposed to when not fully considering the lifecycle of the products they sell or consume.
  4. Ask questions that expose concerning ethical issues in a firm or partner’s products and processes, and that help the manager behave more responsibly.

We should celebrate the great bounty Moore’s Law and the tech industry bestow on our lives. Costs fall, workers become more productive, innovations flourish, and we gorge at a buffet of digital entertainment that includes music, movies, and games. But there is a dark side to this faster and cheaper advancement. A PC has an expected lifetime of three to five years. A cell phone? Two years or less. Rapid obsolescence means the creation of ever-growing mountains of discarded tech junk, known as electronic waste or e-wasteDiscarded, often obsolete technology; also known as electronic waste.. According to the U.S. Environmental Protection Agency (EPA), in 2007 the United States alone generated over 2.5 million tons of e-waste,U.S. Environmental Protection Agency, General Information on E-waste, February 5, 2009. and the results aren’t pretty. Consumer electronics and computing equipment can be a toxic cocktail that includes cadmium, mercury, lead, and other hazardous materials. Once called the “effluent of the affluent,” e-waste will only increase with the rise of living standards worldwide.

The quick answer would be to recycle this stuff. Not only does e-waste contain mainstream recyclable materials we’re all familiar with, like plastics and aluminum, it also contains small bits of increasingly valuable metals such as silver, platinum, and copper. In fact, there’s more gold in one pound of discarded tech equipment than in one pound of mined ore.P. Kovessy, “How to Trash Toxic Tech,” Ottawa Business Journal, May 12, 2008. But as the sordid record of e-waste management shows, there’s often a disconnect between consumers and managers who want to do good and those efforts that are actually doing good. The complexities of the modern value chain, the vagaries of international law, and the nefarious actions of those willing to put profits above principle show how difficult addressing this problem will be.

The process of separating out the densely packed materials inside tech products so that the value in e-waste can be effectively harvested is extremely labor intensive, more akin to reverse manufacturing than any sort of curbside recycling efforts. Sending e-waste abroad can be ten times cheaper than dealing with it at home,C. Bodeen, “In ‘E-waste’ Heartland, a Toxic China,” International Herald Tribune, November 18, 2007. so it’s not surprising that up to 80 percent of the material dropped off for recycling is eventually exported.E. Royte, “E-waste@Large,” New York Times, January 27, 2006. Much of this waste ends up in China, South Asia, or sub-Saharan Africa, where it is processed in dreadful conditions.

Consider the example of Guiyu, China, a region whose poisoning has been extensively chronicled by organizations such as the Silicon Valley Toxics Coalition, the Basel Action Network (BAN), and Greenpeace. Workers in and around Guiyu toil without protective equipment, breathing clouds of toxins generated as they burn the plastic skins off of wires to get at the copper inside. Others use buckets, pots, or wok-like pans (in many cases the same implements used for cooking) to sluice components in acid baths to release precious metals—recovery processes that create even more toxins. Waste sludge and the carcasses of what’s left over are most often dumped in nearby fields and streams. Water samples taken in the region showed lead and heavy metal contamination levels some four hundred to six hundred times greater than what international standards deem safe.E. Grossman, “Where Computers Go to Die—and Kill,” Salon.com, April 10, 2006, http://www.salon.com/news/feature/2006/04/10/ewaste. The area is so polluted that drinking water must be trucked in from eighteen miles away. Pregnancies are six times more likely to end in miscarriage, and 70 percent of the kids in the region have too much lead in their blood.60 Minutes, “Following the Trail of Toxic E-waste,” November 9, 2008.

Figure 5.5 Photos from Guiyu, ChinaJ. Biggs, “Guiyu, the E-waste Capital of China,” CrunchGear, April 4, 2008.

China cares about its environment. The nation has banned the importing of e-waste since 2000.E. Grossman, “Where Computers Go to Die—and Kill,” Salon.com, April 10, 2006. But corruption ensures that e-waste continues to flow into the country. According to one exporter, all that’s required to get e-waste past customs authorities is to tape a one-hundred-dollar bill on the side of the container.C. Bodeen, “In ‘E-waste’ Heartland, a Toxic China,” International Herald Tribune, November 18, 2007. Well-meaning U.S. recyclers, as well as those attempting to collect technology for reuse in poorer countries, are often in the dark as to where their products end up.

The trade is often brokered by middlemen who mask the eventual destination and fate of the products purchased. BAN investigators in Lagos, Nigeria, documented mountains of e-waste with labels from schools, U.S. government agencies, and even some of the world’s largest corporations. And despite Europe’s prohibition on exporting e-waste, many products originally labeled for repair and reuse end up in toxic recycling efforts. Even among those products that gain a second or third life in developing nations, the inevitable is simply postponed, with e-waste almost certain to end up in landfills that lack the protective groundwater barriers and environmental engineering of their industrialized counterparts. The reality is that e-waste management is extraordinarily difficult to monitor and track, and loopholes are rampant.

Thinking deeply about the ethical consequences of a firm’s business is an imperative for the modern manager. A slip up (intentional or not) can, in seconds, be captured by someone with a cell phone, uploaded to YouTube, or offered in a blog posting for the world to see. When Dell was caught using Chinese prison labor as part of its recycling efforts, one blogger chastised the firm with a tweak of its marketing tagline, posting “Dude, you’re getting a cell.”J. Russell, “Dell under Attack over Using Prison Labour,” Inquirer, January 10, 2003. See also http://laughingmeme.org/2003/03/23/dell-recycling-a-ways-to-go-still. The worst cases expose firms to legal action and can tarnish a brand for years. Big firms are big targets, and environmentalists have been quick to push the best-known tech firms and retailers to take back their products for responsible recycling and to eliminate the worst toxins from their offerings.

Consider that even Apple (where Al Gore sits on the firm’s Board of Directors), has been pushed by a coalition of environmental groups on all of these fronts. Critics have shot back that signaling out Apple is unfair. The firm was one of the first computer companies to eliminate lead-lined glass monitors from its product line, and has been a pioneer of reduced-sized packaging that leverage recyclable materials. And Apple eventually claimed the top in Greenpeace’s “Greener Electronics” rankings.J. Dalrymple, “Apple Ranks Highest among Greenpeace’s Top Tech Companies,” The Loop, January 7, 2010. But if the firm that counts Al Gore among its advisors can get tripped up on green issues, all firms are vulnerable.

Environmentalists see this pressure to deal with e-waste as yielding results: Apple and most other tech firms have continually moved to eliminate major toxins from their manufacturing processes. All this demonstrates that today’s business leaders have to be far more attuned to the impact not only of their own actions, but also to those of their suppliers and partners. How were products manufactured? Using which materials? Under what conditions? What happens to items when they’re discarded? Who provides collection and disposal? It also shows the futility of legislative efforts that don’t fully consider and address the problems they are meant to target.

Finding Responsible E-waste Disposers

A recent sting operation led by the U.S. Government Accountability Office (U.S. GAO) found that forty-three American recyclers were willing to sell e-waste illegally to foreign countries, without gaining EPA or foreign country approval. Appallingly, at least three of them held Earth Day electronics-recycling events.U.S. Government Accountability Office (U.S. GAO), Report to the Chairman: Committee on Foreign Affairs, House of Representatives: Electronic Waste, August 2008.

So how can firms and individuals choose proper disposal partners? Several certification mechanisms can help shed light on whether the partner you’re dealing with is a responsible player. The Basel Action Network e-Stewards program certifies firms via a third-party audit, with compliant participants committing to eliminating e-waste export, land dumping, incineration, and toxic recycling via prison labor. The International Association of Electronics Recyclers (IAER) also offers audited electronics recycler certification. And firms certified as ISO 9001 and ISO 14001 compliant attest to quality management and environmental processes. Standards, techniques, and auditing practices are constantly in flux, so consult these organizations for the latest partner lists, guidelines, and audit practices.Basal Action Network e-Stewards program accessed via http://www.e-stewards.org/esteward_certification.html; International Standards Organization accessed via http://www.iso.org/iso/home.htm; the IAER accessed via http://www.iaer.org/search; and G. MacDonald, “Don’t Recycle ‘E-waste’ with Haste, Activists Warn,” USA Today, July 6, 2008.

Which brings us back to Gordon Moore. To his credit, Moore is not just the founder of the world’s largest microprocessor firm and first to identify the properties we’ve come to know as Moore’s Law, he has also emerged as one of the world’s leading supporters of environmental causes. The generosity of the Gordon and Betty Moore foundation includes, among other major contributions, the largest single gift to a private conservation organization. Indeed, Silicon Valley, while being the birthplace of products that become e-waste, also promises to be at the forefront of finding solutions to modern environmental challenges. The Valley’s leading venture capitalists, including Sequoia and Kleiner Perkins (where Al Gore is now a partner), have started multimillion-dollar green investment funds, targeted at funding the next generation of sustainable, environmental initiatives.

Key Takeaways

  • E-waste may be particularly toxic since many components contain harmful materials such as lead, cadmium, and mercury.
  • Managers must consider and plan for the waste created by their products, services, and technology used by the organization. Consumers and governments are increasingly demanding that firms offer responsible methods for the disposal of their manufactured goods and the technology used in their operations.
  • Managers must audit disposal and recycling partners with the same vigor as their suppliers and other corporate partners. If not, an organization’s equipment may end up in environmentally harmful disposal operations.

Questions and Exercises

  1. What is e-waste? What is so dangerous about e-waste?
  2. What sorts of materials might be harvested from e-waste recycling?
  3. Many well-meaning individuals thought that recycling was the answer to the e-waste problem. But why hasn’t e-waste recycling yielded the results hoped for?
  4. What lessons do the challenges of e-waste offer the manager? What issues will your firm need to consider as it consumes or offers products that contain computing components?
  5. Why is it difficult to recycle e-waste?
  6. Why is e-waste exported abroad for recycling rather than processed domestically?
  7. What part does corruption play in the recycling and disposal of e-waste?
  8. What part might product design and production engineering play in the reduction of the impact of technology waste on the environment?
  9. What are the possible consequences should a U.S. firm be deemed “environmentally irresponsible”?
  10. Name two companies that have incurred the wrath of environmental advocates. What might these firms have done to avoid such criticism?

Chapter 4: Netflix: The Making of an E-commerce Giant and the Uncertain Future of Atoms to Bits

4.1 Introduction

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the basics of the Netflix business model.
  2. Recognize the downside the firm may have experienced from an early IPO.
  3. Appreciate why other firms found Netflix’s market attractive, and why many analysts incorrectly suspected Netflix was doomed.

Entrepreneurs are supposed to want to go public. When a firm sells stock for the first time, the company gains a ton of cash to fuel expansion and its founders get rich. Going public is the dream in the back of the mind of every tech entrepreneur. But in 2007, Netflix founder and CEO Reed Hastings told Fortune that if he could change one strategic decision, it would have been to delay the firm’s initial public stock offering (IPO)Also known as “going public.” The first time a firm sells stock to the public.: “If we had stayed private for another two to four years, not as many people would have understood how big a business this could be.”M. Boyle, “Questions for…Reed Hastings,” Fortune, May 23, 2007. Once Netflix was a public company, financial disclosure rules forced the firm to reveal that it was on a money-minting growth tear. Once the secret was out, rivals showed up.

Hollywood’s best couldn’t have scripted a more menacing group of rivals for Hastings to face. First in line with its own DVD-by-mail offering was Blockbuster, a name synonymous with video rental. Some 40 million U.S. families were already card-carrying Blockbuster customers, and the firm’s efforts promised to link DVD-by-mail with the nation’s largest network of video stores. Following close behind was Wal-Mart—not just a big Fortune 500 company but the largest firm in the United States ranked by sales. In Netflix, Hastings had built a great firm, but let’s face it, his was a dot-com, an Internet pure playA firm that focuses on a specific product, service, or business model. An Internet pure play is a firm that only operates through the Internet channel (i.e., with no physical stores or catalogs). without a storefront and with an overall customer base that seemed microscopic compared to these behemoths.

Before all this, Netflix was feeling so confident that it had actually raised prices. Customers loved the service, the company was dominating its niche, and it seemed like the firm could take advantage of a modest price hike, pull in more revenue, and use this to improve and expand the business. But the firm was surprised by how quickly the newcomers mimicked Netflix with cheaper rival efforts. This new competition forced Netflix to cut prices even lower than where they had been before the price increase. To keep pace, Netflix also upped advertising at a time when online ad rates were increasing. Big competitors, a price war, spending on the rise—how could Netflix possibly withstand this onslaught? Some Wall Street analysts had even taken to referring to Netflix’s survival prospects as “The Last Picture Show.”M. Conlin, “Netflix: Flex to the Max,” BusinessWeek, September 24, 2007.

Fast-forward a year later and Wal-Mart had cut and run, dumping their experiment in DVD-by-mail. Blockbuster had been mortally wounded, hemorrhaging billions of dollars in a string of quarterly losses. And Netflix? Not only had the firm held customers, it grew bigger, recording record profits. The dot-com did it. Hastings, a man who prior to Netflix had already built and sold one of the fifty largest public software firms in the United States, had clearly established himself as one of America’s most capable and innovative technology leaders. In fact, at roughly the same time that Blockbuster CEO John Antioco resigned, Reed Hastings accepted an appointment to the Board of Directors of none other than the world’s largest software firm, Microsoft. Like the final scene in so many movies where the hero’s face is splashed across the news, Time named Hastings as one of the “100 most influential global citizens.”

Why Study Netflix?

Studying Netflix gives us a chance to examine how technology helps firms craft and reinforce a competitive advantage. We’ll pick apart the components of the firm’s strategy and learn how technology played a starring role in placing the firm atop its industry. We also realize that while Netflix emerged the victorious underdog at the end of the first show, there will be at least one sequel, with the final scene yet to be determined. We’ll finish the case with a look at the very significant challenges the firm faces as new technology continues to shift the competitive landscape.

How Netflix Works

Figure 4.1

Reed Hastings, a former Peace Corps volunteer with a master’s in computer science, got the idea for Netflix when he was late in returning the movie Apollo 13 to his local video store. The forty-dollar late fee was enough to have bought the video outright with money left over. Hastings felt ripped off, and out of this initial outrage, Netflix was born. The model the firm eventually settled on was a DVD-by-mail service that charged a flat-rate monthly subscription rather than a per-disc rental fee. Customers don’t pay a cent in mailing expenses, and there are no late fees.

Netflix offers nine different subscription plans, starting at less than five dollars. The most popular is a $16.99 option that offers customers three movies at a time and unlimited returns each month. Videos arrive in red Mylar envelopes. After tearing off the cover to remove the DVD, customers reveal prepaid postage and a return address. When done watching videos, consumers just slip the DVD back into the envelope, reseal it with a peel-back sticky-strip, and drop the disc in the mail. Users make their video choices in their “request queue” at Netflix.com.

If a title isn’t available, Netflix simply moves to the next title in the queue. Consumers use the Web site to rate videos they’ve seen, specify their movie preferences, get video recommendations, check out DVD details, and even share their viewing habits and reviews. In 2007, the firm added a “Watch Now” button next to those videos that could be automatically streamed to a PC. Any customer paying at least $8.99 for a DVD-by-mail subscription plan can stream an unlimited number of videos each month at no extra cost.

Key Takeaways

  • Analysts and managers have struggled to realize that dot-com start-up Netflix could actually create sustainable competitive advantage, beating back challenges from Wal-Mart and Blockbuster, among others.
  • Data disclosure required by public companies may have attracted these larger rivals to the firm’s market.
  • Netflix operates via a DVD subscription and video streaming model. Although sometimes referred to as “rental,” the model is really a substitute good for conventional use-based media rental.

Questions and Exercises

  1. How does the Netflix business model work?
  2. Which firms are or have been Netflix’s most significant competitors? How do their financial results or performance of their efforts compare to Netflix’s efforts?
  3. What recent appointment did Reed Hastings accept in addition to his job as Netflix CEO? Why is this appointment potentially important for Netflix?
  4. Why did Wal-Mart and Blockbuster managers, as well as Wall Street analysts, underestimate Netflix? What issues might you advise analysts and managers to consider so that they avoid making these sorts of mistakes in the future?

4.2 Tech and Timing: Creating Killer Assets

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand how many firms have confused brand and advertising, why branding is particularly important for online firms, and the factors behind Netflix’s exceptional brand strength.
  2. Understand the long tail concept, and how it relates to Netflix’s ability to offer the customer a huge (the industry’s largest) selection of movies.
  3. Know what collaborative filtering is, how Netflix uses collaborative filtering software to match movie titles with the customer’s taste, and in what ways this software helps Netflix garner sustainable competitive advantage.
  4. List and discuss the several technologies Netflix uses in its operations to reduce costs and deliver customer satisfaction and enhance brand value.
  5. Understand the role that scale economies play in Netflix’s strategies, and how these scale economies pose an entry barrier to potential competitors.
  6. Understand the role that market entry timing has played in the firm’s success.

To understand Netflix’s strengths, it’s important to view the firm as its customers see it. And for the most part, what they see they like—a lot! Netflix customers are rabidly loyal and rave about the service. The firm repeatedly ranks at the top of customer satisfaction surveys. Ratings agency ForeSee has named Netflix the number one e-commerce site in terms of customer satisfaction nine times in a row (placing it ahead of Apple and Amazon, among others). Netflix has also been cited as the best at satisfying customers by Nielsen and Fast Company, and was also named the Retail Innovator of the Year by the National Retail Federation.

Building a great brand, especially one online, starts with offering exceptional value to the customer. Don’t confuse branding with advertising. During the dot-com era, firms thought brands could be built through Super Bowl ads and expensive television promotion. Advertising can build awareness, but brands are built through customer experience. This is a particularly important lesson for online firms. Have a bad experience at a burger joint and you might avoid that location but try another of the firm’s outlets a few blocks away. Have a bad experience online and you’re turned off by the firm’s one and only virtual storefront. If you click over to an online rival, the offending firm may have lost you forever. But if a firm can get you to stay through quality experience, switching costs and data-driven value might keep you there for a long, long time, even when new entrants try to court you away.

If brand is built through customer experience, consider what this means for the Netflix subscriber. They expect the firm to offer a huge selection, to be able to find what they want, for it to arrive on time, for all of this to occur with no-brainer ease of use and convenience, and at a fair price. Technology drives all of these capabilities, so tech is at the very center of the firm’s brand building efforts. Let’s look at how the firm does it.

Selection: The Long Tail in Action

Customers have flocked to Netflix in part because of the firm’s staggering selection. A traditional video store (and Blockbuster had some 7,800 of them) stocks roughly three thousand DVD titles on its shelves. For comparison, Netflix is able to offer its customers a selection of over one hundred thousand DVD titles, and rising! At traditional brick-and-mortar retailers, shelf space is the biggest constraint limiting a firm’s ability to offer customers what they want when they want it. Just which films, documentaries, concerts, cartoons, TV shows, and other fare make it inside the four walls of a Blockbuster store is dictated by what the average consumer is most likely to be interested in. To put it simply, Blockbuster stocks blockbusters.

Finding the right product mix and store size can be tricky. Offer too many titles in a bigger storefront and there may not be enough paying customers to justify stocking less popular titles (remember, it’s not just the cost of the DVD—firms also pay for the real estate of a larger store, the workers, the energy to power the facility, etc.). You get the picture—there’s a breakeven point that is arrived at by considering the geographic constraint of the number of customers that can reach a location, factored in with store size, store inventory, the payback from that inventory, and the cost to own and operate the store. Anyone who has visited a video store only to find a title out of stock has run up against the limits of the physical store model.

But many online businesses are able to run around these limits of geography and shelf space. Internet firms that ship products can get away with having just a few highly automated warehouses, each stocking just about all the products in a particular category. And for firms that distribute products digitally (think songs on iTunes), the efficiencies are even greater because there’s no warehouse or physical product at all (more on that later).

Offer a nearly limitless selection and something interesting happens: there’s actually more money to be made selling the obscure stuff than the hits. Music service Rhapsody makes more from songs outside of the top ten thousand than it does from songs ranked above ten thousand. At Amazon.com, roughly 60 percent of books sold are titles that aren’t available in even the biggest Borders or Barnes & Noble Superstores.C. Anderson, “The Long Tail,” Wired 12, no. 10 (October 2004), http://www.wired.com/wired/archive/12.10/tail.html. And at Netflix, roughly 75 percent of DVD titles shipped are from back-catalog titles, not new releases (at Blockbuster outlets the equation is nearly flipped, with some 70 percent of business coming from new releases).B. McCarthy, “Netflix, Inc.” (remarks, J. P. Morgan Global Technology, Media, and Telecom Conference, Boston, May 18, 2009). Consider that Netflix sends out forty-five thousand different titles each day. That’s fifteen times the selection available at your average video store! Each quarter, roughly 95 percent of titles are viewed—that means that every few weeks Netflix is able to find a customer for nearly every DVD title that has ever been commercially released.

This phenomenon whereby firms can make money by selling a near-limitless selection of less-popular products is known as the long tailIn this context, refers to an extremely large selection of content or products. The long tail is a phenomenon whereby firms can make money by offering a near-limitless selection.. The term was coined by Chris Anderson, an editor at Wired magazine, who also wrote a best-selling business book by the same name. The “tail” (see Figure 4.2 “The Long Tail”) refers to the demand for less popular items that aren’t offered by traditional brick-and-mortar shops. While most stores make money from the area under the curve from the vertical axis to the dotted line, long tail firms can also sell the less popular stuff. Each item under the right part of the curve may experience less demand than the most popular products, but someone somewhere likely wants it. And as demonstrated from the examples above, the total demand for the obscure stuff is often much larger than what can be profitably sold through traditional stores alone. While some debate the size of the tail (e.g., whether obscure titles collectively are more profitable for most firms), two facts are critical to keep above this debate: (1) selection attracts customers, and (2) the Internet allows large-selection inventory efficiencies that offline firms can’t match.

Figure 4.2 The Long Tail

The long tail works because the cost of production and distribution drop to a point where it becomes economically viable to offer a huge selection. For Netflix, the cost to stock and ship an obscure foreign film is the same as sending out the latest Will Smith blockbuster. The long tail gives the firm a selection advantage (or one based on scale) that traditional stores simply cannot match.

For more evidence that there is demand for the obscure stuff, consider Bollywood cinema—a term referring to films produced in India. When ranked by the number of movies produced each year, Bollywood is actually bigger than Hollywood, but in terms of U.S. demand, even the top-grossing Hindi film might open in only one or two American theaters, and few video stores carry many Bollywood DVDs. Again, we see the limits that geography and shelf space impose on traditional stores. As Anderson puts it, when it comes to traditional methods of distribution, “an audience too thinly spread is the same as no audience at all.”C. Anderson, “The Long Tail,” Wired 12, no. 10 (October 2004), http://www.wired.com/wired/archive/12.10/tail.html. While there are roughly 1.7 million South Asians living in the United States, Bollywood fans are geographically disbursed, making it difficult to offer content at a physical storefront. Fans of foreign films would often find the biggest selection at an ethnic grocery store, but even then, that wouldn’t be much. Enter Netflix. The firm has found the U.S. fans of South Asian cinema, sending out roughly one hundred thousand Bollywood DVDs a month. As geographic constraints go away, untapped markets open up!

The power of Netflix can revive even well-regarded work by some of Hollywood’s biggest names. In between The Godfather and The Godfather Part II, director Francis Ford Coppola made The Conversation, a film starring Gene Hackman that, in 1975, was nominated for a Best Picture Academy Award. Coppola has called The Conversation the finest film he has ever made,D. Leonhardt, “What Netflix Could Teach Hollywood,” New York Times, June 7, 2006. but it was headed for obscurity as the ever-growing pipeline of new releases pushed the film off of video store shelves. Netflix was happy to pick up The Conversation and put it in the long tail. Since then, the number of customers viewing the film has tripled, and on Netflix, this once underappreciated gem became the thirteenth most watched film from its time period.

For evidence on Netflix’s power to make lucrative markets from nonblockbusters, visit the firm’s “Top 100 page.”http://www.netflix.com/Top100. You’ll see a list loaded with films that were notable for their lack of box office success. As of this writing the number one rank had been held for over five years in a row, not by a first-run mega-hit, but by the independent film Crash (an Oscar winner, but box office weakling).R. Elder, “‘Crash’ Remains Top DVD Rental,” Chicago Tribune, April 14, 2009.

Netflix has used the long tail to its advantage, crafting a business model that creates close ties with film studios. In most cases, studios earn a percentage of the subscription revenue for every disk sent out to a Netflix customer. In exchange, Netflix gets DVDs at a very low cost. The movie business is characterized by large fixed costs up front. Studio marketing budgets are concentrated on films when they first appear in theaters, and when they’re first offered on DVD. After that, studios are done promoting a film, focusing instead on their most current titles. But Netflix is able to find an audience for a film without the studios spending a dime on additional marketing. Since so many of the titles viewed on Netflix are in the long tail, revenue sharing is all gravy for the studios—additional income they would otherwise be unlikely to get. It’s a win-win for both ends of the supply chain. These supplier partnerships grant Netflix a sort of soft bargaining power that’s distinctly opposite the strong-arm price bullying that giants like Wal-Mart are often accused of.

The VCR, the Real “Killer App”?

Netflix’s coziness with movie studios is particularly noteworthy, given that the film industry has often viewed new technologies with a suspicion bordering on paranoia. In one of the most notorious incidents, Jack Valenti, the former head of the Motion Picture Association of American (MPAA) once lobbied the U.S. Congress to limit the sale of home video recorders, claiming, “the VCR is to the American film producer and the American public as the Boston strangler is to the woman home alone.”J. Bates, “Formidable Force for Hollywood,” Los Angeles Times, April 27, 2007.

Not only was the statement over the top, Jack couldn’t have been more wrong. Revenue from the sale of VCR tapes would eventually surpass the take from theater box offices, and today, home video brings in about two times box office earnings.

Cinematch: Technology Creates a Data Asset That Delivers Profits

Netflix proves there’s both demand and money to be made from the vast back catalog of film and TV show content. But for the model to work best, the firm needed to address the biggest inefficiency in the movie industry—“audience finding,” that is, matching content with customers. To do this, Netflix leverages some of the industry’s most sophisticated technology, a proprietary recommendation system that the firm calls Cinematch.

Each time a customer visits Netflix after sending back a DVD, the service essentially asks “So, how did you like the movie?” With a single click, each film can be rated on a scale of one to five stars. If you’re new to Netflix, the service can prompt you with a list of movies (or you can search out and rate titles on your own). Love Rushmore but hate The Life Aquatic? Netflix wants to know.

The magic of Cinematch happens not by offering a gross average user rating—user tastes are too varied and that data’s too coarse to be of significant value. Instead, Cinematch develops a map of user ratings and steers you toward titles preferred by people with tastes that are most like yours. Techies and marketers call this trick collaborative filteringA classification of software that monitors trends among customers and uses this data to personalize an individual customer’s experience.. The term refers to a classification of software that monitors trends among customers and uses this data to personalize an individual customer’s experience. Input from collaborative filtering software can be used to customize the display of a Web page for each user so that an individual is greeted only with those items the software predicts they’ll most likely be interested in. The kind of data mining done by collaborative filtering isn’t just used by Netflix; other sites use similar systems to recommend music, books, even news stories. While other firms also employ collaborative filtering, Netflix has been at this game for years, and is constantly tweaking its efforts. The results are considered the industry gold standard.

Figure 4.3 Netflix Community Features

Collaborative filtering software is powerful stuff, but is it a source of competitive advantage? Ultimately it’s just math. Difficult math, to be sure, but nothing prevents other firms from working hard in the lab, running and refining tests, and coming up with software that’s as good, or perhaps one day even better than Netflix’s offering. But what the software has created for the early-moving Netflix is an enormous data advantage that is valuable, results yielding, and impossible for rivals to match. Even if Netflix gave Cinematch to its competitors, they’d be without the over 3 billion ratings that the firm has amassed (according to the firm, users add about a million new ratings to the system each day). More ratings make the system seem smarter, and with more info to go on, Cinematch can make more accurate recommendations than rivals.

Evidence suggests that users trust and value Cinematch. Recommended titles make up over 60 percent of the content users place in their queues—an astonishing penetration rate. Compare that to how often you’ve received a great recommendation from the sullen teen behind the video store counter. While data and algorithms improve the service and further strengthen the firm’s brand, this data is also a switching cost. Drop Netflix for Blockbuster and the average user abandons the two hundred or more films they’ve rated. Even if one is willing to invest the time in recreating their ratings on Blockbuster’s site, the rival will still make less accurate recommendations because there are fewer users and less data to narrow in on similarities across customers.

One way to see how strong these switching costs are is to examine the Netflix churn rateThe rate at which customers leave a product or service.. Churn is a marketing term referring to the rate at which customers leave a product or service. A low churn is usually key to profitability because it costs more to acquire a customer than to keep one. And the longer a customer stays with the firm, the more profitable they become and the less likely they are to leave. If customers weren’t completely satisfied with the Netflix experience, many would be willing to churn out and experiment with rivals offering cheaper service. However, the year after Blockbuster and Wal-Mart launched with copycat efforts, the rate at which customers left Netflix actually fell below 4 percent, an all-time low. And the firm’s churn rates have continued to fall over time. By the middle of 2008, rates for customers in Netflix most active regions of the country were below 3 percent, meaning fewer than three in one hundred Netflix customers canceled their subscriptions each year.Netflix Investor Day presentation, May 2008, accessed via http://ir.netflix.com/events.cfm. To get an idea of how enviable the Netflix churn rates are, consider that a year earlier the mobile phone industry had a churn rate of 38.6 percent, while roughly one in four U.S. banking customers defected that year.“Industry Customer Churn Rate Increases 15%,” GeoConnexion, January 8, 2008. The article contains a summary of the Pittney Bowes G1 finding.

All of this impacts marketing costs, too. Happy customers refer friends (free marketing from a source consumers trust more than a TV commercial). Ninety-four percent of Netflix subscribers say they have recommended the service to someone else, and 71 percent of new subscribers say an existing subscriber has encouraged them to sign up. It’s no wonder subscriber acquisition costs have been steadily falling, further contributing to the firm’s overall profitability.

The Netflix Prize

Netflix isn’t content to stand still with its recommendation engine. Recognizing that there may be useful expertise outside its Los Gatos, California headquarters, the firm launched a crowdsourcingThe act of taking a job traditionally performed by a designated agent (usually an employee) and outsourcing it to an undefined generally large group of people in the form of an open call. effort known as The Netflix Prize (for more on crowdsourcing, see Chapter 7 “Peer Production, Social Media, and Web 2.0”).

Figure 4.4 The Netflix Prize Leader Board

The goal was simple: Offer $1 million to the first group or individual who can improve Cinematch’s ratings accuracy by 10 percent. In order to give developers something to work with, the firm turned over a large ratings database (with customer-identifying information masked, of course). The effort attracted over 30,000 teams from 170 countries. Not bad when you consider that $1 million would otherwise fund just four senior Silicon Valley engineers for about a year. And the effort earned Netflix a huge amount of PR, as newspapers, magazines, and bloggers chatted up the effort.

While Netflix gains access to any of the code submitted as part of the prize, it isn’t exclusive access. The Prize underscores the value of the data asset. Even if others incorporate the same technology as Netflix, the firm still has user data (and attendant customer switching costs) that prevent rivals with equal technology from posing any real threat. Results incorporating many innovations offered by contest participants were incorporated into Cinematch, even before the prize was won.

As the contest dragged on, many participants wondered if the 10 percent threshold could ever be reached. While many teams grew within striking distance, a handful of particularly vexing titles thwarted all algorithms. Perhaps the most notorious title was Napoleon Dynamite. The film is so quirky, and Netflix customers so polarized, that there’s little prior indicator to suggest if you’re in the “love it” or “hate it” camp. One contestant claimed that single film was responsible for 15 percent of the gap between his team’s effort and the million dollars.C. Thompson, “If You Liked This, You’re Sure to Love That,” New York Times, November 21, 2008.

The eventual winner turned out to be a coalition of four teams from four countries—prior rivals who sought to pool their noggins and grab fame and glory (even if their individual prize split was less). BellKor’s Pragmatic Chaos, the first team to cross the 10 percent threshold, included a pair of coders from Montreal; two U.S. researchers from AT&T Labs; a scientist from Yahoo! Research, Israel; and a couple of Austrian consultants.B. Patterson, “Netflix Prize Competitors Join Forces, Cross Magic 10-Percent Mark,” Yahoo! Tech, June 29, 2009. It’s safe to say that without the Netflix Prize, these folks would likely never have met, let alone collaborated.

Patron Saint of the Independent Film Crowd

Many critically acclaimed films that failed to be box office hits have gained a second life on Netflix, netting significant revenue for the studios, with no additional studio marketing. Babel, The Queen, and The Last King of Scotland are among the films that failed to crack the top twenty in the box office, but ranked among the most requested titles on Netflix during the year after their release. Netflix actually delivered more revenue to Fox from The Last King of Scotland than it did from the final X-Men film.Netflix Investor Day presentation, May 2008, accessed via http://ir.netflix.com/events.cfm.

In the true spirit of the long tail, Netflix has occasionally acquired small market titles for exclusive distribution. One of its first efforts involved the Oscar-nominated PBS documentary, Daughters from Danang. PBS hadn’t planned to distribute the disc after the Academy Awards; it was simply too costly to justify producing a run of DVDs that almost no retailer would carry. But in a deal with PBS, Netflix assumed all production costs in exchange for exclusive distribution rights. For months after, the film repeatedly ranked in the Top 15 most requested titles in the documentary category. Cost to PBS—nothing.C. Anderson, “The Long Tail,” Wired 12, no. 10 (October 2004), http://www.wired.com/wired/archive/12.10/tail.html.

A Look at Operations

Tech also lies at the heart of the warehouse operations that deliver customer satisfaction and enhance brand value. As mentioned earlier, brand is built through customer experience, and a critical component of customer experience is for subscribers to get their DVDs as quickly as possible. In order to do this, Netflix has blanketed the country with a network of fifty-eight ultrahigh-tech distribution centers that collectively handle in excess of 1.8 million DVDs a day. These distribution centers are purposely located within driving distance of 119 U.S. Postal Service (USPS) processing and distribution facilities.

By 4:00 a.m. each weekday, Netflix trucks collect the day’s DVD shipments from these USPS hubs and returns the DVDs to the nearest Netflix center. DVDs are fed into custom-built sorters that handle disc volume on the way in and the way out. That same machine fires off an e-mail as soon as it detects your DVD was safely returned (now rate it via Cinematch). Most DVDs never hit the restocking shelves. Scanners pick out incoming titles that are destined for other users and place these titles into a sorted outbound pile with a new, appropriately addressed red envelope. Netflix not only helps out the postal service by picking up and dropping off the DVDs at its hubs, it presorts all outgoing mail for faster delivery. This extra effort has a payoff—Netflix gets the lowest possible postal rates for first-class mail delivery. And despite the high level of automation, 100 percent of all discs are inspected by hand so that cracked ones can be replaced, and dirty ones can be given a wipe down.B. McCarthy, “Netflix, Inc.” (remarks, J. P. Morgan Global Technology, Media, and Telecom Conference, Boston, May 18, 2009). Total in and out turnaround time for a typical Netflix DVD is just eight hours!N. Kenny, “Special Report: Inside Netflix,” WMC TV, July 7, 2009.

First-class mail takes only one day to be delivered within a fifty-mile radius, so the warehouse network allows Netflix to service over 97 percent of its customer base within a two-day window—one day is allotted for receipt; early the next morning the next item in their queue is processed; and the new title arrives at the customer’s address by that afternoon. And in 2009, the firm added Saturday processing. All this means a customer with the firm’s most popular “three disc at a time” plan could watch a movie a day and never be without a fresh title.

Figure 4.5 A Proprietary Netflix Sorting Machine

Figure 4.6 USPS Hubs Serviced by the Netflix Distribution Center Network

Warehouse processes don’t exist in a vacuum; they are linked to Cinematch to offer the firm additional operational advantages. The software recommends movies that are likely to be in stock so users aren’t frustrated by a wait.

Everyone on staff is expected to have an eye on improving the firm’s processes. Every warehouse worker gets a free DVD player and Netflix subscription so that they understand the service from the customer’s perspective and can provide suggestions for improvement. Quality management features are built into systems supporting nearly every process at the firm, allowing Netflix to monitor and record the circumstances surrounding any failures. When an error occurs, a tiger team of quality improvement personnel swoops in to figure out how to prevent any problems from recurring. Each phone call is a cost, not a revenue enhancement, and each error increases the chance that a dissatisfied customer will bolt for a rival.

By paying attention to process improvements and designing technology to smooth operations, Netflix has slashed the number of customer representatives even as subscriptions ballooned. In the early days, when the firm had one hundred and fifteen thousand customers, Netflix had one-hundred phone support reps. By the time the customer base had grown thirtyfold, errors had been reduced to the point where only forty-three reps were needed.J. McGregor, “High Tech Achiever,” Fast Company, October 2005. Even more impressive, because of the firm’s effective use of technology to drive the firm’s operations, fulfillment costs as a percentage of revenue have actually dropped even though postal rates have increased and Netflix has cut prices.

Killer Asset Recap: Understanding Scale

Netflix executives are quite frank that the technology and procedures that make up their model can be copied, but they also realize the challenges that any copycat rival faces. Says the firm’s VP of Operations Andy Rendich, “Anyone can replicate the Netflix operations if they wish. It’s not going to be easy. It’s going to take a lot of time and a lot of money.”Netflix Investor Day presentation, 2008, accessed via http://ir.netflix.com/events.cfm.

While we referred to Netflix as David to the Goliaths of Wal-Mart and Blockbuster, within the DVD-by-mail segment Netflix is now the biggest player by far, and this size gives the firm significant scale advantages. The yearly cost to run a Netflix-comparable nationwide delivery infrastructure is about $300 million.S. Reda and D. Schulz, “Concepts that Clicked,” Stores, May 2008. Think about how this relates to economies of scale. In Chapter 2 “Strategy and Technology: Concepts and Frameworks for Understanding What Separates Winners from Losers”, we said that firms enjoy scale economies when they are able to leverage the cost of an investment across increasing units of production. Even if rivals have identical infrastructures, the more profitable firm will be the one with more customers (see Figure 4.7). And the firm with better scale economies is in a position to lower prices, as well as to spend more on customer acquisition, new features, or other efforts. Smaller rivals have an uphill fight, while established firms that try to challenge Netflix with a copycat effort are in a position where they’re straddling markets, unable to gain full efficiencies from their efforts.

Figure 4.7

Running a nationwide sales network costs an estimated $300 million a year. But Netflix has several times more subscribers than Blockbuster. Which firm has economies of scale?Associated Press, “On the Call: Netflix CEO Reed Hastings,” April 22, 2010; Reuters, “Blockbuster Fourth-Quarter Profit Falls 28 Percent,” February 27, 2010.

For Blockbuster, the arrival of Netflix plays out like a horror film where it is the victim. For several years now, the in-store rental business has been a money loser. Things got worse in 2005 when Netflix pressure forced Blockbuster to drop late fees, costing it about $400 million.T. Mullaney, “Netflix: The Mail-Order House That Clobbered Blockbuster,” BusinessWeek, May 25, 2006. The Blockbuster store network once had the advantage of scale, but eventually its many locations were seen as an inefficient and bloated liability. Between 2006 and 2007, the firm shuttered over 570 stores.A. Farrell, “Blockbuster’s CEO Ousted,” Forbes, July 2, 2007. By 2008, Blockbuster had been in the red for ten of the prior eleven years. During a three-year period that included the launch of its Total Access DVD-by-mail effort, Blockbuster lost over $4 billion.N. MacDonald, “Blockbuster Proves It’s Not Dead Yet,” Maclean’s, March 12, 2008. The firm tried to outspend Netflix on advertising, even running Super Bowl ads for Total Access in 2007, but a money loser can’t outspend its more profitable rival for long, and it has since significantly cut back on promotion. Blockbuster also couldn’t sustain subscription rates below Netflix’s, so it has given up its price advantage. In early 2008, Blockbuster even briefly pursued a merger with another struggling giant, Circuit City, a strategy that has left industry experts scratching their heads. A Viacom executive said about the firm, “Blockbuster will certainly not survive and it will not be missed.”E. Epstein, “Hollywood’s New Zombie: The Last Days of Blockbuster,” Slate, January 9, 2006, http://www.slate.com/id/2133995. This assessment has to sting, given that Viacom was once Blockbuster’s parent (the firm was spun off in 2004).

For Netflix, what delivered the triple scale advantage of the largest selection; the largest network of distribution centers; the largest customer base; and the firm’s industry-leading strength in brand and data assets? Moving first. Timing and technology don’t always yield sustainable competitive advantage, but in this case, Netflix leveraged both to craft what seems to be an extraordinarily valuable pool of assets that continue to grow and strengthen over time. To be certain, competing against a wounded giant like Blockbuster will remain difficult. The latter firm has few options and may spend itself into oblivion, harming Netflix in its collapsing gasp. And as we’ll see in the next section, while technology shifts helped Netflix attack Blockbuster’s once-dominant position, even newer technology shifts may threaten Netflix. As they like to say in the mutual fund industry “Past results aren’t a guarantee of future returns.”

Key Takeaways

  • Durable brands are built through customer experience, and technology lies at the center of the Netflix top satisfaction ratings and hence the firm’s best-in-class brand strength.
  • Physical retailers are limited by shelf space and geography. This limitation means that expansion requires building, stocking, and staffing operations in a new location.
  • Internet retailers serve a larger geographic area with comparably smaller infrastructure and staff. This fact suggests that Internet businesses are more scalable. Firms providing digital products and services are potentially far more scalable, since physical inventory costs go away.
  • The ability to serve large geographic areas through lower-cost inventory means Internet firms can provide access to the long tail of products, potentially earning profits from less popular titles that are unprofitable for physical retailers to offer.
  • Netflix technology revitalizes latent studio assets. Revenue sharing allows Netflix to provide studios with a costless opportunity to earn money from back catalog titles: content that would otherwise not justify further marketing expense or retailer shelf space.
  • The strategically aligned use of technology by this early mover has allowed Netflix to gain competitive advantage through the powerful resources of brand, data and switching costs, and scale.
  • Collaborative filtering technology has been continually refined, but even if this technology is copied, the true exploitable resource created and leveraged through this technology is the data asset.
  • Technology leveraged across the firm’s extensive distribution network offers an operational advantage that allows the firm to reach nearly all of its customers with one-day turnaround.

Questions and Exercises

  1. What are Netflix’s sources of competitive advantage?
  2. Does Netflix have a strong brand? Offer evidence demonstrating why the firm’s brand is or isn’t strong. How is a strong brand built?
  3. Scale advantages are advantages related to size. In what key ways is Netflix “bigger” than the two major competitors who tried to enter the DVD-by-mail market?
  4. What is the long tail? How “long” is the Netflix tail compared to traditional video stores?
  5. What “class” of software does Netflix use to make movie recommendations? Think about Chapter 2 “Strategy and Technology: Concepts and Frameworks for Understanding What Separates Winners from Losers”: Which key competitive resource does this software “create”? What kinds of benefits does this provide to the firm? What benefits does it provide to Netflix’s suppliers?
  6. Could a new competitor match Netflix’s recommendation software? If it did, would this create a threat to Netflix? Why or why not?
  7. What is the Netflix churn rate and what are the reasons behind this rate?
  8. Netflix uses technology to coordinate the process of sorting and dropping off DVDs for the U.S. Postal Service. This application of technology speeds delivery. What other advantage does it give the firm?
  9. How has Netflix improved its customer service operation? Describe the results and impact of this improvement.

4.3 From Atoms to Bits: Opportunity or Threat?

Learning Objectives

After studying this section you should be able to do the following:

  1. Understand the shift from atoms to bits, and how this is impacting a wide range of industries.
  2. Recognize the various key issues holding back streaming video models.
  3. Know the methods that Netflix is using to attempt to counteract these challenges.

Nicholas Negroponte, the former head of MIT’s Media Lab and founder of the One Laptop per Child effort, wrote a now-classic essay on the shift from atoms to bitsThe idea that many media products are sold in containers (physical products, or atoms) for bits (the ones and zeros that make up a video file, song, or layout of a book). As the Internet offers fast wireless delivery to TVs, music players, book readers, and other devices, the “atoms” of the container aren’t necessary. Physical inventory is eliminated, offering great cost savings.. Negroponte pointed out that most media products are created as bits—digital files of ones and zeros that begin their life on a computer. Music, movies, books, and newspapers are all created using digital technology. When we buy a CD, DVD, or even a “dead tree” book or newspaper, we’re buying physical atoms that are simply a container for the bits that were created in software—a sound mixer, a video editor, or a word processor.

The shift from atoms to bits is realigning nearly every media industry. Newspapers struggle as readership migrates online and once-lucrative classified ads and job listings shift to the bits-based businesses of Craigslist, Monster.com, and LinkedIn. Apple dominates music sales, selling not a single “atom” of physical CDs, while most of the atom-selling “record store” chains of a decade ago are bankrupt. Amazon has even begun delivering digital books, developing the Kindle digital reader. Who needs to kill a tree, spill ink, fill a warehouse, and roll a gas-guzzling truck to get you a book? Kindle can slurp your purchases through the air and display them on a device lighter than any college textbook. When Amazon CEO Bezos unveiled the Kindle DX at a press event at Pace University in Spring 2009, he indicated that Kindle book sales were accounting for 35 percent of sales for the two hundred and seventy-five thousand titles available for the deviceA. Penenberg, “Amazon Taps Its Inner Apple,” Fast Company, July 1, 2009.—a jaw-dropping impact for a device many had thought to be an expensive, niche product for gadget lovers.

Video is already going digital, but Netflix became a profitable business by handling the atoms of DVDs. The question is, will the atoms to bits shift crush Netflix and render it as irrelevant as Hastings did Blockbuster? Or can Reed pull off yet another victory and recast his firm for the day that DVDs disappear?

Concerns over the death of the DVD and the relentless arrival of new competitors are probably the main cause for Netflix’s stock volatility these past few years. Through the first half of 2010, the firm’s growth, revenue, and profit graphs all go up and to the right, but the stock has experienced wild swings as pundits have mostly guessed wrong about the firm’s imminent demise (one well-known Silicon Valley venture capitalist even referred to the firm as “an ice cube in the sun,”M. Copeland, “Netflix Lives! Video Downloads Haven’t Made the DVD-by-Mail Business Obsolete,” Fortune, April 21, 2008. a statement Netflix countered with five years of record-breaking growth and profits). The troughs on the Netflix stock graph have proven great investment opportunities for the savvy. The firm broke all previous growth and earnings records and posted its lowest customer churn ever, even as a deep recession and the subprime crisis hammered many other firms. The firm continued to enjoy its most successful quarters as a public company, and subscriber growth rose even as DVD sales fell. But even the most bullish investor knows there’s no stopping the inevitable shift from atoms to bits, and the firm’s share price swings continue. When the DVD dies, the high-tech shipping and handling infrastructure that Netflix has relentlessly built will be rendered worthless.

Reed Hastings clearly knows this, and he has a plan: “We named the company Netflix for a reason; we didn’t name it DVDs-by-mail.”M. Boyle, “Questions for…Reed Hastings,” Fortune, May 23, 2007. But he also prepared the public for a first-cut service that was something less than we’d expect from the long tail poster child. When speaking about the launch of the firm’s Internet video streaming offering in January 2007, Hastings said it would be “underwhelming.” The two biggest limitations of this initial service? As we’ll see below, not enough content, and figuring out how to squirt the bits to a television.

Access to Content

First the content. Three years after the launch of Netflix streaming option (enabled via a “Watch Now” button next to movies that can be viewed online), only 17,000 videos were offered, just 17 percent of the firm’s long tail. And not the best 17 percent. Why so few titles? It’s not just studio reluctance or fear of piracy. There are often complicated legal issues involved in securing the digital distribution rights for all of the content that makes up a movie. Music, archival footage, and performer rights may all hold up a title from being available under “Watch Now.” The 2007 Writers Guild strike occurred largely due to negotiations over digital distribution, showing just how troublesome these issues can be.

Add to that the exclusivity contracts negotiated by key channels, in particular the so-called premium television networks. Film studios release their work in a system called windowingIndustry practice whereby content (usually a motion picture) is available to a given distribution channel for a specified time period or “window,” usually under a different revenue model (e.g., ticket sale, purchase, license fee).. Content is available to a given distribution channel (in theaters, through hospitality channels like hotels and airlines, on DVD, via pay-per-view, via pay cable, then broadcast commercial TV) for a specified time window, usually under a different revenue model (ticket sales, disc sales, license fees for broadcast). Pay television channels in particular have negotiated exclusive access to content as they strive to differentiate themselves from one another. This exclusivity means that even when a title becomes available for streaming by Netflix, it may disappear when a pay TV window opens up. If HBO or Showtime has an exclusive for a film, it’s pulled from the Netflix streaming service until the exclusive pay TV time window closes. A 2008 partnership with the Starz network helped provide access to some content locked up inside pay television windows, and deals with Disney and CBS allow for streaming of current-season shows.S. Portnoy, “Netflix News: Starz Catalog Added to Online Service, Streaming to PS3, Xbox 360 through PlayOn Beta Software,” ZDNet, October 2, 2008, http://blogs.zdnet.com/home-theater/?p=120. But the firm still has a long way to go before the streaming tail seems comparably long when compared against its disc inventory.

While studios embrace the audience-finding and revenue-sharing advantages of Netflix, they also don’t want to undercut higher-revenue early windows. Fox, Universal, and Warner have all demanded that Netflix delay sending DVDs to customers until twenty-eight days after titles go on sale. In exchange, Netflix has received guarantees that these studios will offer more content for digital streaming.

There’s also the influence of the king of DVD sales: Wal-Mart. The firm accounts for about 40 percent of DVD sales—a scale that delivers a lot of the bargaining power it has used to “encourage” studios to hold content from competing windows or to limit offering digital titles at competitive pricing during the peak new release period.R. Grover, “Wal-Mart and Apple Battle for Turf,” BusinessWeek, August 31, 2006. Apparently, Wal-Mart isn’t ready to yield ground in the shifts from atoms to bits, either. In February 2010, the retail giant spent an estimated $100 million to buy the little-known video streaming outfit VUDU.B. Stone, “Wal-Mart Adds Clout to Streaming,” New York Times, February 22, 2010. Wal-Mart’s negotiating power with studios may help it gain special treatment for VUDU. As an example, VUDU was granted exclusive high-definition streaming rights for the hit movie Avatar, offering the title online the same day the DVD appeared for sale.J. Jacobson, “VUDU/Wal-Mart Gets Avatar HD Streaming Exclusive,” Electronic House, April 22, 2010.

Studios may also be wary of the increasing power Netflix has over product distribution, and as such, they may be motivated to keep rivals around. Studios have granted Blockbuster more favorable distribution terms than Netflix. In many cases, Blockbuster can now distribute DVDs the day of release instead of waiting nearly a month, as Netflix does.J. Birchall, “Blockbuster Strikes Deal to Ensure DVD Supply,” Financial Times, April 8, 2010. Studios are likely concerned that Netflix may be getting so big that it will one day have Wal-Mart-like negotiating leverage.

Supplier Power and Atoms to Bits

The winner-take-all, winner-take-most dynamics of digital distribution can put suppliers at a disadvantage. If firms rely on one channel partner for a large portion of sales, that partner has an upper hand in negotiations. For years, record labels and movie studios complained that Apple’s dominance of iTunes allowed them little negotiating room in price setting. A boycott where NBC temporarily lifted TV shows from iTunes is credited with loosening Apple’s pricing policies. Similarly, when Amazon’s Kindle dominated the e-book reader market, Amazon enforced a $9.99 price on electronic editions, even as publishers lobbied for higher rates. It wasn’t until Apple arrived with a creditable e-book rival in the iPad that Amazon’s leverage was weakened to the point where publishers were allowed to set their own e-book prices.M. Rich and B. Stone, “Publisher Wins Fight with Amazon over E-Books,” New York Times, January 31, 2010.

Taken together, all these factors make it clear that shifting the long tail from atoms to bits will be significantly more difficult than buying DVDs and stacking them in a remote warehouse.

Figure 4.8 Film Release Windows

But How Does It Get to the TV?

The other major problem lies in getting content to the place where most consumers want to watch it: the living room TV. Netflix’s “Watch Now” button first worked only on Windows PCs. Although the service was introduced in January 2007, the months before were fueled with speculation that the firm would partner with TiVo. Just one month later, TiVo announced its partner—Amazon.com. At that point Netflix found itself up against a host of rivals that all had a path to the television: Apple had its own hardware solution in Apple TV (not to mention the iPod and iPhone for portable viewing), the cable companies delivered OnDemand through their set-top boxes, and now Amazon had TiVo.

An internal team at Netflix developed a prototype set top box that Hastings himself supported offering. But most customers aren’t enthusiastic about purchasing yet another box for their set top, the consumer electronics business is brutally competitive, and selling hardware would introduce an entirely new set of inventory, engineering, marketing, distribution, and competitive complexities.

The solution Netflix eventually settled on was to think beyond one hardware alternative and instead recruit others to provide a wealth of choice. The firm developed a software platform and makes this available to firms seeking to build Netflix access into their devices. Today, Netflix streaming is baked into televisions and DVD players from LG, Panasonic, Samsung, Sony, Toshiba, and Vizio, among others. It’s also available on all major video game consoles. A Netflix app for Apple’s iPad was available the day the device shipped. Even TiVo now streams Netflix. And that internally developed Netflix set-top box? The group was spun out to form Roku, an independent firm that launched their own $99 Netflix streamer.

The switch to Blu-ray movies may offer the most promise. Blu-ray players are on the fast track to commoditization. If consumer electronics firms incorporate Netflix access into their players as a way to attract more customers with an additional, differentiating feature, Hastings’s firm could end up with more living room access than either Amazon or Apple. There are 73 million households in the United States that have a DVD player and an Internet connection. Should a large portion of these homes end up with a Netflix-ready Blu-ray player, Hastings will have built himself an enviable base through which to grow the video streaming business.

Disintermediation and Digital Distribution

The purchase of NBC/Universal by Comcast, the largest cable television provider in the United States, has consolidated content and distribution in a single firm. The move can be described as both vertical integration (when an organization owns more than one layer of its value chain) and disintermediationRemoving an organization from a firm’s distribution channel. Disintermediation collapses the path between supplier and customer. (removing an organization from a firm’s distribution channel).J. Gallaugher, “E-Commerce and the Undulating Distribution Channel,” Communications of the ACM, July 2002. Disintermediation in the video industry offers two potentially big benefits. First, studios don’t need to share revenue with third parties; they can keep all the money generated through new windows. Also critically important, studios keep the interface with their customers. Remember, in the digital age data is valuable; if another firm sits between a supplier and its customers, the supplier loses out on a key resource for competitive advantage. For more on the value of the data asset in maintaining and strengthening customer relationships, see Chapter 11 “The Data Asset: Databases, Business Intelligence, and Competitive Advantage”.

Who’s going to win the race for delivering bits to the television is still very much an uncertain bet. The models all vary significantly. Apple’s early efforts were limited, with the firm offering only video purchases for Apple TV, but eventually moving to online “rentals” that can also play on the firm’s entire line of devices. Movie studios are now all in Apple’s camp, although the firm did temporarily lose NBC’s television content in a dispute over pricing. Amazon and Microsoft also have online rentals and purchase services, and can get their content to the television via TiVo and Xbox, respectively (yes, this makes Microsoft both a partner and a sort of competitor, a phenomenon often referred to as coopetitionWhen firms find themselves in situations where they are both competitors and collaborators, or “frenemies.” Technology is increasingly redefining products, pushing the boundaries of delivery channels, and creating situations where firms often find themselves in these collaborative/rival relationships., or frenemiesA. Brandenberger and B. Nalebuff, Co-opetition: A Revolution Mindset that Combines Competition and Cooperation: The Game Theory Strategy That’s Changing the Game of Business (New York: Broadway Business, 1997); and S. Johnson, “The Frenemy Business Relationship,” Fast Company, November 25, 2008.). Hulu, a joint venture backed by NBC, Fox, and other networks, is free, earning money from ads that run like TV commercials. While Hulu has also received glowing reviews, the venture has lagged in offering a method to get streaming content to the television. Netflix pioneered “all-you-can-eat” subscription streaming. Anyone who has at least the $8.99 subscription plan can view an unlimited number of video streams. And Blockbuster isn’t dead yet. It also streams over TiVo and has other offerings in the works.

There’s a clear upside to the model when it shifts to streaming: it will eliminate a huge chunk of costs associated with shipping and handling. Postage represents one-third of the firm’s expenses. A round-trip DVD mailing, even at the deep discounts Netflix receives from the U.S. Postal Service, runs about eighty cents. The bandwidth and handling costs to send bits to a TV set are around a nickel.B. McCarthy, “Netflix, Inc.” (remarks, J. P. Morgan Global Technology, Media, and Telecom Conference, Boston, May 18, 2009). At some point, if postage goes away, Netflix may be in a position to offer even greater profits to its studio suppliers, and to make more money itself, too.

Wrangling licensing costs presents a further challenge. Estimates peg Netflix 2009 streaming costs at about $100 million, up 250 percent in three years. But these expenses still deliver just a fraction of the long tail. Streaming licensing deals are tricky because they’re so inconsistent even when titles are available. Rates vary, with some offered via a flat rate for unlimited streams, a per-stream rate, a rate for a given number of streams, and various permutations in between. Some vendors have been asking as much as four dollars per stream for more valuable contentD. Rayburn, “Netflix Streaming Costs,” Streaming Media, June/July 2009.—a fee that would quickly erase subscriber profits, making any such titles too costly to add to the firm’s library. Remember, Netflix doesn’t charge more for streaming—it’s built into the price of its flat-rate subscriptions.

Any extra spending doesn’t come at the best time. The switch to Blu-ray movies means that Netflix will be forced into the costly proposition of carrying two sets of video inventory: standard and high-def. Direct profits may not be the driver. Rather, the service may be a feature that attracts new customers to the firm and helps prevent subscriber flight to rival video-on-demand efforts. The stealth arrival of a Netflix set-top box, in the form of upgraded Blu-ray players, might open even more customer acquisition opportunities to the firm. Bought a Blu-ray player? For just nine dollars per month you can get a ticket to the all-you-can-eat Netflix buffet. And more customers ready to watch content streamed by Netflix may prime the pump for studios to become more aggressive in licensing more of their content. Many TV networks and movie studios are leery of losing bargaining power to a dominant firm, having witnessed how Apple now dictates pricing terms to music labels. The goodwill Netflix has earned over the years may pay off if it can become the studios’ partner of first choice.

While one day the firm will lose the investment in its warehouse infrastructure, nearly all assets have a limited lifespan. That’s why corporations depreciate assets, writing their value down over time. The reality is that the shift from atoms to bits won’t flick on like a light switch; it will be a hybrid transition that takes place over several years. If the firm can grab long-tail content, grow its customer base, and lock them in with the switching costs created by Cinematch (all big “ifs”), it just might emerge as a key player in a bits-only world.

Is the hybrid strategy a dangerous straddling gambit or a clever ploy to remain dominant? Netflix really doesn’t have a choice but to try. Hastings already has a long history as one of the savviest strategic thinkers in tech. As the networks say, stay tuned!

Key Takeaways

  • The shift from atoms to bits is impacting all media industries, particularly those relying on print, video, and music content. Content creators, middlemen, retailers, consumers, and consumer electronics firms are all impacted.
  • Netflix’s shift to a streaming model (from atoms to bits) is limited by access to content and in methods to get this content to televisions.
  • Windowing and other licensing issues limit available content, and inconsistencies in licensing rates make profitable content acquisitions a challenge.

Questions and Exercises

  1. What do you believe are the most significant long-term threats to Netflix? How is Netflix trying to address these threats? What obstacles does the firm face in dealing with these threats?
  2. Who are the rivals to Netflix’s “Watch Now” effort? Do any of these firms have advantages that Netflix lacks? What are these advantages?
  3. Why would a manufacturer of DVD players be motivated to offer the Netflix “Watch Now” feature in its products?
  4. Describe various revenue models available as video content shifts from atoms to bits. What are the advantages and disadvantages to each—for consumers, for studios, for middlemen like television networks and Netflix?
  5. Wal-Mart backed out of the DVD-by-mail industry. Why does the firm continue to have so much influence with the major film studios? What strategic asset is Wal-Mart leveraging?
  6. Investigate the firm Red Box. Do you think they are a legitimate threat to Netflix? Why or why not?

Chapter 3: Zara: Fast Fashion from Savvy Systems

3.1 Introduction

Learning Objective

After studying this section you should be able to do the following:

  1. Understand how Zara’s parent company Inditex leveraged a technology-enabled strategy to become the world’s largest fashion retailer.

The poor, ship-building town of La Coruña in northern Spain seems an unlikely home to a tech-charged innovator in the decidedly ungeeky fashion industry, but that’s where you’ll find “The Cube,” the gleaming, futuristic central command of the Inditex Corporation (Industrias de Diseño Textil), parent of game-changing clothes giant, Zara. The blend of technology-enabled strategy that Zara has unleashed seems to break all of the rules in the fashion industry. The firm shuns advertising and rarely runs sales. Also, in an industry where nearly every major player outsources manufacturing to low-cost countries, Zara is highly vertically integrated, keeping huge swaths of its production process in-house. These counterintuitive moves are part of a recipe for success that’s beating the pants off the competition, and it has turned the founder of Inditex, Amancio Ortega, into Spain’s wealthiest man and the world’s richest fashion executive.

Figure 3.1

Zara’s operations are concentrated in Spain, but they have stores around the world like these in Manhattan and Shanghai.

The firm tripled in size between 1996 and 2000, then its earnings skyrocketed from $2.43 billion in 2001 to $13.6 billion in 2007. By August 2008, sales edged ahead of Gap, making Inditex the world’s largest fashion retailer.J. Hall, “Zara Is Now Bigger Than Gap,” Telegraph, August 18, 2008. Table 3.1 “Gap versus Inditex at a Glance” compares the two fashion retailers. While Inditex supports eight brands, Zara is unquestionably the firm’s crown jewel and growth engine, accounting for roughly two-thirds of sales.R. Murphy, “Expansion Boosts Inditex Net,” Women’s Wear Daily, April 1, 2008.

Table 3.1 Gap versus Inditex at a Glance

Gap Inditex
Revenue $14.5 billion $14.7 billion
Net Income $967 million $1.68 billion
Number of Stores 3,149 4,359
Number of Countries 6 73
Biggest Brand Gap Zara
Number of Other Brads 4 7
Based in San Francisco, USA Arteixo (near La Coruña), Spain
First Store Opened 1969 1975

Why Study Zara?

While competitors falter, Zara is undergoing one of the fastest global expansions the fashion world has ever seen, opening one store per day and entering new markets worldwide—seventy-three countries so far. The chain’s profitability is among the highest in the industry.D. Sull and S. Turconi, “Fast Fashion Lessons,” Business Strategy Review, Summer 2008. The fashion director for luxury goods maker LVMH calls Zara “the most innovative and devastating retailer in the world.”J. Surowiecki, “The Most Devastating Retailer in the World,” New Yorker, September 18, 2000.

Zara’s duds look like high fashion but are comparatively inexpensive (average item price is $27, although prices vary by country).C. Rohwedder, “Zara Grows as Retail Rivals Struggle,” Wall Street Journal, March 26, 2009. A Goldman analyst has described the chain as “Armani at moderate prices,” while another industry observer suggests that while fashions are more “Banana Republic,” prices are more “Old Navy.”J. Folpe, “Zara Has a Made-to-Order Plan for Success,” Fortune, September 4, 2000. Legions of fans eagerly await “Z-day,” the twice-weekly inventory delivery to each Zara location that brings in the latest clothing lines for women, men, and children.

In order to understand and appreciate just how counterintuitive and successful Zara’s strategy is, and how technology makes all of this possible, it’s important to first examine the conventional wisdom in apparel retail. To do that we’ll look at former industry leader—Gap.

Gap: An Icon in Crisis

Most fashion retailers place orders for a seasonal collection months before these lines make an appearance in stores. While overseas contract manufacturers may require hefty lead times, trying to guess what customers want months in advance is a tricky business. In retail in general and fashion in particular, there’s a saying: inventory equals death. Have too much unwanted product on hand and you’ll be forced to mark down or write off items, killing profits. For years, Gap sold most of what it carried in stores. Micky Drexler, a man with a radar-accurate sense of style and the iconic CEO who helped turn Gap’s button-down shirts and khakis into America’s business casual uniform, led the way. Drexler’s team had spot-on tastes throughout the 1990s, but when sales declined in the early part of the following decade, Drexler was left guessing on ways to revitalize the brand, and he guessed wrong—disastrously wrong. Chasing the youth market, Drexler filled Gap stores with miniskirts, low-rise jeans, and even a much-ridiculed line of purple leather pants.J. Boorstein, “Fashion Victim,” Fortune, April 13, 2006. The throngs of teenagers he sought to attract never showed up, and the shift in offerings sent Gap’s mainstay customers to retailers that easily copied the styles that Gap had made classic.

The inventory hot potato Drexler was left with crushed the firm. Gap’s same-store sales declined for twenty-nine months straight. Profits vanished. Gap founder and chairman Dan Fisher lamented, “It took us thirty years to get to $1 billion in profits and two years to get to nothing.”P. Sellers, “Gap’s New Guy Upstairs,” Fortune, April 14, 2003. The firm’s debt was downgraded to junk status. Drexler was out and for its new head the board chose Paul Pressler, a Disney executive who ran theme parks and helped rescue the firm’s once ailing retail effort.

Pressler shut down hundreds of stores, but the hemorrhaging continued largely due to bad bets on colors and styles.L. Lee, “Paul Pressler’s Fall from The Gap,” BusinessWeek, February 26, 2007. During one holiday season, Gap’s clothes were deemed so off target that the firm scrapped its advertising campaign and wrote off much of the inventory. The marketing model used by Gap to draw customers in via big-budget television promotion had collapsed. Pressler’s tenure saw same-store sales decline in eighteen of twenty-four months.J. Boorstein, “Fashion Victim,” Fortune, April 13, 2006. A Fortune article on Pressler’s leadership was titled “Fashion Victim.” BusinessWeek described his time as CEO as a “Total System Failure,”L. Lee, “Paul Pressler’s Fall from the Gap,” BusinessWeek, February 26, 2007. and Wall Street began referring to him as DMW for Dead Man Walking. In January 2007, Pressler resigned, with Gap hoping its third chief executive of the decade could right the ailing giant.

Contract Manufacturing: Lower Costs at What Cost?

Conventional wisdom suggests that leveraging cheap contract manufacturingOutsourcing production to third-party firms. Firms that use contract manufacturers don’t own the plants or directly employ the workers who produce the requested goods. in developing countries can keep the cost of goods low. Firms can lower prices and sell more product or maintain higher profit margins—all good for the bottom line. But many firms have also experienced the ugly downside to this practice. Global competition among contract firms has led to race-to-the-bottom cost-cutting measures. Too often, this means that in order to have the low-cost bid, contract firms skimp on safety, ignore environmental concerns, employ child labor, and engage in other ghastly practices.

The apparel industry in particular has been plagued by accusations of employing sweatshop labor to keep costs down. Despite the fact that Gap audits contract manufacturers and has a high standard for partner conduct, the firm has repeatedly been taken to task by watchdog groups, the media, and its consumers, who have exposed unacceptable contract manufacturing conditions that Gap failed to catch. This negative exposure includes the October 2007 video showing Gap clothes made by New Delhi children as young as ten years old in what were described as “slave labor” conditions.E. Cho, “Gap: Report of Kids’ Sweatshop ‘Deeply Disturbing,’” CNN.com, October 29, 2007, http://www.cnn.com/2007/WORLD/asiapcf/10/29/gap.labor/index.html#cnnSTCVideo.

Gap is not alone; Nike, Wal-Mart, and many other apparel firms have been tarnished in similar incidents. Big firms are big targets and those that fail to adequately ensure their products are made under acceptable labor conditions risk a brand-damaging backlash that may turn off customers, repel new hires, and leave current staff feeling betrayed. Today’s manager needs to think deeply not only about their own firm’s ethical practices, but also those of all of their suppliers and partners.

Tech for Good: The Fair Factories Clearinghouse

The problem of sweatshop labor has plagued the clothing industry for years. Managers often feel the pressure to seek ever-lower costs and all too often end up choosing suppliers with unacceptably poor practices. Even well-meaning firms can find themselves stung by corner-cutting partners that hide practices from auditors or truck products in from unmonitored off-site locations. The results can be tragic for those exploited, and can carry lasting negative effects for the firm. The sweatshop moniker continues to dog Nike years after allegations were uncovered and the firm moved aggressively to deal with its problems.

Nike rival Reebok (now part of Adidas) has always taken working conditions seriously. The firm even has a Vice President of Human Rights and has made human dignity a key platform for its philanthropic efforts. Reebok invested millions in developing an in-house information system to track audits of its hundreds of suppliers along dimensions such as labor, safety, and environmental practices. The goal in part was to identify any bad apples, so that one division, sporting goods, for example, wouldn’t use a contractor identified as unacceptable by the sneaker line.

The data was valuable to Reebok, particularly given that the firm has hundreds of contract suppliers. But senior management realized the system would do even more good if the whole industry could share and contribute information. Reebok went on to donate the system and provided critical backing to help create the nonprofit organization Fair Factories Clearinghouse. With management that includes former lawyers for Amnesty International, Fair Factories (FairFactories.org) provides systems where apparel and other industries can share audit information on contract manufacturers. Launching the effort wasn’t as easy as sharing the technology. The U.S. Department of Justice needed to provide a special exemption, and had to be convinced the effort wouldn’t be used by buyers to collude and further squeeze prices from competitors (the system is free of pricing data).

Suppliers across industries now recognize that if they behave irresponsibly the Fair Factories system will carry a record of their misdeeds, notifying all members to avoid the firm. As more firms use the system, its database becomes broader and more valuable. To their credit, both Gap and Nike have joined the Fair Factories Clearinghouse.

Key Takeaways

  • Zara has used technology to dominate the retail fashion industry as measured by sales, profitability, and growth.
  • Excess inventory in the retail apparel industry is the kiss of death. Long manufacturing lead times require executives to guess far in advance what customers will want. Guessing wrong can be disastrous, lowering margins through markdowns and write-offs.
  • Contract manufacturing can offer firms several advantages, including lower costs and increased profits. But firms have also struggled with the downside of cost-centric contract manufacturing when partners have engaged in sweatshop labor and environmental abuse.

Questions and Exercises

  1. Has anyone shopped at Zara? If so, be prepared to share your experiences and observations with your class. What did you like about the store? What didn’t you like? How does Zara differ from other clothing retailers in roughly the same price range? If you’ve visited Zara locations in different countries, what differences did you notice in terms of offerings, price, or other factors?
  2. What is the “conventional wisdom“ of the fashion industry with respect to design, manufacturing, and advertising?
  3. What do you suppose are the factors that helped Gap to at one point rise to be first in sales in the fashion industry?
  4. Who ran Gap in the 1990s? How did the executive perform prior to leaving Gap? Describe what happened to sales. Why?
  5. Who was the Gap’s second CEO of this decade? How did sales fare under him? Why?
  6. Where do Gap clothes come from? Who makes them? Why? Are there risks in this approach?
  7. Describe the Fair Factories Clearinghouse. Which firm thought of this effort? Why did they give the effort away? What happens as more firms join this effort and share their data?

3.2 Don’t Guess, Gather Data

Learning Objective

After studying this section you should be able to do the following:

  1. Contrast Zara’s approach with the conventional wisdom in fashion retail, examining how the firm’s strategic use of information technology influences design and product offerings, manufacturing, inventory, logistics, marketing, and ultimately profitability.

Having the wrong items in its stores hobbled Gap for nearly a decade. But how do you make sure stores carry the kinds of things customers want to buy? Try asking them. Zara’s store managers lead the intelligence-gathering effort that ultimately determines what ends up on each store’s racks. Armed with personal digital assistants (PDAs)Handheld computing devices meant largely for mobile use outside an office setting. Examples include devices from Palm, Apple’s iPhone, and devices running the Windows Pocket PC operating system.—handheld computing devices meant largely for mobile use outside an office setting—to gather customer input, staff regularly chat up customers to gain feedback on what they’d like to see more of. A Zara manager might casually ask, “What if this skirt were in a longer length?” “Would you like it in a different color?” “What if this V-neck blouse were available in a round neck?” Managers are motivated because they have skin in the game. The firm is keen to reward success—as much as 70 percent of salaries can come from commissions.K. Capell, “Zara Thrives by Breaking All the Rules,” BusinessWeek, October 9, 2008.

Another level of data gathering starts as soon as the doors close. Then the staff turns into a sort of investigation unit in the forensics of trendspotting, looking for evidence in the piles of unsold items that customers tried on but didn’t buy. Are there any preferences in cloth, color, or styles offered among the products in stock?D. Sull and S. Turconi, “Fast Fashion Lessons,” Business Strategy Review, Summer 2008.

PDAs are also linked to the store’s point-of-sale (POS) systemTransaction processing systems that capture customer purchases. Cash registers and store checkout systems are examples of point-of-sale systems. These systems are critical for capturing sales data and are usually linked to inventory systems to subtract out any sold items.—a transaction process that captures customer purchase information—showing how garments rank by sales. In less than an hour, managers can send updates that combine the hard data captured at the cash register with insights on what customers would like to see.C. Rohwedder and K. Johnson, “Pace-Setting Zara Seeks More Speed to Fight Its Rising Cheap-Chic Rivals,” Wall Street Journal, February 20, 2008. All this valuable data allows the firm to plan styles and issue rebuy orders based on feedback rather than hunches and guesswork. The goal is to improve the frequency and quality of decisions made by the design and planning teams.

Design

Rather than create trends by pushing new lines via catwalk fashion shows, Zara designs follow evidence of customer demand. Data on what sells and what customers want to see goes directly to “The Cube” outside La Coruña, where teams of some three hundred designers crank out an astonishing thirty thousand items a year versus two to four thousand items offered up at big chains like H&M (the world’s third largest fashion retailer) and Gap.M. Pfeifer, “Fast and Furious,” Latin Trade, September 2007; and “The Future of Fast Fashion,” Economist, June 18, 2005. While H&M has offered lines by star designers like Stella McCartney and Karl Lagerfeld, as well as celebrity collaborations with Madonna and Kylie Minogue, the Zara design staff consists mostly of young, hungry Project Runway types fresh from design school. There are no prima donnas in “The Cube.” Team members must be humble enough to accept feedback from colleagues and share credit for winning ideas. Individual bonuses are tied to the success of the team, and teams are regularly rotated to cross-pollinate experience and encourage innovation.

Manufacturing and Logistics

In the fickle world of fashion, even seemingly well-targeted designs could go out of favor in the months it takes to get plans to contract manufacturers, tool up production, then ship items to warehouses and eventually to retail locations. But getting locally targeted designs quickly onto store shelves is where Zara really excels. In one telling example, when Madonna played a set of concerts in Spain, teenage girls arrived to the final show sporting a Zara knockoff of the outfit she wore during her first performance.“The Future of Fast Fashion,” Economist, June 18, 2005. The average time for a Zara concept to go from idea to appearance in store is fifteen days versus their rivals who receive new styles once or twice a season. Smaller tweaks arrive even faster. If enough customers come in and ask for a round neck instead of a V neck, a new version can be in stores with in just ten days.J. Tagliabue, “A Rival to Gap That Operates Like Dell,” New York Times, May 30, 2003. To put that in perspective, Zara is twelve times faster than Gap despite offering roughly ten times more unique products!M. Helft, “Fashion Fast Forward,” Business 2.0, May 2002. At H&M, it takes three to five months to go from creation to delivery—and they’re considered one of the best. Other retailers need an average of six months to design a new collection and then another three months to manufacture it. VF Corp (Lee, Wrangler) can take nine months just to design a pair of jeans, while J. Jill needs a year to go from concept to store shelves.L. Sullivan, “Designed to Cut Time,” InformationWeek, February 28, 2005. At Zara, most of the products you see in stores didn’t exist three weeks earlier, not even as sketches.J. Surowiecki, “The Most Devastating Retailer in the World,” New Yorker, September 18, 2000.

The firm is able to be so responsive through a competitor-crushing combination of vertical integrationWhen a single firm owns several layers in its value chain. and technology-orchestrated coordination of suppliers, just-in-time manufacturing, and finely tuned logistics. Vertical integration is when a single firm owns several layers in its value chainThe set of activities through which a product or service is created and delivered to customers..Definition from the “father” of the value chain, Michael Porter. See M. Porter, “Strategy and the Internet,” Harvard Business Review 79, no. 3 (March 2001): 62–78, among others. While H&M has nine hundred suppliers and no factories, nearly 60 percent of Zara’s merchandise is produced in-house, with an eye on leveraging technology in those areas that speed up complex tasks, lower cycle time, and reduce error. Profits from this clothing retailer come from blending math with a data-driven fashion sense. Inventory optimization models help the firm determine how many of which items in which sizes should be delivered to each specific store during twice-weekly shipments, ensuring that each store is stocked with just what it needs.C. Gentry, “European Fashion Stores Edge Past U.S. Counterparts,” Chain Store Age, December 2007. Outside the distribution center in La Coruña, fabric is cut and dyed by robots in twenty-three highly automated factories. Zara is so vertically integrated, the firm makes 40 percent of its own fabric and purchases most of its dyes from its own subsidiary. Roughly half of the cloth arrives undyed so the firm can respond as any midseason fashion shifts occur. After cutting and dying, many items are stitched together through a network of local cooperatives that have worked with Inditex so long they don’t even operate with written contracts. The firm does leverage contract manufacturers (mostly in Turkey and Asia) to produce staple items with longer shelf lives, such as t-shirts and jeans, but such goods account for only about one-eighth of dollar volume.N. Tokatli, “Global Sourcing: Insights from the Global Clothing Industry—The Case of Zara, a Fast Fashion Retailer,” Journal of Economic Geography 8, no. 1 (2008): 21–38.

All of the items the firm sells end up in a five-million-square-foot distribution center in La Coruña, or a similar facility in Zaragoza in the northeast of Spain. The La Coruña facility is some nine times the size of Amazon’s warehouse in Fernley, Nevada, or about the size of ninety football fields.M. Helft, “Fashion Fast Forward,” Business 2.0, May 2002. The facilities move about two and a half million items every week, with no item staying in-house for more than seventy-two hours. Ceiling-mounted racks and customized sorting machines patterned on equipment used by overnight parcel services, and leveraging Toyota-designed logistics, whisk items from factories to staging areas for each store. Clothes are ironed in advance and packed on hangers, with security and price tags affixed. This system means that instead of wrestling with inventory during busy periods, employees in Zara stores simply move items from shipping box to store racks, spending most of their time on value-added functions like helping customers find what they want. Efforts like this help store staff regain as much as three hours in prime selling time.C. Rohwedder and K. Johnson, “Pace-Setting Zara Seeks More Speed to Fight Its Rising Cheap-Chic Rivals,” Wall Street Journal, February 20, 2008; and K. Capell, “Zara Thrives by Breaking All the Rules,” BusinessWeek, October 9, 2008.

Trucks serve destinations that can be reached overnight, while chartered cargo flights serve farther destinations within forty-eight hours.K. Capell, “Zara Thrives by Breaking All the Rules,” BusinessWeek, October 9, 2008. The firm recently tweaked its shipping models through Air France–KLM Cargo and Emirates Air so flights can coordinate outbound shipment of all Inditex brands with return legs loaded with raw materials and half-finished clothes items from locations outside of Spain. Zara is also a pioneer in going green. In fall 2007, the firm’s CEO unveiled an environmental strategy that includes the use of renewable energy systems at logisticsCoordinating and enabling the flow of goods, people, information, and other resources among locations. centers including the introduction of biodiesel for the firm’s trucking fleet.

Stores

Most products are manufactured for a limited production run. While running out of bestsellers might be seen as a disaster at most retailers, at Zara the practice delivers several benefits.

First, limited runs allow the firm to cultivate the exclusivity of its offerings. While a Gap in Los Angeles carries nearly the same product line as one in Milwaukee, each Zara store is stocked with items tailored to the tastes of its local clientele. A Fifth Avenue shopper quips, “At Gap, everything is the same,” while a Zara shopper in Madrid says, “You’ll never end up looking like someone else.”K. Capell, “Fashion Conquistador,” BusinessWeek, September 4, 2006. Upon visiting a Zara, the CEO of the National Retail Federation marveled, “It’s like you walk into a new store every two weeks.”M. Helft, “Fashion Fast Forward,” Business 2.0, May 2002.

Second, limited runs encourage customers to buy right away and at full price. Savvy Zara shoppers know the newest items arrive on black plastic hangers, with store staff transferring items to wooden ones later on. Don’t bother asking when something will go on sale; if you wait three weeks the item you wanted has almost certainly been sold or moved out to make room for something new. Says one twenty-three year-old Barcelona shopper, “If you see something and don’t buy it, you can forget about coming back for it because it will be gone.”K. Capell, “Fashion Conquistador,” BusinessWeek, September 4, 2006. A study by consulting firm Bain & Company estimated that the industry average markdown ratio is approximately 50 percent, while Zara books some 85 percent of its products at full price.D. Sull and S. Turconi, “Fast Fashion Lessons,” Business Strategy Review, Summer 2008; and K. Capell, “Fashion Conquistador,” BusinessWeek, September 4, 2006.

The constant parade of new, limited-run items also encourages customers to visit often. The average Zara customer visits the store seventeen times per year, compared with only three annual visits made to competitors.N. Kumar and S. Linguri, “Fashion Sense,” Business Strategy Review, Summer 2006. Even more impressive—Zara puts up these numbers with almost no advertising. The firm’s founder has referred to advertising as a “pointless distraction.” The assertion carries particular weight when you consider that during Gap’s collapse, the firm increased advertising spending but sales dropped.P. Bhatnagar, “How Do You Ad(dress) the Gap?” Fortune, October 11, 2004. Fashion retailers spend an average of 3.5 percent of revenue promoting their products, while ad spending at Inditex is just 0.3 percent.“Zara, A Spanish Success Story,” CNN.com, June 15, 2001, http://edition.cnn.com/BUSINESS/programs/yourbusiness/stories2001/zara.

Finally, limited production runs allow the firm to, as Zara’s CEO once put it, “reduce to a minimum the risk of making a mistake, and we do make mistakes with our collections.”C. Vitzthum, “Zara’s Success Lies in Low-Cost Lines and a Rapid Turnover of Collections,” Wall Street Journal, May 18, 2001. Failed product introductions are reported to be just 1 percent, compared with the industry average of 10 percent.N. Kumar and S. Linguri, “Fashion Sense,” Business Strategy Review, Summer 2006. So even though Zara has higher manufacturing costs than rivals, Inditex gross margins are 56.8 percent compared to 37.5 percent at Gap.C. Rohwedder, “Zara Grows as Retail Rivals Struggle,” Wall Street Journal, March 26, 2009. For labor cost comparison, K. Capell, “Zara Thrives by Breaking All the Rules,” BusinessWeek, October 9, 2008, reports that workers in Spain earn an average of $1,650/month versus $206/month in China’s Guangdong Province.

While stores provide valuable front-line data, headquarters plays a major role in directing in-store operations. Software is used to schedule staff based on each store’s forecasted sales volume, with locations staffing up at peak times such as lunch or early evening. The firm claims these more flexible schedules have shaved staff work hours by 2 percent. This constant refinement of operations throughout the firm’s value chain has helped reverse a prior trend of costs rising faster than sales.C. Rohwedder and K. Johnson, “Pace-Setting Zara Seeks More Speed to Fight Its Rising Cheap-Chic Rivals,” Wall Street Journal, February 20, 2008.

Even the store displays are directed from “The Cube,” where a basement staging area known as “Fashion Street” houses a Potemkin village of bogus storefronts meant to mimic some of the chain’s most exclusive locations throughout the world. It’s here that workers test and fine-tune the chain’s award-winning window displays, merchandise layout, and even determine the in-store soundtrack. Every two weeks, new store layout marching orders are forwarded to managers at each location.C. Rohwedder and K. Johnson, “Pace-Setting Zara Seeks More Speed to Fight Its Rising Cheap-Chic Rivals,” Wall Street Journal, February 20, 2008.

Technology ≠ Systems. Just Ask Prada

Here’s another interesting thing about Zara. Given the sophistication and level of technology integration within the firm’s business processes, you’d think that Inditex would far outspend rivals on tech. But as researchers Donald Sull and Stefano Turconi discovered, “Whether measured by IT workers as a percentage of total employees or total spending as a percentage of sales, Zara’s IT expenditure is less than one-fourth the fashion industry average.”D. Sull and S. Turconi, “Fast Fashion Lessons,” Business Strategy Review, Summer 2008. Zara excels by targeting technology investment at the points in its value chain where it will have the most significant impact, making sure that every dollar spent on tech has a payoff.

Contrast this with high-end fashion house Prada’s efforts at its flagship Manhattan location. The firm hired the Pritzker Prize–winning hipster architect Rem Koolhaas to design a location Prada would fill with jaw-dropping technology. All items for sale in the store would sport with radio frequency identification (RFID) tagsSmall chip-based tags that wirelessly emit a unique identifying code for the item that they are attached to. Think of RFID systems as a next-generation bar code. (small chip-based tags that wirelessly emit a unique identifying code for the item that they are attached to). Walk into a glass dressing room and customers could turn the walls opaque, then into a kind of combination mirror and heads-up display. By wirelessly reading the tags on each garment, dressing rooms would recognize what was brought in and make recommendations of matching accessories as well as similar products that patrons might consider. Customers could check inventory, and staff sporting PDAs could do the same. A dressing room camera would allow clients to see their front and back view side-by-side as they tried on clothes.

It all sounded slick, but execution of the vision was disastrous. Customers didn’t understand the foot pedals that controlled the dressing room doors and displays. Reports surfaced of fashionistas disrobing in full view, thinking the walls went opaque when they didn’t. Others got stuck in dressing rooms when pedals failed to work, or doors broke, unable to withstand the demands of the high-traffic tourist location. The inventory database was often inaccurate, regularly reporting items as out of stock even though they weren’t. As for the PDAs, staff reported that they “don’t really use them anymore” and that “we put them away so tourists don’t play with them.” The investment in Prada’s in-store technology was also simply too high, with estimates suggesting the location took in just one-third the sales needed to justify expenses.G. Lindsay, “Prada’s High-Tech Misstep,” Business 2.0, March 1, 2004.

The Prada example offers critical lessons for managers. While it’s easy to get seduced by technology, an information system (IS)An integrated solution that combines five components: hardware, software, data, procedures, and the people who interact with and are impacted by the system. is actually made up of more than hardware and software. An IS also includes data used or created by the system, as well as the procedures and the people who interact with the system.A. Sanchenko, “Foundations of Information Systems in Business” (lecture, October 13, 2007), http://www.scribd.com/doc/396076/Foundations-of-Information-Systems-in-Business. Getting the right mix of these five components is critical to executing a flawless information system rollout. Financial considerations should forecast the return on investment (ROI)The amount earned from an expenditure.—the amount earned from an expenditure—of any such effort (i.e., what will we get for our money and how long will it take to receive payback?). And designers need to thoroughly test the system before deployment. At Prada’s Manhattan flagship store, the effort looked like tech chosen because it seemed fashionable rather than functional.

Key Takeaways

  • Zara store management and staff use PDAs and POS systems to gather and analyze customer preference data to plan future designs based on feedback, rather than on hunches and guesswork.
  • Zara’s combination of vertical integration and technology-orchestrated supplier coordination, just-in-time manufacturing, and logistics allows it to go from design to shelf in days instead of months.
  • Advantages accruing to Inditex include fashion exclusivity, fewer markdowns and sales, lower marketing expenses, and more frequent customer visits.
  • Zara’s IT expenditures are low by fashion industry standards. The spectacular benefits reaped by Zara from the deployment of technology have resulted from targeting technology investment at the points in the value chain where it has the greatest impact, and not from the sheer magnitude of the investment. This is in stark contrast to Prada’s experience with in-store technology deployment.
  • While information technology is just hardware and software, information systems also include data, people, and procedures. It’s critical for managers to think about systems, rather than just technologies, when planning for and deploying technology-enabled solutions.

Questions and Exercises

  1. In what ways is the Zara model counterintuitive? In what ways has Zara’s model made the firm a better performer than Gap and other competitors?
  2. What factors account for a firm’s profit margin? What does Gap focus on? What factors does Zara focus on to ensure a strong profit margin?
  3. How is data captured in Zara stores? Using what types or classifications of information systems? How does the firm use this data?
  4. What role does technology play in enabling the other elements of Zara’s counterintuitive strategy? Could the firm execute its strategy without technology? Why or why not?
  5. How does technology spending at Zara compare to that of rivals? Advertising spending? Failed product percentages? Markdowns?
  6. What risks are inherent in the conventional practices in the fashion industry? Is Zara susceptible to these risks? Is Zara susceptible to different risks? If so, what are these?
  7. Consider the Prada case mentioned in the sidebar “Technology ≠ Systems.” What did Prada fail to consider when it rolled out the technology in its flagship location? Could this effort have been improved for better results? If you were put in charge of this kind of effort, what factors would you consider? What would determine whether you’d go forward with the effort or not? If you did go forward, what factors would you consider and how might you avoid some of the mistakes made by Prada?

3.3 Moving Forward

Learning Objectives

After studying this section you should be able to do the following:

  1. Detail how Zara’s approach counteracts specific factors that Gap has struggled with for over a decade.
  2. Identify the environmental threats that Zara is likely to face, and consider options available to the firm for addressing these threats.

The holy grail for the strategist is to craft a sustainable competitive advantage that is difficult for competitors to replicate. And for nearly two decades Zara has delivered the goods. But that’s not to say the firm is done facing challenges.

Consider the limitations of Zara’s Spain-centric, just-in-time manufacturing model. By moving all of the firm’s deliveries through just two locations, both in Spain, the firm remains hostage to anything that could create a disruption in the region. Firms often hedge risks that could shut down operations—think weather, natural disaster, terrorism, labor strife, or political unrest—by spreading facilities throughout the globe. If problems occur in northern Spain, Zara has no such fallback.

In addition to the operationsThe organizational activities that are required to produce goods or services. Operations activities can involve the development, execution, control, maintenance, and improvement of an organization’s service and manufacturing procedures. vulnerabilities above, the model also leaves the firm potentially more susceptible to financial vulnerabilities during periods when the euro strengthens relative to the dollar. Many low-cost manufacturing regions have currencies that are either pegged to the dollar or have otherwise fallen against the euro. This situation means Zara’s Spain-centric costs rise at higher rates compared to competitors, presenting a challenge in keeping profit margins in check. Rising transportation costs are another concern. If fuel costs rise, the model of twice-weekly deliveries that has been key to defining the Zara experience becomes more expensive to maintain.

Still, Zara is able to make up for some cost increases by raising prices overseas (in the United States, Zara items can cost 40 percent or more than they do in Spain). Zara reports that all North American stores are profitable, and that it can continue to grow its presence, serving forty to fifty stores with just two U.S. jet flights a week.J. Tagliabue, “A Rival to Gap That Operates Like Dell,” New York Times, May 30, 2003. Management has considered a logistics center in Asia, but expects current capacity will suffice until 2013.C. Rohwedder and K. Johnson, “Pace-Setting Zara Seeks More Speed to Fight Its Rising Cheap-Chic Rivals,” Wall Street Journal, February 20, 2008. Another possibility might be a center in the Maquiladora region of northern Mexico, which could serve the U.S. markets via trucking capacity similar to the firm’s Spain-based access to Europe, while also providing a regional center to serve expansion throughout the Western Hemisphere.

Rivals have studied the Zara recipe, and while none have attained the efficiency of Amancio Ortega’s firm, many are trying to learn from the master. There is precedent for contract firms closing the cycle time gap with vertically integrated competitors that own their own factories. Dell (a firm that builds its own PCs while nearly all its competitors use contract labor) has recently seen its manufacturing advantage from vertical integration fall as the partners that supply rivals have mimicked its techniques and have become far more efficient.T. Friscia, K. O’Marah, D. Hofman, and J. Souza, “The AMR Research Supply Chain Top 25 for 2009,” AMR Research, May 28, 2009, http://www.amrresearch.com/Content/View.aspx?compURI=tcm:7-43469. In terms of the number of new models offered, clothing is actually more complex than computing, suggesting that Zara’s value chain may be more difficult to copy. Still, H&M has increased the frequency of new items in stores, Forever 21 and Uniqlo get new looks within six weeks, and Renner, a Brazilian fast fashion rival, rolls out mini collections every two months.M. Pfeifer, “Fast and Furious,” Latin Trade, September 2007; and C. Rohwedder and K. Johnson, “Pace-Setting Zara Seeks More Speed to Fight Its Rising Cheap-Chic Rivals,” Wall Street Journal, February 20, 2008. Rivals have a keen eye on Inditex, with the CFO of luxury goods firm Burberry claiming the firm is a “fantastic case study” and “we’re mindful of their techniques.”C. Rohwedder and K. Johnson, “Pace-Setting Zara Seeks More Speed to Fight Its Rising Cheap-Chic Rivals,” Wall Street Journal, February 20, 2008.

Finally, firm financial performance can also be impacted by broader economic conditions. When the economy falters, consumers simply buy less and may move a greater share of their wallet to less-stylish and lower-cost offerings from deep discounters like Wal-Mart. Zara is particularly susceptible to conditions in Spain, since the market accounts for nearly 40 percent of Inditex sales,J. Hall, “Zara Is Now Bigger Than Gap,” Telegraph, August 18, 2008. as well as to broader West European conditions (which with Spain make up 79 percent of sales).C. Rohwedder, “Zara Grows as Retail Rivals Struggle,” Wall Street Journal, March 26, 2009. Global expansion will provide the firm with a mix of locations that may be better able to endure downturns in any single region. Recent Spanish and European financial difficulties have made clear the need to decrease dependence on sales within one region.

Zara’s winning formula can only exist through management’s savvy understanding of how information systems can enable winning strategies (many tech initiatives were led by José Maria Castellano, a “technophile” business professor who became Ortega’s right-hand man in the 1980s).C. Rohwedder and K. Johnson, “Pace-Setting Zara Seeks More Speed to Fight Its Rising Cheap-Chic Rivals,” Wall Street Journal, February 20, 2008. It is technology that helps Zara identify and manufacture the clothes customers want, get those products to market quickly, and eliminate costs related to advertising, inventory missteps, and markdowns. A strategist must always scan the state of the market as well as the state of the art in technology, looking for new opportunities and remaining aware of impending threats. With systems so highly tuned for success, it may be unwise to bet against “The Cube.”

Key Takeaway

  • Zara’s value chain is difficult to copy; but it is not invulnerable, nor is future dominance guaranteed. Zara management must be aware of the limitations in its business model, and must continually scan its environment and be prepared to react to new threats and opportunities.

Questions and Exercises

  1. The Zara case shows how information systems can impact every single management discipline. Which management disciplines were mentioned in this case? How does technology impact each?
  2. Would a traditional Internet storefront work well with Zara’s business model? Why or why not?
  3. Zara’s just-in-time, vertically integrated model has served the firm well, but an excellent business is not a perfect business. Describe the limitations of Zara’s model and list steps that management might consider to minimize these vulnerabilities.