The History of Computing
Computers touch all most every aspect of our lives today. We take the way they work for granted and the unsung heroes who built the technology, protocols, philosophies, and circuit boards, patched them all together - and sometimes willed amazingness out of nothing. Not in this podcast. Welcome to the History of Computing. Let's get our nerd on!
info_outline
Lotus: From Yoga to Software
06/27/2023
Lotus: From Yoga to Software
Nelumbo nucifera, or the sacred lotus, is a plant that grows in flood plains, rivers, and deltas. Their seeds can remain dormant for years and when floods come along, blossom into a colony of plants and flowers. Some of the oldest seeds can be found in China, where they’re known to represent longevity. No surprise, given their level of nitrition and connection to the waters that irrigated crops by then. They also grow in far away lands, all the way to India and out to Australia. The flower is sacred in Hinduism and Buddhism, and further back in ancient Egypt. Padmasana is a Sanskrit term meaning lotus, or Padma, and Asana, or posture. The Pashupati seal from the Indus Valley civilization shows a diety in what’s widely considered the first documented yoga pose, from around 2,500 BCE. 2,700 years later (give or take a century), the Hindu author and mystic Patanjali wrote a work referred to as the Yoga Sutras. Here he outlined the original asanas, or sitting yoga poses. The Rig Veda, from around 1,500 BCE, is the oldest currently known Vedic text. It is also the first to use the word “yoga”. It describes songs, rituals, and mantras the Brahmans of the day used - as well as the Padma. Further Vedic texts explore how the lotus grew out of Lord Vishnu with Brahma in the center. He created the Universe out of lotus petals. Lakshmi went on to grow out of a lotus from Vishnu as well. It was only natural that humans would attempt to align their own meditation practices with the beautiful meditatios of the lotus. By the 300s, art and coins showed people in the lotus position. It was described in texts that survive from the 8th century. Over the centuries contradictions in texts were clarified in a period known as Classical Yoga, then Tantra and and Hatha Yoga were developed and codified in the Post-Classical Yoga age, and as empires grew and India became a part of the British empire, Yoga began to travel to the west in the late 1800s. By 1893, Swami Vivekananda gave lectures at the Parliament of Religions in Chicago. More practicioners meant more systems of yoga. Yogendra brought asanas to the United States in 1919, as more Indians migrated to the United States. Babaji’s kriya yoga arrived in Boston in 1920. Then, as we’ve discussed in previous episodes, the United States tightened immigration in the 1920s and people had to go to India to get more training. Theos Bernard’s Hatha Yoga: The Report of a Personal Experience brought some of that knowledge home when he came back in 1947. Indra Devi opened a yoga studio in Hollywood and wrote books for housewives. She brought a whole system, or branch home. Walt and Magana Baptiste opened a studio in San Francisco. Swamis began to come to the US and more schools were opened. Richard Hittleman began to teach yoga in New York and began to teach on television in 1961. He was one of the first to seperate the religious aspect from the health benefits. By 1965, the immigration quotas were removed and a wave of teachers came to the US to teach yoga. The Beatles went to India in 1966 and 1968, and for many Transcendental Meditation took root, which has now grown to over a thousand training centers and over 40,000 teachers. Swamis opened meditation centers, institutes, started magazines, and even magazines. Yoga became so big that Rupert Holmes even poked fun of it in his song “Escape (The Piña Colada Song)” in 1979. Yoga had become part of the counter-culture, and the generation that followed represented a backlash of sorts. A common theme of the rise of personal computers is that the early pioneers were a part of that counter-culture. Mitch Kapor graduated high school in 1967, just in time to be one of the best examples of that. Kapor built his own calculator in as a kid before going to camp to get his first exposure to programming on a Bendix. His high school got one of the 1620 IBM minicomputers and he got the bug. He went off to Yale at 16 and learned to program in APL and then found Computer Lib by Ted Nelson and learned BASIC. Then he discovered the Apple II. Kapor did some programming for $5 per hour as a consultant, started the first east coast Apple User Group, and did some work around town. There are generations of people who did and do this kind of consulting, although now the rates are far higher. He met a grad student through the user group named Eric Rosenfeld who was working on his dissertation and needed some help programming, so Kapor wrote a little tool that took the idea of statistical analysis from the Time Shared Reactive Online Library, or TROLL, and ported it to the microcomputer, which he called Tiny Troll. Then he enrolled in the MBA program at MIT. He got a chance to see VisiCalc and meet Bob Frankston and Dan Bricklin, who introduced him to the team at Personal Software. Personal Software was founded by Dan Fylstra and Peter Jennings when they published Microchips for the KIM-1 computer. That led to ports for the 1977 Trinity of the Commodore PET, Apple II, and TRS-80 and by then they had taken Bricklin and Franston’s VisiCalc to market. VisiCalc was the killer app for those early PCs and helped make the Apple II successful. Personal Software brought Kapor on, as well as Bill Coleman of BEA Systems and Electronic Arts cofounder Rich Mellon. Today, software developers get around 70 percent royalties to publish software on app stores but at the time, fees were closer to 8 percent, a model pulled from book royalties. Much of the rest went to production of the box and disks, the sales and marketing, and support. Kapor was to write a product that could work with VisiCalc. By then Rosenfeld was off to the world of corporate finance so Kapor moved to Silicon Valley, learned how to run a startup, moved back east in 1979, and released VisiPlot and VisiTrend in 1981. He made over half a million dollars in the first six months in royalties. By then, he bought out Rosenfeld’s shares in what he was doing, hired Jonathan Sachs, who had been at MIT earlier, where he wrote the STOIC programming language, and then went to work at Data General. Sachs worked on spreadsheet ideas at Data General with a manager there, John Henderson, but after they left Data General, and the partnership fell apart, he worked with Kapor instead. They knew that for software to be fast, it needed to be written in a lower level language, so they picked the Intel 8088 assembly language given that C wasn’t fast enough yet. The IBM PC came in 1981 and everything changed. Mitch Kapor and Jonathan Sachs started Lotus in 1982. Sachs got to work on what would become Lotus 1-2-3. Kapor turned out to be a great marketer and product manager. He listened to what customers said in focus groups. He pushed to make things simpler and use less jargon. They released a new spreadsheet tool in 1983 and it worked flawlessly on the IBM PC and while Microsoft had Multiplan and VisCalc was the incumbent spreadsheet program, Lotus quickly took market share from then and SuperCalc. Conceptually it looked similar to VisiCalc. They used the letter A for the first column, B for the second, etc. That has now become a standard in spreadsheets. They used the number 1 for the first row, the number 2 for the second. That too is now a standard. They added a split screen, also now a standard. They added macros, with branching if-then logic. They added different video modes, which could give color and bitmapping. They added an underlined letter so users could pull up a menu and quickly select the item they wanted once they had those orders memorized, now a standard in most menuing systems. They added the ability to add bar charts, pie charts, and line charts. One could even spread their sheet across multiple monitors like in a magazine. They refined how fields are calculated and took advantage of the larger amounts of memory to make Lotus far faster than anything else on the market. They went to Comdex towards the end of the year and introduced Lotus 1-2-3 to the world. The software could be used as a spreadsheet, but the 2 and 3 referred to graphics and database management. They did $900,000 in orders there before they went home. They couldn’t even keep up with the duplication of disks. Comdex was still invitation only. It became so popular that it was used to test for IBM compatibility by clone makers and where VisiCalc became the app that helped propel the Apple II to success, Lotus 1-2-3 became the app that helped propel the IBM PC to success. Lotus was rewarded with $53 million in sales for 1983 and $156 million in 1984. Mitch Kapor found himself. They quickly scaled from less than 20 to 750 employees. They brought in Freada Klein who got her PhD to be the Head of Employee Relations and charged her with making them the most progressive employer around. After her success at Lotus, she left to start her own company and later married. Sachs left the company in 1985 and moved on to focus solely on graphics software. He still responds to requests on the phpBB forum at dl-c.com. They ran TV commercials. They released a suite of Mac apps they called Lotus Jazz. More television commercials. Jazz didn’t go anywhere and only sold 20,000 copies. Meanwhile, Microsoft released Excel for the Mac, which sold ten times as many. Some blamed the lack os sales on the stringent copy protection. Others blamed the lack of memory to do cool stuff. Others blamed the high price. It was the first major setback for the young company. After a meteoric rise, Kapor left the company in 1986, at about the height of their success. He replaced himself with Jim Manzi. Manzi pushed the company into network applications. These would become the center of the market but were just catching on and didn’t prove to be a profitable venture just yet. A defensive posture rather than expanding into an adjacent market would have made sense, at least if anyone knew how aggressive Microsoft was about to get it would have. Manzi was far more concerned about the millions of illegal copies of the software in the market than innovation though. As we turned the page to the 1990s, Lotus had moved to a product built in C and introduced the ability to use graphical components in the software but not wouldn’t be ported to the new Windows operating system until 1991 for Windows 3. By then there were plenty of competitors, including Quattro Pro and while Microsoft Excel began on the Mac, it had been a showcase of cool new features a windowing operating system could provide an application since released for Windows in 1987. Especially what they called 3d charts and tabbed spreadsheets. There was no catching up to Microsoft by then and sales steadily declined. By then, Lotus released Lotus Agenda, an information manager that could be used for time management, project management, and as a database. Kapor was a great product manager so it stands to reason he would build a great product to manage products. Agenda never found commercial success though, so was later open sourced under a GPL license. Bill Gross wrote Magellan there before he left to found GoTo.com, which was renamed to Overture and pioneered the idea of paid search advertising, which was acquired by Yahoo!. Magellan cataloged the internal drive and so became a search engine for that. It sold half a million copies and should have been profitable but was cancelled in 1990. They also released a word processor called Manuscript in 1986, which never gained traction and that was cancelled in 1989, just when a suite of office automation apps needed to be more cohesive. Ray Ozzie had been hired at Software Arts to work on VisiCalc and then helped Lotus get Symphony out the door. Symphony shipped in 1984 and expanded from a spreadsheet to add on text with the DOC word processor, and charts with the GRAPH graphics program, FORM for a table management solution, and COM for communications. Ozzie dutifully shipped what he was hired to work on but had a deal that he could build a company when they were done that would design software that Lotus would then sell. A match made in heaven as Ozzie worked on PLATO and borrowed the ideas of PLATO Notes, a collaboration tool developed at the University of Illinois Champagne-Urbana to build what he called Lotus Notes. PLATO was more more than productivity. It was a community that spanned decades and Control Data Corporation had failed to take it to the mass corporate market. Ozzie took the best parts for a company and built it in isolation from the rest of Lotus. They finally released it as Lotus Notes in 1989. It was a huge success and Lotus bought Iris in 1994. Yet they never found commercial success with other socket-based client server programs and IBM acquired Lotus in 1995. That product is now known as Domino, the name of the Notes 4 server, released in 1996. Ozzie went on to build a company called Groove Networks, which was acquired by Microsoft, who appointed him one of their Chief Technology Officers. When Bill Gates left Microsoft, Ozzie took the position of Chief Software Architect he vacated. He and Dave Cutler went on to work on a project called Red Dog, which evolved into what we now know as Microsoft Azure. Few would have guessed that Ozzie and Kapor’s handshake agreement on Notes could have become a real product. Not only could people not understand the concept of collaboration and productivity on a network in the late 1980s but the type of deal hadn’t been done. But Kapor by then realized that larger companies had a hard time shipping net-new software properly. Sometimes those projects are best done in isolation. And all the better if the parties involved are financially motivated with shares like Kapor wanted in Personal Software in the 1970s before he wrote Lotus 1-2-3. VisiCalc had sold about a million copies but that would cease production the same year Excel was released. Lotus hung on longer than most who competed with Microsoft on any beachhead they blitzkrieged. Microsoft released Exchange Server in 1996 and Notes had a few good years before Exchange moved in to become the standard in that market. Excel began on the Mac but took the market from Lotus eventually, after Charles Simonyi stepped in to help make the product great. Along the way, the Lotus ecosystem created other companies, just as they were born in the Visi ecosystem. Symantec became what we now call a “portfolio” company in 1985 when they introduced NoteIt, a natural language processing tool used to annotate docs in Lotus 1-2-3. But Bill Gates mentioned Lotus by name multiple times as a competitor in his Internet Tidal Wave memo in 1995. He mentioned specific features, like how they could do secure internet browsing and that they had a web publisher tool - Microsoft’s own FrontPage was released in 1995 as well. He mentioned an internet directory project with Novell and AT&T. Active Directory was released a few years later in 1999, after Jim Allchin had come in to help shepherd LAN Manager. Notes itself survived into the modern era, but by 2004 Blackberry released their Exchange connector before they released the Lotus Domino connector. That’s never a good sign. Some of the history of Lotus is covered in Scott Rosenberg’s 2008 book, Dreaming in Code. Others are documented here and there in other places. Still others are lost to time. Kapor went on to invest in UUNET, which became a huge early internet service provider. He invested in Real Networks, who launched the first streaming media service on the Internet. He invested in the creators of Second Life. He never seemed vindictive with Microsoft but after AOL acquired Netscape and Microsoft won the first browser war, he became the founding chair of the Mozilla Foundation and so helped bring Firefox to market. By 2006, Firefox took 10 percent of the market and went on to be a dominant force in browsers. Kapor has also sat on boards and acted as an angel investor for startups ever since leaving the company he founded. He also flew to Wyoming in 1990 after he read a post on The WELL from John Perry Barlow. Barlow was one of the great thinkers of the early Internet. They worked with Sun Microsystems and GNU Debugging Cypherpunk John Gilmore to found the Electronic Frontier Foundation, or EFF. The EFF has since been the nonprofit who leads the fight for “digital privacy, free speech, and innovation.” So not everything is about business.
/episode/index/show/thehistoryofcomputing/id/27298917
info_outline
Section 230 and the Concept of Internet Exceptionalism
06/05/2023
Section 230 and the Concept of Internet Exceptionalism
We covered computer and internet copyright law in a previous episode. That type of law began with interpretations that tried to take the technology out of cases so they could be interpreted as though what was being protected was a printed work, or at least it did for a time. But when it came to the internet, laws, case law, and their knock-on effects, the body of jurisprudence work began to diverge. Safe Harbor mostly refers to the Online Copyright Infringement Liability Limitation Act, or OCILLA for short, was a law passed in the late 1980s that shields online portals and internet service providers from copyright infringement. Copyright infringement is one form of immunity, but more was needed. Section 230 was another law that protects those same organizations from being sued for 3rd party content uploaded on their sites. That’s the law Trump wanted overturned during his final year in office but given that the EU has Directive 2000/31/EC, Australia has the Defamation Act of 2005, Italy has the Electronic Commerce Directive 2000, and lots of other countries like England and Germany have had courts find similarly, it is now part of being an Internet company. Although the future of “big tech” cases (and the damage many claim is being done to democracy) may find it refined or limited. That’s because the concept of Internet Exceptionalism itself is being reconsidered now that the internet is here to stay. Internet Exceptionalism is a term that notes that laws that diverge from precedents for other forms of media distribution. For example, a newspaper can be sued for liable or defamation, but a website is mostly shielded from such suits because the internet is different. Pages are available instantly, changes be made instantly, and the reach is far greater than ever before. The internet has arguably become the greatest tool to spread democracy and yet potentially one of its biggest threats. Which some might have argued about newspapers, magazines, and other forms of print media in centuries past. The very idea of Internet Exceptionalism has eclipsed the original intent. Chris Cox and Ron Widen initially intended to help fledgling Internet Service Providers (ISPs) jumpstart content on the internet. The internet had been privatized in 1995 and companies like CompuServe, AOL, and Prodigy were already under fire for the content on their closed networks. Cubby v CompuServe in 1991 had found that online providers weren’t considered publishers of content and couldn’t be held liable for free speech practiced on their platforms in part because they did not exercise editorial control of that content. Stratton Oakmont v Prodigy found that Prodigy did have editorial control (and in fact advertised themselves as having a better service because of it) and so could be found liable like a newspaper would. Cox and Widen were one of the few conservative and liberal pairs of lawmakers who could get along in the decisive era when Newt Gingrich came to power and tried to block everything Bill Clinton tried to do. Yet there were aspects of the United States that were changing outside of politics. Congress spent years negotiating a telecommunications overhaul bill that came to be known as The Telecommunications Act of 1996. New technology led to new options. Some saw content they found to be indecent and so the Communications Decency Act (or Title V of the Telecommunications Act) was passed in 1996, but in Reno v ACLU found to be a violation of the first amendment, and struck down by the Supreme Court in 1997. Section 230 of that act was specifically about the preservation of free speech and so severed from the act and stood alone. It would be adjudicated time and time and eventually became an impenetrable shield that protects online providers from the need to scan every message posted to a service to see if it would get them sued. Keep in mind that society itself was changing quickly in the early 1990s. Tipper Gore wanted to slap a label on music to warn parents that it had explicit lyrics. The “Satanic Panic” as it’s called by history reused tropes such as cannibalism and child murder to give the moral majority an excuse to try to restrict that which they did not understand. Conservative and progressive politics have always been a 2 steps forward and 1 step back truce. Heavy metal would seem like nothin’ once parents heard the lyrics of gagster rap. But Section 230 continued on. It stated that “No provider or user of an interactive computer service shall be treated as the publisher or speaker of any information provided by another information content provider.” It only took 27 words to change the world. They said that the people that host the content can’t be sued for the content because, as courts interpreted it, it’s free speech. Think of a public forum like a hall on a college campus that might restrict one group from speaking and so suppress speech or censer a group. Now, Section 230 didn’t say it wasn’t allowed to screen material but instead shielded providers from being held liable for that material. The authors of the bill felt that if providers would be held liable for any editing that they wouldn’t do any. Now providers could edit some without reviewing every post. And keep in mind the volume of posts in message boards and of new websites had already become too much in the late 1990s to be manually monitored. Further, as those companies became bigger business they became more attractive to law suits. Section 230 had some specific exclusions. Any criminal law could still be applied, as could state, sex trafficking, and privacy laws. Intellectual property laws also remained untouched, thus OCILLA. To be clear, reading the law, the authors sought to promote the growth of the internet - and it worked. Yelp gets sued over revues but cases are dismissed. Twitter can get sued over a Tweet when someone doesn’t like what is said, but it’s the poster and not Twitter who is liable. Parody sites, whistleblower sites, watchdog sites, revue sites, blogs, and an entire industry was born, which each player of what would later be known as the Web 2.0 market could self-regulate themselves. Those businesses grew far beyond the message boards of the 1990s. This was also a time when machine learning became more useful. A site like Facebook could show a feed of posts not in reverse chronological order, but instead by “relevance.” Google could sell ads and show them based on the relevance of a search term. Google could buy YouTube and they could have ads on videos. Case after case poked at the edges of what could be used to hold a site liable. The fact that the courts saw a post on Reddit as free speech, no matter how deplorable the comments, provided a broad immunity to liability that was, well, exceptional in a way. Some countries could fine or imprison people if they posted something negative about the royal family or party in charge. Some of those countries saw the freedom of speech so important as a weapon that could be used against the US in a way. The US became a safe haven in a way to free speech and many parts of the internet were anonymous. In this way (as was previously done with films and other sources of entertainment and news) the US began to export the culture of free speech. But every country also takes imports. Some of those were real, true ideas homegrown or brought in from abroad. Early posters of message boards maybe thought the Armenian Genocide was a hoax - or the Holocaust. A single post could ruin a career. Craigslist allowed for sex trafficking and while they eventually removed that, sites like Backpage have received immunity. So even some of the exceptions are, um, not. Further, extremist groups use pages to spread propaganda and even recruit soldiers to spread terror. The courts found that sites were immune to suits over fake profiles on dating sites - even if it was a famous person and the person was getting threatening calls. The courts initially found sites needed to take down content if they were informed it was libelous - but have received broad immunity even when they don’t due to the sheer amount of content. Batzel v Smith saw a lawyers firm ruined over false reports she was the granddaughter of Nazi Heinrich Himmler and the beneficiary of Nazi art theft, even though she wasn’t - she too lost her case. Sites provide neutral tools and so are shielded from defamation - even if they’re neutralish you rarely see them held to account. In Goddard v. Google, the Google Keyword Tool recommended that advertisers include the word “free” in mobile content, which Goddard claimed led to fraudulent subscription service recruitment. This was machine learning-based recommendations. The court again found provided the Keyword Tool was neutral that advertisers could adopt or reject the recommendation. Still, time and time again the idea of safe harbor for internet companies and whether internet exceptionalism should continue comes up. The internet gave a voice to the oppressed, but also to the oppressors. That’s neutrality in a way, except that the oppressors (especially when state sponsored actors are involved) often have more resources to drown out other voices, just like in real life. Some have argued a platform like Facebook should be held accountable for their part in the Capitol riots, which is to say as a place where people practiced free speech. Others look to Backpage as facilitating the exploitation of children or as a means of oppression. Others still see terrorist networks as existing and growing because of the ability to recruit online. The Supreme Court is set to hear docket number 21-1333 in 2022. Gonzalez v. Google was brought by Reynaldo Gonzalez, and looks at whether 230 can immunize Google even though they have made targeted recommendations - in this case when ISIS used YouTube vides to recruit new members - through the recommendation algorithm. An algorithm that would be neutral. But does a platform as powerful have a duty to do more, especially when there’s a chance that Section 230 bumps up against anti-terrorism legislation. Again and again the district courts in the United States have found section 230 provides broad immunization to online content providers. Now, the Supreme Court will weigh in. After that, billions of dollars may have to be pumped into better content filtration or they may continue to apply broad first amendment guidance. The Supreme Court is packed with “originalists”. They still have phones, which the framers did not. The duty that common law places on those who can disseminate negligent or reckless content has lost the requirement for reasonable care due to the liability protections afforded purveyors of content by Section 230. This has given rise to hate speech and misinformation. John Perry Barlow’s infamous A Declaration of the Independence of Cyberspace in protest of the CDA was supported by Section 230 of that same law. But the removal of the idea and duty of reasonable care and the exemptions have now removed any accountability from what seems like any speech. Out of the ashes of accountability the very concept of free speech and where the duty of reasonable care lies may be reborn. We now have the ability to monitor via machine learning, we’ve now redefined what it means to moderate, and there’s now a robust competition for eyeballs on the internet. We’ve also seen how a lack of reasonable standards can lead to real life consequences and that an independent cyberspace can bleed through into the real world. If the Supreme Court simply upholds findings from the past then the movement towards internet sovereignty may accelerate or may stay the same. Look to where venture capital flows for clues as to how the First Amendment will crash into the free market, and see if its salty waters leave data and content aggregators with valuations far lower than where they once were. The asset of content may some day become a liability with injuries that could provide an existential threat to the owner. The characters may walk the astral plane but eventually must return to the prime material plane along their tether to take a long rest or face dire consequences. The world simply can’t continue to become more and more toxic - and yet there’s a reason the First Amendment is, well, first. Check out Twenty-Six Words Created the Internet. What Will It Take to Save It?
/episode/index/show/thehistoryofcomputing/id/27052317
info_outline
Bluetooth: From Kings to Personal Area Networks
05/17/2023
Bluetooth: From Kings to Personal Area Networks
Bluetooth The King Ragnar Lodbrok was a legendary Norse king, conquering parts of Denmark and Sweden. And if we’re to believe the songs, he led some of the best raids against the Franks and the the loose patchwork of nations Charlemagne put together called the Holy Roman Empire. We use the term legendary as the stories of Ragnar were passed down orally and don’t necessarily reconcile with other written events. In other words, it’s likely that the man in the songs sung by the bards of old are likely in fact a composite of deeds from many a different hero of the norse. Ragnar supposedly died in a pit of snakes at the hands of the Northumbrian king and his six sons formed a Great Heathen Army to avenge their father. His sons ravaged modern England int he wake of their fathers death before becoming leaders of various lands they either inherited or conquered. One of those sons, Sigurd Snake-in-the-Eye, returned home to rule his lands and had children, including Harthacnut. He in turn had a son named Gorm. Gorm the Old was a Danish king who lived to be nearly 60 in a time when life expectancy for most was about half that. Gorm raised a Jelling stone in honor of his wife Thyra. As did his son, in the honor of his wife. That stone is carved with runes that say: “King Haraldr ordered this monument made in memory of Gormr, his father, and in memory of Thyrvé, his mother; that Haraldr who won for himself all of Denmark and Norway and made the Danes Christian.” That stone was erected by a Danish king named Herald Gormsson. He converted to Christianity as part of a treaty with the Holy Roman Emperor of the day. He united the tribes of Denmark into a kingdom. One that would go on to expand the reach and reign of the line. Just as Bluetooth would unite devices. Even the logo is a combination of runes that make up his initials HB. Once united, their descendants would go on to rule Denmark, Norway, and England. For a time. Just as Bluetooth would go on to be an important wireless protocol. For a time. Personal Area Networks Many early devices shipped with infrared so people could use a mouse or keyboard. But those never seemed to work so great. And computers with a mouse and keyboard and drawing pad and camera and Zip drive and everything else meant that not only did devices have to be connected to sync but they also had to pull a lot of power and create an even bigger mess on our desks. What the world needed instead was an inexpensive chip that could communicate wirelessly and not pull a massive amount of power since some would be in constant communication. And if we needed a power cord then might as well just use USB or those RS-232 interfaces (serial ports) that were initially developed in 1960 - that were slow and cumbersome. And we could call this a Personal Area Network, or PAN. The Palm Pilot was popular, but docking and pluging in that serial port was not exactly optimal. Yet every ATX motherboard had a port or two. So a Bluetooth Special Interest Group was formed to conceive and manage the standard in 1988 and while initially had half a dozen companies now has over 30,000. The initial development started in the late 1990s with Ericcson. It would use short-range UHF radio waves in the 2.402 GHz and 2.48 GHz bands to exchange data with computers and cell phones, which were evolving into mobile devices at the time. The technology was initially showcased at COMDEX in 1999. Within a couple of years there were phones that could sync, kits for cars, headsets, and chips that could be put into devices - or cards or USB adapters, to get a device to sync 721 Kbps. We could add 2 to 8 Bluetooth secondary devices that paired to our primary. They then frequency hopped using their Bluetooth device address provided by the primary, which sends a radio signal to secondaries with a range of addresses to use. The secondaries then respond with the frequency and clock state. And unlike a lot of other wireless technologies, it just kinda’ worked. And life seemed good. Bluetooth went to the IEEE, which had assigned networking the 802 standard with Ethernet being 802.3 and Wi-Fi being 802.11. So Personal Area Networks became 802.15, with Bluetooth 1.1 becoming 802.15.1. And the first phone shipped in 2001, the Sony Ericsson T39. Bluetooth 2 came in 2005 and gave us 2.1 Mbps speeds and increased the range from 10 to 30 meters. By then, over 5 million devices were shipping every week. More devices mean we have a larger attack surface space. And security researchers were certainly knocking at the door. Bluetooth 2.1 added secure simple pairing. Then Bluetooth 3 in 2009 bringing those speeds up to 24 Mbps and once connected allowing Wi-Fi to pick up connections once established. But we were trading speed for energy and this wasn’t really the direction Bluetooth needed to go. Even if a billion devices had shipped by the end of 2006. Bluetooth 4 The mobility era was upon us and it was increasingly important, not just for the ARM chips, but also for the rest of the increasing number of devices, to use less power. Bluetooth 4 came along in 2010 and was slower at 1 Mbps, but used less energy. This is when the iPhone 4S line fully embraced the technology, helping make it a standard. While not directly responsible for the fitness tracker craze, it certainly paved the way for a small coin cell battery to run these types of devices for long periods of time. And it allowed for connecting devices 100 meters, or well over 300 feet away. So leave the laptop in one room and those headphones should be fine in the next. And while we’re at it, maybe we want those headphones to work on two different devices. This is where Multipoint comes into play. That’s the feature of Bluetooth 4 that allows those devices to pass seamlessly between the phone and the laptop, maintaining a connection to each. Apple calls their implementation of this feature Handoff. Bluetooth 5 came in 2016, allowing for connections up to 240 meters, or around 800 feet. Well, according to what’s between us and our devices, as with other protocols. We also got up to 2 Mbps, which dropped as we moved further away from devices. Thus we might get buffering issues or slower transfers with weaker connections. But not outright dropping the connection. Bluetooth Evolves Bluetooth was in large part developed to allow our phones to sync to our computers. Most don’t do that any more. And the developers wanted to pave the way for wireless headsets. But it also allowed us to get smart scales, smart bulbs, wearables like smart watches and glasses, Bluetooth printers, webcams, keyboards, mice, GPS devices, thermostats, and even a little device that tells me when I need to water the plants. Many home automation devices, or IoT as we seem to call them these days began as Bluetooth but given that we want them to work when we take all our mostly mobile computing devices out of the home, many of those have moved over to Wi-Fi these days. Bluetooth was initially conceived as a replacement for the serial port. Higher throughput needs moved to USB and USB-C. Lower throughput has largely moved to Bluetooth, with the protocol split between Low Energy and higher bandwidth application which with high definition audio now includes headphones. Once the higher throughput needs went to parallel and SCSI but now there are so many other options. And the line is blurred between what goes where. Billions of routers and switches have been sold, billions of wireless access points. Systems on a Chip now include Wi-Fi and Bluetooth together on the same chip. The programming languages for native apps have also given us frameworks and APIs where we can establish a connection over 5G, Wi-Fi, or Bluetooth, and then hand them off where the needs diverge. Seamless to those who use our software and elegant when done right. Today over four billion bluetooth devices ship per year, growing at about 10 percent a year. The original needs that various aspects of Bluetooth was designed for have moved to other protocols and the future of the Personal Area Network may be at least in part moved to Wi-Fi or 5G. But for now it’s a standard that has aged well and continues to make life easier for those who use it.
/episode/index/show/thehistoryofcomputing/id/26869728
info_outline
One History Of 3D Printing
05/03/2023
One History Of 3D Printing
One of the hardest parts of telling any history, is which innovations are significant enough to warrant mention. Too much, and the history is so vast that it can't be told. Too few, and it's incomplete. Arguably, no history is ever complete. Yet there's a critical path of innovation to get where we are today, and hundreds of smaller innovations that get missed along the way, or are out of scope for this exact story. Children have probably been placing sand into buckets to make sandcastles since the beginning of time. Bricks have survived from round 7500BC in modern-day Turkey where humans made molds to allow clay to dry and bake in the sun until it formed bricks. Bricks that could be stacked. And it wasn’t long before molds were used for more. Now we can just print a mold on a 3d printer. A mold is simply a block with a hollow cavity that allows putting some material in there. People then allow it to set and pull out a shape. Humanity has known how to do this for more than 6,000 years, initially with lost wax casting with statues surviving from the Indus Valley Civilization, stretching between parts of modern day Pakistan and India. That evolved to allow casting in gold and silver and copper and then flourished in the Bronze Age when stone molds were used to cast axes around 3,000 BCE. The Egyptians used plaster to cast molds of the heads of rulers. So molds and then casting were known throughout the time of the earliest written works and so the beginning of civilization. The next few thousand years saw humanity learn to pack more into those molds, to replace objects from nature with those we made synthetically, and ultimately molding and casting did its part on the path to industrialization. As we came out of the industrial revolution, the impact of all these technologies gave us more and more options both in terms of free time as humans to think as well as new modes of thinking. And so in 1868 John Wesley Hyatt invented injection molding, patenting the machine in 1872. And we were able to mass produce not just with metal and glass and clay but with synthetics. And more options came but that whole idea of a mold to avoid manual carving and be able to produce replicas stretched back far into the history of humanity. So here we are on the precipice of yet another world-changing technology becoming ubiquitous. And yet not. 3d printing still feels like a hobbyists journey rather than a mature technology like we see in science fiction shows like Star Trek with their replicators or printing a gun in the Netflix show Lost In Space. In fact the initial idea of 3d printing came from a story called Things Pass By written all the way back in 1945! I have a love-hate relationship with 3D printing. Some jobs just work out great. Others feel very much like personal computers in the hobbyist era - just hacking away until things work. It’s usually my fault when things go awry. Just as it was when I wanted to print things out on the dot matrix printer on the Apple II. Maybe I fed the paper crooked or didn’t check that there was ink first or sent the print job using the wrong driver. One of the many things that could go wrong. But those fast prints don’t match with the reality of leveling and cleaning nozzles and waiting for them to heat up and pulling filament out of weird places (how did it get there, exactly)! Or printing 10 add-ons for a printer to make it work the way it probably should have out of the box. Another area where 3d printing is similar to the early days of the personal computer revolution is that there are a few different types of technology in use today. These include color-jet printing (CJP), direct metal printing (DMP), fused deposition modeling (FDM), Laser Additive Manufacturing (LAM, multi-jet printing (MJP), stereolithography (SLA), selective laser melting (SLM), and selective laser sintering (SLS). Each could be better for a given type of print job to be done. Some forms have flourished while others are either their infancy or have been abandoned like extinct languages. Language isolates are languages that don’t fit into other families. Many are the last in a branch of a larger language family tree. Others come out of geographically isolated groups. Technology also has isolates. Konrad Zuse built computers in pre-World War II Germany and after that aren’t considered to influence other computers. In other words, every technology seems to have a couple of false starts. Hideo Kodama filed the first patent to 3d print in 1980 - but his method of using UV lights to harden material doesn’t get commercialized. Another type of 3d printing includes printers that were inkjets that shot metal alloys onto surfaces. Inkjet printing was invented by Ichiro Endo at Canon in the 1950s, supposedly when he left a hot iron on a pen and ink bubbled out. Thus the “Bubble jet” printer. And Jon Vaught at HP was working on the same idea at about the same time. These were patented and used to print images from computers over the coming decades. Johannes Gottwald patented a printer like this in 1971. Experiments continued through the 1970s when companies like Exxon were trying to improve various prototyping processes. Some of their engineers joined an inventor Robert Howard in the early 1980s to found a company called Howtek and they produced the Pixelmaster, using hot-melt inks to increment the ink jet with solid inks, which then went on to be used by Sanders Prototype, which evolved into a company called Solidscape to market the Modelmaker. And some have been used to print solar cells, living cells, tissue, and even edible birthday cakes. That same technique is available with a number of different solutions but isn’t the most widely marketable amongst the types of 3D printers available. SLA There’s often a root from which most technology of the day is derived. Charles, or Chuck, Hull coined the term stereolithography, where he could lay down small layers of an object and then cure the object with UV light, much as the dentists do with fillings today. This is made possibly by photopolymers, or plastics that are easily cured by an ultraviolet light. He then invented the stereolithography apparatus, or SLA for short, a machine that printed from the bottom to the top by focusing a laser on photopolymer while in a liquid form to cure the plastic into place. He worked on it in 1983, filed the patent in 1984, and was granted the patent in 1986. Hull also developed a file format for 3D printing called STL. STL files describe the surface of a three-dimensional object, geometrically using Cartesian coordinates. Describing coordinates and vectors means we can make objects bigger or smaller when we’re ready to print them. 3D printers print using layers, or slices. Those can change based on the filament on the head of a modern printer, the size of the liquid being cured, and even the heat of a nozzle. So the STL file gets put into a slicer that then converts the coordinates on the outside to the polygons that are cured. These are polygons in layers, so they may appear striated rather than perfectly curved according to the size of the layers. However, more layers take more time and energy. Such is the evolution of 3D printing. Hull then founded a company called 3D Systems in Valencia California to take his innovation to market. They sold their first printer, the SLA-1 in 1988. New technologies start out big and expensive. And that was the case with 3D Systems. They initially sold to large engineering companies but when solid-state lasers came along in 1996 they were able to provide better systems for cheaper. Languages also have other branches. Another branch in 3d printing came in 1987, just before the first SLA-1 was sold. Carl Deckard and his academic adviser Joe Beaman at the University of Texas worked on a DARPA grant to experiment with creating physical objects with lasers. They formed a company to take their solution to market called DTM and filed a patent for what they called selective laser sintering. This compacts and hardens a material with a heat source without having to liquify it. So a laser, guided by a computer, can move around a material and harden areas to produce a 3D model. Now in addition to SLA we had a second option, with the release of the Sinterstation 2500plus. Then 3D Systems then acquired DTM for $45 million in 2001. FDM After Hull published his findings for SLA and created the STL format, other standards we use today emerged. FDM is short for Fused Deposition Modeling and was created by Scott Crump in 1989. He then started a company with his wife Lisa to take the product to market, taking the company public in 1994. Crump’s first patent expired in 2009. In addition to FDM, there are other formats and techniques. AeroMat made the first 3D printer that could produce metal in 1997. These use a laser additive manufacturing process, where lasers fuse powdered titanium alloys. Some go the opposite direction and create out of bacteria or tissue. That began in 1999, when Wake Forest Institute of Regenerative medicine grew a 3D printed urinary bladder in a lab to be used as a transplant. We now call this bioprinting and can take tissue and lasers to rebuild damaged organs or even create a new organ. Organs are still in their infancy with success trials on smaller animals like rabbits. Another aspect is printing dinner using cell fibers from cows or other animals. There are a number of types of materials used in 3D printing. Most printers today use a continuous feed of one of these filaments, or small coiled fibers of thermoplastics that melt instead of burn when they’re heated up. The most common in use today is PLA, or polylactic acid, is a plastic initially created by Wall Carothers of DuPont, the same person that brought us nylon, neoprene, and other plastic derivatives. It typically melts between 200 and 260 degrees Celsius. Printers can also take ABS filament, which is short for acrylonitrile-butadien-styerene. Other filament types include HIPS, PET, CPE, PVA, and their derivative forms. Filament is fed into a heated extruder assembly that melts the plastic. Once melted, filament extrudes into place through a nozzle as a motor sends the nozzle on a x and y axis per layer. Once a layer of plastic is finished being delivered to the areas required to make up the desired slice, the motor moves the extruder assembly up or down on a z axis between layers. Filament is just between 1.75 millimeters and 3 millimeters and comes in spools between half a kilogram and two kilograms. These thermoplastics cool very quickly. Once all of the slices are squirted into place, the print is removed from the bed and the nozzle cools off. Filament comes in a number of colors and styles. For example, wood fibers can be added to filament to get a wood-grained finish. Metal can be added to make prints appear metallic and be part metal. Printing isn’t foolproof, though. Filament often gets jammed or the spool gets stuck, usually when something goes wrong. Filament also needs to be stored in a temperature and moisture controlled location or it can cause jobs to fail. Sometimes the software used to slice the .stl file has an incorrect setting, like the wrong size of filament. But in general, 3D printing using the FDM format is pretty straight forward these days. Yet this is technology that should have moved faster in terms of adoption. The past 10 years have seen more progress than the previous ten though. Primarily due to the maker community. Enter the Makers The FDM patent expired in 2009. In 2005, a few years before the FDM patent expired, Dr. Adrian Bowyer started a project to bring inexpensive 3D printers to labs and homes around the world. That project evolved into what we now call the Replicating Rapid Prototyper, or RepRap for short. RepRap evolved into an open source concept to create self-replicating 3D printers and by 2008, the Darwin printer was the first printer to use RepRap. As a community started to form, more collaborators designed more parts. Some were custom parts to improve the performance of the printer, or replicate the printer to become other printers. Others held the computing mechanisms in place. Some even wrote code to make the printer able to boot off a MicroSD card and then added a network interface so files could be uploaded to the printer wirelessly. There was a rising tide of printers. People were reading about what 3D printers were doing and wanted to get involved. There was also a movement in the maker space, so people wanted to make things themselves. There was a craft to it. Part of that was wanting to share. Whether that was at a maker space or share ideas and plans and code online. Like the RepRap team had done. One of those maker spaces was NYC Resistor, founded in 2007. Bre Pettis, Adam Mayer, and Zach Smith from there took some of the work from the RepRap project and had ideas for a few new projects they’d like to start. The first was a site that Zach Smith created called Thingiverse. Bre Pettis joined in and they allowed users to upload .stl files and trade them. It’s now the largest site for trading hundreds of thousands of designs to print about anything imaginable. Well, everything except guns. Then comes 2009. The patent for FDM expires and a number of companies respond by launching printers and services. Almost overnight the price for a 3D printer fell from $10,000 to $1,000 and continued to drop. Shapeways had created a company the year before to take files and print them for people. Pettis, Mayer, and Smith from NYC Resistor also founded a company called MakerBot Industries. They’d already made a little bit of a name for themselves with the Thingiverse site. They knew the mind of a maker. And so they decided to make a kit to sell to people that wanted to build their own printers. They sold 3,500 kits in the first couple of years. They had a good brand and knew the people who bought these kinds of devices. So they took venture funding to grow the company. So they raised $10M in funding in 2011 in a round led by the Foundry Group, along with Bezos, RRE, 500 Startups and a few others. They hired and grew fast. Smith left in 2012 and they were getting closer and closer with Stratasys, who if we remember were the original creators of FDM. So Stratasys ended up buying out the company in 2013 for $403M. Sales were disappointing so there was a changeup in leadership, with Pettis leaving and they’ve become much more about additive manufacturing than a company built to appeal to makers. And yet the opportunity to own that market is still there. This was also an era of Kickstarter campaigns. Plenty of 3D printing companies launched through kickstarter including some to take PLA (a biodegradable filament) and ABS materials to the next level. The ExtrusionBot, the MagicBox, the ProtoPlant, the Protopasta, Mixture, Plybot, Robo3D, Mantis, and so many more. Meanwhile, 3D printing was in the news. 2011 saw the University of Southhampton design a 3d printed aircraft. Ecologic printing cars, and practically every other car company following suit that they were fabricating prototypes with 3d printers, even full cars that ran. Some on their own, some accidentally when parts are published in .stl files online violating various patents. Ultimaker was another RepRap company that came out of the early Darwin reviews. Martijn Elserman, Erik de Bruin, and Siert Wijnia who couldn’t get the Darwin to work so they designed a new printer and took it to market. After a few iterations, they came up with the Ultimaker 2 and have since been growing and releasing new printers A few years later, a team of Chinese makers, Jack Chen, Huilin Liu, Jingke Tang, Danjun Ao, and Dr. Shengui Chen took the RepRap designs and started a company to manufacturing (Do It Yourself) kits called Creality. They have maintained the open source manifesto of 3D printing that they inherited from RepRap and developed version after version, even raising over $33M to develop the Ender6 on Kickstarter in 2018, then building a new factory and now have the capacity to ship well over half a million printers a year. The future of 3D Printing We can now buy 3D printing pens, over 170 3D Printer manufacturers including 3D systems, Stratasys, and Ceality but also down-market solutions like Fusion3, Formlabs, Desktop Metal, Prusa, and Voxel8. There’s also a RecycleBot concept and additional patents expiring every year. There is little doubt that at some point, instead of driving to Home Depot to get screws or basic parts, we’ll print them. Need a new auger for the snow blower? Just print it. Cover on the weed eater break? Print it. Need a dracolich mini for the next Dungeons and Dragons game? Print it. Need a new pinky toe. OK, maybe that’s a bit far. Or is it? In 2015, Swedish Cellink releases bio-ink made from seaweed and algae, which could be used to print cartilage and later released the INKREDIBLE 3D printer for bio printing. The market in 2020 was valued at $13.78 billion with 2.1 million printers shipped. That’s expected to grow at a compound annual growth rate of 21% for the next few years. But a lot of that is healthcare, automotive, aerospace, and prototyping still. Apple made the personal computer simple and elegant. But no Apple has emerged for 3D printing. Instead it still feels like the Apple II era, where there are 3D printers in a lot of schools and many offer classes on generating files and printing. 3D printers are certainly great for prototypers and additive manufacturing. They’re great for hobbyists, which we call makers these days. But there will be a time when there is a printer in most homes, the way we have electricity, televisions, phones, and other critical technologies. But there are a few things that have to happen first, to make the printers easier to use. These include: Every printer needs to automatically level. This is one of the biggest reasons jobs fail and new users become frustrated. More consistent filament. Spools are still all just a little bit different. Printers need sensors in the extruder that detect if a job should be paused because the filament is jammed, humid, or caught. This adds the ability to potentially resume print jobs and waste less filament and time. Automated slicing in the printer microcode that senses the filament and slices. Better system boards (e.g. there’s a tool called Klipper that moves the math from the system board on a Creality Ender 3 to a Raspberry Pi). Cameras on the printer should watch jobs and use TinyML to determine if they are going to fail as early as possible to halt printing so it can start over. Most of the consumer solutions don’t have great support. Maybe users are limited to calling a place in a foreign country where support hours don’t make sense for them or maybe the products are just too much of a hacker/maker/hobbyist solution. There needs to be an option for color printing. This could be a really expensive sprayer or ink like inkjet printers use at first We love to paint minis we make for Dungeons and Dragons but could get amazingly accurate resolutions to create amazing things with automated coloring. For a real game changer, the RecycleBot concept needs to be merged with the printer. Imagine if we dropped our plastics into a recycling bin that 3D printers of the world used to create filament. This would help reduce the amount of plastics used in the world in general. And when combined with less moving around of cheap plastic goods that could be printed at home, this also means less energy consumed by transporting goods. The 3D printing technology is still a generation or two away from getting truly mass-marketed. Most hobbyists don’t necessarily think of building an elegant, easy-to-use solution...
/episode/index/show/thehistoryofcomputing/id/26734395
info_outline
Adobe: From Pueblos to Fonts and Graphics to Marketing
04/16/2023
Adobe: From Pueblos to Fonts and Graphics to Marketing
The Mogollon culture was an indigenous culture in the Western United States and Mexico that ranged from New Mexico and Arizona to Sonora, Mexico and out to Texas. They flourished from around 200 CE until the Spanish showed up and claimed their lands. The cultures that pre-existed them date back thousands more years, although archaeology has yet to pinpoint exactly how those evolved. Like many early cultures, they farmed and foraged. As they farmed more, their homes become more permanent and around 800 CE they began to create more durable homes that helped protect them from wild swings in the climate. We call those homes adobes today and the people who lived in those peublos and irrigated water, often moving higher into mountains, we call the Peubloans - or Pueblo Peoples. Adobe homes are similar to those found in ancient cultures in what we call Turkey today. It’s an independent evolution. Adobe Creek was once called Arroyo de las Yeguas by the monks from Mission Santa Clara and then renamed to San Antonio Creek by a soldier Juan Prado Mesa when the land around it was given to him by the governor of Alto California at the time, Juan Bautista Alvarado. That’s the same Alvarado as the street if you live in the area. The creek runs for over 14 miles north from the Black Mountain and through Palo Alto, California. The ranchers built their adobes close to the creeks. American settlers led the Bear Flag Revolt in 1846, and took over the garrison of Sonoma, establishing the California Republic - which covered much of the lands of the Peubloans. There were only 33 of them at first, but after John Fremont (yes, he of whom that street is named after as well) encouraged the Americans, they raised an army of over 100 men and Fremont helped them march on Sutter’s fort, now with the flag of the United States, thanks to Joseph Revere of the US Navy (yes, another street in San Francisco bears his name). James Polk had pushed to expand the United States. Manfiest Destiny. Remember The Alamo. Etc. The fort at Monterey fell, the army marched south. Admiral Sloat got involved. They named a street after him. General Castro surrendered - he got a district named after him. Commodore Stockton announced the US had taken all of Calfironia soon after that. Manifest destiny was nearly complete. He’s now basically the patron saint of a city, even if few there know who he was. The forts along the El Camino Real that linked the 21 Spanish Missions, a 600-mile road once walked by their proverbial father, Junípero Serra following the Portolá expedition of 1769, fell. Stockton took each, moving into Los Angeles, then San Diego. Practically all of Alto California fell with few shots. This was nothing like the battles for the independence of Texas, like when Santa Anna reclaimed the Alamo Mission. Meanwhile, the waters of Adobe Creek continued to flow. The creek was renamed in the 1850s after Mesa built an adobe on the site. Adobe Creek it was. Over the next 100 years, the area evolved into a paradise with groves of trees and then groves of technology companies. The story of one begins a little beyond the borders of California. Utah was initialy explored by Francisco Vázquez de Coronado in 1540 and settled by Europeans in search of furs and others who colonized the desert, including those who established the Church of Jesus Christ of Latter-day Saints, or the Mormons - who settled there in 1847, just after the Bear Flag Revolt. The United States officially settled for the territory in 1848 and Utah became a territory and after a number of map changes wher ethe territory got smaller, was finally made a state in 1896. The University of Utah had been founded all the way back in 1850, though - and re-established in the 1860s. 100 years later, the University of Utah was a hotbed of engineers who pioneered a number of graphical advancements in computing. John Warnock went to grad school there and then went on to co-found Adobe and help bring us PostScript. Historically, PS, or Postscript was a message to be placed at the end of a letter, following the signature of the author. The PostScript language was a language to describe a page of text computationally. It was created by Adobe when Warnock, Doug Brotz, Charles Geschke, Bill Paxton (who worked on the Mother of All Demos with Doug Englebart during the development of Online System, or NLS in the late 70s and then at Xerox PARC), and Ed Taft. Warnock invented the Warnock algorithm while working on his PhD and went to work at Evans & Sutherland with Ivan Sutherland who effectively created the field of computer graphics. Geschke got his PhD at Carnegie Melon in the early 1970s and then went of to Xerox PARC. They worked with Paxton at PARC and before long, these PhDs and mathematicians had worked out the algorithms and then the languages to display images on computers while working on InterPress graphics at Xerox and Gerschke left Xerox and started Adobe. Warnock joined them and they went to market with Interpress as PostScript, which became a foundation for the Apple LaswerWriter to print graphics. Not only that, PostScript could be used to define typefaces programmatically and later to display any old image. Those technologies became the foundation for the desktop publishing industry. Apple released the 1984 Mac and other vendors brought in PostScript to describe graphics in their proprietary fashion and by 1991 they released PostScript Level 2 and then PostScript 3 in 1997. Other vendors made their own or furthered standards in their own ways and Adobe could have faded off into the history books of computing. But Adobe didn’t create one product, they created an industry and the company they created to support that young industry created more products in that mission. Steve Jobs tried to buy Adobe before that first Mac as released, for $5,000,000. But Warnock and Geschke had a vision for an industry in mind. They had a lot of ideas but development was fairly capital intensive, as were go to market strategies. So they went public on the NASDAQ in 1986. They expanded their PostScript distribution and sold it to companies like Texas Instruments for their laser printer, and other companies who made IBM-compatible companies. They got up to $16 million in sales that year. Warnock’s wife was a graphic designer. This is where we see a diversity of ideas help us think about more than math. He saw how she worked and could see a world where Ivan Sutherland’s Sketchpad was much more given how far CPUs had come since the TX-0 days at MIT. So Adobe built and released Illustrator in 1987. By 1988 they broke even on sales and it raked in $19 million in revenue. Sales were strong in the universities but PostScript was still the hot product, selling to printer companies, typesetters, and other places were Adobe signed license agreements. At this point, we see where the math, cartesian coordinates, drawn by geometric algorithms put pixels where they should be. But while this was far more efficient than just drawing a dot in a coordinate for larger images, drawing a dot in a pixel location was still the easier technology to understand. They created Adobe Screenline in 1989 and Collectors Edition to create patterns. They listened to graphic designers and built what they heard humans wanted. Photoshop Nearly every graphic designer raves about Adobe Photoshop. That’s because Photoshop is the best selling graphics editorial tool that has matured far beyond most other traditional solutions and now has thousands of features that allow users to manipulate images in practically any way they want. Adobe Illustrator was created in 1987 and quickly became the de facto standard in vector-based graphics. Photoshop began life in 1987 as well, when Thomas and John Knoll, wanted to build a simpler tool to create graphics on a computer. Rather than vector graphics they created a raster graphical editor. They made a deal with Barneyscan, a well-known scanner company that managed to distribute over two hundred copies of Photoshop with their scanners and Photoshop became a hit as it was the first editing software people heard about. Vector images are typically generated with Cartesian coordinates based on geometric formulas and so scale out more easily. Raster images are comprised of a grid of dots, or pixels, and can be more realistic. Great products are rewarded with competitions. CorelDRAW was created in 1989 when Michael Bouillon and Pat Beirne built a tool to create vector illustrations. The sales got slim after other competitors entered the market and the Knoll brothers got in touch with Adobe and licensed the product through them. The software was then launched as Adobe Photoshop 1 in 1990. They released Photoshop 2 in 1991. By now they had support for paths, and given that Adobe also made Illustrator, EPS and CMYK rasterization, still a feature in Photoshop. They launched Adobe Photoshop 2.5 in 1993, the first version that could be installed on Windows. This version came with a toolbar for filters and 16-bit channel support. Photoshop 3 came in 1994 and Thomas Knoll created what was probably one of the most important features added, and one that’s become a standard in graphical applications since, layers. Now a designer could create a few layers that each had their own elements and hide layers or make layers more transparent. These could separate the subject from the background and led to entire new capabilities, like an almost faux 3 dimensional appearance of graphics.. Then version four in 1996 and this was one of the more widely distributed versions and very stable. They added automation and this was later considered part of becoming a platform - open up a scripting language or subset of a language so others built tools that integrated with or sat on top of those of a product, thus locking people into using products once they automated tasks to increase human efficiency. Adobe Photoshop 5.0 added editable type, or rasterized text. Keep in mind that Adobe owned technology like PostScript and so could bring technology from Illustrator to Photoshop or vice versa, and integrate with other products - like export to PDF by then. They also added a number of undo options, a magnetic lasso, improved color management and it was now a great tool for more advanced designers. Then in 5.5 they added a save for web feature in a sign of the times. They could created vector shapes and continued to improve the user interface. Adobe 5 was also a big jump in complexity. Layers were easy enough to understand, but Photoshop was meant to be a subset of Illustrator features and had become far more than that. So in 2001 they released Photoshop Elements. By now they had a large portfolio of products and Elements was meant to appeal to the original customer base - the ones who were beginners and maybe not professional designers. By now, some people spent 40 or more hours a day in tools like Photoshop and Illustrator. Adobe Today Adobe had released PostScript, Illustrator, and Photoshop. But they have one of the most substantial portfolios of products of any company. They also released Premiere in 1991 to get into video editing. They acquired Aldus Corporation to get into more publishing workflows with PageMaker. They used that acquisition to get into motion graphics with After Effects. They acquired dozens of companies and released their products as well. Adobe also released the PDF format do describe full pages of information (or files that spread across multiple pages) in 1993 and Adobe Acrobat to use those. Acrobat became the de facto standard for page distribution so people didn’t have to download fonts to render pages properly. They dabbled in audio editing when they acquired Cool Edit Pro from Syntrillium Software and so now sell Adobe Audition. Adobe’s biggest acquisition was Macromedia in 2005. Here, they added a dozen new products to the portfolio, which included Flash, Fireworks, WYSYWIG web editor Dreamweaver, ColdFusion, Flex, and Breeze, which is now called Adobe Connect. By now, they’d also created what we call Creative Suite, which are packages of applications that could be used for given tasks. Creative Suite also signaled a transition into a software as a service, or SaaS mindset. Now customers could pay a monthly fee for a user license rather than buy large software packages each time a new version was released. Adobe had always been a company who made products to create graphics. They expanded into online marketing and web analytics when they bought Omniture in 2009 for $1.8 billion. These products are now normalized into the naming convention used for the rest as Adobe Marketing Cloud. Flash fell by the wayside and so the next wave of acquisitions were for more mobile-oriented products. This began with Day Software and then Nitobi in 2011. And they furthered their Marketing Cloud support with an acquisition of one of the larger competitors when they acquired Marketo in 2018 and acquiring Workfront in 2020. Given how many people started working from home, they also extended their offerings into pure-cloud video tooling with an acquisition of Frame.io in 2021. And here we see a company started by a bunch of true computer sciencists from academia in the early days of the personal computer that has become far more. They could have been rolled into Apple but had a vision of a creative suite of products that could be used to make the world a prettier place. Creative Suite then Creative Cloud shows a move of the same tools into a more online delivery model. Other companies come along to do similar tasks, like infinite digital whiteboard Miro - so they have to innovate to stay marketable. They have to continue to increase sales so they expand into other markets like the most adjacent Marketing Cloud. At 22,500+ employees and with well over $12 billion in revenues, they have a lot of families dependent on maintaining that growth rate. And so the company becomes more than the culmination of their software. They become more than graphic design, web design, video editing, animation, and visual effects. Because in software, if revenues don’t grow at a rate greater than 10 percent per year, the company simply isn’t outgrowing the size of the market and likely won’t be able to justify stock prices at an inflated earnings to price ratio that shows explosive growth. And yet once a company saturates sales in a given market they have shareholders to justify their existence to. Adobe has survived many an economic downturn and boom time with smart, measured growth and is likely to continue doing so for a long time to come.
/episode/index/show/thehistoryofcomputing/id/26553366
info_outline
The Evolution of Fonts on Computers
04/10/2023
The Evolution of Fonts on Computers
Gutenburg shipped the first working printing press around 1450 and typeface was born. Before then most books were hand written, often in blackletter calligraphy. And they were expensive. The next few decades saw Nicolas Jensen develop the Roman typeface, Aldus Manutius and Francesco Griffo create the first italic typeface. This represented a period where people were experimenting with making type that would save space. The 1700s saw the start of a focus on readability. William Caslon created the Old Style typeface in 1734. John Baskerville developed Transitional typefaces in 1757. And Firmin Didot and Giambattista Bodoni created two typefaces that would become the modern family of Serif. Then slab Serif, which we now call Antique, came in 1815 ushering in an era of experimenting with using type for larger formats, suitable for advertisements in various printed materials. These were necessary as more presses were printing more books and made possible by new levels of precision in the metal-casting. People started experimenting with various forms of typewriters in the mid-1860s and by the 1920s we got Frederic Goudy, the first real full-time type designer. Before him, it was part of a job. After him, it was a job. And we still use some of the typefaces he crafted, like Copperplate Gothic. And we saw an explosion of new fonts like Times New Roman in 1931. At the time, most typewriters used typefaces on the end of a metal shaft. Hit a kit, the shaft hammers onto a strip of ink and leaves a letter on the page. Kerning, or the space between characters, and letter placement were often there to reduce the chance that those metal hammers jammed. And replacing a font would have meant replacing tons of precision parts. Then came the IBM Selectric typewriter in 1961. Here we saw precision parts that put all those letters on a ball. Hit a key, the ball rotates and presses the ink onto the paper. And the ball could be replaced. A single document could now have multiple fonts without a ton of work. Xerox exploded that same year with the Xerox 914, one of the most successful products of all time. Now, we could type amazing documents with multiple fonts in the same document quickly - and photocopy them. And some of the numbers on those fancy documents were being spat out by those fancy computers, with their tubes. But as computers became transistorized heading into the 60s, it was only a matter of time before we put fonts on computer screens. Here, we initially used bitmaps to render letters onto a screen. By bitmap we mean that a series, or an array of pixels on a screen is a map of bits and where each should be displayed on a screen. We used to call these raster fonts, but the drawback was that to make characters bigger, we needed a whole new map of bits. To go to a bigger screen, we probably needed a whole new map of bits. As people thought about things like bold, underline, italics, guess what - also a new file. But through the 50s, transistor counts weren’t nearly high enough to do something different than bitmaps as they rendered very quickly and you know, displays weren’t very high quality so who could tell the difference anyways. Whirlwind was the first computer to project real-time graphics on the screen and the characters were simple blocky letters. But as the resolution of screens and the speed of interactivity increased, so did what was possible with drawing glyphs on screens. Rudolf Hell was a German, experimenting with using cathode ray tubes to project a CRT image onto paper that was photosensitive and thus print using CRT. He designed a simple font called Digital Grotesk, in 1968. It looked good on the CRT and the paper. And so that font would not only be used to digitize typesetting, loosely based on Neuzeit Book. And we quickly realized bitmaps weren’t efficient to draw fonts to screen and by 1974 moved to outline, or vector, fonts. Here a Bézier curve was drawn onto the screen using an algorithm that created the character, or glyph using an outline and then filling in the space between. These took up less memory and so drew on the screen faster. Those could be defined in an operating system, and were used not only to draw characters but also by some game designers to draw entire screens of information by defining a character as a block and so taking up less memory to do graphics. These were scalable and by 1979 another German, Peter Karow, used spline algorithms wrote Ikarus, software that allowed a person to draw a shape on a screen and rasterize that. Now we could graphically create fonts that were scalable. In the meantime, the team at Xerox PARC had been experimenting with different ways to send pages of content to the first laser printers. Bob Sproull and Bill Newman created the Press format for the Star. But this wasn’t incredibly flexible like what Karow would create. John Gaffney who was working with Ivan Sutherland at Evans & Sutherland, had been working with John Warnock on an interpreter that could pull information from a database of graphics. When he went to Xerox, he teamed up with Martin Newell to create J&M, which harnessed the latest chips to process graphics and character type onto printers. As it progressed, they renamed it to Interpress. Chuck Geschke started the Imaging Sciences Laboratory at Xerox PARC and eventually left Xerox with Warnock to start a company called Adobe in Warnock’s garage, which they named after a creek behind his house. Bill Paxton had worked on “The Mother of All Demos” with Doug Engelbart at Stanford, where he got his PhD and then moved to Xerox PARC. There he worked on bitmap displays, laser printers, and GUIs - and so he joined Adobe as a co-founder in 1983 and worked on the font algorithms and helped ship a page description language, along with Chuck Geschke, Doug Brotz, and Ed Taft. Steve Jobs tried to buy Adobe in 1982 for $5 million. But instead they sold him just shy of 20% of the company and got a five-year license for PostScript. This allowed them to focus on making the PostScript language more extensible, and creating the Type 1 fonts. These had 2 parts. One that was a set of bit maps And another that was a font file that could be used to send the font to a device. We see this time and time again. The simpler an interface and the more down-market the science gets, the faster we see innovative industries come out of the work done. There were lots of fonts by now. The original 1984 Mac saw Susan Kare work with Jobs and others to ship a bunch of fonts named after cities like Chicago and San Francisco. She would design the fonts on paper and then conjure up the hex (that’s hexadecimal) for graphics and fonts. She would then manually type the hexadecimal notation for each letter of each font. Previously, custom fonts were reserved for high end marketing and industrial designers. Apple considered licensing existing fonts but decided to go their own route. She painstakingly created new fonts and gave them the names of towns along train stops around Philadelphia where she grew up. Steve Jobs went for the city approach but insisted they be cool cities. And so the Chicago, Monaco, New York, Cairo, Toronto, Venice, Geneva, and Los Angeles fonts were born - with her personally developing Geneva, Chicago, and Cairo. And she did it in 9 x 7. I can still remember the magic of sitting down at a computer with a graphical interface for the first time. I remember opening MacPaint and changing between the fonts, marveling at the typefaces. I’d certainly seen different fonts in books. But never had I made a document and been able to set my own typeface! Not only that they could be in italics, outline, and bold. Those were all her. And she inspired a whole generation of innovation. Here, we see a clean line from Ivan Sutherland and the pioneering work done at MIT to the University of Utah to Stanford through the oNLine System (or NLS) to Xerox PARC and then to Apple. But with the rise of Windows and other graphical operating systems. As Apple’s 5 year license for PostScript came and went they started developing their own font standard as a competitor to Adobe, which they called TrueType. Here we saw Times Roman, Courier, and symbols that could replace the PostScript fonts and updating to Geneva, Monaco, and others. They may not have gotten along with Microsoft, but they licensed TrueType to them nonetheless to make sure it was more widely adopted. And in exchange they got a license for TrueImage, which was a page description language that was compatible with PostScript. Given how high resolution screens had gotten it was time for the birth of anti-aliasing. He we could clean up the blocky “jaggies” as the gamers call them. Vertical and horizontal lines in the 8-bit era looked fine but distorted at higher resolutions and so spatial anti-aliasing and then post-processing anti-aliasing was born. By the 90s, Adobe was looking for the answer to TrueImage. So 1993 brought us PDF, now an international standard in ISO 32000-1:2008. But PDF Reader and other tools were good to Adobe for many years, along with Illustrator and then Photoshop and then the other products in the Adobe portfolio. By this time, even though Steve Jobs was gone, Apple was hard at work on new font technology that resulted in Apple Advanced Typography, or AAT. AAT gave us ligature control, better kerning and the ability to write characters on different axes. But even though Jobs was gone, negotiations between Apple and Microsoft broke down to license AAT to Microsoft. They were bitter competitors and Windows 95 wasn’t even out yet. So Microsoft started work on OpenType, their own font standardized language in 1994 and Adobe joined the project to ship the next generation in 1997. And that would evolve into an open standard by the mid-2000s. And once an open standard, sometimes the de facto standard as opposed to those that need to be licensed. By then the web had become a thing. Early browsers and the wars between them to increment features meant developers had to build and test on potentially 4 or 5 different computers and often be frustrated by the results. So the WC3 began standardizing how a lot of elements worked in Extensible Markup Language, or XML. Images, layouts, colors, even fonts. SVGs are XML-based vector image. In other words the browser interprets a language that displays the image. That became a way to render Web Open Format or WOFF 1 was published in 2009 with contributions by Dutch educator Erik van Blokland, Jonathan Kew, and Tal Leming. This built on the CSS font styling rules that had shipped in Internet Explorer 4 and would slowly be added to every browser shipped, including Firefox since 3.6, Chrome since 6.0, Internet Explorer since 9, and Apple’s Safari since 5.1. Then WOFF 2 added Brotli compression to get sizes down and render faster. WOFF has been a part of the W3C open web standard since 2011. Out of Apple’s TrueType came TrueType GX, which added variable fonts. Here, a single font file could contain a number or range of variants to the initial font. So a family of fonts could be in a single file. OpenType added variable fonts in 2016, with Apple, Microsoft, and Google all announcing support. And of course the company that had been there since the beginning, Adobe, jumped on board as well. Fewer font files, faster page loads. So here we’ve looked at the progression of fonts from the printing press, becoming more efficient to conserve paper, through the advent of the electronic typewriter to the early bitmap fonts for screens to the vectorization led by Adobe into the Mac then Windows. We also see rethinking the font entirely so multiple scripts and character sets and axes can be represented and rendered efficiently. I am now converting all my user names into pig Latin for maximum security. Luckily those are character sets that are pretty widely supported. The ability to add color to pig Latin means that OpenType-SVG will allow me add spiffy color to my glyphs. It makes us wonder what’s next for fonts. Maybe being able to design our own, or more to the point, customize those developed by others to make them our own. We didn’t touch on emoji yet. But we’ll just have to save the evolution of character sets and emoji for another day. In the meantime, let’s think on the fact that fonts are such a big deal because Steve Jobs took a caligraphy class from a Trappist monk named Robert Palladino while enrolled at Reed College. Today we can painstakingly choose just the right font with just the right meaning because Palladino left the monastic life to marry and have a son. He taught jobs about serif and san serif and kerning and the art of typography. That style and attention to detail was one aspect of the original Mac that taught the world that computers could have style and grace as well. It’s not hard to imagine if entire computers still only supported one font or even one font per document. Palladino never owned or used a computer though. His influence can be felt through the influence his pupil Jobs had. And it’s actually amazing how many people who had such dramatic impacts on computing never really used one. Because so many smaller evolutions came after them. What evolutions do we see on the horizon today? And how many who put a snippet of code on a service like GitHub may never know the impact they have on so many?
/episode/index/show/thehistoryofcomputing/id/26472642
info_outline
Flight Part II: From Balloons to Autopilot to Drones
04/03/2023
Flight Part II: From Balloons to Autopilot to Drones
In our previous episode, we looked at the history of flight - from dinosaurs to the modern aircraft that carry people and things all over the world. Those helped to make the world smaller, but UAVs and drones have had a very different impact in how we lead our lives - and will have an even more substantial impact in the future. That might not have seemed so likely in the 1700s, though - when unmann Unmanned Aircraft Napoleon conquered Venice in 1797 and then ceded control to the Austrians the same year. He then took it as part of a treaty in 1805 and established the first Kingdom of Italy. Then lost it in 1814. And so they revolted in 1848. One of the ways the Austrians crushed the revolt, in part employing balloons, which had been invented in 1783, that were packed with explosives. 200 balloons packed with bombs later, one found a target. Not a huge surprise that such techniques didn’t get used again for some time. The Japanese tried a similar tactic to bomb the US in World War II - then there were random balloons in the 2020s, just for funsies. A few other inventions needed to find one another in order to evolve into something entirely new. Radio was invented in the 1890s. Nikola Tesla built a radio controlled boat in 1898. Airplanes came along in 1903. Then came airships moved by radio. So it was just a matter of time before the cost of radio equipment came down enough to match the cost of building smaller airplanes that could be controlled with remote controls as well. The first documented occurrence of that was in 1907 when Percy Sperry filed a patent for a kite fashioned to look and operate like a plane, but glide in the wind. The kite string was the first remote control. Then electrical signals went through those strings and eventually the wire turned into radio - the same progress we see with most manual machinery that needs to be mobile. Technology moves upmarket, so Sperry Corporation the aircraft with autopilot features in 1912. At this point, that was just a gyroscopic heading indicator and attitude indicator that had been connected to hydraulically operated elevators and rudders but over time would be able to react to all types of environmental changes to save pilots from having to constantly manually react while flying. That helped to pave the way for longer and safer flights, as automation often does. Then came World War I. Tesla discussed aerial combat using unmanned aircraft in 1915 and Charles Kettering (who developed the electric cash register and the electric car starter) gave us The Kettering Bug, a flying, remote controlled torpedo of sorts. Elmer Sperry worked on a similar device. British war engineers like Archibald Low were also working on attempts but the technology didn’t evolve fast enough and by the end of the war there wasn’t much interest in military funding. But a couple of decades can do a lot. Both for miniaturization and maturity of technology. 1936 saw the development of the first navy UAV aircraft by the name of Queen Bee by Admiral William H. Stanley then the QF2. They was primarily used for aerial target practice as a low-cost radio-controlled drone. The idea was an instant hit and later on, the military called for the development of similar systems, many of which came from Hollywood of all places. Reginald Denny was a British gunner in World War I. They shot things from airplanes. After the war he moved to Hollywood to be an actor. By the 1930s he got interested in model airplanes that could fly and joined up with Paul Whittier to open a chain of hobby shops. He designed a few planes and eventually grew them to be sold to the US military as targets. The Radioplane as they would be known even got joysticks and they sold tens of thousands during World War II. War wasn’t the only use for UAVs. Others were experimenting and by 1936 we got the first radio controlled model airplane competition in 1936, a movement that continued to grow and evolve into the 1970s. We got the Academy of Model Aeronautics (or AMA) in 1936, who launched a magazine called Model Aviation and continues to publish, provide insurance, and act as the UAV, RC airplane, and drone community representative to the FAA. Their membership still runs close to 200,000. Most of these model planes were managed from the ground using radio remote controls. The Federal Communications Commission, or FCC, was established in 1934 to manage the airwaves. They stepped in to manage what frequencies could be used for different use cases in the US, including radio controlled planes. Where there is activity, there are stars. The Big Guff, built by brothers Walt and Bill Guff, was the first truly successful RC airplane in that hobbiest market. Over the next decades solid state electronics got smaller, cheaper, and more practical. As did the way we could transmit bits over those wireless links. 1947 saw the first radar-guided missile, the subsonic Firebird, which over time evolved into a number of programs. Electro-mechanical computers had been used to calculate trajectories for ordinances during World War II so with knowledge of infrared, we got infrared homing then television cameras mounted into missiles and when combined with the proximity fuse, which came with small pressure, magnetic, acoustic, radio, then optical transmitters. We got much better at blowing things up. Part of that was studying the German V-2 rocket programs. They used an analog computer to control the direction and altitude of missiles. The US Polaris and Minuteman missile programs added transistors then microchips to missiles to control the guidance systems. Rockets had computers and so they showed up in airplanes to aid humans in guiding those, often replacing Sperry’s original gyroscopic automations. The Apollo Guidance Computer from the 1969 moon landing was an early example of times when humans even put their lives in the hands of computers - with manual override capabilities of course. Then as the price of chips fell in the 1980s we started to see them in model airplanes. Modern Drones By now, radio controlled aircraft had been used for target practice, to deliver payloads and blow things up, and even for spying. Aircraft without humans to weight them down could run on electric motors rather than combustable engines. Thus they were quieter. This technology allowed the UAVs to fly undetected thus laying the very foundation for the modern depiction of drones used by the military for covert operations. As the costs fell and carrying capacity increased, we saw them used in filmmaking, surveying, weather monitoring, and anywhere else a hobbyist could use their hobby in their career. But the cameras weren’t that great yet. Then Fairchild developed the charge-coupled device, or CCD, in 1969. The first digital camera arguably came out of Eastman Kodak in 1975 when Steven Sasson built a prototype using a mixture of batteries, movie camera lenses, Fairchild CCD sensors, and Motorola parts. Sony came out with the Magnetic Video Camera in 1981 and Canon put the RC701 on the market in 1986. Fuji, Dycam, even the Apple QuickTake, came out in the next few years. Cameras were getting better resolution, and as we turned the page into the 1990s, those cameras got smaller and used CompactFlash to store images and video files. The first aerial photograph is attributed to Gaspar Tournachon, but the militaries of the world used UAVs that were B-17 and Grumman Hellcats from World War II that had been converted to drones full of sensors to study nuclear radiation clouds when testing weapons. Those evolved into Reconnaisance drones like the Aerojet SD-2, with mounted analog cameras in the 50s and 60s. During that time we saw the Ryan Firebees and DC-130As run thousands of flights snapping photos to aid intelligence gathering. Every country was in on it. The USSR, Iran, North Korea, Britain. And the DARPA-instigated Amber and then Predator drones might be considered the modern precursor to drones we play with today. Again, we see the larger military uses come down market once secrecy and cost meet a cool factor down-market. DARPA spent $40 million on the Amber program. Manufacturers of consumer drones have certainly made far more than that. Hobbyists started to develop Do It Yourself (DIY) drone kits in the early 2000s. Now that there were websites, we didn’t have to wait for magazines to show up, we could take to the World Wide Web forums and trade ideas for how to do what the US CIA had done when they conducted the first armed drone strike in 2001 - just maybe without the weapon systems since this was in the back yard. Lithium-ion batteries were getting cheaper and lighter. As were much faster chips. Robotics had come a long way as well, and moving small parts of model aircraft was much simpler than avoiding all the chairs in a room at Stanford. Hobbyists turned into companies that built and sold drones of all sizes, some of which got in the way of commercial aircraft. So the FAA started issuing drone permits in 2006. Every technology had a point, where the confluence of all these technologies meets into a truly commercially viable product. We had Wi-Fi, RF (or radio frequency), iPhones, mobile apps, tiny digital cameras in our phones, and even in spy teddy bears, we understood flight, propellers, plastics were heavier-than-air, but lighter than metal. So in 2010 we got the Parrot AR Drone. This was the first drone that was sold to the masses that was just plug and play. And an explosion of drone makers followed, with consumer products ranging from around $20 to hundreds now. Drone races, drone aerogymnastics, drone footage on our Apple and Google TV screens, and with TinyML projects for every possible machine learning need we can imagine, UAVs that stabilize cameras, can find objects based on information we program into it, and any other use we can imagine. The concept of drones or unmanned aerial vehicles (UAV) has come a long way since the Austrians tried to bomb the Venetians into submission. Today there are mini drones, foldable drones, massive drones that can carry packages, racing drones, and even military drones programmed to kill. In fact, right now there are debates raging in the UN around whether to allow drones to autonomously kill. Because Skynet. We’re also experimenting with passenger drone technology. Because autonomous driving is another convergence just waiting in the wings. Imagine going to the top of a building and getting in a small pod then flying a few buildings over - or to the next city. Maybe in our lifetimes, but not as soon as some of the companies who have gone public to do just this thought.
/episode/index/show/thehistoryofcomputing/id/26430762
info_outline
Flight: From Dinosaurs to Space
03/25/2023
Flight: From Dinosaurs to Space
Humans have probably considered flight since they found birds. As far as 228 million years ago, the Pterosaurs used flight to reign down onto other animals from above and eat them. The first known bird-like dinosaur was the Archaeopteryx, which lived around 150 million years ago. It’s not considered an ancestor of modern birds - but other dinosaurs from the same era, the theropods, are. 25 million years later, in modern China, the Confuciusornis sanctus had feathers and could have flown. The first humans wouldn’t emerge from Africa until 23 million years later. By the 2300s BCE, the Summerians depicted shepherds riding eagles, as humanity looked to the skies in our myths and legends. These were creatures, not vehicles. The first documented vehicle of flight was as far back as the 7th century BCE when the Rāmāyana told of the Pushpaka Vimāna, a palace made by Vishwakarma for Brahma, complete with chariots that flew the king Rama high into the atmosphere. The Odyssey was written around the same time and tells of the Greek pantheon of Gods but doesn’t reference flight as we think of it today. Modern interpretations might move floating islands to the sky, but it seems more likely that the floating island of Aeollia is really the islands off Aeolis, or Anatolia, which we might refer to as the modern land of Turkey. Greek myths from a few hundred years later introduced more who were capable of flight. Icarus flew into the sun with wings that had been fashioned by Daedalus. By then, they could have been aware, through trade routes cut by Alexander and later rulers, of kites from China. The earliest attempts at flight trace their known origins to 500 BCE in China. Kites were, like most physical objects, heavier than air and could still be used to lift an object into flight. Some of those early records even mention the ability to lift humans off the ground with a kite. The principle used in kites was used later in the development of gliders and then when propulsion was added, modern aircraft. Any connection between any of these is conjecture as we can’t know how well the whisper net worked in those ages. Many legends are based on real events. The history of humanity is vast and many of our myths are handed down through the generations. The Greeks had far more advanced engineering capabilities than some of the societies that came after. They were still weary of what happened if they flew too close to the sun. In fact, emperors of China are reported to have forced some to leap from cliffs on a glider as a means of punishment. Perhaps that was where the fear of flight for some originated from. Chinese emperor Wang Mang used a scout with bird features to glide on a scouting mission around the same time as the Icarus myth might have been documented. Whether this knowledge informed the storytellers Ovid documented in his story of Icarus is lost to history, since he didn’t post it to Twitter. Once the Chinese took the string off the kite and they got large enough to fly with a human, they had also developed hang gliders. In the third century BCE, Chinese inventors added the concept of rotors for vertical flight when they developed helicopter-style toys. Those were then used to frighten off enemies. Some of those evolved into the beautiful paper lanterns that fly when lit.There were plenty of other evolutions and false starts with flight after that. Abbas ibn Ferns also glided with feathers in the 9th century. A Benedictine monk did so again in the 11th century. Both were injured when they jumped out of towers in the Middle Ages that spanned the Muslim Golden Age to England. Leonardo da Vinci studied flight for much of his life. His studies produced another human-power ornithopter and other contraptions; however he eventually realized that humans would not be able to fly on their own power alone. Others attempted the same old wings made of bird feathers, wings that flapped on the arms, wings tied to legs, different types of feathers, finding higher places to jump from, and anything they could think of. Many broke bones, which continued until we found ways to supplement human power to propel us into the air. Then a pair of brothers in the Ottoman Empire had some of the best luck. Hezarafen Ahmed Çelebi crossed the Bosphorus strait on a glider. That was 1633, and by then gunpowder already helped the Ottomans conquer Constantinople. That ended the last vestiges of ancient Roman influence along with the Byzantine empire as the conquerers renamed the city to Instanbul. That was the power of gunpowder. His brother then built a rocket using gunpowder and launched himself high in the air, before he glided back to the ground. The next major step was the hot air balloon. The modern hot air balloon was built by the Montgolfier brothers in France and first ridden in 1783 and (Petrescu & Petrescu, 2013). 10 days later, the first gas balloon was invented by Nicholas Louis Robert and Jacques Alexander Charles. The gas balloon used hydrogen and in 1785, used to cross the English Channel. That trip sparked the era of dirigibles. We built larger balloons to lift engines with propellers. That began a period that culminated with the Zeppelin. From the 1700s and on, much of what da Vinci realized was rediscovered, but this time published, and the body of knowledge built out. The physics of flight were then studied as new sciences emerged. Sir George Cayley started to actually apply physics to flight in the 1790s. Powered Flight We see this over and over in history; once we understand the physics and can apply science, progress starts to speed up. That was true when Archimedes defined force multipliers with the simple machines in the 3rd century BCE, true with solid state electronics far later, and true with Cayley’s research. Cayley conducted experiments, documented his results, and proved hypotheses. He finally got to codifying bird flight and why it worked. He studied the Chinese tops that worked like modern helicopters. He documented glided flight and applied math to why it worked. He defined drag and measured the force of windmill blades. In effect, he got to the point that he knew how much power was required based on the ratio of weight to actually sustain flight. Then to achieve that, he explored the physics of fixed-wing aircraft, complete with an engine, tail assembly, and fuel. His work culminated in a work called “On Aerial Navigation” that was published in 1810. By the mid-1850s, there was plenty of research that flowed into the goal for sustained air travel. Ideas like rotors led to rotor crafts. Those were all still gliding. Even with Cayley’s research, we had triplane gliders, gliders launched from balloons. After that, the first aircrafts that looked like the modern airplanes we think of today were developed. Cayley’s contributions were profound. He even described how to mix air with gasoline to build an engine. Influenced by his work, others built propellers. Some of those were steam powered and others powered by tight springs, like clockworks. Aeronautical societies were created, wing counters and cambering were experimented with, and wheels were added to try to lift off. Some even lifted a little off the ground. By the 1890s, the first gasoline powered biplane gliders were developed and flown, even if those early experiments crashed. Humanity was finally ready for powered flight. The Smithsonian housed some of the earliest experiments. They hired their third director, Samuel Langley, in 1887. He had been interested in aircraft for decades and as with many others had studied the Cayley work closely. He was a consummate tinkerer and had already worked in solar physics and developed the Allegheny Time System. The United States War department gave him grants to pursue his ideas to build an airplane. By then, there was enough science that humanity knew it was possible to fly and so there was a race to build powered aircraft. We knew the concepts of drag, rudders, thrust from some of the engineering built into ships. Some of that had been successfully used in the motorcar. We also knew how to build steam engines, which is what he used in his craft. He called it the Aerodrome and built a number of models. He was able to make it further than anyone at the time. He abandoned flight in 1903 when someone beat him to the finish line. That’s the year humans stepped beyond gliding and into the first controlled, sustained, and powered flight. There are reports that Gustave Whitehead beat the Wright Brothers, but he didn’t keep detailed notes or logs, and so the Wrights are often credited with the discovery. They managed to solve the problem of how to roll, built steerable rudders, and built the first biplane with an internal combustion engine. They flew their first airplane out of North Carolina when Orville Wright went 120 feet and his brother went 852 feet later that day. That plane now lives at the National Air and Space Museum in Washington DC and December 17th, 1903 represents the start of the age of flight. The Wright’s spent two years testing gliders and managed to document their results. They studied in wind tunnels, tinkered with engines, and were methodical if not scientific in their approach. They didn’t manage to have a public demonstration until 1908 though and so there was a lengthy battle over the patents they filed. Turns out it was a race and there were a lot of people who flew within months of one another. Decades of research culminated into what had to be: airplanes. Innovation happened quickly. Flight improved enough that planes could cross English Channel by 1909. There were advances after that, but patent wars over the invention drug on and so investors stayed away from the unproven technology. Flight for the Masses The superpowers of the world were at odds for the first half of the 1900s. An Italian pilot flew a reconnaissance mission in Libya in the Italo-Turkish war in 1911. It took only 9 days before they went from just reconnaissance and dropped grenades on Turkish troops from the planes. The age of aerial warfare had begun. The Wrights had received an order for the first plane from the military back in 1908. Military powers took note and by World War I there was an air arm of every military power. Intelligence wins wars. The innovation was ready for the assembly lines, so during and after the war, the first airplane manufacturers were born. Dutch engineer Anthony Fokker was inspired by Wilbur Wright’s exhibition in 1908. He went on to start a company and design the Fokker M.5, which evolved into the Fokker E.I. after World War I broke out in 1914. They mounted a machine gun and synchronized it to the propeller in 1915. Manfred von Richthofen, also known as the Red Baron, flew one before he upgraded to the Fokker D.VII and later an Albatros. Fokker made it all the way into the 1990s before they went bankrupt. Albatros was founded in 1909 by Enno Huth, who went on to found the German Air Force before the war. The Bristol Aeroplane Company was born in 1910 after Sir George White, who was involved in transportation already, met Wilbur Wright in France. Previous companies were built to help hobbyists, similar to how many early PC companies came from inventors as well. This can be seen with people like Maurice Mallet, who helped design gas balloons and dirigibles. He licensed airplane designs to Bristol who later brought in Frank Barnwell and other engineers that helped design the Scout. They based the Bristol Fighters that were used in World War I on those designs. Another British manufacturer was Sopwith, started by Thomas Sopwith, who taught himself to fly and then started a company to make planes. They built over 16,000 by the end of the war. After the war they pivoted to make ABC motorcycles and eventually sold to Hawker Aircraft in 1920, which later sold to Raytheon. The same paradigm played out elsewhere in the world, including the United States. Once those patent disputes were settled, plenty knew flight would help change the world. By 1917 the patent wars in the US had to end as the countries contributions to flight suffered. No investor wanted to touch the space and so there was a lack of capital to expand. Orville Write passed away in 1912 and Wilbur sold his rights to the patents, so the Assistant Secretary of the Navy, Franklin D. Roosevelt, stepped in and brought all the parties to the table to develop a cross-licensing organization. After almost 25 years, we could finally get innovation in flight back on track globally. In rapid succession, Loughead Aircraft, Lockheed, and Douglas Aircraft were founded. Then Jack Northrop left those and started his own aircraft company. Boeing was founded in 1957 as Aero Products and then United Aircraft, which was spun off into United Airlines as a carrier in the 1930s with Boeing continuing to make planes. United was only one of many a commercial airline that was created. Passenger air travel started after the first air flights with the first airline ferrying passengers in 1914. With plenty of airplanes assembled at all these companies, commercial travel was bound to explode into its own big business. Delta started as a cropdusting service in Macon, Georgia in 1925 and has grown into an empire. The worlds largest airline at the time of this writing is American Airlines, which started in 1926 when a number of smaller airlines banded together. Practically every country had at least one airline. Pan American (Panam for short) in 1927, Ryan Air started in 1926, Slow-Air in 1924, Finnair in 1923, Quantus in 1920, KLM in 1919, and the list goes on. Enough that the US passed the Air Commerce Act in 1926, which over time led to the department of Air Commerce, which evolved into the Federal Aviation Administration, or FAA we know today. Aircrafts were refined and made more functional. World War I brought with it the age of aerial combat. Plenty of supply after the war and then the growth of manufacturers Brough further innovation to compete with one another, and commercial aircraft and industrial uses (like cropdusting) enabled more investment into R&D In 1926, the first flying boat service was inaugurated from New York to Argentina. Another significant development in aviation was in the 1930s when the jet engine was invented. This invention was done by Frank Whittle who registered a turbojet engine patent. A jet plane was also developed by Hans von Ohain and was called the Heinkel He 178 (Grant, 2017). The plane first flew in 1939, but the Whittle jet engine is the ancestor of those found in planes in World War II and beyond. And from there to the monster airliners and stealth fighters or X-15 becomes a much larger story. The aerospace industry continued to innovate both in the skies and into space. The history of flight entered another phase in the Cold War. Rand corporation developed the concept of Intercontinental Ballistic Missiles (or ICBMs) and the Soviet Union launched the first satellite into space in 1957. Then in 1969, Neil Armstrong and Buzz Aldrin made the first landing on the moon and we continued to launch into space throughout the 1970s to 1990s, before opening up space travel to private industry. Those projects got bigger and bigger and bigger. But generations of enthusiasts and engineers were inspired by devices far smaller, and without pilots in the device.
/episode/index/show/thehistoryofcomputing/id/26339883
info_outline
SABRE and the Travel Global Distribution System
03/16/2023
SABRE and the Travel Global Distribution System
Computing has totally changed how people buy and experience travel. That process seemed to start with sites that made it easy to book travel, but as with most things we experience in our modern lives, it actually began far sooner and moved down-market as generations of computing led to more consumer options for desktops, the internet, and the convergence of these technologies. Systems like SABRE did the original work to re-think travel - to take logic and rules out of the heads of booking and travel agents and put them into a digital medium. In so doing, they paved the way for future generations of technology and to this day retain a valuation of over $2 billion. SABRE is short for Semi-Automated Business Research Environment. It’s used to manage over a third of global travel, to the tune of over a quarter trillion US dollars a year. It’s used by travel agencies and travel services to reserve car rentals, flights, hotel rooms, and tours. Since Sabre was released services like Amadeus and Travelport were created to give the world a Global Distribution System, or GDS. Passenger air travel began when airlines ferrying passengers cropped up in 1914 but the big companies began in the 1920s, with KLM in 1919, Finnair in 1923, Delta in 1925, American Airlines and Ryan Air in 1926, Pan American in 1927, and the list goes on. They grew quickly and by 1926 the Air Commerce Act led to a new department in the government called Air Commerce, which evolved into the FAA, or Federal Aviation Administration in the US. And each country, given the possible dangers these aircraft posed as they got bigger and loaded with more and more fuel, also had their own such departments. The aviation industry blossomed in the roaring 20s as people traveled and found romance and vacation. At the time, most airlines were somewhat regional and people found travel agents to help them along their journey to book travel, lodgings, and often food. The travel agent naturally took over air travel much as they’d handled sea travel before. But there were dangers in traveling in those years between the two World Wars. Nazis rising to power in Germany, Mussolini in Italy, communist cleansings in Russia and China. Yet, a trip to the Great Pyramid of Giza could now be a week instead of months. Following World War II, there was a fracture in the world between Eastern and Western powers, or those who aligned with the former British empire and those who aligned with the former Russian empire, now known as the Soviet Union. Travel within the West exploded as those areas were usually safe and often happy to accept the US dollar. Commercial air travel boomed not just for the wealthy, but for all. People had their own phones now, and could look up a phone number in a phone book and call a travel agent. The travel agents then spent hours trying to build the right travel package. That meant time on the phone with hotels and time on the phone with airlines. Airlines like American head. To hire larger and larger call centers of humans to help find flights. We didn’t just read about Paris, we wanted to go. Wars had connected the world and now people wanted to visit the places they’d previously just seen in art books or read about in history books. But those call centers grew. A company like American Airlines couldn’t handle all of its ticketing needs and the story goes that the CEO was sitting beside a seller from IBM when they came up with the idea of a computerized reservation system. And so SABRE was born in the 1950s, when American Airlines agreed to develop a real-time computing platform. Here, we see people calling in and pressing buttons to run commands on computers. The tones weren’t that different than a punch card, really. The system worked well enough for American that they decided to sell access to other firms. The computers used were based loosely after the IBM mainframes used in the SAGE missile air defense system. Here we see the commercial impacts of the AN/FSQ-7 the US government hired IBM to build as IBM added the transistorized options to the IBM 704 mainframe in 1955. That gave IBM the interactive computing technology that evolved into the 7000 series mainframes. Now that IBM had the interactive technology, and a thorough study had been done to evaluate the costs and impacts of a new reservation system, American and IBM signed a contract to build the system in 1957. They went live to test reservation booking shortly thereafter. But it turns out there was a much bigger opportunity here. See, American and other airlines had paper processes to track how many people were on a flight and quickly find open seats for passengers, but it could take an hour or three to book tickets. This was fairly common before software ate the world. Everything from standing in line at the bank, booking dinner at a restaurant, reserving a rental car, booking hotel rooms, and the list goes on. There were a lot of manual processes in the world - people weren’t just going to punch holes in a card to program their own flight and wait for some drum storage to tell them if there was an available seat. That was the plan American initially had in 1952 with the Magnetronic Reservisor. That never worked out. American had grown to one of the largest airlines and knew the perils and costs of developing software and hardware like this. Their system cost $40 million in 1950s money to build with IBM. They also knew that as other airlines grew to accommodate more people flying around the world, that the more flights, the longer that hour or three took. So they should of course sell the solution they built to other airlines. Thus, parlaying the SAGE name, famous as a Cold War shield against the nuclear winter, Sabre Corporation began. It was fairly simple at first, with a pair of IBM 7090 mainframes that could take over 80,000 calls a day in 1960. Some travel agents weren’t fans of the new system, but those who embraced it found they could get more done in less time. Sabre sold reservation systems to airlines and soon expanded to become the largest data-processor in the world. Far better than the Reservisor would have been and now able to help bring the whole world into the age of jet airplane travel. That exploded to thousands of flights an hour in the 1960s and even turned over all booking to the computer. The system got busy and over the years IBM upgraded the computers to the S/360. They also began to lease systems to travel agencies in the 1970s after Max Hopper joined the company and began the plan to open up the platform as TWA had done with their PARS system. Then they went international, opened service bureaus in other cities (given that we once had to pay for a toll charge to call a number). And by the 1980s Sabre was how the travel agents booked flights. The 1980s brought easysabjre, so people could use their own computers to book flights and by then - and through to the modern era, a little over a third of all reservations are made on Sabre. By the mid-1980s, United had their own system called Apollo, Delta had one called Datas, and other airlines had their own as well. But SABRE could be made to be airline neutral. IBM had been involved with many American competitors, developing Deltamatic for Delta, PANAMAC for Pan Am, and other systems. But SABRE could be hooked to thee new online services for a whole new way to connect systems. One of these was CompuServe in 1980, then Prodigy’s GEnie and AOL as we turned the corner into the 1990s. Then they started a site called Travelocity in 1996 which was later sold to Expedia. In the meantime, they got serious competition, which eventually led to a slew of acquisitions to remain compeititve. The competition included Amadeus, Galileo International, and Worldspan on provider in the Travelport GDS. The first of them originated from United Airlines, and by 1987 was joined by Aer Lingus, Air Portugal, Alitalia, British Airways, KLM, Olympic, Sabena, and Swissair to create Galileo, which was then merged with the Apollo reservation system. The technology was acquired through a company called Videcom International, which initially started developing reservation software in 1972, shortly after the Apollo and Datas services went online. They focused on travel agents and branched out into reservation systems of all sorts in the 1980s. As other systems arose they provided an aggregation to them by connecting to Amadeus, Galileo, and Worldspan. Amadeus was created in 1987 to be a neutral GDS after the issues with Sabre directing reservations to American Airlines. That was through a consortium of Air France, Iberia, Lufthansa, and SAS. They acquired the assets of the bankrupt System One and they eventually added other travel options including hotels, cars rentals, travel insurance, and other amenities. They went public in 1999 just before Sabre did and then were also taken private just before Sabre was. Worldspan was created in 1990 and the result of merging or interconnecting the systems of Delta, Northwest Airlines, and TWA, which was then acquired by Travelport in 2007. By then, SABRE had their own programming languages. While the original Sabre languages were written in assembly, they wrote their own language on top of C and C++ called SabreTalk and later transitioned to standard REST endpoints. They also weren’t a part of American any longer. There were too many problems with manipulating how flights were displayed to benefit American Airlines and they had to make a clean cut. Especially after Congress got involved in the 1980s and outlawed that type of bias for screen placement. Now that they were a standalone company, Sabre went public then was taken private by private equity firms in 2007, and relisted on NASDAQ in 2014. Meanwhile, travel aggregators had figured out they could hook into the GDS systems and sell discount airfare without a percentage going to travel agents. Now that the GDS systems weren’t a part of the airlines, they were able to put downward pressure on prices. Hotwire, which used Sabre and a couple of other systems, and TripAdvisor, which booked travel through Sabre and Amadeus, were created in 2000 and Microsoft launched Expedia in 1996, which had done well enough to get spun off into its own public company by 2000. Travelocity operated inside Sabre until sold, and so the airlines put together a site of their own that they called Orbitz, which in 2001 was the biggest e-commerce site to have ever launched. And out of the bursting of the dot com bubble came online travel bookings. Kayak came in 2004 Sabre later sold Travelocity to Expedia, which uses Sabre to book travel. That allowed Sabre to focus on providing the back end travel technology. They now do over $4 billion in revenue in their industry. American Express had handled travel for decades but also added flights and hotels to their site, integrating with Sabre and Amadeus as well. Here, we see a classic paradigm in play. First the airlines moved their travel bookings from paper filing systems to isolated computer systems - what we’d call mainframes today. The airlines then rethink the paradigm and aggregate other information into a single system, or a system intermixed with other data. In short, they enriched the data. Then we expose those as APIs to further remove human labor and put systems on assembly lines. Sites hook into those and the GDS systems, as with many aggregators, get spun off into their own companies. The aggregated information then benefits consumers (in this case travelers) with more options and cheaper fares. This helps counteract the centralization of the market where airlines acquire other airlines but in some way also cheapen the experience. Gone are the days when a travel agent guides us through our budgets and helps us build a killer itinerary. But in a way that just makes travel much more adventurous.
/episode/index/show/thehistoryofcomputing/id/26250987
info_outline
The Story of Intel
03/07/2023
The Story of Intel
We’ve talked about the history of microchips, transistors, and other chip makers. Today we’re going to talk about Intel in a little more detail. Intel is short for Integrated Electronics. They were founded in 1968 by Robert Noyce and Gordon Moore. Noyce was an Iowa kid who went off to MIT to get a PhD in physics in 1953. He went off to join the Shockley Semiconductor Lab to join up with William Shockley who’d developed the transistor as a means of bringing a solid-state alternative to vacuum tubes in computers and amplifiers. Shockley became erratic after he won the Nobel Prize and 8 of the researchers left, now known as the “traitorous eight.” Between them came over 60 companies, including Intel - but first they went on to create a new company called Fairchild Semiconductor where Noyce invented the monolithic integrated circuit in 1959, or a single chip that contains multiple transistors. After 10 years at Fairchild, Noyce joined up with coworker and fellow traitor Gordon Moore. Moore had gotten his PhD in chemistry from Caltech and had made an observation while at Fairchild that the number of transistors, resistors, diodes, or capacitors in an integrated circuit was doubling every year and so coined Moore’s Law, that it would continue to to do so. They wanted to make semiconductor memory cheaper and more practical. They needed money to continue their research. Arthur Rock had helped them find a home at Fairchild when they left Shockley and helped them raise $2.5 million in backing in a couple of days. The first day of the company, Andy Grove joined them from Fairchild. He’d fled the Hungarian revolution in the 50s and gotten a PhD in chemical engineering at the University of California, Berkeley. Then came Leslie Vadász, another Hungarian emigrant. Funding and money coming in from sales allowed them to hire some of the best in the business. People like Ted Hoff , Federico Faggin, and Stan Mazor. That first year they released 64-bit static random-access memory in the 3101 chip, doubling what was on the market as well as the 3301 read-only memory chip, and the 1101. Then DRAM, or dynamic random-access memory in the 1103 in 1970, which became the bestselling chip within the first couple of years. Armed with a lineup of chips and an explosion of companies that wanted to buy the chips, they went public within 2 years of being founded. 1971 saw Dov Frohman develop erasable programmable read-only memory, or EPROM, while working on a different problem. This meant they could reprogram chips using ultraviolet light and electricity. In 1971 they also created the Intel 4004 chip, which was started in 1969 when a calculator manufacturer out of Japan ask them to develop 12 different chips. Instead they made one that could do all of the tasks of the 12, outperforming the ENIAC from 1946 and so the era of the microprocessor was born. And instead of taking up a basement at a university lab, it took up an eight of an inch by a sixth of an inch to hold a whopping 2,300 transistors. The chip didn’t contribute a ton to the bottom line of the company, but they’d built the first true microprocessor, which would eventually be what they were known for. Instead they were making DRAM chips. But then came the 8008 in 1972, ushering in an 8-bit CPU. The memory chips were being used by other companies developing their own processors but they knew how and the Computer Terminal Corporation was looking to develop what was a trend for a hot minute, called programmable terminals. And given the doubling of speeds those gave way to microcomputers within just a few years. The Intel 8080 was a 2 MHz chip that became the basis of the Altair 8800, SOL-20, and IMSAI 8080. By then Motorola, Zilog, and MOS Technology were hot on their heals releasing the Z80 and 6802 processors. But Gary Kildall wrote CP/M, one of the first operating systems, initially for the 8080 prior to porting it to other chips. Sales had been good and Intel had been growing. By 1979 they saw the future was in chips and opened a new office in Haifa, Israiel, where they designed the 8088, which clocked in at 4.77 MHz. IBM chose this chip to be used in the original IBM Personal Computer. IBM was going to use an 8-bit chip, but the team at Microsoft talked them into going with the 16-bit 8088 and thus created the foundation of what would become the Wintel or Intel architecture, or x86, which would dominate the personal computer market for the next 40 years. One reason IBM trusted Intel is that they had proven to be innovators. They had effectively invented the integrated circuit, then the microprocessor, then coined Moore’s Law, and by 1980 had built a 15,000 person company capable of shipping product in large quantities. They were intentional about culture, looking for openness, distributed decision making, and trading off bureaucracy for figuring out cool stuff. That IBM decision to use that Intel chip is one of the most impactful in the entire history of personal computers. Based on Microsoft DOS and then Windows being able to run on the architecture, nearly every laptop and desktop would run on that original 8088/86 architecture. Based on the standards, Intel and Microsoft would both market that their products ran not only on those IBM PCs but also on any PC using the same architecture and so IBM’s hold on the computing world would slowly wither. On the back of all these chips, revenue shot past $1 billion for the first time in 1983. IBM bought 12 percent of the company in 1982 and thus gave them the Big Blue seal of approval, something important event today. And the hits kept on coming with the 286 to 486 chips coming along during the 1980s. Intel brought the 80286 to market and it was used in the IBM PC AT in 1984. This new chip brought new ways to manage addresses, the first that could do memory management, and the first Intel chip where we saw protected mode so we could get virtual memory and multi-tasking. All of this was made possible with over a hundred thousand transistors. At the time the original Mac used a Motorola 68000 but the sales were sluggish while they flourished at IBM and slowly we saw the rise of the companies cloning the IBM architecture, like Compaq. Still using those Intel chips. Jerry Sanders had actually left Fairchild a little before Noyce and Moore to found AMD and ended up cloning the instructions in the 80286, after entering into a technology exchange agreement with Intel. This led to AMD making the chips at volume and selling them on the open market. AMD would go on to fast-follow Intel for decades. The 80386 would go on to simply be known as the Intel 386, with over 275,000 transistors. It was launched in 1985, but we didn’t see a lot of companies use them until the early 1990s. The 486 came in 1989. Now we were up to a million transistors as well as a math coprocessor. We were 50 times faster than the 4004 that had come out less than 20 years earlier. I don’t want to take anything away from the phenomenal run of research and development at Intel during this time but the chips and cores and amazing developments were on autopilot. The 80s also saw them invest half a billion in reinvigorating their manufacturing plants. With quality manufacturing allowing for a new era of printing chips, the 90s were just as good to Intel. I like to think of this as the Pentium decade with the first Pentium in 1993. 32-bit here we come. Revenues jumped 50 percent that year closing in on $9 billion. Intel had been running an advertising campaign around Intel Inside. This represented a shift from the IBM PC to the Intel. The Pentium Pro came in 1995 and we’d crossed 5 million transistors in each chip. And the brand equity was rising fast. More importantly, so was revenue. 1996 saw revenues pass $20 billion. The personal computer was showing up in homes and on desks across the world and most had Intel Inside - in fact we’d gone from Intel inside to Pentium Inside. 1997 brought us the Pentium II with over 7 million transistors, the Xeon came in 1998 for servers, and 1999 Pentium III. By 2000 they introduced the first gigahertz processor at Intel and they announced the next generation after Pentium: Itanium, finally moving the world to the 64 bit processor. As processor speeds slowed they were able to bring multi-core processors and massive parallelism out of the hallowed halls of research and to the desktop computer in 2005. 2006 saw Intel go from just Windows to the Mac. And we got 45 nanometer logic technology in 2006 using hafnium-based high-k for transistor gates represented a shift from the silicon-gated transistors of the 60s and allowed them to move to hundreds of millions of transistors packed into a single chip. i3, i5, i7, an on. The chips now have over a couple hundred million transistors per core with 8 cores on a chip potentially putting us over 1.7 or 1.8 transistors per chip. Microsoft, IBM, Apple, and so many others went through huge growth and sales jumps then retreated dealing with how to run a company of the size they suddenly became. This led each to invest heavily into ending a lost decade effectively with R&D - like when IBM built the S/360 or Apple developed the iMac and then iPod. Intel’s strategy had been research and development. Build amazing products and they sold. Bigger, faster, better. The focus had been on power. But mobile devices were starting to take the market by storm. And the ARM chip was more popular on those because with a reduced set of instructions they could use less power and be a bit more versatile. Intel coined Moore’s Law. They know that if they don’t find ways to pack more and more transistors into smaller and smaller spaces then someone else will. And while they haven’t been huge in the RISC-based System on a Chip space, they do continue to release new products and look for the right product-market fit. Just like they did when they went from more DRAM and SRAM to producing the types of chips that made them into a powerhouse. And on the back of a steadily rising revenue stream that’s now over $77 billion they seem poised to be able to whether any storm. Not only on the back of R&D but also some of the best manufacturing in the industry. Chips today are so powerful and small and contain the whole computer from the era of those Pentiums. Just as that 4004 chip contained a whole ENIAC. This gives us a nearly limitless canvas to design software. Machine learning on a SoC expands the reach of what that software can process. Technology is moving so fast in part because of the amazing work done at places like Intel, AMD, and ARM. Maybe that positronic brain that Asimov promised us isn’t as far off as it seems. But then, I thought that in the 90s as well so I guess we’ll see.
/episode/index/show/thehistoryofcomputing/id/26156919
info_outline
AI Hype Cycles And Winters On The Way To ChatGPT
02/22/2023
AI Hype Cycles And Winters On The Way To ChatGPT
Carlota Perez is a researcher who has studied hype cycles for much of her career. She’s affiliated with the University College London, the University of Sussex, The Tallinn University of Technology in Astonia and has worked with some influential organizations around technology and innovation. As a neo-Schumpeterian, she sees technology as a cornerstone of innovation. Her book Technological Revolutions and Financial Capital is a must-read for anyone who works in an industry that includes any of those four words, including revolutionaries. Connecticut-based Gartner Research was founded by GideonGartner in 1979. He emigrated to the United States from Tel Aviv at three years old in 1938 and graduated in the 1956 class from MIT, where he got his Master’s at the Sloan School of Management. He went on to work at the software company System Development Corporation (SDC), the US military defense industry, and IBM over the next 13 years before starting his first company. After that failed, he moved into analysis work and quickly became known as a top mind in the technology industry analysts. He often bucked the trends to pick winners and made banks, funds, and investors lots of money. He was able to parlay that into founding the Gartner Group in 1979. Gartner hired senior people in different industry segments to aid in competitive intelligence, industry research, and of course, to help Wall Street. They wrote reports on industries, dove deeply into new technologies, and got to understand what we now call hype cycles in the ensuing decades. They now boast a few billion dollars in revenue per year and serve well over 10,000 customers in more than 100 countries. Gartner has developed a number of tools to make it easier to take in the types of analysis they create. One is a Magic Quadrant, reports that identify leaders in categories of companies by a vision (or a completeness of vision to be more specific) and the ability to execute, which includes things like go-to-market activities, support, etc. They lump companies into a standard four-box as Leaders, Challengers, Visionaries, and Niche Players. There’s certainly an observer effect and those they put in the top right of their four box often enjoy added growth as companies want to be with the most visionary and best when picking a tool. Another of Gartner’s graphical design patterns to display technology advances is what they call the “hype cycle”. The hype cycle simplifies research from career academics like Perez into five phases. * The first is the technology trigger, which is when a breakthrough is found and PoCs, or proof-of-concepts begin to emerge in the world that get press interested in the new technology. Sometimes the new technology isn’t even usable, but shows promise. * The second is the Peak of Inflated Expectations, when the press picks up the story and companies are born, capital invested, and a large number of projects around the new techology fail. * The third is the Trough of Disillusionment, where interest falls off after those failures. Some companies suceeded and can show real productivity, and they continue to get investment. * The fourth is the Slope of Enlightenment, where the go-to-market activities of the surviving companies (or even a new generation) begin to have real productivity gains. Every company or IT department now runs a pilot and expectations are lower, but now achievable. * The fifth is the Plateau of Productivity, when those pilots become deployments and purchase orders. The mainstream industries embrace the new technology and case studies prove the promised productivity increases. Provided there’s enough market, companies now find success. There are issues with the hype cycle. Not all technologies will follow the cycle. The Gartner approach focuses on financials and productivity rather than true adoption. It involves a lot of guesswork around subjective, synthetical, and often unsystematic research. There’s also the ever-resent observer effect. However, more often than not, the hype is seperated from the tech that can give organizations (and sometimes all of humanity) real productivity gains. Further, the term cycle denotes a series of events when it should in fact be cyclical as out of the end of the fifth phase a new cycle is born, or even a set of cycles if industries grow enough to diverge. ChatGPT is all over the news feeds these days, igniting yet another cycle in the cycles of AI hype that have been prevalent since the 1950s. The concept of computer intelligence dates back to the 1942 with Alan Turing and Isaac Asimov with “Runaround” where the three laws of robotics initially emerged from. By 1952 computers could play themselves in checkers and by 1955, Arthur Samuel had written a heuristic learning algorthm he called “temporal-difference learning” to play Chess. Academics around the world worked on similar projects and by 1956 John McCarthy introduced the term “artificial intelligence” when he gathered some of the top minds in the field together for the McCarthy workshop. They tinkered and a generation of researchers began to join them. By 1964, Joseph Weizenbaum’s "ELIZA" debuted. ELIZA was a computer program that used early forms of natural language processing to run what they called a “DOCTOR” script that acted as a psychotherapist. ELIZA was one of a few technologies that triggered the media to pick up AI in the second stage of the hype cycle. Others came into the industry and expectations soared, now predictably followed by dilsillusionment. Weizenbaum wrote a book called Computer Power and Human Reason: From Judgment to Calculation in 1976, in response to the critiques and some of the early successes were able to then go to wider markets as the fourth phase of the hype cycle began. ELIZA was seen by people who worked on similar software, including some games, for Apple, Atari, and Commodore. Still, in the aftermath of ELIZA, the machine translation movement in AI had failed in the eyes of those who funded the attempts because going further required more than some fancy case statements. Another similar movement called connectionism, or mostly node-based artificial neural networks is widely seen as the impetus to deep learning. David Hunter Hubel and Torsten Nils Wiesel focused on the idea of convultional neural networks in human vision, which culminated in a 1968 paper called "Receptive fields and functional architecture of monkey striate cortex.” That built on the original deep learning paper from Frank Rosenblatt of Cornell University called "Principles of Neurodynamics: Perceptrons and the Theory of Brain Mechanisms" in 1962 and work done behind the iron curtain by Alexey Ivakhnenko on learning algorithms in 1967. After early successes, though, connectionism - which when paired with machine learning would be called deep learning when Rina Dechter coined the term in 1986, went through a similar trough of disillusionment that kicked off in 1970. Funding for these projects shot up after the early successes and petered out ofter there wasn’t much to show for them. Some had so much promise that former presidents can be seen in old photographs going through the models with the statiticians who were moving into computing. But organizations like DARPA would pull back funding, as seen with their speech recognition projects with Cargegie Mellon University in the early 1970s. These hype cycles weren’t just seen in the United States. The British applied mathemetician James Lighthill wrote a report for the British Science Research Council, which was published in 1973. The paper was called “Artificial Intelligence: A General Survey” and analyzed the progress made based on the amount of money spent on artificial intelligence programs. He found none of the research had resulted in any “major impact” in fields that the academics had undertaken. Much of the work had been done at the University of Edinbourgh and funding was drastically cut, based on his findings, for AI research around the UK. Turing, Von Neumann, McCarthy, and others had either intentially or not, set an expectation that became a check the academic research community just couldn’t cash. For example, the New York Times claimed Rosenblatt’s perceptron would let the US Navy build computers that could “walk, talk, see, write, reproduce itself, and be conscious of its existence” in the 1950s - a goal not likely to be achieved in the near future even seventy years later. Funding was cut in the US, the UK, and even in the USSR, or Union of the Soviet Socialist Republic. Yet many persisted. Languages like Lisp had become common in the late 1970s, after engineers like Richard Greenblatt helped to make McCarthy’s ideas for computer languages a reality. The MIT AI Lab developed a Lisp Machine Project and as AI work was picked up at other schools like Stanford began to look for ways to buy commercially built computers ideal to be Lisp Machines. After the post-war spending, the idea that AI could become a more commercial endeavor was attractive to many. But after plenty of hype, the Lisp machine market never materialized. The next hype cycle had begun in 1983 when the US Department of Defense pumped a billion dollars into AI, but that spending was cancelled in 1987, just after the collapse of the Lisp machine market. Another AI winter was about to begin. Another trend that began in the 1950s but picked up steam in the 1980s was expert systems. These attempt to emulate the ways that humans make decisions. Some of this work came out of the Stanford Heuristic Programming Project, pioneered by Edward Feigenbaum. Some commercial companies took the mantle and after running into barriers with CPUs, by the 1980s those got fast enough. There were inflated expectations after great papers like Richard Karp’s “Reducibility among Combinatorial Problems” out of UC Berkeley in 1972. Countries like Japan dumped hundreds of millions of dollars (or yen) into projects like “Fifth Generation Computer Systems” in 1982, a 10 year project to build up massively parallel computing systems. IBM spent around the same amount on their own projects. However, while these types of projects helped to improve computing, they didn’t live up to the expectations and by the early 1990s funding was cut following commercial failures. By the mid-2000s, some of the researchers in AI began to use new terms, after generations of artificial intelligence projects led to subsequent AI winters. Yet research continued on, with varying degrees of funding. Organizations like DARPA began to use challenges rather than funding large projects in some cases. Over time, successes were found yet again. Google Translate, Google Image Search, IBM’s Watson, AWS options for AI/ML, home voice assistants, and various machine learning projects in the open source world led to the start of yet another AI spring in the early 2010s. New chips have built-in machine learning cores and programming languages have frameworks and new technologies like Jupyter notebooks to help organize and train data sets. By 2006, academic works and open source projects had hit a turning point, this time quietly. The Association of Computer Linguistics was founded in 1962, initially as the Association for Machine Translation and Computational Linguistics (AMTCL). As with the ACM, they have a number of special interest groups that include natural language learning, machine translation, typology, natural language generation, and the list goes on. The 2006 proceedings on the Workshop of Statistical Machine Translation began a series of dozens of workshops attended by hundreds of papers and presenters. The academic work was then able to be consumed by all, inlcuding contributions to achieve English-to-German and Frnech tasks from 2014. Deep learning models spread and become more accessible - democratic if you will. RNNs, CNNs, DNNs, GANs. Training data sets was still one of the most human intensive and slow aspects of machine learning. GANs, or Generative Adversarial Networks were one of those machine learning frameworks, initially designed by Ian Goodfellow and others in 2014. GANs use zero-sum game techniques from game theory to generate new data sets - a genrative model. This allowed for more unsupervised training of data. Now it was possible to get further, faster with AI. This brings us into the current hype cycle. ChatGPT was launched in November of 2022 by OpenAI. OpenAI was founded as a non-profit in 2015 by Sam Altman (former cofounder of location-based social network app Loopt and former president of Y Combinator) and a cast of veritable all-stars in the startup world that included: * Reid Hoffman, former Paypal COO, LinkedIn founder and venture capitalist. * Peter Thiel, former cofounder of Paypal and Palantir, as well as one of the top investors in Silicon Valley. * Jessica Livingston, founding partner at Y Combinator. * Greg Brockman, an AI researcher who had worked on projects at MIT and Harvard OpenAI spent the next few years as a non-profit and worked on GPT, or Generative Pre-trained Transformer autoregression models. GPT uses deep learning models to process human text and produce text that’s more human than previous models. Not only is it capable of natural language processing but the generative pre-training of models has allowed it to take a lot of unlabeled text so people don’t have to hand label weights, thus automated fine tuning of results. OpenAI dumped millions into public betas by 2016 and were ready to build products to take to market by 2019. That’s when they switched from a non-profit to a for-profit. Microsoft pumped $1 billion into the company and they released DALL-E to produce generative images, which helped lead to a new generation of applications that could produce artwork on the fly. Then they released ChatGPT towards the end of 2022, which led to more media coverage and prognostication of world-changing technological breakthrough than most other hype cycles for any industry in recent memory. This, with GPT-4 to be released later in 2023. ChatGPT is most interesting through the lens of the hype cycle. There have been plenty of peaks and plateaus and valleys in artificial intelligence over the last 7+ decades. Most have been hyped up in the hallowed halls of academia and defense research. ChatGPT has hit mainstream media. The AI winter following each seems to be based on the reach of audience and depth of expectations. Science fiction continues to conflate expectations. Early prototypes that make it seem as though science fiction will be in our hands in a matter of weeks lead media to conjecture. The reckoning could be substantial. Meanwhile, projects like TinyML - with smaller potential impacts for each use but wider use cases, could become the real benefit to humanity beyond research, when it comes to everyday productivity gains. The moral of this story is as old as time. Control expectations. Undersell and overdeliver. That doesn’t lead to massive valuations pumped up by hype cycles. Many CEOs and CFOs know that a jump in profits doesn’t always mean the increase will continue. Some intentially slow expectations in their quarterly reports and calls with analysts. Those are the smart ones.
/episode/index/show/thehistoryofcomputing/id/26015742
info_outline
Hackers and Chinese Food: Origins of a Love Affair
12/30/2022
Hackers and Chinese Food: Origins of a Love Affair
Research into the history of computers sometimes leads down some interesting alleys - or wormholes even. My family would always go out to eat Chinese food, or pick it up, on New Year’s day. None of the one Chinese restaurants in the area actually closed, so it just made sense. The Christmas leftovers were gone by then and no one really wanted to cook. My dad mentioned there were no Chinese restaurants in our area in the 1970s - so it was a relatively new entrant to the cuisine of my North Georgia town. Whether it’s the Tech Model Railroad or hobbyists from Cambridge, stories abound of young engineers debating the merits of this programming technique or chipset or that. So much so that while reading Steven Levy’s Hackers or Tom Lean’s Electronic Dreams, I couldn’t help but hop on Door Dash and order up some yummy fried rice. Then I started to wonder, why this obsession? For one, many of these hackers didn’t have a ton of money. Chinese food was quick and cheap. The restaurants were often family-owned and small. There were higher end restaurants but concepts like P.F. Chang’s hadn’t sprung up yet. That wouldn’t come until 1993. Another reason it was cheap is that many of the proprietors of the restaurants were recent immigrants. Some were from Hunan, others from Taipei or Sichuan, Shanghai, or Peking (the Romanized name for Beijing). Chinese immigrants began to flow into the United States during the Gold Rush of California in the late 1840s and early 1850s. The Qing Empire had been at its height at the end of the 1700s and China ruled over a third of humans in the world. Not only that - it was one of the top economies in the world. But rapid growth in population meant less farmland for everyone - less jobs to go around. Poverty spread, just as colonial powers began to pick away at parts of the empire. Britain had banned the slave trade in 1807 and Chinese laborers had been used to replace the slaves. The use of opium spread throughout the colonies and with the laborers, back into China. The Chinese tried to ban the opium trade and seized opium in Canton. The British had better ships, better guns, and when the First Opium War broke out, China was forced to give up Hong Kong to the British in 1842, which began what some historians refer to as a century of humiliation while China gave up land until they were able to modernize. Hong Kong became a British colony under Queen Victoria and the Victorian obsession with China grew. Art, silks (as with the Romans), vases, and anything the British could get their hands on flowed through Hong Kong. Then came the Taiping Rebellion, which lasted from 1851 to 1864. A Christian was named theocrat and China was forced to wage a war internally with around 20 million people dying and scores more being displaced. The scent of an empire in decay was in the air. Set against a backdrop of more rebellions, the Chinese army was weakened to the point that during the First Sino-Japanese War in 1894, and more intervention from colonial powers. By 1900, the anti-colonial and anti-Christian Boxer Uprising saw missionaries slaughtered and foreigners expelled. Great powers of the day sent ships and troops to retrieve their peoples and soon declared war on the empire and seized Beijing. This was all expensive, led to reparations, a prohibition on importing arms, razing of forts, and more foreign powers occupying areas of China. The United States put over $10 million of its take from the Boxer Indemnity as they called it, to help support Chinese students who came to the United States. The Qing court had lost control and by 1911 the Wuchang Uprising began and by 1912 2,000 years of Chinese dynasties was over with the Republic of China founded in 1912, and internal conflicts for power continuing until Mao Zedong and his followers finally seized power, established the People’s Republic of China as a communist nation, and cleansed the country of detractors during what they called the Great Leap Forward, resulting in 45 million dead. China itself was diplomatically disconnected with the United States at the time, who had backed the government now in exile in the capital city of Taiwan, Taipei - or the Republic of China as they were called during the Civil War. The food, though. Chinese food began to come into the United States during the Gold Rush. Cantonese merchants flowed into the sparkling bay of San Francisco, and emigrants could find jobs mining, laying railroad tracks, and in agriculture. Hard work means you get real hungry, and they cooked food like they had at home. China had a better restaurant and open market cooking industry than the US at the time (and arguably still does). Some of he Chinese who settled in San Francisco started restaurants - many better than those run by Americans. The first known restaurant owned by a Chinese proprietor was Canton Restaurant in 1849. As San Francisco grew, so grew the Chinese food industry. Every group of immigrants faces xenophobia or racism. The use of the Chinese laborers had led to laws in England that attempted to limit their use. In some cases they were subjugated into labor. The Chinese immigrants came into the California Gold Rush and many stayed. More restaurants were opened and some catered to white people more than the Chinese. The Transcontinental Railroad was completed in 1869 and tourists began to visit San Francisco from the east. China Towns began to spring up in other major cities across the United States. Restaurants, laundries, and other even eastern pharmacies. New people bring new ways and economies go up and down. Prejudice reared its ugly head. There was an economic recession in the 1870s. There were fears that the Chinese were taking jobs, causing wages to go down, and crime. Anti-Chinese sentiment became law in the Chinese Exclusion Act in 1882, which halted immigration into the US. That would be repealed in 1943. Conservative approaches to immigration did nothing to limit the growing appeal of Chinese food in the United States. Merchants, like those who owned Chinese restaurants, could get special visas. They could bring relatives and workers. Early Chinese restaurants had been called “chow chow houses” and by the early 1900s there were new Chop Suey restaurants in big cities, that were affordable. Chop Suey basically means “odds and ends” and most of the dishes were heavily westernized but still interesting and delicious. The food was fried in ways it hadn’t been in China, and sweeter. Ideas from other asian nations also began to come in, like fortune cookies, initially from Japan. Americans began to return home from World War II in the late 1940s. Many had experienced new culinary traditions in lands they visited. Initially Cantonese-inspired, more people flowed in from other parts of China like Taiwan and they brought food inspired from their native lands. Areas like New York and San Francisco got higher end restaurants. Once the Chinese Exclusion Act was repealed, plenty of immigrants fled wars and cleansing in China. Meanwhile, Americans embraced access to different types of foods - like Italian, Chinese, and fast food. Food became a part of the national identity. Further, new ways to preserve food became possible as people got freezers and canneries helped spread foods - like pasta sauce. This was the era of the spread of Spam and other types of early processed foods. The military helped spread the practice - as did Jen Paulucci, who bought Chun King Corporation in 1947. The Great Depression proved there needed to be new ways to distribute foods. Some capitalized on that. 4,000+ Chinese restaurants in the US in the 1940s meant there were plenty of companies to buy those goods rather than make them fresh. Chop Suey, possibly created by the early Chinese migrants. A new influx of immigrants would have new opportunities to diversify the American pallate. The 1960s saw an increase in legislation to protect human rights. Amidst the civil rights movement, the Hart-Celler Act of 1965 stopped the long-standing practice of controlling immigration effectively by color. The post-war years saw shifting borders and wars throughout the world - especially in Eastern Europe and Asia. The Marshal Plan helped rebuild the parts of Asia that weren’t communist, and opened the ability for more diverse people to move to the US. Many that we’ve covered went into computing and helped develop a number of aspects of computing. They didn’t just come from China - they came from Russia, Poland, India, Japan, Korea, Vietnam, Thailand, and throughout. Their food came with them. This is the world the Hackers that Steven Levy described lived in. The first Chinese restaurant opened in London in 1907 and as well when people who lived in Hong Kong moved to the UK, especially after World War II. That number of Chinese restaurants in the US grew to tens of thousands in the decades since Richard Nixon visited Beijing in 1972 to open relations back up with China. But the impact at the time was substantial, even on technologists. It wasn’t just those hackers from MIT that loved their Chinese food, but those in Cambridge as well in the 1980s, who partook in a more Americanized Chinese cuisine, like “Chow mein” - which loosely translates from “fried noodles” and emerged in the US in the early 1900s. Not all dishes have such simple origins to track down. Egg rolls emerged in the 1930s, a twist on the more traditional Chinese sprint roll. Ding Baozhen, a governor of the Sichuan province in the Qing Dynasty, discovered a spicy marinated chicken dish in the mid-1800s that spread quickly. He was the Palace Guardian, or Kung Pao, as the dish is still known. Zuo Zongtang, better known as General Tso, was a Qing Dynasty statesman and military commander who helped put down the Taiping Rebellion in the later half of the 1800s. Chef Peng Chang-kuei escaped communist China to Taiwan, where he developed General Tso’s chicken and named it after the war hero. It came to New York in the 1970s. Sweet and Sour pork also got its start in the Qing era, in 18th century Cantonese cuisine and spread to the US with the Gold Rush. Some dishes are far older. Steamed dumplings were popular from Afghanistan to Japan and go back to the Han Dynasty - possibly invented by the Chinese doctor Zhang Zhongjing in the centuries before or after the turn of the millennia. Peking duck is far older, getting its start in 1300s Ming Dynasty, or Yuan - but close to Shanghai. Otto Reichardt brought the ducks to San Francisco to be served in restaurants in 1901. Chinese diplomats helped popularize the dish in the 1940s as some of their staffs stayed in the US and the dish exploded in popularity in the 1970s - especially after Nixon’s trip to China, which included a televised meal on Tiananmen Square where he and Henry Kissinger ate the dish. There are countless stories of Chinese-born immigrants bringing their food to the world. Some are emblematic of larger population shifts globally. Cecilia Chiang grew up in Shanghai until Japan invaded, when she and her sister fled to Chengdu, only to flee the Chinese Communists and emigrate to the US in 1959. She opened The Mandarin in 1960 in San Francisco and a second location in 1967. It was an upscale restaurant and introduced a number of new dishes to the US from China. She went on to serve everyone from John Lennon to Julia Child - and her son Philip replaced her in 1989 before starting a more mainstream chain of restaurants he called P.F. Chang’s in 1993. The American dream, as it had come to be known. Plenty of other immigrants from countries around the world were met with open arms. Chemists, biologists, inventors, spies, mathematicians, doctors, physicists, and yes, computer scientists. And of course, chefs. Diversity of thought, diversity of ideas, and diversity-driven innovation can only come from diverse peoples. The hackers innovated over their Americanized versions of Chinese food - many making use of technology developed by immigrants from China, their children, or those who came from other nations. Just as those from nearly every industry did.
/episode/index/show/thehistoryofcomputing/id/25464210
info_outline
The Silk Roads: Then And Now...
10/28/2022
The Silk Roads: Then And Now...
The Silk Road, or roads more appropriately, has been in use for thousands of years. Horses, jade, gold, and of course silk flowed across the trade routes. As did spices - and knowledge. The term Silk Road was coined by a German geographer named Ferdinand van Richthofen in 1870 to describe a network of routes that was somewhat formalized in the second century that some theorize date back 3000 years, given that silk has been found on Egyptian mummies from that time - or further. The use of silk itself in China in fact dates back perhaps 8,500 years. Chinese silk has been found in Scythian graves, ancient Germanic graves, and along mountain ranges and waterways around modern India gold and silk flowed between east and west. These gave way to empires along the Carpathian Mountains or Kansu Corridor. There were Assyrian outposts in modern Iran and the Sogdia built cities around modern Samarkand in Uzbekistan, an area that has been inhabited since the 4th millennium BCE. The Sogdians developed trading networks that spanned over 1,500 miles - into ancient China. The road expanded with he Persian Royal Road from the 5th century BCE across Turkey and with the conquests of Alexander the Great in the 300s BCE, the Macedonian Empire pushed into Central Asia into modern Uzbekistan. The satrap Diodotus I claimed independence of one of those areas between the Hindu Kush, Pamirs, and Tengri Tagh mountains, which became known as the Hellenized name Bactria and called the Greco-Bactrian and then Into-Greek Kingdoms by history. Their culture also dates back thousands of years further. The Bactrians became powerful enough to push into the Indus Valley, west along the Caspian Sea, and north to the Syr Darya river, known as the Jaxartes at the time and to the Aral Sea. They also pushed south into modern Pakistan and Afghanistan, and east to modern Kyrgyzstan. To cross the Silk Road was to cross through Bactria, and they were considered a Greek empire in the east. The Han Chinese called them Daxia in the third century BCE. They grew so wealthy from the trade that they became the target of conquest by neighboring peoples once the thirst for silk could not be unquenched in the Roman Empire. The Romans consumed so much silk that silver reserves were worn thin and they regulated how silk could be used - something some of the Muslim’s would do over the next generations. Meanwhile, the Chinese hadn’t known where their silk was destined, but had been astute enough to limit who knew how silk was produced. The Chinese general Pan Chao in the first century AD and attempted to make contact with the Roman’s only to be thwarted by Parthians, who acted as the middlemen on many a trade route. It wasn’t until the Romans pushed East enough to control the Persian Gulf that an envoy was sent by Marcus Aurelius that made direct contact with China in 166 AD and from there, spread throughout the kingdom. Justinian even sent monks to bring home silkworm eggs but they were never able to reproduce silk, in part because they didn’t have mulberry trees. Yet, the west had perpetrated industrial espionage on the east, a practice that would be repeated in 1712 when a Jesuit priest found how the Chinese created porcelain. The Silk Road was a place where great fortunes could be found or lost. The Dread Pirate Roberts was a character from a movie called the Princess Bride, who had left home to make his fortune, so he could spend his life with his love, Buttercup. The Silk Road had made many a fortune, so Ross Ulbricht used that name on a site he created called the Silk Road, along with Frosty and Attoid. He’d gotten his Bachelors at the University of Texas and Masters at Penn State University before he got the idea to start a website he called the Silk Road in 2011. Most people connected to the site via ToR and paid for items in bitcoins. After he graduated from Penn State, he’d started a couple of companies that didn’t do that well. Given the success of Amazon, he and a friend started a site to sell used books, but Ulbricht realized it was more profitable to be the middle man, as the Parthians had thousands of years earlier. The new site would be Underground Brokers and later changed to The Silk Road. Cryptocurrencies allowed for anonymous transactions. He got some help from others, including two that went by the pseudonyms Smedley (later suspected to be Mike Wattier) and Variety Jones (later suspected to be Thomas Clark). They started to facilitate transactions in 2011. Business was good almost from the beginning. Then Gawker published an article about the site and more and more attention was paid to what was sold through this new darknet portal. The United States Department of Justice and other law enforcement agencies got involved. When bitcoins traded at less than $80 each, the United States Drug Enforcement Agency (DEA) seized 11 bitcoins, but couldn’t take the site down for good. It was actually an IRS investigator named Gary Alford who broke the case when he found the link between the Dread Pirate Roberts and Attoid and then a post that included Ulbricht’s name and phone number. Ulbricht was picked up in San Francisco and 26,000 bitcoins were seized, along with another 144,000 from Ulbricht’s personal wallets. Two federal agents were arrested when it was found they traded information about the investigation to Ulbricht. Ulbricht was also accused of murder for hire, but those charges never led to much. Ulbricht now servers a life sentence. The Silk Road of the darknet didn’t sell silk. 70% of the 10,000 things sold were drugs. There were also fake identities, child pornography, and through a second site, firearms. There were scammers. Tens of millions of dollars flowed over this new Silk Road. But the secrets weren’t guarded well enough and a Silk Road 2 was created in 2013, which only lasted a year. Others come and go. It’s kinda’ like playing whack-a-mole. The world is a big place and the reach of law enforcement agencies limited, thus the harsh sentence for Ulbricht.
/episode/index/show/thehistoryofcomputing/id/24835719
info_outline
Simulmatics: Simulating Advertising, Data, Democracy, and War in the 1960s
10/14/2022
Simulmatics: Simulating Advertising, Data, Democracy, and War in the 1960s
Dassler shoes was started by Adolf Dassler in 1924 in Germany, after he came home from World War I. His brother Rudolph joined him. They made athletic shoes and developed spikes to go on the bottom of the shoes. By 1936, they convinced Jesse Owens to wear their shoes on the way to his gold medals. Some of the American troops who liked the shoes during World War II helped spread the word. The brothers had a falling out soon after the war was over. Adolph founded Adidas while Rudolph created a rival shoe company called Puma. This was just in time for the advertising industry to convince people that if they bought athletic shoes that they would instantly be, er, athletic. The two companies became a part of an ad-driven identity that persists to this day. One most who buy the products advertised hardly understand themselves. A national identity involves concentric circles of understanding. The larger a nation, the more concentric circles and the harder it is to nail down exactly who has what identity. Part of this is that people spend less time thinking about who they are and more time being told who they should want to be like. Woven into the message of who a person should be is a bunch of products that a person has to buy to become the ideal. That’s called advertising. James White founded the first modern advertising agency called ‘R. F. White & Son' in Warwick Square, London in 1800. The industry evolved over the next hundred or so years as more plentiful supplies led to competition and so more of a need to advertise goods. Increasingly popular newspapers from better printing presses turned out a great place to advertise. The growth of industrialism meant there were plenty of goods and so competition between those who manufactured or trafficked those goods. The more efficient the machines of industry became, the more the advertising industry helped sell what the world might not yet know it needed. Many of those agencies settled into Madison Avenue in New York as balances of global power shifted and so by the end of World War II, Madison Avenue became a synonym for advertising. Many now-iconic brands were born in this era. Manufacturers and distributors weren’t the only ones to use advertising. People put out ads to find loves in personals and by the 1950s advertising even began to find its way into politics. Iconic politicians could be created. Dwight D Eisenhower served as the United States president from 1953 to 1961. He oversaw the liberation of Northern Africa in World War II, before he took command to plan the invasion of Normandy on D Day. He was almost universally held as a war hero in the United States. He had not held public office but the ad men of Madison Avenue were able to craft messages that put him into the White House. Messages like “I like Ike.” These were the early days of television and the early days of computers. A UNIVAC was able to predict that Eisenhower would defeat Adlai Stevenson in a landslide election in 1952. The country was not “Madly for Adlai” as his slogan went. ENIAC had first been used in 1945. MIT Whirlwind was created in 1951, and the age of interactive computing was upon us. Not only could a computer predict who might win an election but new options in data processing allowed for more granular ways to analyze data. A young Senator named John F. Kennedy was heralded as a “new candidate for the 1960s.” Just a few years later Stephenson had lambasted Ike for using advertising, but this new generation was willing to let computers help build a platform - just as the advertisers were starting to use computers to help them figure out the best way to market a product. It turns out that words mattered. At the beginning of that 1960 election, many observed they couldn’t tell much difference between the two candidates: Richard Nixon and John Kennedy. Kennedy’s democrats were still largely factored between those who believed in philosophies dating back to the New Deal and segregationists. Ike presided over the early days of the post-World War II new world order. This new generation, like new generations before and since, was different. They seemed to embrace the new digital era. Someone like JFK wasn’t punching cards and feeding them into a computer, writing algorithms, or out surveying people to collect that data. That was done by a company that was founded in 1959 called Simulmatics. Jill Lepore called them the What If men in her book called If/Then - a great read that goes further into the politics of the day. It’s a fascinating read. The founder of the company was a Madison Avenue ad man named Ed Greenfield. He surrounded himself with a cast of characters that included people from John Hopkins University, MIT, Yale, and IBM. Ithiel de Sola Pool had studied Nazi and Soviet propaganda during World War II. He picked up on work from Hungarian Frigyes Karinthy and with students ran Monte Carlo simulations on people’s acquaintances to formulate what would later become The Small World Problem or the Six Degrees of Separation, a later inspiration for the social network of the same name and even later, for Facebook. The social sciences had become digital. Political science could then be used to get at the very issues that could separate Kennedy from Nixon. The People Machine as one called it was a computer simulation, thus the name of the company. It would analyze voting behaviors. The previous Democratic candidate Stevenson had long-winded, complex speeches. They analyzed the electorate and found that “I Like Ike” resonated with more people. It had, after all, been developed by the same ad man who came up with “Melts in your mouth, not in your hands” for M&Ms. They called the project Project Microscope. They recruited some of the best liberal minds in political science and computer science. They split the electorate into 480 groups. A big focus was how to win the African-American vote. Turns out Gallup polls didn’t study that vote because Southern newspapers had blocked doing so. Civil rights, and race relations in general wasn’t unlike a few other issues. There was anti-Catholic, anti-Jew, and anti-a lot. The Republicans were the party of Lincoln and had gotten a lot of votes over the last hundred years for that. But factions within the party had shifted. Loyalties were shifting. Kennedy was a Catholic but many had cautioned he should down-play that issue. The computer predicted civil rights and anti-Catholic bigotry would help him, which became Kennedy’s platform. He stood for what was right but were they his positions or just what the nerds thought? He gained votes at the last minute. Turns out the other disenfranchised groups saw the bigotry against one group as akin to bigotry against their own; just like the computers thought they would. Kennedy became an anti-segregationist, as that would help win the Black vote in some large population centers. It was the most aggressive, or liberal, civil-rights plank the Democrats had ever taken up. Civil rights are human rights. Catholic rights are as well. Kennedy offered the role of Vice President to Lyndon B Johnson, the Senate Majority Leader and was nominated to the Democratic candidate. Project Microscope from Simulmatics was hired in part to shore up Jewish and African-American votes. They said Kennedy should turn the fact that he was a Catholic into a strength. Use the fact he was Catholic to give up a few votes here and there in the South but pick up other votes. He also took the Simulmatics information as it came out of the IBM 704 mainframe to shore up his stance on other issues. That confidence helped him out-perform Nixon in televised debates. They used teletypes and even had the kids rooms converted into temporary data rooms. CBS predicted Nixon would win. Less than an hour later they predicted Kennedy would win. Kennedy won the popular vote by .1 percent of the country even after two recounts. The Black vote hat turned out big for Kennedy. News leaked about the work Simulmatics had done for Kennedy. Some knew that IBM had helped Hitler track Jews as has been written about in the book IBM and the Holocaust by Edwin Black. Others still had issues with advertising in campaigns and couldn’t fathom computers. Despite Stalin’s disgust for computers some compared the use of computers to Stalinistic propaganda. Yet it worked - even if in retrospect the findings were all things we could all take for granted. They weren’t yet. The Kennedy campaign at first denied the “use of an electronic brain and yet their reports live on in the Kennedy Library. A movement against the use of the computer seemed to die after Kennedy was assassinated. Books of fiction persisted, like The 480 from Eugene Burdick, which got its title from the number of groups Simulmatics used. The company went on to experiment with every potential market their computer simulation could be used in. The most obvious was the advertising industry. But many of those companies went on to buy their own computers. They already had what many now know is the most important aspect of any data analytics project: the data. Sometimes they had decades of buying data - and could start over on more modern computers. They worked with the Times to analyze election results in 1962, to try and catch newspapers up with television. The project was a failure and newspapers leaned into more commentary and longer-term analysis to remain a relevant supplier of news in a world of real-time television. They applied their brand of statistics to help simulate the economy of Venezuela in a project called Project Camelot, which LBJ later shot down. Their most profitable venture became working with the defense department to do research in Vietnam. They collected data, analyzed data, punched data into cards, and fed it into computers. Pool was unabashedly pro-US and it’s arguable that they saw what they wanted to see. So did the war planners in the pentagon, who followed Robert McNamara. McNamara had been one of the Quiz Kids who turned around the Ford Motor Company with a new brand of data-driven management to analyze trends in the car industry, shore up supply chains, and out-innovate the competition. He became the first president of the company who wasn’t a Ford. His family had moved to the US from Ireland to flee the Great Irish Famine. Not many generations later he got an MBA from Harvard before he became a captain in the United States Army Air Forces during World War II primarily as an analyst. Henry Ford the second hired his whole group to help with the company. As many in politics and the military learn, companies and nations are very different. They did well at first, reducing the emphasis on big nuclear first strike capabilities and developing other military capabilities. One of those was how to deal with guerrilla warfare and counterinsurgencies. That became critical in Vietnam, a war between the communist North Vietnamese and the South Vietnamese. The North was backed by North Korea, China, and the Soviet Union, the South backed by the United States, South Korea, Australia. Others got involved but those were the main parties. We can think of McNamara’s use of computers to provide just in time provisioning of armed forces and move spending to where it could be most impactful, which slashed over $10 billion in military spending. As the Vietnam war intensified, statistically the number of troops killed by Americans vs American casualties made it look computationally like the was was being won. In hindsight we know it was not. Under McNamara, ARPA hired Simulmatics to study the situation on the ground. They would merge computers, information warfare, psychological warfare, and social sciences. The Vietnamese that they interviewed didn’t always tell them the truth. After all, maybe they were CIA agents. Many of the studies lacked true scholars as the war was unpopular back home. People who collected data weren’t always skilled at the job. They spoke primarily with those they didn’t get shot at as much while going to see. In general, the algorithms might have worked or might not have worked - but they had bad data. Yet Simulmatics sent reports that the operations were going well to McNamara. Many in the military would remember this as real capabilities at cyber warfare and information warfare were developed in the following decades. Back home, Simulmatics also became increasingly tied up in things Kennedy might have arguably fought against. There were riots, civil rights protests, and Simulatics took contracts to simulate racial riots. Some felt they could riot or go die in in the jungles of Vietnam. The era of predictive policing had begun as the hope of the early 1960s turned into the apathy of the late 1960s. Martin Luther King Jr spoke out again riot prediction, yet Simulmatics pushed on. Whether their insights were effective in many of the situations, just like in Vietnam - was dubious. They helped usher in the era of Surveillance capitalism, in a way. But the arrival of computers in ad agencies meant that if they hadn’t of, someone else would have. People didn’t take kindly to being poked, prodded, and analyzed intellectually. Automation took jobs, which Kennedy had addressed in rhetoric if not in action. The war was deeply unpopular as American soldiers came home from a far off land in caskets. The link between Simulmatics and academia was known. Students protested against them and claimed they were war criminals. The psychological warfare abroad, being on the wrong side of history at home with the race riots, and the disintegrating military-industrial-university complex didn’t help. There were technical issues. The technology had changed away from languages like FORTRAN. Further, the number of data points required and how they were processed required what we now call “Big Data” and “machine learning.” Those technologies showed promise early but more mathematics needed to be developed to fully weaponize the surveillance everything. More code and libraries needed to be developed to crunch the large amounts of statistics. More work needed to be done to get better data and process it. The computerization of the social sciences was just beginning and while people like Pool predicted the societal impacts we could expect, people at ARPA doubted the results and the company they created could not be saved as all these factors converged to put them into bankruptcy in 1970. Their ideas and research lived on. Pool and others published some of their findings. Books opened the minds to the good and bad of what technology could do. The Southern politicians, or Dixiecrats, fell apart. Nixon embraced a new brand of conservatism as he lost the race to be the Governor of California to Pat Brown in 1962. There were charges of voter fraud from the 1960 election. The Mansfeld Amendment restricted military funding of basic research in 1969 and went into effect in 1970. Ike had warned of the growing links between universities as the creators of weapons of war like what Simulmatics signified and the amendment helped pull back funding for such exploits. As Lepore points out in her book, mid-century liberalism was dead. Nixon tapped into the silent majority who countered the counterculture of the 1960s. Crime rose and the conservatives became the party of law and order. He opened up relations with China, spun down the Vietnam war, negotiated with the Soviet leader Brezhnev to warm relations, and rolled back Johnson’s attempts at what had been called The Great Society to get inflation back in check. Under him the incarceration rate in the United States exploded. His presidency ended with Watergate and under Ford, Carter, Reagan, and Bush, the personal computer became prolific and the internet, once an ARPA project began to take shape. They all used computers to find and weigh issues, thaw the Cold War, and build a new digitally-driven world order. The Clinton years saw an acceleration of the Internet and by the early 2000s companies like PayPal were on the rise. One of their founders was Peter Thiel. Peter Thiel founded Palantir in 2003 then invested in companies like Facebook with his PayPal money. Palantir received backing from In-Q-Tel “World-class, cutting-edge technologies for National Security”. In-Q-Tel was founded in 1999 as the global technological evolution began to explode. While the governments of the world had helped build the internet, it wasn’t long before they realized it gave an asymmetrical advantage to newcomers. The more widely available the internet, the more far reaching attacks could go, the more subversive economic warfare could be. Governmental agencies like the United States Central Intelligence Agency (CIA) needed more data and the long promised artificial intelligence technologies to comb through that data. Agencies then got together and launched their own venture capital fund, similar to those in the private sector - one called In-Q-Tel. Palantir has worked to develop software for the US Immigration and Customers Enforcement, or ICE, to investigate criminal activities and allegedly used data obtained from Cambridge Analytica along with Facebook data. The initial aim of the company was to take technology developed for PayPal’s fraud detection and apply it to other areas like terrorism, with help from intelligence agencies. They help fight fraud for nations and have worked with the CIA, NSA, FBI, CDC, and various branches of the United States military on various software projects. Their Gotham project is the culmination of decades of predictive policing work. There are dozens of other companies like Palantir. Just as Pool’s work on Six Degrees of Separation, social networks made the amount of data that could be harvested all the greater. Companies use that data to sell products. Nations use that data for propaganda. Those who get elected to run nations use that data to find out what they need to say to be allowed to do so. The data is more accurate with every passing year. Few of the ideas are all that new, just better executed. The original sin mostly forgotten, we still have to struggle with the impact and ethical ramifications. Politics has always had a bit of a ruse in a rise to power. Now it’s less about personal observation and more about the observations and analyses that can be gleaned from large troves of data. The issues brought up in books like The 480 are as poignant today as they were in the 1950s.
/episode/index/show/thehistoryofcomputing/id/24691455
info_outline
Taiwan, TSMC, NVIDIA, and Foundries
09/30/2022
Taiwan, TSMC, NVIDIA, and Foundries
Taiwan is a country about half the size of Maine with about 17 times the population of that state. Taiwan sits just over a hundred miles off the coast of mainland China. It’s home to some 23 and a half million humans, roughly half way between Texas and Florida or a few more than live in Romania for the Europeans. Taiwan was connected to mainland China by a land bridge in the Late Pleistocene and human remains have been found dating back to 20,000 to 30,000 years ago. About half a million people on the island nation are aboriginal, or their ancestors are from there. But the population became more and more Chinese in recent centuries. Taiwan had not been part of China during the earlier dynastic ages but had been used by dynasties in exile to attack one another and so became a part of the Chinese empire in the 1600s. Taiwan was won by Japan in the late 1800s and held by the Japanese until World War II. During that time, a civil war had raged on the mainland of China with the Republic of China eventually formed as the replacement government for the Qing dynasty following a bloody period of turf battles by warlords and then civil war. Taiwan was in martial law from the time the pre-communist government of China retreated there during the exit of the Nationalists from mainland China in the 1940s to the late 1980. During that time, just like the exiled Han dynasty, they orchestrated war from afar. They stopped fighting, much like the Koreans, but have still never signed a peace treaty. And so large parts of the world remained in stalemate. As the years became decades, Taiwan, or the Republic of China as they still call themselves, has always had an unsteady relationship with the People’s Republic of China, or China as most in the US calls them. The Western world recognized the Republic of China and the Soviet and Chines countries recognized the mainland government. US President Richard Nixon visited mainland China in 1972 to re-open relations with the communist government there and relations slowly improved. The early 1970s was a time when much of the world still recognized the ruling government of Taiwan as the official Chinese government and there were proxy wars the two continued to fight. The Taiwanese and Chinese still aren’t besties. There are deep scars and propaganda that keep relations from being repaired. Before World War II, the Japanese also invaded Hong Kong. During the occupation there, Morris Chang’s family became displaced and moved to a few cities during his teens before he moved Boston to go to Harvard and then MIT where he did everything to get his PhD except defend his thesis. He then went to work for Sylvania Semiconductor and then Texas Instruments, finally getting his PhD from Stanford in 1964. He became a Vice President at TI and helped build an early semiconductor designer and foundry relationship when TI designed a chip and IBM manufactured it. The Premier of Taiwan at the time, Sun Yun-suan, who played a central role in Taiwan’s transformation from an agrarian economy to a large exporter. His biggest win was when to recruit Chang to move to Taiwan and found TSCM, or Taiwan Semiconductor Manufacturing Company. Some of this might sound familiar as it mirrors stories from companies like Samsung in South Korea. In short, Japanese imperialism, democracies versus communists, then rapid economic development as a massive manufacturing powerhouse in large part due to the fact that semiconductor designers were split from semiconductor foundry’s or where chips are actually created. In this case, a former Chinese national was recruited to return as founder and led TSMC for 31 years before he retired in 2018. Chang could see from his time with TI that more and more companies would design chips for their needs and outsource manufacturing. They worked with Texas Instruments, Intel, AMD, NXP, Marvell, MediaTek, ARM, and then the big success when they started to make the Apple chips. The company started down that path in 2011 with the A5 and A6 SoCs for iPhone and iPad on trial runs but picked up steam with the A8 and A9 through A14 and the Intel replacement for the Mac, the M1. They now sit on a half trillion US dollar market cap and are the largest in Taiwan. For perspective, their market cap only trails the GDP of the whole country by a few billion dollars. Nvidia TSMC is also a foundry Nvidia uses. As of the time of this writing, Nvidia is the 8th largest semiconductor company in the world. We’ve already covered Broadcom, Qualcomm, Micron, Samsung, and Intel. Nvidia is a fabless semiconductor company and so design chips that vendors like TSMC manufacture. Nvidia was founded by Jensen Huang, Chris Malachowsky, and Curtis Priem in 1993 in Santa Clara, California (although now incorporated in Delaware). Not all who leave the country they were born in due to war or during times of war return. Huang was born in Taiwan and his family moved to the US right around the time Nixon re-established relations with mainland China. Huang then went to grad school at Stanford before he became a CPU designer at AMD and a director at LSI Logic, so had experience as a do-er, a manager, and a manager’s manager. He was joined by Chris Malachowsky and Curtis Priem, who had designed the IBM Professional Graphics Adapter and then the GX graphics chip at Sun. because they saw this Mac and Windows and Amiga OS graphical interface, they saw the games one could play on machines, and they thought the graphics cards would be the next wave of computing. And so for a long time, Nvidia managed to avoid competition with other chip makers with a focus on graphics. That initially meant gaming and higher end video production but has expanded into much more like parallel programming and even cryptocurrency mining. They were more concerned about the next version of the idea or chip or company and used NV in the naming convention for their files. When it came time to name the company, they looked up words that started with those letters, which of course don’t exist - so instead chose invidia or Nvidia for short, as it’s latin for envy - what everyone who saw those sweet graphics the cards rendered would feel. They raised $20 million in funding and got to work. First with SGS-Thomson Microelectronics in 1994 to manufacture what they were calling a graphical-user interface accelerator that they packaged on a single chip. They worked with Diamond Multimedia Systems to install the chips onto the boards. In 1995 they released NV1. The PCI card was sold as Diamond Edge 3D and came with a 2d/3d graphics core with quadratic texture mapping. Screaming fast and Virtual Fighter from Sega ported to the platform. DirectX had come in 1995. So Nviia released DirectX drivers that supported Direct3D, the api that Microsoft developed to render 3d graphics. This was a time when 3d was on the rise for consoles and desktops. Nvidia timed it perfectly and reaped the rewards when they hit a million sold in the first four months for the RIVA, a 128-bit 3d processor that got used as an OEM in 1997. Then the 1998 RIVAZX with RIVATNT for multi-texture 3D processing. They also needed more manufacturing support at this point and entered into a strategic partnership with TSMC to manufacture their boards. A lot of vendors had a good amount of success in their niches. By the late 1990s there were companies who made memory, or the survivors of the DRAM industry after ongoing price dumping issues. There were companies that made central processors like Intel. Nvidia led the charge for a new type of chip, the GPU. They invented the GPU in 1999 when they released the GeForce 256. This was the first single-chip GPU processor. This means integrated lightings, triangle setups, rendering, like the old math coprocessor but for video. Millions of polygons could be drawn on screens every second. They also released the Quadro Pro GPU for professional graphics and went public in 1999 at an IPO of $12 per share. Nvidia used some of the funds from the IPO to scale operations, organically and inorganically. In 2000 they released the GeForce2 Go for laptops and acquired 3dfx, closing deals to get their 3d chips in devices from OEM manufacturers who made PCs and in the new Microsoft Xbox. By 2001 they hit $1 billion in revenues and released the GeForce 3 with a programmable GPU, using APIs to make their GPU a platform. They also released the nForce integrated graphics and so by 2002 hit 100 million processors out on the market. They acquired MediaQ in 2003 and partnered with game designer Blizzard to make Warcraft. They continued their success in the console market when the GeForce platform was used in the PS 3 in 2005 and by 2006 had sold half a billion processors. They also added the CUDA architecture that year to put a general purpose GPU on the market and acquired Hybrid Graphics who develops 2D and 3D embedded software for mobile devices. In 2008 they went beyond the consoles and PCs when Tesla used their GPUs in cars. They also acquired PortalPlayer, who supplies semiconductors and software for personal media players and launched the Tegra mobile processor to get into the exploding mobile market. More acquisitions in 2008 but a huge win when the GeForce 9400M was put into Apple MacBooks. Then more smaller chips in 2009 when the Tegra processors were used in Android devices. They also continued to expand how GPUs were used. They showed up in Ultrasounds and in 2010 the Audi. By then they had the Tianhe-1A ready to go, which showed up in supercomputers and the Optimus. All these types of devices that could use a GPU meant they hit a billion processors sold in 2011, which is when they went dual core with the Tegra 2 mobile processor and entered into cross licensing deals with Intel. At this point TSMC was able to pack more and more transistors into smaller and smaller places. This was a big year for larger jobs on the platform. By 2012, Nvidia got the Kepler-based GPUs out by then and their chips were used in the Titan supercomputer. They also released a virtualized GPU GRID for cloud processing. It wasn’t all about large-scale computing efforts. The Tegra-3 and GTX 600 came out in 2012 as well. Then in 2013 the Tegra 4, a quad-core mobile processor, a 4G LTE mobile processor, Nvidia Shield for portable gaming, the GTX Titan, a grid appliance. In 2014 the Tegra K1 192, a shield tablet, and Maxwell. In 2015 came the TegraX1 with deep learning with 256 cores and Titan X and Jetson TX1 for smart machines, and the Nvidia Drive for autonomous vehicles. They continued that deep learning work with an appliance in 2016 with the DGX-1. The Drive got an update in the form of PX 2 for in-vehicle AI. By then, they were a 20 year old company and working on the 11th generation of the GPU and most CPU architectures had dedicated cores for machine learning options of various types. 2017 brought the Volta, Jetson TX2, and SHIELD was ported over to the Google Assistant. 2018 brought the Turing GPU architecture, the DGX-2, AGX Xavier, Clara, 2019 brought AGX Orin for robots and autonomous or semi-autonomous piloting of various types of vehicles. They also made the Jetson Nano and Xavier, and EGX for Edge Computing. At this point there were plenty of people who used the GPUs to mine hashes for various blockchains like with cryptocurrencies and the ARM had finally given Intel a run for their money with designs from the ARM alliance showing up in everything but a Windows device (so Apple and Android). So they tried to buy ARM from SoftBank in 2020. That deal fell through eventually but would have been an $8 billion windfall for Softbank since they paid $32 billion for ARM in 2016. We probably don’t need more consolidation in the CPU sector. Standardization, yes. Some of top NVIDIA competitors include Samsung, AMD, Intel Corporation Qualcomm and even companies like Apple who make their own CPUs (but not their own GPUs as of the time of this writing). In their niche they can still make well over $15 billion a year. The invention of the MOSFET came from immigrants Mohamed Atalla, originally from Egypt, and Dawon Kahng, originally from from Seoul, South Korea. Kahng was born in Korea in 1931 but immigrated to the US in 1955 to get his PhD at THE Ohio State University and then went to work for Bell Labs, where he and Atalla invented the MOSFET, and where Kahng retired. The MOSFET was an important step on the way to a microchip. That microchip market with companies like Fairchild Semiconductors, Intel, IBM, Control Data, and Digital Equipment saw a lot of chip designers who maybe had their chips knocked off, either legally in a clean room or illegally outside of a clean room. Some of those ended in legal action, some didn’t. But the fact that factories overseas could reproduce chips were a huge part of the movement that came next, which was that companies started to think about whether they could just design chips and let someone else make them. That was in an era of increasing labor outsourcing, so factories could build cars offshore, and the foundry movement was born - or companies that just make chips for those who design them. As we have covered in this section and many others, many of the people who work on these kinds of projects moved to the United States from foreign lands in search of a better life. That might have been to flee Europe or Asian theaters of Cold War jackassery or might have been a civil war like in Korea or Taiwan. They had contacts and were able to work with places to outsource too and given that these happened at the same time that Hong Kong, Singapore, South Korea, and Taiwan became safe and with no violence. And so the Four Asian Tigers economies exploded, fueled by exports and a rapid period of industrialization that began in the 1960s and continues through to today with companies like TSMC, a pure play foundry, or Samsung, a mixed foundry - aided by companies like Nvidia who continue to effectively outsource their manufacturing operations to companies in the areas. At least, while it’s safe to do so. We certainly hope the entire world becomes safe. But it currently is not. There are currently nearly a million Rohingya refugees fleeing war in Myanmar. Over 3.5 million have fled the violence in Ukraine. 6.7 million have fled Syria. 2.7 million have left Afghanistan. Over 3 million are displaced between Sudan and South Sudan. Over 900,000 have fled Somalia. Before Ukranian refugees fled to mostly Eastern European countries, they had mainly settled in Turkey, Jordan, Lebanon, Pakistan, Uganda, Germany, Iran, and Ethiopia. Very few comparably settled in the 2 largest countries in the world: China, India, or the United States. It took decades for the children of those who moved or sent their children abroad to a better life to be able to find a better life. But we hope that history teaches us to get there faster, for the benefit of all.
/episode/index/show/thehistoryofcomputing/id/24545652
info_outline
The History of Zynga and founder Mark Pincus
08/19/2022
The History of Zynga and founder Mark Pincus
Mark Pincus was at the forefront of mobile technology when it was just being born. He is a recovering venture capitalist who co-founded his first company with Sunil Paul in 1995. FreeLoader was at the forefront of giving people the news through push technology, just as the IETF was in the process of ratifying HTTP2. He sold that for $38 million only to watch it get destroyed. But he did invest in a startup that one of the interns founded when he gave Sean Parker $100,000 to help found Napster. Pincus then started Support.com, which went public in 2000. Then Tribe.net, which Cisco acquired. As a former user, it was fun while it lasted. Along the way, Pincus teamed up with Reid Hoffman, former PayPal executive and founder of LinkedIn and bought the Six Degrees patent that basically covered all social networking. Along the way, he invested in Friendster, Buddy Media, Brightmail, JD.com, Facebook, Snapchat, and Twitter. Investing in all those social media properties gave him a pretty good insight into what trends were on the way. Web 2.0 was on the rise and social networks were spreading fast. As they spread, each attempted to become a platform by opening APIs for third-party developers. This led to an opening to create a new company that could build software that sat on top of these social media companies. Meanwhile, the gaming industry was in a transition from desktop and console games to hyper-casual games that are played on mobile devices. So Pincus recruited conspirators to start yet another company and with Michael Luxton, Andrew Trader, Eric Schiermeyer, Steve Schoettler, and Justin Waldron, Zinga was born in 2007. Actually Zinga is the dog. The company Zynga was born in 2007. Facebook was only three years old at the time, but was already at 14 million users to start 2007. That’s when they opened up APIs for integration with third party products through FBML, or Facebook Markup Language. They would have 100 million within a year. Given his track record selling companies and picking winners, Zynga easily raised $29 million to start what amounts to a social game studio. They make games that people access through social networks. Luxton, Schiermeyer, and Waldron created the first game, Zynga Poker in 2007. It was a simple enough Texas hold ’em poker game but rose to include tens of millions of players at its height, raking in millions in revenue. They’d proven the thesis. Social networks, especially Facebook, were growing.. The iPhone came out in 2007. That only hardened their resolve. They sold poker chips in 2008. Then came FarmVille. FarmVille was launched in 2009 and an instant hit. The game went viral and had a million daily users in a week. It was originally written in flash and later ported to iPhones and other mobile platforms. It’s now been installed over 700 million times and ran until 2020, when Flash support was dropped by Facebook. FarmVille was free-to-play and simple. It had elements of a 4x game like Civilization, but was co-op, meaning players didn’t exterminate one another but instead earned points and thus rankings. In fact, players could help speed up tasks for one another. Players began with a farm - an empty plot of land. They earned experience points by doing routine tasks. Things like growing crops, upgrading items, plowing more and more land. Players took their crops to the market and sold them for coins. Coins could also be bought. If a player didn’t harvest their crops when they were mature, the crops would die. Thus, they had players coming back again and again. Push notifications helped remind people about the state of their farm. Or the news in FreeLoader-speak. Some players became what we called dolphins, or players that spent about what they would on a usual game. Maybe $10 to $30. Others spent thousands, which we referred to as whales. They became the top game on Facebook and the top earner. They launched sequels as well, with FarmVille 2 and FarmVille 3. They bought Challenge Games in 2010, which was founded by Andrew Busy to develop casual games a well. They bought 14 more companies. They grew to 750 employees. They opened offices in Bangalore, India and Ireland. They experimented with other platforms, like Microsoft’s MSN gaming environment and Google TV. They released CastleVille. And they went public towards the end of 2011. It was a whirlwind ride, and just really getting started. They released cute FarmVille toys. They also released Project Z, Mafia Wars, Hanging with Friends, Adventure World, and Hidden Chronicles. And along the way they became a considerable advertising customer for Facebook, with ads showing up for Mafia Wars and Project Z constantly. Not only that, but their ads flooded other mobile ad networks, as The Sims Social and other games caught on and stole eyeballs. And players were rewarded for spamming the walls of other players, which helped to increase the viral nature of the early Facebook games. Pincus and the team built a successful, vibrant company. They brought in Jeff Karp and launched Pioneer Trail. Then another smash hit, Words with Friends. They bought Newtoy for $53.3 million to get it, after Paul and David Bettner who wrote a game called Chess with Friends a few years earlier. But revenues dropped as the Facebook ride they’d been on began to transition from people gaming in a web browser to mobile devices. All this growth and the company was ready for the next phase. In 2013, Zynga hired Donald Mattrick to be the CEO and Pincus moved to the role of Chief Product Officer. The brought in Alex Garden, the General Manager for Xbox Music , Video, and Reading, who had founded the Homeward creator Relic Entertainment back in the 1990s. The new management didn’t fix the decline. The old games continued to lose market share and Pincus came back to run the company as CEO and cut the staff by 18 percent. In 2015 they brought in Frank Gibeau to the board and by 2016 moved him to CEO of the company. One challenge with the move to mobile was who got the processing payments. Microtransactions had gone through Facebook for years. They moved to Stripe in 2020. They acquired Gram Games, to get Merge Dragons! They bought Small Giant Games to get Empires & Puzzles. They bought Peak Games to get Toon Blast and Toy Blast. They picked up Rollic to get a boatload of actions and puzzle games. They bought Golf Rival by acquiring StarLark. And as of the time of this writing they have nearly 200 million players actively logging into their games. There are a few things to take from the story of Zynga. One is that a free game doesn’t put $2.8 billion in revenues on the board, which is what they made in 2021. Advertising amounts for just north of a half billion, but the rest comes from in app purchases. The next is that the transition from owner-operators is hard. Pincus and the founding team had a great vision. They executed and were rewarded by taking the company to a gangbuster IPO. The market changed and it took a couple of pivots to get there. That led to a couple of management shakeups and a transition to more of a portfolio mindset with the fleet of games they own. Another lesson is that larger development organizations don’t necessarily get more done. That’s why Zynga has had to acquire companies to get hits since around the time that they bought Words with Friends. Finally, when a company goes public the team gets distracted. Not only is going through an IPO expensive and the ensuing financial reporting requirements a hassle to deal with, but it’s distracting. Employees look at stock prices during the day. Higher ranking employees have to hire a team of accountants to shuffle their money around in order to take advantage of tax loopholes. Growth leads to political infighting and power grabbing. There are also regulatory requirements with how we manage our code and technology that slow down innovation. But it all makes us better run and a safer partner eventually. All companies go through this. Those who navigate towards a steady state fastest have the best chance of surviving one more lesson: when the first movers prove a monetization thesis the ocean will get red fast. Zynga became the top mobile development company again after weathering the storm and making a few solid acquisitions. But as Bill Gates pointed out in the 1980s, gaming is a fickle business. So Zynga agreed to be acquired for $12.7 billion in 2022 by Take-Two Interactive, who now owns the Civilization, Grand Theft Auto, Borderlands, WWE, Red Dead, Max Payne, NBA 2K, PGA 2K, Bioshock, Duke Nukem, Rainbow Six: Rogue Spear, Battleship, Centipede, and the list goes on and on. They’ve been running a portfolio for a long time. Pincus took away nearly $200 million in the deal and about $350 million in Take-Two equity. Ads and loot boxes can be big business. Meanwhile, Pincus and Hoffman from LinkedIn work well together, apparently. They built Reinvent Capital, an investment firm that shows that venture capital has quite a high recidivism rate. They had a number of successful investments and SPACs. Zynga was much more. They exploited Facebook to shoot up to hundreds of millions in revenue. That was revenue Facebook then decided they should have a piece of in 2011, which cut those Zynga revenues in half over time. This is an important lesson any time a huge percentage of revenue is dependent on another party who can change the game (no pun intended) at any time. Diversify.
/episode/index/show/thehistoryofcomputing/id/24103038
info_outline
The Evolution Of Unix, Mac, and Chrome OS Shells
07/15/2022
The Evolution Of Unix, Mac, and Chrome OS Shells
In the beginning was the command line. Actually, before that were punch cards and paper tape. But at Multics and RSTS and DTSS came out, programmers and users needed a way to interface with the computer through the teletypes and other terminals that appeared in the early age of interactive computing. Those were often just a program that sat on a filesystem eventually as a daemon, listening for input on keyboards. This was one of the first things the team that built Unix needed, once they had a kernel that could compile. And from the very beginning it was independent of the operating system. Due to the shell's independence from the underlying operating system, numerous shells have been developed during Unix’s history, albeit only a few have attained widespread use. Shells are also referred to as Command-Line Interpreters (or CLIs), processes commands a user sends from a teletype, then a terminal. This provided a simpler interface for common tasks, rather than interfacing with the underlying C programming. Over the years, a number of shells have come and gone. Some of the most basic and original commands came from Multics, but the shell as we know it today was introduced as the Thompson shell in the first versions of Unix. Ken Thompson introduced the first Unix shell in 1971 with the Thompson Shell, the ancestor of the shell we still find in /bin/sh. The shell ran in the background and allowed for a concise syntax for redirecting the output of commands to one another. For example, pass the output to a file with > or read input from a file with <. We could also pipe output to a second command using a |. This simple structure has persisted into modern shells, even as the kernels and capabilities have become infinitely more complex. The Thompson shell didn’t have options for scripting, as Ken Thompson was more of a C programmer. So future shells would add those options. But when it came to basic operations in Unix, like running commands and viewing their output, he didn’t need to give users more privileged forms of access through the shell. Others built tools for Unix as well. Bill Joy wrote a different text editor when Berkeley had Thompson out to install Unix on their PDP. And 1977 saw the earliest forms of what we would later call the Bourne Shell, written by Steve Bourne. This shell. The Bourne shell was designed with two key aims: to act as a command interpreter for interactively executing operating system commands and to facilitate scripting. One of the more important aspects of going beyond piping output into other commands and into a more advanced scripting language is the ability to perform conditional if/then statements, loops, and variables. And thus rather than learn C to write simple programmers, generations of engineers and end users could now use basic functional programming at a bourne shell. Bill Joy created the C shell in 1978 while a graduate student at the University of California, Berkeley. It was designed for Berkeley Software Distribution (BSD) Unix machines. One of the main design goals of the C shell was to build a scripting language that seemed like C. Joy added one of my favorite features of every shell made after that one: command history. I’ve written many shell scripts by just cut-copy-pasting a few commands from my bash history and piping or variabalizing the output. Add to that the ability to use the up or down arrow to re-run previous commands and we got a huge productivity gain for people that did the same tasks, like editing a file. Simply scroll up through previous commands to run the same vi editor. That vi command also shipped first with BSD. There was another huge time saver out there in another operating system. An operating system called Tenex had name and command completion. The Tenex OS first shipped out of BBN, or Bolt, Beranek, and Newman, for PDPs. Unix had as well and so a number of early users had experience with both. Tenex had command completion, just hit the tab and the command being typed would automatically complete if the text started matched the text of a command in a path for commands. That project was started by Ken Greer at Carnegie Mellon University in 1975 and got merged into the C shell in 1981, adding the t for Tenex to the C for C shell and gave us tcsh. Thus tcsh had backwards compatibility with csh. David Korn at Bell labs added the korn shell, or ksh in the early 1980s. He added the idea that the user interface could provide a number of editors. For example, use emacs or vi to edit files. He borrowed ideas from the c shell and made minor tweaks that provided outsized impacts to productivity. Even Microsoft added a Korn shell option into Windows NT, as though Dave Cutler was paying homage to another great programmer. Brian Fox then added on to the Bourne shell with bash. He was working with the Free Software Foundation with Richard Stallman, and they wanted a shell that could do more advanced scripting but whose source code was open source. They started the project in 1988 and shipped bash in 1989. Bash then went on to become the most widely used and distributed shell in the arsenal of the Unix programmer. Bash stands for Bourne Again Shell and so was backwards compatible with bourne shell but also added features from tcsh, korn, and C shell, staying mostly backwards compatible with other shells. Due to the licensing, bash became the de facto standard (and often default shell) for GNU/Linux distributions and serves as the standard interactive shell for users, located at /bin/bash location. Now we had command history, tabbed auto-completion, command-line editing, multiple paths, multiple options for interpreters, a directory stack, full environment variables, and the modern command line environment. Paul Falstad created the initial version of zsh, or the Z Shell, in 1990. The Z shell (zsh) can be used interactively as a login shell or as more sophisticated command interpreter for shell scripting. As with previous shells, it is an optimized Bourne shell that incorporates several features from bash and tcsh and is mostly backwards compatible. Zsh comes with the tabbed auto-completion, regex integration (in addition to the standard glowing options available since the 1970s, additional shorthand for command scoping, but with a number of security features. The ability to limit memory and privilege escalations became critical in order to keep from having some of the same issues we’ve seen for decades with Windows and other operating systems as they evolved to meet Unix scripting, borrowing many a feature for Powershell from cousins in the Unix and Linux worlds. These are just the big ones. Sometimes it feels like every developer with a decent grasp of C and a workflow divergent from the norm (which is most developers), has taken a stab at developing their own shell. This is one of the great parts of having access to source code. The options are endless. At this point, we just take these productivity gains for granted. But it was decades of innovative approaches as Unix and then Linux and now MacOS and Android reached out to the rest of the world to change how we work.
/episode/index/show/thehistoryofcomputing/id/23750525
info_outline
St Jude, Felsenstein, and Community Memory
06/25/2022
St Jude, Felsenstein, and Community Memory
Lee Felsenstein went to the University of California, Berkeley in the 1960s. He worked at the tape manufacturer Ampex, where Oracle was born out of before going back to Berkeley to finish his degree. He was one of the original members of the Homebrew Computer Club, and as with so many inspired by the Altair S-100 bus, designed the Sol-20, arguably the first microcomputer that came with a built-in keyboard that could be hooked up to a television in 1976. The Apple II was introduced the following year. Adam Osborne was another of the Homebrew Computer Club regulars who wrote An Introduction to Microcomputers and sold his publishing company to McGraw-Hill in 1979. Flush with cash, he enlisted Felsenstein to help create another computer, which became the Osborne 1. The first commercial portable computer, although given that it weighed almost 25 pounds, is more appropriate to call a luggable computer. Before Felsensten built computers, though, he worked with a few others on a community computing project they called Community Memory. Judith Milhon was an activist in the 1960s Civil Rights movement who helped organize marches and rallies and went to jail for civil disobedience. She moved to Ohio, where she met Efrem Lipkin, and as with many in what we might think of as the counterculture now, they moved to San Francisco in 1968. St Jude, as she became called learned to program in 1967 and ended up at the Berkeley Computer Company after the work on the Berkeley timesharing projects was commercialized. There, she met Pam Hardt at Project One. Project One was a technological community built around an alternative high school founded by Ralph Scott. They brought together a number of non-profits to train people in various skills and as one might expect in the San Francisco area counterculture they had a mix of artists, craftspeople, filmmakers, and people with deep roots in technology. So much so that it became a bit of a technological commune. They had a warehouse and did day care, engineering, film processing, documentaries, and many participated in anti-Vietnam war protests. They had all this space and Hardt called around to find the computer. She got an SDS-940 mainframe donated by TransAmerica in 1971. Xerox had gotten out of the computing business and TransAmerica’s needs were better suited for other computers at the time. They had this idea to create a bulletin board system for the community and created a project at Project One they called Resource One. Plenty thought computers were evil at the time, given their rapid advancements during the Cold War era, and yet many also thought there was incredible promise to democratize everything. Peter Deutsch then donated time and an operating system he’d written a few years before. She then published a request for help in the People’s Computer Computer magazine and got a lot of people who just made their own things. An early precursor to maybe micro-services, where various people tinkered with data and programs. They were able to do so because of the people who could turn that SDS into a timesharing system. St Jude’s partner Lipkin took on the software part of the project. Chris Macie wrote a program that digitized information on social services offered in the area that was maintained by Mary Janowitz, Sherry Reson, and Mya Shone. That was eventually taken over by the United Way until the 1990s. Felsenstein helped with the hardware. They used teletype terminals to connect a video terminal and keyboard built into a wooden cabinet so real humans could access the system. The project then evolved into what was referred to as Community Memory. Community Memory Community Memory became the first public computerized bulletin board system established in 1973 in Berkeley, California. The first Community Memory terminal was located at Leopard’s Record in Berkeley. This was the first opportunity for people who were not studying the scientific subject to be able to use computers. It became very popular but soon was shut down by the founders because they face hurdles to replicate the equipment and languages being used. They were unable to expand the project. This allowed them to expand the timesharing system into the community and became a free online community-based resource used to share knowledge, organize, and grow. The initial stage of Community Memory from 1973 to 1975, was an experiment to see how people would react to using computers to share information. Operating from 1973 to 1992, it went from minicomputers to microcomputers as those became more prevelant. Before Resource One and Community Memory, computers weren’t necessarily used for people. They were used for business, scientific research, and military purposes. After Community Memory, Felsenstein and others in the area and around the world helped make computers personal. Commun tty Memory was one aspect of that process but there were others that unfolded in the UK, France, Germany and even the Soviet Union - although those were typically impacted by embargoes and a lack of the central government’s buy-in for computing in general. After the initial work was done, many of the core instigators went in their own directions. For example, Felsenstein went on to create the SOL and pursue his other projects in personal computing. Many had families or moved out of the area after the Vietnam War ended in 1975. The economy still wasn’t great, but the technical skills made them more employable. Some of the developers and a new era of contributors regrouped and created a new non-profit in 1977. They started from scratch and developed their own software, database, and communication packages. It was very noisy so they encased it in a card box. It had a transparent plastic top so they could see what was being printed out. This program ran from 1984 to 1989. After more research, a new terminal was released in 1989 in Berkeley. By then it had evolved into a pre-web social network. The modified keyboard had brief instructions mounted on it, which showed the steps to send a message, how to attach keywords to messages, and how to search those keywords to find messages from others. Ultimately, the design underwent three generations, ending in a network of text-based browsers running on basic IBM PCs accessing a Unix server. It was never connected to the Internet, and closed in 1992. By then, it was large, unpowered, and uneconomical to run in an era where servers and graphical interfaces were available. A booming economy also ironically meant a shortage of funding. The job market exploded for programmers in the decade that led up to the dot com bubble and with inconsistent marketing and outreach, Community Memory shut down in 1992. Many of the people involved with Resource One and Community memory went on to have careers in computing. St Jude helped found the cypherpunks and created Mondo 2000 magazine, a magazine dedicated to that space where computers meet culture. She also worked with Efrem Lipkin on CoDesign, and he was a CTO for many of the dot coms in the late 1990s. Chris Neustrup became a programmer for Agilent. The whole operation had been funded by various grants and donations and while there haven’t been any studies on the economic impact due to how hard it is to attribute inspiration rather than direct influence, the payoff was nonetheless considerable.
/episode/index/show/thehistoryofcomputing/id/23538881
info_outline
Research In Motion and the Blackberry
06/17/2022
Research In Motion and the Blackberry
Lars Magnus Ericsson was working for the Swedish government that made telegraph equipment in the 1870s when he started a little telegraph repair shop in 1976. That was the same year the telephone was invented. After fixing other people’s telegraphs and then telephones he started a company making his own telephone equipment. He started making his own equipment and by the 1890s was shipping gear to the UK. As the roaring 20s came, they sold stock to buy other companies and expanded quickly. Early mobile devices used radios to connect mobile phones to wired phone networks and following projects like ALOHANET in the 1970s they expanded to digitize communications, allowing for sending early forms of text messages, the way people might have sent those telegraphs when old Lars was still alive and kicking. At the time, the Swedish state-owned Televerket Radio was dabbling in this space and partnered with Ericsson to take first those messages then as email became a thing, email, to people wirelessly using the 400 to 450 MHz range in Europe and 900 MHz in the US. That standard went to the OSI and became a 1G wireless packet switching network we call Mobitex. Mike Lazaridis was born in Istanbul and moved to Canada in 1966 when he was five, attending the University of Waterloo in 1979. He dropped out of school to take a contract with General Motors to build a networked computer display in 1984. He took out a loan from his parents, got a grant from the Canadian government, and recruited another electrical engineering student, Doug Fregin from the University of Windsor, who designed the first circuit boards. to join him starting a company they called Research in Motion. Mike Barnstijn joined them and they were off to do research. After a few years doing research projects, they managed to build up a dozen employees and a million in revenues. They became the first Mobitex provider in America and by 1991 shipped the first Mobitex device. They brought in James Balsillie as co-CEO, to handle corporate finance and business development in 1992, a partnership between co-CEOs that would prove fruitful for 20 years. Some of those work-for-hire projects they’d done involved reading bar codes so they started with point-of-sale, enabling mobile payments and by 1993 shipped RIMGate, a gateway for Mobitex. Then a Mobitex point-of-sale terminal and finally with the establishment of the PCMCIA standard, a PCMCIP Mobitex modem they called Freedom. Two-way paging had already become a thing and they were ready to venture out of PoS systems. So in 1995, they took a $5 million investment to develop the RIM 900 OEM radio modem. They also developed a pager they called the Inter@ctive Pager 900 that was capable of two-way messaging the next year. Then they went public on the Toronto Stock Exchange in 1997. The next year, they sold a licensing deal to IBM for the 900 for $10M dollars. That IBM mark of approval is always a sign that a company is ready to play in an enterprise market. And enterprises increasingly wanted to keep executives just a quick two-way page away. But everyone knew there was a technology convergence on the way. They worked with Ericsson to further the technology and over the next few years competed with SkyTel in the interactive pager market. Enter The Blackberry They knew there was something new coming. Just as the founders know something is coming in Quantum Computing and run a fund for that now. They hired a marketing firm called Lexicon Branding to come up with a name and after they saw the keys on the now-iconic keyboard, the marketing firm suggested BlackBerry. They’d done the research and development and they thought they had a product that was special. So they released the first BlackBerry 850 in Munich in 1999. But those were still using radio networks and more specifically the DataTAC network. The age of mobility was imminent, although we didn’t call it that yet. Handspring and Palm each went public in 2000. In 2000, Research In Motion brought its first cellular phone product in the BlackBerry 957, with push email and internet capability. But then came the dot com bubble. Some thought the Internet might have been a fad and in fact might disappear. But instead the world was actually ready for that mobile convergence. Part of that was developing a great operating system for the time when they released the BlackBerry OS the year before. And in 2000 the BlackBerry was named Product of the Year by InfoWorld. The new devices took the market by storm and shattered the previous personal information manager market, with shares of that Palm company dropping by over 90% and Palm OS being setup as it’s own corporation within a couple of years. People were increasingly glued to their email. While the BlackBerry could do web browsing and faxing over the internet, it was really the integrated email access, phone, and text messaging platform that companies like General Magic had been working on as far back as the early 1990s. The Rise of the BlackBerry The BlackBerry was finally the breakthrough mobile product everyone had been expecting and waiting for. Enterprise-level security, integration with business email like Microsoft’s Exchange Server, a QWERTY keyboard that most had grown accustomed to, the option to use a stylus, and a simple menu made the product an instant smash success. And by instant we mean after five years of research and development and a massive financial investment. The Palm owned the PDA market. But the VII cost $599 and the BlackBerry cost $399 at the time (which was far less than the $675 Inter@ctive Pager had cost in the 1990s). The Palm also let us know when we had new messages using the emerging concept of push notifications. 2000 had seen the second version of the BlackBerry OS and their AOL Mobile Communicator had helped them spread the message that the wealthy could have access to their data any time. But by 2001 other carriers were signing on to support devices and BlackBerry was selling bigger and bigger contracts. 5,000 devices, 50,000 devices, 100,000 devices. And a company called Kasten Chase stepped in to develop a secure wireless interface to the Defense Messaging System in the US, which opened up another potential two million people in the defense industry They expanded the service to cover more and more geographies in 2001 and revenues doubled, jumping to 164,000 subscribers by the end of the year. That’s when they added wireless downloads so could access all those MIME attachments in email and display them. Finally, reading PDFs on a phone with the help of GoAmerica Communications! And somehow they won a patent for the idea that a single email address could be used on both a mobile device and a desktop. I guess the patent office didn’t understand why IMAP was invented by Mark Crispin at Stanford in the 80s, or why Exchange allowed multiple devices access to the same mailbox. They kept inking contracts with other companies. AT&T added the BlackBerry in 2002 in the era of GSM. The 5810 was the first truly convergent BlackBerry that offered email and a phone in one device with seamless SMS communications. It shipped in the US and the 5820 in Europe and Cingular Wireless jumped on board in the US and Deutsche Telekom in Germany, as well as Vivendi in France, Telecom Italia in Italy, etc. The devices had inched back up to around $500 with service fees ranging from $40 to $100 plus pretty limited data plans. The Tree came out that year but while it was cool and provided a familiar interface to the legions of Palm users, it was clunky and had less options for securing communications. The NSA signed on and by the end of the year they were a truly global operation, raking in revenues of nearly $300 million. The Buying Torndado They added web-based application in 2003, as well as network printing. They moved to a Java-based interface and added the 6500 series, adding a walkie-talkie function. But that 6200 series at around $200 turned out to be huge. This is when they went into that thing a lot of companies do - they started suing companies like Good and Handspring for infringing on patents they probably never should have been awarded. They eventually lost the cases and paid out tens of millions of dollars in damages. More importantly they took their eyes off innovating, a common mistake in the history of computing companies. Yet there were innovations. They released Blackberry Enterprise Server in 2004 then bolted on connectors to Exchange, Lotus Domino, and allowed for interfacing with XML-based APIs in popular enterprise toolchains of the day. They also later added support for GroupWise. That was one of the last solutions that worked with symmetric key cryptography I can remember using and initially required the devices be cradled to get the necessary keys to secure communications, which then worked over Triple-DES, common at the time. One thing we never liked was that messages did end up living at Research in Motion, even if encrypted at the time. This is one aspect that future types of push communications would resolve. And Microsoft Exchange’s ActiveSync. By 2005 there were CVEs filed for BlackBerry Enterprise Server, racking up 17 in the six years that product shipped up to 5.0 in 2010 before becoming BES 10 and much later Blackberry Enterprise Mobility Management, a cross-platform mobile device management solution. Those BES 4 and 5 support contracts, or T-Support, could cost hundreds of dollars per incident. Microsoft had Windows Mobile clients out that integrated pretty seamlessly with Exchange. But people loved their Blackberries. Other device manufacturers experimented with different modes of interactivity. Microsoft made APIs for pens and keyboards that flipped open. BlackBerry added a trackball in 2006, that was always kind of clunky. Nokia, Ericsson, Motorola, and others were experimenting with new ways to navigate devices, but people were used to menus and even styluses. And they seemed to prefer a look and feel that seemed like what they used for the menuing control systems on HVAC controls, video games, and even the iPod. The Eye Of The Storm A new paradigm was on the way. Apple's iPhone was released in 2007 and Google's Android OS in 2008. By then the BlackBerry Pearl was shipping and it was clear which devices were better. No one saw the two biggest threats coming. Apple was a consumer company. They were slow to add ActiveSync policies, which many thought would be the corporate answer to mobile management as group policies in Active Directory had become for desktops. Apple and Google were slow to take the market, as BlackBerry continued to dominate the smartphone industry well into 2010, especially once then-president Barack Obama strong-armed the NSA into allowing him to use a special version of the BlackBerry 8830 World Edition for official communiques. Other world leaders followed suit, as did the leaders of global companies that had previously been luddites when it came to constantly being online. Even Eric Schmidt, then chairman of google loved his Crackberry in 2013, 5 years after the arrival of Android. Looking back, we can see a steady rise in iPhone sales up to the iPhone 4, released in 2010. Many still said they loved the keyboard on their BlackBerries. Organizations had built BES into their networks and had policies dating back to NIST STIGs. Research in Motion owned the enterprise and held over half the US market and a fifth of the global market. That peaked in 2011. BlackBerry put mobility on the map. But companies like AirWatch, founded in 2003 and MobileIron, founded in 2007, had risen to take a cross-platform approach to the device management aspect of mobile devices. We call them Unified Endpoint Protection products today and companies could suddenly support BlackBerry, Windows Mobile, and iPhones from a single console. Over 50 million Blackberries were being sold a year and the stock was soaring at over $230 a share. Today, they hold no market share and their stock performance shows it. Even though they’ve pivoted to more of a device management company, given their decades of experience working with some of the biggest and most secure companies and governments in the world. The Fall Of The BlackBerry The iPhone was beautiful. It had amazing graphics and a full touch screen. It was the very symbol of innovation. The rising tide of the App Store also made it a developers playground (no pun intended). It was more expensive than the Blackberry, but while they didn’t cater to the enterprise, they wedged their way in there with first executives and then anyone. Initially because of ActiveSync, which had come along in 1996 mostly to support Windows Mobile, but by Exchange Server 2003 SP 2 could do almost anything Outlook could do - provided software developers like Apple could make the clients work. So by 2011, Exchange clients could automatically locate a server based on an email address (or more to the point based on DNS records for the domain) and work just as webmail, which was open in almost every IIS implementation that worked with Exchange. And Office365 was released in 2011, paving the way to move from on-prem Exchange to what we now call “the cloud.” And Google Mail had been around for 7 years by then and people were putting it on the BlackBerry as well, blending home and office accounts on the same devices at times. In fact, Google licensed Exchange ActiveSync, or EAS in 2009 so support for Gmail was showing up on a variety of devices. BlackBerry had everything companies wanted. But people slowly moved to that new iPhone. Or Androids when decent models of phones started shipping with the OS on them. BlackBerry stuck by that keyboard, even though it was clear that people wanted full touchscreens. The BlackBerry Bold came out in 2009. BlackBerry had not just doubled down with the keyboard instead of full touchscreen, but they tripled down on it. They had released the Storm in 2008 and then the Storm in 2009 but they just had a different kind of customer. Albeit one that was slowly starting to retire. This is the hard thing about being in the buying tornado. We’re so busy transacting that we can’t think ahead to staying in the eye that we don’t see how the world is changing outside of it. As we saw with companies like Amdahl and Control Data, when we only focus on big customers and ignore the mass market we leave room for entrants in our industries who have more mass appeal. Since the rise of the independent software market following the IBM anti-trust cases, app developers have been a bellwether of successful platforms. And the iPhone revenue split was appealing to say the least. Sales fell off fast. By 2012, the BlackBerry represented less than 6 percent of smartphones sold and by the start of 2013 that number dropped in half, falling to less than 1 percent in 2014. That’s when the White House tested replacements for the Blackberry. There was a small bump in sales when they finally released a product that had competitive specs to the iPhone, but it was shortly lived. The Crackberry craze was officially over. BlackBerry shot into the mainstream and brought the smartphone with them. They made the devices secure and work seamlessly in corporate environments and for those who could pay money to run BES or BIS. They proved the market and then got stuck in the Innovator’s Dilemna. They became all about features that big customers wanted and needed. And so they missed the personal part of personal computing. Apple, as they did with the PC and then graphical user interfaces saw a successful technology and made people salivate over it. They saw how Windows had built a better sandbox for developers and built the best app delivery mechanism the world has seen to date. Google followed suit and managed to take a much larger piece of the market with more competitive pricing. There is so much we didn’t discuss, like the short-lived Playbook tablet from BlackBerry. Or the Priv. Because for the most part, they a device management solution today. The founders are long gone, investing in the next wave of technology: Quantum Computing. The new face of BlackBerry is chasing device management, following adjacencies into security and dabbling in IoT for healthcare and finance. Big ticket types of buys that include red teaming to automotive management to XDR. Maybe their future is in the convergence of post-quantum security, or maybe we’ll see their $5.5B market cap get tasty enough for one of those billionaires who really, really, really wants their chicklet keyboard back. Who knows but part of the fun of this is it’s a living history.
/episode/index/show/thehistoryofcomputing/id/23461355
info_outline
Colossal Cave Adventure
06/02/2022
Colossal Cave Adventure
Imagine a game that begins with a printout that reads: You are standing at the end of a road before a small brick building. Around you is a forest. A small stream flows out of the building and down a gully. In the distance there is a tall gleaming white tower. Now imagine typing some information into a teletype and then reading the next printout. And then another. A trail of paper lists your every move. This is interactive gaming in the 1970s. Later versions had a monitor so a screen could just show a cursor and the player needed to know what to type. Type N and hit enter and the player travels north. “Search” doesn’t work but “look” does. “Take water” works as does “Drink water” but it takes hours to find dwarves and dragons and figure out how to battle or escape. This is one of the earliest games we played and it was marvelous. The game was called Colossal Cave Adventure and it was one of the first conversational adventure games. Many came after it in the 70s and 80s, in an era before good graphics were feasible. But the imagination was strong. The Oregon Trail was written before it, in 1971 and Trek73 came in 1973, both written for HP minicomputers. Dungeon was written in 1975 for a PDP-10. The author, Don Daglow, went on the work on games like Utopia and Neverwinter Nights Another game called Dungeon showed up in 1975 as well, on the PLATO network at the University of Illinois Champagne-Urbana. As the computer monitor spread, so spread games. William Crowther got his degree in physics at MIT and then went to work at Bolt Baranek and Newman during the early days of the ARPANET. He was on the IMP team, or the people who developed the Interface Message Processor, the first nodes of the packet switching ARPANET, the ancestor of the Internet. They were long hours, but when he wasn’t working, he and his wife Pat explored caves. She was a programmer as well. Or he played the new Dungeons & Dragons game that was popular with other programmers. The two got divorced in 1975 and like many suddenly single fathers he searched for something for his daughters to do when they were at the house. Crowther combined exploring caves, Dungeons & Dragons, and FORTRAN to get Colossal Cave Adventure, often just called Adventure. And since he worked on the ARPANET, the game found its way out onto the growing computer network. Crowther moved to Palo Alto and went to work for Xerox PARC in 1976 before going back to BBN and eventually retiring from Cisco. Crowther loosely based the game mechanics on the ELIZA natural language processing work done by Joseph Weizenbaum at the MIT Artificial Intelligence Laboratory in the 1960s. That had been a project to show how computers could be shown to understand text provided to computers. It was most notably used in tests to have a computer provide therapy sessions. And writing software for the kids or gaming can be therapeutic as well. As can replaying happier times. Crowther explored Mammoth Cave National Park in Kentucky in the early 1970s. The characters in the game follow along his notes about the caves, exploring the area around it using natural language while the computer looked for commands in what was entered. It took about 700 lines to do the original Fortran code for the PDP-10 he had at his disposal at BBN. When he was done he went off on vacation, and the game spread. Programmers in that era just shared code. Source needed to be recompiled for different computers, so they had to. Another programmer was Don Woods, who also used a PDP-10. He went to Princeton in the 1970s and was working at the Stanford AI Lab, or SAIL, at the time. He came across the game and asked Crowther if it would be OK to add a few features and did. His version got distributed through DECUS, or the Digital Equipment Computer Users Society. A lot of people went there for software at the time. The game was up to 3,000 lines of code when it left Woods. The adventurer could now enter the mysterious cave in search of the hidden treasures. The concept of the computer as a narrator began with Collosal Cave Adventure and is now widely used. Although we now have vast scenery rendered and can point and click where we want to go so don’t need to type commands as often. The interpreter looked for commands like “move”, “interact” with other characters, “get” items for the inventory, etc. Woods went further and added more words and the ability to interpret punctuation as well. He also added over a thousand lines of text used to identify and describe the 40 locations. Woods continued to update that game until the mid-1990s. James Gillogly of RAND ported the code to C so it would run on the newer Unix architecture in 1977 and it’s still part of many a BSD distribution. Microsoft published a version of Adventure in 1979 that was distributed for the Apple II and TRS-80 and followed that up in 1981 with a version for Microsoft DOS or MS-DOS. Adventure was now a commercial product. Kevin Black wrote a version for IBM PCs. Peter Gerrard ported it to Amiga Bob Supnik rose to a Vice President at Digital Equipment, not because he ported the game, but it didn’t hurt. And throughout the 1980s, the game spread to other devices as well. Peter Gerrard implemented the version for the Tandy 1000. The Original Adventure was a version that came out of Aventuras AD in Spain. They gave it one of the biggest updates of all. Colossal Cave Adventure was never forgotten, even though it was Zork was replaced. Zork came along in 1977 and Adventureland in 1979. Ken and Roberta Williams played the game in 1979. Ken had bounced around the computer industry for awhile and had a teletype terminal at home when he came across Colossal Cave Adventure in 1979. The two became transfixed and opened their own company to make the game they released the next year called Mystery House. And the text adventure genre moved to a new level when they sold 15,000 copies and it became the first hit. Rogue, and others followed, increasingly interactive, until fully immersive graphical games replaced the adventure genre in general. That process began when Warren Robinett of Atari created the 1980 game, Adventure. Robinett saw Colossal Cave Adventure when he visited the Stanford Artificial Intelligence Laboratory in 1977. He was inspired into a life of programming by a programming professor he had in college named Ken Thompson while he was on sabbatical from Bell Labs. That’s where Thompason, with Dennis Ritchie and one of the most amazing teams of programmers ever assembled, gave the world Unix and the the C programming language at Bell Labs. Adventure game went on to sell over a million copies and the genre of fantasy action-adventure games moved from text to video.
/episode/index/show/thehistoryofcomputing/id/23308790
info_outline
MySpace And My First Friend, Tom
05/14/2022
MySpace And My First Friend, Tom
Before Facebook, there was MySpace. People logged into a web page every day to write to friends, show off photos, and play music. Some of the things we still do on social networks. The world had been shifting to personal use of computers since the early days when time sharing systems were used in universities. Then came the Bulletin Board Systems of the 80s. But those were somewhat difficult to use and prone to be taken over by people like the ones who went on to found DefCon and hacking collectives. Then in the 1990s computers and networks started to get easier to use. We got tools like AOL Instant Messenger and a Microsoft knockoff called Messenger. It’s different ‘cause it doesn’t say Instant. The rise of the World Wide Web meant that people could build their own websites in online communities. We got these online communities like Geocities in 1994, where users could build their own little web page. Some were notes from classes at universities; others how to be better at dressing goth. They tried to sort people by communities they called cities, and then each member got an address number in their community. They grew fast and even went public before being acquired by Yahoo! in 1999. Tripod showed up the year after Geocities came out and got acquired by Yahoo! competitor Lycos in 1998, signaling that portal services in a pre-modern search engine world would be getting into more content to show ads to eyeballs. Angelfire was another that started in 1996 and ended up in the Lycos portfolio as well. More people had more pages and that meant more eyeballs to show ads to. No knowledge of HTML was really required but it did help to know some. The GeoCities idea about communities was a good one. Turns out people liked hanging out with others like themselves online. People liked reading thoughts and ideas and seeing photos if they ever bothered to finish downloading. But forget to bookmark a page and it could be lost in the cyberbits or whatever happened to pages when we weren’t looking at them. The concept of six agrees of Kevin Bacon had been rolling around a bit, so Andrew Weinreich got the idea to do something similar to Angelfire and the next year created SixDegrees.com. It was easy to evolve the concept to bookmark pages by making connections on the site. Except to get people into the site and signing up the model appeared to be the flip side: enter real world friends and family and they were invited to join up. Accepted contacts could then post on each others bulletin boards or send messages to one another. We could also see who our connections were connected to, thus allowing us to say “oh I met that person at a party.” Within a few years the web of contacts model was so successful that it had a few million users and was sold for over $100 million. By 2000 it was shut down but had proven there was a model there that could work. Xanga came along the next year as a weblog and social networking site but never made it to the level of success. Classmates.com is still out there as well, having been founded in 1995 to build a web of contacts for finding those friends from high school we lost contact with. Then came Friendster and MySpace in 2003. Friendster came out of the gate faster but faded away quicker. These took the concepts of SixDegrees.com where users invited friends and family but went a little further, allowing people to post on one another boards. MySpace went a little further. They used some of the same concepts Geocities used and allowed people to customize their own web pages. When some people learned HTML to edit their pages, they got the bug to create. And so a new generation of web developers was created as people learned to layout pages and do basic web programming in order to embed files, flash content, change backgrounds, and insert little DHTML or even JavaScript snippets. MySpace was co-founded by Chris DeWolfe, Uber Whitcomb, Josh Berman, and Tom Anderson while working at an incubator or software holding company called eUniverse, which was later renamed to Intermix Media. Brad Greenspan founded that after going to UCLA and then jumping headfirst into the startup universe. He created Entertainment Universe, then raised $2M in capital from Lehman Brothers, another $5M from others and bought a young site called CD Universe, which was selling Compact Disks online. He reverse merged that into an empty public shell company, like a modern SPAC works, and was suddenly the CEO of a public company, expanding into online DVD sales. Remember, these were the days leading up to the dot com bubble. There was a lot of money floating around. They expanded into dating sites and other membership programs. We’d think of monthly member fees as Monthly Recurring Revenue now, but at the time there was so much free stuff on the internet that those most sites just gave it away and built revenue streams on advertising revenues. CDs and DVDs have data on them. Data can be shared. Napster proved how lucrative that could be by then. Maybe that was something eUniverse should get into. DeWolfe created a tool called Sitegeist, which was a site with a little dating, a little instant messaging, and a little hyper localized search. It was just a school project but got him thinking. Then, like millions of us were about to do, he met Tom. Tom was a kid from the valley who’d been tinkering with computers for years, as “Lord Flathead” who’d been busted hacking as a kid before going off to the University of California at Berkeley before coming home to LA to do software QA for an online storage company. The company he worked for got acquired as a depressed asset by eUniverse in 2002, along with Josh Berman. They got matched up with DeWolfe, and saw this crazy Friendster coming out of nowhere and decided to build something like it. They had a domain they weren’t using called MySpace.com, which they were going to use for another online storage project. So they grabbed Aber Whitcomb, fired up a ColdFusion IDE and given the other properties eUniverse was sitting on had the expertise to get everything up and running fairly quickly. So they launched MySpace internally first and then had little contests to see who could get the most people to sign up. eUniverse had tens of millions of users on the other properties so they emailed them too. Within two years they had 20 million users and were the centerpiece of the eUniverse portfolio. Wanting in on what the young kids were doing these days, Rupert Murdoch and News Corporation, or NewsCorp for short, picked up the company for $580 Million in cash. It’s like an episode of Succession, right? After the acquisition of Myspace by news corporation, Myspace continued its exponential growth. Later in the year, the site started signing up 200,000 new users every day. About a year later, it was registering approx. 320,000 users each day. They localized into different languages and became the biggest website in the US. So they turned on the advertising machine, paying back their purchase price by doing $800 million in revenue back to NewsCorp. MySpace had become the first big social media platform that was always free that allowed users to freely express their minds and thoughts with millions of other users, provided they were 13 years or older. They restricted access to profiles of people younger than 16 years in such a way that they couldn’t be viewed by people over 18 years old. That was to keep sexual predators from accessing the profile of a minor. Kids turned out to be a challenge. In 2006, during extensive research the company began detecting and deleting profiles of registered sex offenders which had started showing up on the platform. Myspace partnered with Sentinel Tech Holdings Corporation to build a searchable, national database containing names, physical descriptions, and other identity details known as the Sentinel Safe which allowed them to keep track of over half a million registered sex offenders from U.S. government records. This way they developed the first national database of convicted sex offenders to protect kids on the platform, which they then provided to state attorney generals when the sex offenders tried to use MySpace. Facebook was created in 2004 and Twitter was created in 2006. They picked up market share, but MySpace continued to do well in 2007 then not as well in 2008. By 2009, Facebook surpassed Myspace in the number of unique U.S. visitors. Myspace began a rapid decline and lost members fast. Network effects can disappear as quickly as they are created. They kept the site simple and basic; people would log in, make new friends, and share music, photos, and chat with people. Facebook and Twitter constantly introduced new features for users to explore; this kept the existing users on the site and attracted more users. Then social media companies like twitter began to target users on Myspace. New and more complicated issues kept coming up. Pages were vandalized, there were phishing attacks, malware got posted to the site, and there were outages as the ColdFusion code had been easy to implement but proved harder to hyperscale. In fact, few had needed to scale a site like MySpace had in that era. Not only were users abandoning the platform, but employees at Myspace started to leave. The changes to MySpace’s executive ranks went down quicky in June 2009 by a layoff of 37.5% of its workforce reducing, the employees went down from 1,600 to 1,000. Myspace attempted to rebrand itself as primarily a music site to try and gain the audience they lost. They changed the layout to make it look more attractive but continued a quick decline just as Facebook and Twitter were in the midst of a meteoric rise. In 2011 News Corporation sold Myspace to Specific Media and Justin Timberlake for around $35 million. Timberlake wanted to make a platform where fans could go and communicate with their favorite entertainers, listen to new music, watch videos, share music, and connect with others who liked the same things. Like Geocities but for music lovers. They never really managed to turn things around. In 2016, Myspace and its parent company were acquired by Time Inc. and later Time inc. was in turn purchased by the Meredith Corporation. A few months later the news cycle on and about the platform became less positive. A hacker retrieved 427 million Myspace passwords and tried to sell them for $2,800. In 2019, Myspace accidentally deleted over 50 million digital files including photos, songs, and videos during a server migration. Everything up to 2015 was erased. In some ways that’s not the worst thing, considering some of the history left on older profiles. MySpace continues to push music today, with shows that include original content, like interviews with artists. It’s more of a way for artists to project their craft than a social network. It’s featured content, either sponsored by a label or artist, or from artists so popular or with such an intriguing story their label doesn’t need to promote them. There are elements of a social network left, but nothing like the other social networks of the day. And there’s some beauty in that simplicity. MySpace was always more than just a social networking website; it was the social network that kickstarted the web 2.0 experience we know today. Tom was everyone who joined the networks first friend. So he became the first major social media star. MySpace became the most visited social networking site in the world, often surpassing Google in number of visitors. Then the network effect moved elsewhere, and those who inherited the users analyzed what caused them to move away from MySpace and either through copying features, out innovating, or acquisition, have managed to remain dominant for over a decade. But there’s always something else right around the corner. One of the major reasons people abandoned MySpace was to be with those who thought just like them. When Facebook was only available to college kids it had a young appeal. It slowly leaked into the mainstream and my grandmother started typing the word like when I posted pictures of my kid. Because we grew up. They didn’t attempt to monetize too early. They remained stable. They didn’t spend more than they needed to keep the site going, so never lost control to investors. Meanwhile, MySpace grew to well over a thousand people to support a web property that would take a dozen to support today. Facebook may move fast and break things. But they do so because they saw what happens when we don’t.
/episode/index/show/thehistoryofcomputing/id/23108078
info_outline
Gateway 2000, and Sioux City
05/09/2022
Gateway 2000, and Sioux City
Theophile Bruguier was a fur trader who moved south out of Monreal after a stint as an attorney in Quebec before his fiancé died. He became friends with Chief War Eagle of the Yankton Sioux. We call him Chief, but he left the Santee rather than have a bloody fight over who would be the next chief. The Santee were being pushed down from the Great Lakes area of Minnesota and Wisconsin by the growing Ojibwe and were pushing further and further south. There are two main divisions of the Sioux people: the Dakota and the Lakota. There are two main ethnic groups of the Dakota, the Eastern, sometimes called the Santee and the Western, or the Yankton. After the issues with the his native Santee, he was welcomed by the Yankton, where he had two wives and seven children. Chief War Eagle then spent time with the white people moving into the area in greater and greater numbers. They even went to war and he acted as a messenger for them in the War of 1812 and then became a messenger for the American Fur Company and a guide along the Missouri. After the war, he was elected a chief and helped negotiate peace treaties. He married two of his daughters off to Theophile Bruguier, who he sailed the Missouri with on trips between St Louis and Fort Pierre in the Dakota territory. The place where Theophile settled was where the Big Sioux and Missouri rivers meet. Two water ways for trade made his cabin a perfect place to trade, and the chief died a couple of years later and was buried in what we now call War Eagle Park, a beautiful hike above Sioux City. His city. Around the same time, the Sioux throughout the Minnesota River were moved to South Dakota to live on reservations, having lost their lands and war broke out in the 1860s. Back at the Bruguier land, more French moved into the area after Bruguier opened a trading post and was one of the 17 white people that voted in the first Woodbury County election, once Wahkaw County was changed to Woodbury to honor Levi Woodbury, a former Supreme Court Justice. Bruguier sold some of his land to Joseph Leonais in 1852. He sold it to a land surveyor, Dr. John Cook, who founded Sioux City in 1854. By 1860, with the westward expansion of the US, the population had already risen to 400. Steamboats, railroads, livestock yards, and by 1880 they were over 7,000 souls, growing to 6 times that by the time Bruguier died in 1896. Seemingly more comfortable with those of the First Nations, his body is interred with Chief War Eagle and his first two wives on the bluffs overlooking Sioux City, totally unrecognizable by then. The goods this new industry brought had to cross the rivers. Before there were bridges to cross the sometimes angry rivers, ranchers had to ferry cattle across. Sometimes cattle fell off the barges and once they were moving, they couldn’t stop for a single head of cattle. Ted Waitt’s ancestors rescued cattle and sold them, eventually homesteading their own ranch. And that ranch is where Ted started Gateway Computers in 1985 with his friend Mike Hammond. Michael Dell started Dell computers in 1984 and grew the company on the backs of a strong mail order business. He went from selling repair services and upgrades to selling full systems. He wasn’t the only one to build a company based on a mail and phone order business model in the 1980s and 1990s. Before the internet that was the most modern way to transact business. Ted Waitt went to the University of Iowa in Iowa City a couple of years before Michael Dell went to the University of Texas. He started out in marketing and then spent a couple of years working for a reseller and repair store in Des Moines before he decided to start his own company. Gateway began life in 1985 as the Texas Instruments PC Network, or TIPC Network for short. They sold stuff for Texas Instruments computers like modems, printers, and other peripherals. The TI-99/4A had been released in 1979 and was discontinued a year before. It was a niche hobbyist market even by then, but the Texas Instruments Personal Computer had shipped in 1983 and came with an 8088 CPU. It was similar to an IBM PC and came with a DOS. But Texas Instruments wasn’t a clone maker and the machines weren’t fully Personal Computer compatible. Instead, there were differences. They found some success and made more than $100,000 in just a few months, so brought in Tedd’s brother Norm. Compaq, Dell, and a bunch of other companies were springing up to build computers. Anyone who had sold parts for an 8088 and used DOS on it knew how to build a computer. And after a few years of supplying parts, they had a good idea how to find inexpensive components to build their own computers. They could rescue parts and sell them to meatpacking plants as full-blown computers. They just needed some Intel chips, some boards, which were pretty common by then, some RAM, which was dirt cheap due to a number of foreign companies dumping RAM into the US market. They built some computers and got up to $1 million in revenue in 1986. Then they became an IBM-compatible personal computer when they found the right mix of parts. It was close to what Texas Instruments sold, but came with a color monitor and two floppy disk drives, which were important in that era before all the computers came with spinning hard drives. Their first computer sold for just under $2,000, which made it half what a Texas Instruments computer cost. They found the same thing that Dell had found: the R&D and marketing overhead at big companies meant they could be more cost-competitive. They couldn’t call the computers a TIPC Network though. Sioux City, Iowa became the Gateway to the Dakotas, and beyond, so they changed their name to Gateway 2000. Gateway 2000 then released an 80286, which we lovingly called the 286, in 1988 and finally left the ranch to move into the city. They also put Waitt’s marketing classes to use and slapped a photo of the cows from the ranch in a magazine that said “Computers from Iowa?” and one of the better tactics for long-term loyalty, they gave cash bonuses to employees based on their profits. Within a year, they jumped to $12 million in sales. Then $70 million in 1989, and moved to South Dakota in 1990 to avoid paying state income tax. The cow turned out to be popular, so they kept Holstein cows in their ads and even added them to the box. Everyone knew what those Gateway boxes looked like. Like Dell, they hired great tech support who seemed to love their jobs at Gateway and would help with any problems people found. They brought in the adults in 1990. Executives from big firms. They had been the first to Mae color monitors standard and now, with the release of Windows they became the first big computer seller to standardize on the platform. They released a notebook computer in 1992. The HandBook was their first computer that didn’t do well. It could have been the timing, but in the midst of a recession in a time when most households were getting computers, a low cost computer sold well and sales hit $1 billion. Yet they had trouble scaling to their ship hundreds of computers a day. They opened an office in Ireland and ramped up sales overseas. Then they went public in 1993, raising $150 million. The Wiatt’s hung on to 85% of the company and used the capital raised in the IPO to branch into other areas to complete the Gateway offering: modems, networking equipment, printers, and more support representatives. Sales in 1994 hit $2.7 billion a year. They added another support center a few hours down the Missouri River in Kansas City. They opened showrooms. They added a manufacturing plant in Malaysia. They bought Osborne Computer. They opened showrooms and by 1996 Gateway spent tens of millions a year in advertising. The ads worked and they became a household name. They became a top ten company in computing with $5 billion in sales. Dell was the only direct personal computer supplier who was bigger. They opened a new sales channel: the World Wide Web. Many still called after they looked up prices at first but by 1997 they did hundreds of millions in sales on the web. By then, Ethernet had become the standard network protocol so they introduced the E-Series, which came with networks. They bought Advanced Logic Research to expand into servers. They launched a dialup provider called gateway.net. By the late 1990s, the ocean of companies who sold personal computers was red. Anyone could head down to the local shop, buy some parts, and build their own personal computer. Dell, HP, Compaq, and others dropped their prices and Gateway was left needing a new approach. Three years before Apple opened their first store, Gateway launched Gateway Country, retail stores that sold the computer, the dialup service, and they went big fast, launching 58 stores in 26 states in a short period of time. With 2000 right around the corner, they also changed their name to Gateway, Inc. Price pressure continued to hammer away at them and they couldn’t find talent so they moved to San Diego. 1999 proved a pivotal year for many in technology. The run-up to the dot com bubble meant new web properties popped up constantly. AOL had more capital than they could spend and invested heavily into Gateway to take over the ISP business, which had grown to over half a million subscribers. They threw in free Internet access with the computers, opened more channels into different sectors, and expanded the retail stores to over 200. Some thought Waitt needed to let go and let someone with more executive experience come in. So long-time AT&T exec Jeff Weitzen, who had joined the company in 1998 took over as CEO. By then Waitt was worth billions and it made sense that maybe he could go run a cattle ranch. By then his former partner Mike Hammond had a little business fixing up cars so why not explore something new. Waitt stayed on as chairman as Weitzen reorganized the company. But the prices of computers continued to fall. To keep up, Gateway released the Astro computer in 2000. This was an affordable, small desktop that had a built-in monitor, CPU, and speakers. It ran a 400 MHz Intel Celeron, had a CD-ROM, and a 4.3 GB hard drive, with 64 Megabytes of memory, a floppy, a modem, Windows 98 Second Edition, Norton Anti-Virus, USB ports, and the Microsoft Works Suite. All this came in at $799. Gateway had led the market with Windows and other firsts they jumped on board with. They had been aggressive. The first iMac had been released in 1998 and this seemed like they were following that with a cheaper computer. Gateway Country stores grew over 400+ stores. But the margins had gotten razor thin. That meant profits were down. Waitt came back to run the company, the US Securities and Exchange Commission filed charges for fraud against Weitzen, the former controller, and the former CFO, and that raged on for years. In that time, Gateway got into TVs, cameras, MP3 players, and in 2004 acquired eMachines, a rapidly growing economy PC manufacturer. Their CEO, Wayne Inouye then came in to run Gateway. He had been an executive at The Good Guys! and Best Buy before taking the helm of eMachines in 2001, helping them open sales channels in retail stores. But Gateway didn’t get as much a foothold in retail. That laptop failure from the 1980s stuck with Gateway. They never managed to ship a game-changing laptop. Then the market started to shift to laptops. Other companies left on that market but Gateway never seemed able to ship the right device. They instead branched into consumer electronics. The dot com bubble burst and they never recovered. The financial woes with the SEC hurt trust in the brand. The outsourcing hurt the trust in the brand. The acquisition of a budget manufacturer hurt the brand. Apple managed to open retail stores to great success, while preserving relationships with big box retailers. But Gateway lost that route to market when they opened their own stores. Then Acer acquired Gateway in 2007. They can now be found at Walmart, having been relaunched as a budget brand of Acer, a company who the big American firms once outsourced to, but who now stands on their own two feed as a maker of personal computers.
/episode/index/show/thehistoryofcomputing/id/23056988
info_outline
The WYSIWYG Web
04/29/2022
The WYSIWYG Web
/episode/index/show/thehistoryofcomputing/id/22948157
info_outline
Whistling Our Way To Windows XP
04/25/2022
Whistling Our Way To Windows XP
Microsoft had confusion in the Windows 2000 marketing and disappointment with Millennium Edition, which was built on a kernel that had run its course. It was time to phase out the older 95, 98, and Millennium code. So in 2001, Microsoft introduced Windows NT 5.1, known as Windows XP (eXperience). XP came in a Home or Professional edition. Microsoft built a new interface they called Whistler for XP. It was sleeker and took more use of the graphics processors of the day. Jim Allchin was the Vice President in charge of the software group by then and helped spearhead development. XP had even more security options, which were simplified in the home edition. They did a lot of work to improve the compatibility between hardware and software and added the option for fast user switching so users didn’t have to log off completely and close all of their applications when someone else needed to use the computer. They also improved on the digital media experience and added new libraries to incorporate DirectX for various games. Professional edition also added options that were more business focused. This included the ability to join a network and Remote Desktop without the need of a third party product to take control of the keyboard, video, and mouse of a remote computer. Users could use their XP Home Edition computer to log into work, if the network administrator could forward the port necessary. XP Professional also came with the ability to support multiple processors, send faxes, an encrypted file system, more granular control of files and other objects (including GPOs), roaming profiles (centrally managed through Active Directory using those GPOs), multiple language support, IntelliMirror (an oft forgotten centralized management solution that included RIS and sysprep for mass deployments), an option to do an Automated System Recovery, or ASR restore of a computer. Professional also came with the ability to act as a web server, not that anyone should run one on a home operating system. XP Professional was also 64-bit given the right processor. XP Home Edition could be upgraded to from Windows 98, Windows 98 Second Edition, Millineum, and XP Professional could be upgraded to from any operating system since Windows 98 was released., including NT 4 and Windows 2000 Professional. And users could upgrade from Home to Professional for an additional $100. Microsoft also fixed a few features. One that had plagued users was that they had to gracefully unmount a drive before removing it; Microsoft got in front of this when they removed the warning that a drive was disconnected improperly and had the software take care of that preemptively. They removed some features users didn’t really use like NetMeeting and Phone Dialer and removed some of the themes options. The 3D Maze was also sadly removed. Other options just cleaned up the interface or merged technologies that had become similar, like Deluxe CD player and DVD player were removed in lieu of just using Windows Media Player. And chatty network protocols that caused problems like NetBEUI and AppleTalk were removed from the defaults, as was the legacy Microsoft OS/2 subsystem. In general, Microsoft moved from two operating system code bases to one. Although with the introduction of Windows CE, they arguably had no net-savings. However, to the consumer and enterprise buyer, it was a simpler licensing scheme. Those enterprise buyers were more and more important to Microsoft. Larger and larger fleets gave them buying power and the line items with resellers showed it with an explosion in the number of options for licensing packs and tiers. But feature-wise Microsoft had spent the Microsoft NT and Windows 2000-era training thousands of engineers on how to manage large fleets of Windows machines as Microsoft Certified Systems Engineers (MCSE) and other credentials. Deployments grew and by the time XP was released, Microsoft had the lions’ share of the market for desktop operating systems and productivity apps. XP would only cement that lead and create a generation of systems administrators equipped to manage the platform, who never knew a way other than the Microsoft way. One step along the path to the MCSE was through servers. For the first couple of years, XP connected to Windows 2000 Servers. Windows Server 2003, which was built on the Windows NT 5.2 kernel, was then released in 2003. Here, we saw Active Directory cement a lead created in 2000 over servers from Novell and other vendors. Server 2003 became the de facto platform for centralized file, print, web, ftp, software time, DHCP, DNS, event, messeging, and terminal services (or shared Remote Desktop services through Terminal Server). Server 2003 could also be purchased with Exchange 2003. Given the integration with Microsoft Outlook and a number of desktop services, Microsoft Exchange. The groupware market in 2003 and the years that followed were dominated by Lotus Notes, Novell’s GroupWise, and Exchange. Microsoft was aggressive. They were aggressive on pricing. They released tools to migrate from Notes to Exchange the week before IBM’s conference. We saw some of the same tactics and some of the same faces that were involved in Microsoft’s Internet Explorer anti-trust suit from the 1990s. The competition to Change never recovered and while Microsoft gained ground in the groupware space through the Exchange Server 4.0, 5.0, 5.5, 2000, 2003, 2007, 2010, 2013, and 2016 eras, by Exchange 2019 over half the mailboxes formerly hosted by on premises Exchange servers had moved to the cloud and predominantly Microsoft’s Office 365 cloud service. Some still used legacy Unix mail services like sendmail or those hosted by third party providers like GoDaddy with their domain or website - but many of those ran on Exchange as well. The only company to put up true competition in the space has been Google. Other companies had released tools to manage Windows devices en masse. Companies like Altiris sprang out of needs for companies who did third party software testing to manage the state of Windows computers. Microsoft had a product called Systems Management Server but Altiris built a better product, so Microsoft built an even more robust solution called System Center Configuration Management server, or SCCM for short, and within a few years Altiris lost so much business they were acquired by Symantec. Other similar stories played out across other areas where each product competed with other vendors and sometimes market segments - and usually won. To a large degree this was because of the tight hold Windows had on the market. Microsoft had taken the desktop metaphor and seemed to own the entire stack by the end of the Windows XP era. However, the technology we used was a couple of years after the product management and product development teams started to build it. And by the end of the XP era, Bill Gates had been gone long enough, and many of the early stars that almost by pure will pushed products through development cycles were as well. Microsoft continued to release new versions of the operating systems but XP became one of the biggest competitors to later operating systems rather than other companies. This reluctance to move to Vista and other technologies was the main reason extended support for XP through to 2012, around 11 years after it was released.
/episode/index/show/thehistoryofcomputing/id/22901609
info_outline
Windows NT 5 becomes Windows 2000
04/17/2022
Windows NT 5 becomes Windows 2000
Microsoft Windows 2000 was the successor to Windows NT 4.0, which had been released in 1997. Windows 2000 didn’t have a code name (supposedly because Jim Allchin didn’t like codenames), although its service packs did; Service Pack 1 and Windows 2000 64-bit were codenamed "Asteroid" and "Janus," respectively. 2000 began as NT 5.0 but Microsoft announced the name change in 1998, in a signal with when customer might expect the OS. Some of the enhancements were just to match the look and feel of the consumer Windows 98 counterpart. For example, the logo in the boot screens was cleaned up and they added new icons. Some found Windows 2000 to be more reliable, others claimed it didn’t have enough new features. But what it might have lacked in features from a cursory glance, Windows 2000 made up for in stability, scalability, and reliability. This time around, Microsoft had input from some of their larger partners. They released the operating system to partners in 1999, after releasing three release candidates or developer previews earlier that year. They needed to, if only so third parties could understand what items needed to be sold to customers. There were enough editions now, that it wasn’t uncommon for resellers to have to call the licensing desk at a distributor (similar to a wholesaler for packaged goods) in order to figure out what line items the reseller needed to put on a bid, or estimate. Reporters hailed it as the most stable product ever produced by Microsoft. It was also the most secure version. 2000 brought Group Policies forward from NT and enhanced what could be controlled from a central system. The old single line domain concept for managing domains was enhanced to become what Microsoft called Active Directory, a modern directory service that located resources in a database and allowed for finely grained controls of those resources. Windows 2000 also introduced NTFS 3, an Encrypted File System that was built on top of layers of APIs, each with their own controls. Still, Windows 98 was the most popular operating system in the world by then and it was harder to move people to it than initially expected. Microsoft released Windows 98 Second Edition in 1999 and then Windows Millennium Edition, or Me, in 2000. Millennium was a flop and helped move more people into 2000, even though 2000 was marketed as a business or enterprise operating system. Windows 2000 Professional was the workstation workhorse. Active Directory and other server services ran on Windows 2000 Server Edition. They also released Advanced Server and Datacenter Server for even more advanced environments, with Datacenter able to support up to 32 CPUs. Professional borrowed many features from both NT and 98 Second Edition, including the Outlook Express email client, expanded file system support, WebDAV support, Windows Media Player, WDM (Windows Driver Model), the Microsoft Management Console (MMC) for making it easier to manage those GPOs, support for new mass storage devices like Firewire, hibernation and passwords to wake up from hibernation, the System File Checker, new debugging options, better event logs, Windows Desktop Update (which gave us “Patch Tuesday”), a new Windows Installer, Windows Management Instrumentation (WMI), Plug and Play hardware (installing new hardware in Windows NT was a bit more like doing so in Unix than Windows 95), and all the transitions and animations of the Windows shell like an Explorer integrated with Internet Explorer. Some of these features were abused. We got Code Red, Nimbda, and other malware that became high profile attacks against vulnerable binaries. These were unprecedented in terms of how quickly a flaw in the code could get abused en masse. Hundreds of thousands of computers could be infected in a matter of days with a well crafted exploit. Even some of the server services were exploited such as the IIS, or Internet Information Services server. Microsoft responded with security bulletins but buffer overflows and other vulnerabilities allows mass infections. So much so that the US and other governments got involved. This wasn’t made any easier by the fact that the source code for parts of 2000 was leaked on the Internet and had been used to help find new exploits. Yet Windows 2000 was still the most secure operating system Microsoft had put out. Imagine how many viruses and exploits would have appeared on all those computers if it hadn’t of been. And within Microsoft, Windows 2000 was a critical step toward mass adoption of the far more stable, technically sophisticated Windows NT platform. It demonstrated that a technologically powerful Windows operating system could also have a user-friendly interface and multimedia capabilities.
/episode/index/show/thehistoryofcomputing/id/22819385
info_outline
The R Programming Language
04/01/2022
The R Programming Language
R is the 18th level of the Latin alphabet. It represents the rhotic consonant, or the r sound. It goes back to the Greek Rho, the Phoenician Resh before that and the Egyptian rêš, which is the same name the Egyptians had for head, before that. R appears in about 7 and a half percent of the words in the English dictionary. And R is probably the best language out there for programming around various statistical and machine learning tasks. We may use tools like Tensorflow imported to languages like python to prototype but R is incredibly performant for all the maths. And so it has become an essential piece of software for data scientists. The R programming language was created in 1993 by two statisticians Robert Gentleman, and Ross Ihaka at the University of Auckland, New Zealand. It has since been ported to practically every operating system and is available at r-project.org. Initially called "S," the name changed to "R" to avoid a trademark issue with a commercial software package that we’ll discuss in a bit. R was primarily written in C but used Fortran and since even R itself. And there have been statistical packages since the very first computers were used for math. IBM in fact packaged up BMDP when they first started working on the idea at UCLA Health Computing Facility. That was 1957. Then came SPSS out of the University of Chicago in 1968. And the same year, John Sall and others gave us SAS, or Statistical Analysis System) out of North Carolina State University. And those evolved from those early days through into the 80s with the advent of object oriented everything and thus got not only windowing interfaces but also extensibility, code sharing, and as we moved into the 90s, acquisition’s. BMDP was acquired by SPSS who was then acquired by IBM and the products were getting more expensive but not getting a ton of key updates for the same scientific and medical communities. And so we saw the upstarts in the 80s, Data Desk and JMP and others. Tools built for windowing operating systems and in object oriented languages. We got the ability to interactively manipulate data, zoom in and spin three dimensional representations of data, and all kinds of pretty aspects. But they were not a programmers tool. S was begun in the seventies at Bell Labs and was supposed to be a statistical MATLAB, a language specifically designed for number crunching. And the statistical techniques were far beyond where SPSS and SAS had stopped. And with the breakup of Ma Bell, parts of Bell became Lucent, which sold S to Insightful Corporation who released S-PLUS and would later get bought by TIBCO. Keep in mind, Bell was testing line quality and statistics and going back to World War II employed some of the top scientists in those fields, ones who would later create large chunks of the quality movement and implementations like Six Sigma. Once S went to a standalone software company basically, it became less about the statistics and more about porting to different computers to make more money. Private equity and portfolio conglomerates are, by nature, after improving the multiples on a line of business. But sometimes more statisticians in various feels might feel left behind. And this is where R comes into the picture. R gained popularity among statisticians because it made it easier to write complicated statistical algorithms without learning an entire programming language. Its popularity has grown significantly since then. R has been described as a cross between MATLAB and SPSS, but much faster. R was initially designed to be a language that could handle statistical analysis and other types of data mining, an offshoot of which we now call machine learning. R is also an open-source language and as with a number of other languages has plenty of packages available through a package repository - which they call CRAN (Comprehensive R Archive Network). This allows R to be used in fields outside of statistics and data science or to just get new methods to do math that doesn’t belong in the main language. There are over 18,000 packages for R. One of the more popular is ggplot2, an open-source data visualization package. data.table is another that performs programmatic data manipulation operations. dplyr provides functions designed to enable data frame manipulation in an intuitive manner. tidyr helps create tidier data. Shiny generates interactive web apps. And there are plenty of packages to make R easier, faster, and more extensible. By 2015, more than 10 million people used R every month and it’s now the 13th most popular language in use. And the needs have expanded. We can drop r scripts into other programs and tools for processing. And some of the workloads are huge. This led to the development of parallel computing, specifically using MPI (Message Passing Interface). R programming is one of the most popular languages used for statistical analysis, statistical graphics generation, and data science projects. There are other languages or tools for specific uses but it’s even started being used in those. The latest version, R 4.1.2, was released on 21/11/01. R development, as with most thriving open source solutions, is guided by a group of core developers supported by contributions from the broader community. It became popular because it provides all essential features for data mining and graphics needed for academic research and industry applications and because of the pluggable and robust and versatile nature. And projects like tensorflow and numpy and sci-kit have evolved for other languages. And there are services from companies like Amazon that can host and process assets from both, both using unstructured databases like NoSQL or using Jupyter notebooks. A Jupyter Notebook is a JSON document, following a versioned schema that contains an ordered list of input/output cells which can contain code, text (using Markdown), formulas, algorithms, plots and even media like audio or video. Project Jupyter was a spin-off of iPython but the goal was to create a language-agnostic tool where we could execute aspects in Ruby or Haskel or Python or even R. This gives us so many ways to get our data into the notebook, in batches or deep learning environments or whatever pipeline needs to be built based on an organization’s stack. Especially if the notebook has a frontend based on Amazon SageMaker Notebooks, Google's Colaboratory and Microsoft's Azure Notebook. Think about this. 25% of the languages lack a rhotic consonant. Sometimes it seems like we’ve got languages that do everything or that we’ve built products that do everything. But I bet no matter the industry or focus or sub-specialty, there’s still 25% more automation or instigation into our own data to be done. Because there always will be.
/episode/index/show/thehistoryofcomputing/id/22649180
info_outline
The Earliest Days of Microsoft Windows NT
03/24/2022
The Earliest Days of Microsoft Windows NT
The first operating systems as we might think of them today (or at least anything beyond a basic task manager) shipped in the form of Multics in 1969. Some of the people who worked on that then helped created Unix at Bell Labs in 1971. Throughout the 1970s and 1980s, Unix flowed to education, research, and corporate environments through minicomputers and many in those environments thought a flavor of BSD, or Berkeley Software Distribution, might become the operating system of choice on microcomputers. But the microcomputer movement had a while other plan if only in spite of the elder minicomputers. Apple DOS was created in 1978 in a time when most companies who made computers had to mail their own DOS as well, if only so software developers could built disks capable of booting the machines. Microsoft created their Disk Operating System, or MS-DOS, in 1981. They proceeded to Windows 1 to sit on top of MS-DOS in 1985, which was built in Intel’s 8086 assembler and called operating system services via interrupts. That led to poor programmers locking down points in order to access memory addresses and written assuming a single-user operating system. Then came Windows 2 in 1987, Windows 3 in 1992, and released one of the most anticipated operating systems of all time in 1995 with Windows 95. 95 turned into 98, and then Millineum in 2000. But in the meantime, Microsoft began work on another generation of operating systems based on a fusion of ideas between work they were doing with IBM, work architects had done at Digital Equipment Corporation (DEC), and rethinking all of it with modern foundations of APIs and layers of security sitting atop a kernel. Microsoft worked on OS/2 with IBM from 1985 to 1989. This was to be the IBM-blessed successor of the personal computer. But IBM was losing control of the PC market with the rise of cloned IBM architectures. IBM was also big, corporate, and the small, fledgeling Microsoft was able to move quicker. Really small companies that find success often don’t mesh well with really big companies that have layers of bureaucracy. The people Microsoft originally worked with were nimble and moved quickly. The ones presiding over the massive sales and go to market efforts and the explosion in engineering team size was back to the old IBM. OS/2 had APIs for most everything the computer could do. This meant that programmers weren’t just calling assembly any time they wanted and invading whatever memory addresses they wanted. They also wanted preemptive multitasking and threading. And a file system since by then computers had internal hard drives. The Microsoft and IBM relationship fell apart and Microsoft decided to go their own way. Microsoft realized that DOS was old and building on top of DOS was going to some day be a big, big problem. Windows 3 was closer, as was 95, so they continued on with that plan. But they started something similar to what we’d call a fork of OS/2 today. So Gates went out to recruit the best in the industry. He hired Dave Cutler from Digital Equipment to take on the architecture of the new operating system. Cutler had worked on the VMS operating system and helped lead efforts for next-generation operating system at DEC that they called MICA. And that moment began the march towards a new operating system called NT, which borrowed much of the best from VMS, Microsoft Windows, and OS/2 - and had little baggage. Microsoft was supposed to make version 3 of OS/2 but NT OS/2 3.0 would become just Windows NT when Microsoft stopped developing on OS/2. It took 12 years, because um, they had a loooooot of customers after the wild success of first Windows 3 and then Windows 95, but eventually Cutler and team’s NT would replace all other operating systems in the family with the release of Windows 2000. Cutler wanted to escape the confines of what was by then the second largest computing company in the world. Cutler worked on VMS and RSX-12 before he got to Microsoft. There were constant turf battles and arguments about microkernels and system architecture and meetings weren’t always conducive with actually shipping code. So Cutler went somewhere he could. At least, so long as they kept IBM at bay. Cutler brought some of the team from Digital with him and they got to work on that next generation of operating systems in 1988. They sat down to decide what they wanted to build, using the NS OS/2 operating system they had a starting point. Microsoft had sold Xenix and the team knew about most every operating system on the market at the time. They wanted a multi-user environment like a Unix. They wanted programming APIs, especially for networking, but different than what BSD had. In fact, many of the paths and structures of networking commands in Windows still harken back to emulating those structures. The system would be slow on the 8086 processor, but ever since the days of Xerox PARC, everyone knew Moore’s Law was real and that the processors would double in speed every other year. Especially since Moore was still at Intel and could make his law remain true with the 286 and 386 chips in the pipeline. They also wanted the operating system to be portable since IBM selected the Intel CPU but there were plenty of other CPU architectures out there as well. The original name for NT was to be OS/2 3.0. But the IBM and Microsoft relationship fell apart and the two companies took their operating systems in different directions. OS/2 became went the direction of Warp and IBM never recovered. NT went in a direction where some ideas came over from Windows 95 or 3.1 but mostly the team just added layers of APIs and focused on making NT a fully 32-bit version of Windows that could that could be ported to other platforms including ARM, PowerPC, and the DEC Alpha that Cutler had exposure to from his days at Digital. The name became Windows NT and NT began with version 3, as it was in fact the third installment of OS/2. The team began with Cutler and a few others, grew to eight and by the time it finally shipped as NT 3.1 in 1993 there were a few hundred people working on the project. Where Windows 95 became the mass marketed operating system, NT took lessons learned from the Unix, IBM mainframe, and VMS worlds and packed them into an operating system that could run on a corporate desktop computer, as microcomputers were called by then. The project cost $150 million, about the same as the first iPhone. It was a rough start. But that core team and those who followed did what Apple couldn’t in a time when a missing modern operating system nearly put Apple out of business. Cutler inspired, good managers drove teams forward, some bad managers left, other bad managers stayed, and in an almost agile development environment they managed to break through the conflicts and ship an operating system that didn’t actually seem like it was built by a committee. Bill Gates knew the market and was patient enough to let NT 3 mature. They took the parts of OS/2 like LAN Manager. They took parts of Unix like ping. But those were at the application level. The microkernel was the most important part. And that was a small core team, like it always is. The first version they shipped to the public was Windows NT 3.1. The sales people found it easiest to often say that NT was the business-oriented operating system. Over time, the Windows NT series was slowly enlarged to become the company’s general-purpose OS product line for all PCs, and thus Microsoft abandoned the Windows 9x family, which might or might not have a lot to do with the poor reviews Millennium Edition had. Other aspects of the application layer the original team didn’t do much with included the GUI, which was much more similar to Windows 3.x. But based on great APIs they were able to move faster than most, especially in that era where Unix was in weird legal territory, changing hands from Bell to Novell, and BSD was also in dubious legal territory. The Linux kernel had been written in 1991 but wasn’t yet a desktop-class operating system. So the remaining choices most business considered were really Mac, which had serious operating system issues at the time and seemed to lack a vision since Steve Jobs left the company, or Windows. Windows NT 3.5 was introduced in 1994, followed by 3.51 a year later. During those releases they shored up access control lists for files, functions, and services. Services being similar in nearly every way to a process in Unix. It sported a TCP/IP network stack but also NetBIOS for locating computers to establish a share and a file sharing stack in LAN Manager based on the Server Message Block, or SMB protocol that Barry Feigenbaum wrote at IBM in 1983 to turn a DOS computer into a file server. Over the years, Microsoft and 3COM add additional functionality and Microsoft added the full Samba with LDAP out of the University of Michigan as a backend and Kerberos (out of MIT) to provide single sign-on services. 3.51 also brought a lot of user-mode components from Windows 95. That included the Windows 95 common control library, which included the rich edit control, and a number of tools for developers. NT could run DOS software, now they were getting it to run Windows 95 software without sacrificing the security of the operating system where possible. It kinda’ looked like a slightly more boring version of 95. And some of the features were a little harder to use, like configuring a SCSI driver to get a tape drive to work. But they got the ability to run Office 95 and it was the last version that ran the old Program Manager graphical interface. Cutler had been joined by Moshe Dunie, who led the management side of NT 3.1, through NT 4 and became the VP of the Windows Operating System Division so also had responsibility for Windows 98 and 2000. For perspective, that operating system group grew to include 3,000 badged Microsoft employees and about half that number of contractors. Mark Luovsky and Lou Perazzoli joined from Digital. Jim Alchin came in from Banyan Vines. Windows NT 4.0 was released in 1996, with a GUI very similar to Windows 95. NT 4 became the workhorse of the field that emerged for large deployments of computers we now refer to as enterprise computing. It didn’t have all the animation-type bells and whistles of 95 but did perform about as well as any operating system could. It had the NT Explorer to browse files, a Start menu, for which many of us just clicked run and types cmd. It had a Windows Desktop Update and a task scheduler. They released a number of features that would take years for other vendors to catch up with. The DCOM, or Distributed Component Object Modeling and Object Linking & Embedding (or OLE) was a core aspect any developer had to learn. The Telephony API (or TAPI) allowed access to the modem. The Microsoft Transaction Server allowed developers to build network applications on their own sockets. The Crypto API allowed developers to encrypt information in their applications. The Microsoft Message Queuing service allowed queuing data transfer between services. They also built in DirectX support and already had OpenGL support. The Task Manager in NT 4 was like an awesome graphical version of the top command on Unix. And it came with Internet Explorer 2 built in. NT 4 would be followed by a series of service packs for 4 years before the next generation of operating system was ready. That was Windows 5, or more colloquially called Windows 2000. In those years NT became known as NT Workstation, the server became known as NT Server, they built out Terminal Server Edition in collaboration with Citrix. And across 6 service packs, NT became the standard in enterprise computing. IBM released OS/2 Warp version 4.52 in 2001, but never had even a fraction of the sales Microsoft did. By contrast, NT 5.1 became Windows XP and 6 became Vista in while OS/2 was cancelled in 2005.
/episode/index/show/thehistoryofcomputing/id/22533272
info_outline
Qualcomm: From Satellites to CDMA to Snapdragons
03/17/2022
Qualcomm: From Satellites to CDMA to Snapdragons
Qualcomm is the world's largest fabless semiconductor designer. The name Qualcomm is a mashup of Quality and Communications and communications has been a hallmark of the company since its founding. They began in satellite communications and today most every smartphone has a Qualcomm chip. The ubiquity of communications in our devices and everyday lives has allowed them a $182 billion market cap as of the time of this writing. Qualcomm began with far humbler beginnings. They emerged out of a company called Linkabit in 1985. Linkabit was started by Irwin Jacobs, Leonard Kleinrock, and Andrew Viterbi - all three former graduate students at MIT. Viterbi moved to California to take a job with JPL in Pasadena, where he worked on satellites. He then went off to UCLA where he developed what we now call the Viterti algorithm, for encoding and decoding digital communications. Jacobs worked on a book called Principles of Communication Engineering after getting his doctorate at MIT. Jacobs then took a year of leave to work at JPL after he met Viterbi in the early 1960s and the two hit it off. By 1966, Jacobs was a professor at the University of California, San Diego. Kleinrock was at UCLA by then and the three realized they had too many consulting efforts between them, but if they consolidated the request they could pool their resources. Eventually Jacobs and Viterbi left and Kleinrock got busy working on the first ARPANET node when it was installed at UCLA. Jerry Heller, Andrew Cohen, Klein Gilhousen, and James Dunn eventually moved into the area to work at Linkabit and by the 1970s Jacobs was back to help design telecommunications for satellites. They’d been working to refine the theories from Claude Shannon’s time at MIT and Bell Labs and were some of the top names in the industry on the work. And the space race needed a lot of this type of work. They did their work on Scientific Data Systems computers in an era before that company was acquired by Xerox. Much as Claude Shannon got started thinking of data loss as it pertains to information theory while trying to send telegraphs over barbed wire, they refined that work thinking about sending images from mars to earth. Others from MIT worked on other space projects as a part of missions. Many of those early employees were Viterbi’s PhD students and they were joined by Joseph Odenwalder, who took Viterbi’s decoding work and combined it with a previous dissertation out of MIT when he joined Linkabit. That got used in the Voyager space probes and put Linkabit on the map. They were hiring some of the top talent in digital communications and were able to promote not only being able to work with some of the top minds in the industry but also the fact that they were in beautiful San Diego, which appealed to many in the Boston or MIT communities during harsh winters. As solid state electronics got cheaper and the number of transistors more densely packed into those wafers, they were able to exploit the ability to make hardware and software for military applications by packing digital signal processors that had previously taken a Sigma from SDS into smaller and smaller form factors, like the Linkabit Microprocessor, which got Viterbi’s algorithm for encoding data into a breadboard and a chip. The work continued with defense contractors and suppliers. They built modulation and demodulation for UHF signals for military communications. That evolved into a Command Post Modem/Processor they sold, or CPM/P for short. They made modems for the military in the 1970s, some of which remained in production until the 1990s. And as they turned their way into the 1980s, they had more than $10 million in revenue. The UC San Diego program grew in those years, and the Linkabit founders had more and more local talent to choose from. Linkabit developed tools to facilitate encoded communications over commercial satellites as well. They partnered with companies like IBM and developed smaller business units they were able to sell off. They also developed a tool they called VideoCipher to encode video, which HBO and others used to do what we later called scrambling on satellite signals. As we rounded the corner into the 1990s, though, they turned their attention to cellular services with TDMA (Time-Division Multiple Access), an early alternative to CDMA. Along the way, Linkabit got acquired by a company called MACOM in 1980 for $25 million. The founders liked that the acquirer was a fellow PhD from MIT and Linkabit stayed separate but grew quickly with the products they were introducing. As with most acquisitions, the culture changed and by 1985 the founders were gone. The VideoCipher and other units were sold off, spun off, or people just left and started new companies. Information theory was decades old at this point, plenty of academic papers had been published, and everyone who understood the industry knew that digital telecommunications was about to explode; a perfect storm for defections. Qualcomm Over the course of the next few years over two dozen companies were born as the alumni left and by 2003, 76 companies were founded by Linkabit alumni, including four who went public. One of the companies that emerged included the Linkabit founders Irwin Jacobs and Andrew Viterbi, Begun in 1985, Qualcomm is also based in San Diego. The founders had put information theory into practice at Linkabit and seen that the managers who were great at finance just weren’t inspiring to scientists. Qualcomm began with consulting and research, but this time looked for products to take to market. They merged with a company called Omninet and the two released the OmniTRACS satellite communication system for trucking and logistical companies. They landed Schneider National and a few other large customers and grew to over 600 employees in those first five years. It remained a Qualcomm subsidiary until recently. Even with tens of millions in revenue, they operated at a loss while researching what they knew would be the next big thing. Code-Division Multiple Acces, or CDMA, is a technology that allows for sending information over multiple channels so users can share not just a single frequency of the radio band, but multiple frequencies without a lot of interference. The original research began all the way back in the 1930s when Dmitry Ageyev in the Soviet Union researched the theory of code division of signals at Leningrad Electrotechnical Institute of Communications. That work and was furthered during World War II by German researchers like Karl Küpfmüller and Americans like Claude Shannon, who focused more on the information theory of communication channels. People like Lee Yuk-wing then took the cybernetics work from pioneers like Norbert Weiner and helped connect those with others like Qualcomm’s Jacobs, a student of Yuk-wing’s when he was a professor at MIT. They were already working on CDMA jamming in the early 1950s at MIT’s Lincoln Lab. Another Russian named Leonid Kupriyanovich put the concept of CMDA into practice in the later 1950s so the Soviets could track people using a service they called Altai. That made it perfect for perfect for tracking trucks and within a few years was released in 1965 as a pre-cellular radiotelephone network that got bridged to standard phone lines. The Linkabit and then Qualcomm engineers had worked closely with satellite engineers at JPL then Hughes and other defense then commercial contractors. They’d come in contact with work and built their own intellectual property for decades. Bell was working on mobile, or cellular technologies. Ameritech Mobile Communications, or Advanced Mobile Phone System (AMPS) as they were known at the time, launched the first 1G network in 1983 and Vodaphone launched their first service in the UK in 1984. Qualcomm filed their first patent for CDMA the next year. That patent is one of the most cited documents in all of technology. Qualcomm worked closely with the Federal Communications Commission (FCC) in the US and with industry consortiums, such as the CTIA, or Cellular Telephone Industries Association. Meanwhile Ericsson promoted the TDMA standard as they claimed it was more standard; however, Qualcomm worked on additional patents and got to the point that they licensed their technology to early cell phone providers like Ameritech, who was one of the first to switch from the TDMA standard Ericsson promoted to CDMA. Other carriers switched to CDMA as well, which gave them data to prove their technology worked. The OmniTRACS service helped with revenue, but they needed more. So they filed for an initial public offering in 1991 and raised over $500 billion in funding between then and 1995 when they sold another round of shares. By then, they had done the work to get CDMA encoding on a chip and it was time to go to the mass market. They made double what they raised back in just the first two years, reaching over $800 million in revenue in 1996. Qualcomm and Cell Phones One of the reasons Qualcomm was able to raise so much money in two substantial rounds of public funding is that the test demonstrations were going so well. They deployed CDMA in San Diego, New York, Honk Kong, Los Angeles, and within just a few years had over a dozen carriers running substantial tests. The CTIA supported CDMA as a standard in 1993 and by 1995 they went from tests to commercial networks. The standard grew in adoption from there. South Korea standardized on CDMA between 1993 to 116. The CDMA standard was embraced by Primeco in 1995, who used the 1900 MHz PCS band. This was a joint venture between a number of vendors including two former regional AT&T spin-offs from before the breakup of AT&T and represented interests from Cox Communications, Sprint, and turned out to be a large undertaking. It was also the largest cellular launch with services going live in 19 cities and the first phones were from a joint venture between Qualcomm and Sony. Most of PrimeCo’s assets were later merged with AirTouch Cellular and the Bell Atlantic Mobile to form what we now know as Verizon Wireless. Along the way, there were a few barriers to mass proliferation of the Qualcomm CDMA standards. One is that they made phones. The Qualcomm Q cost them a lot to manufacture and it was a market with a lot of competition who had cheaper manufacturing ecosystems. So Qualcomm sold the manufacturing business to Kyocera, who continued to license Qualcomm chips. Now they could shift all of their focus on encoding bits of data to be carried over multiple radio channels to do their part in paving the way for 2G and 3G networks with the chips that went into most phones of the era. Qualcomm couldn’t have built out a mass manufacturing ecosystem to supply the world with every phone needed in the 2G and 3G era. Nor could they make the chips that went in those phones. The mid and late 1990s saw them outsource then just license their patents and know-how to other companies. A quarter of a billion 3G subscribers across over a hundred carriers in dozens of countries. They got in front of what came after CDMA and worked on multiple other standards, including OFDMA, or Orthogonal frequency-Division Multiple Access. For those they developed the Qualcomm Flarion Flash-OFDM and 3GPP 5G NR, or New Radio. And of course a boatload of other innovative technologies and chips. Thus paving the way to have made Qualcomm instrumental in 5G and beyond. This was really made possible by this hyper-specialization. Many of the same people who developed the encoding technology for the Voyager satellite decades prior helped pave the way for the mobile revolution. They ventured into manufacturing but as with many of the designers of technology and chips, chose to license the technology in massive cross-licensing deals. These deals are so big Apple sued Qualcomm recently for a billion in missed rebates. But there were changes happening in the technology industry that would shake up those licensing deals. Broadcom was growing into a behemoth. Many of their designs sent from stand-alone chips to being a small part of a SoC, or system on a chip. Suddenly, cross-licensing the ARM gave Qualcomm the ability to make full SoCs. Snapdragon has been the moniker of the current line of SoCs since 2007. Qualcomm has an ARM Architectural License and uses the ARM instruction set to create their own CPUs. The most recent incarnation is known as Krait. They also create their own Graphics Processor (GPU) and Digital Signal Processors (DSPs) known as Adreno and Hexagon. They recently acquired Arteris' technology and engineering group, and they used Arteris' Network on Chip (NoC) technology. Snapdragon chips can be found in the Samsung Galaxy, Vivo, Asus, and Xiaomi phones. Apple designs their own chips that are based on the ARM architecture, so in some ways compete with the Snapdragon, but still use Qualcomm modems like every other SoC. Qualcomm also bought a new patent portfolio from HP, including the Palm patents and others, so who knows what we’ll find in the next chips - maybe a chip in a stylus. Their slogan is "enabling the wireless industry," and they’ve certainly done that. From satellite communications that required a computer the size of a few refrigerators to battlefield communications to shipping trucks with tracking systems to cell towers, and now the full processor on a cell phone. They’ve been with us since the beginning of the mobile era and one has to wonder if the next few generations of mobile technology will involve satellites, so if Qualcomm will end up right back where they began: encoding bits of information theory into silicon.
/episode/index/show/thehistoryofcomputing/id/22483913
info_outline
The Short But Sweet History Of The Go Programming Language
03/13/2022
The Short But Sweet History Of The Go Programming Language
The Go Programming Language Go is an open-source programming language with influences from Limbo, C, APL, Modular, Oberon, Pascal, Alex, Erlang, and most importantly, C. While relatively young compared to many languages, there are over 365,000 repositories of Go projects on Github alone. There are a few reason it gained popularity so quickly: it’s fast and efficient in the right hands, simple to pick up, doesn’t have some of the baggage of some more mature languages, and the name Ken Thompson. The seamless way we can make calls from Go into C and the fact that Ken Thompson was one of the parties responsible for C, makes it seem in part like a modern web enabled language that can stretch between the tasks C is still used for all the way to playing fart sounds in an app. And it didn’t hurt that co-author Rob Pike had whelped write books, co-created UTF-8, and was part of the distributed operating system Plan 9 team at Bell Labs and had worked on the Limbo programming language there. And Robert Griesemer was another co-author. He’d begun his career studying under Niklaus Wirth, the greater of Pascal, Modula, and Oberon. So it’s no surprise that he’d go on to write compilers and design languages. Before go, he’d worked on the V8 JavaScript engine at Google and a compiler for the Java HotSpot Virtual Machine. So our intrepid heroes assembled (pun intended) at Google in 2009. But why? Friends don’t let friends write in C. Thompson had done something amazing for the world with C. But that was going on 50 years ago. And others had picked up the mantle with C++. But there were shortcomings the team wanted to address. And so Go has the ability to concatenate string variables without using a preprocessor, has many similarities to languages like BASIC from the Limbo influences, but the most impressive feature about this programming language is its support for concurrent execution. And probably the best garbage collection facility I’ve ever seen. The first version of the language wasn't released to the public and wouldn’t be for a few years. The initial compiler was written in C but over time they got to where it can be self-hosted, which is to say that Go is compiled in Go. Go is a compiled language that can run on a command line, in a browser, on the server, or even be used to compile itself. Go compiles fast and has no global variables to clutter memory. This simplicity makes it easy to read through Go code line by line without consulting any parsing tools or syntax charts. Let’s look at a quick Hello World: // A basic Go program that demonstrates "Hello World!"
package main
import "fmt"
func main() {
fmt.Println("Hello World!")
} The output would be a simple Hello World! Fairly straight forward but the power gets into more of the scripting structures - especially given that a micro service is just a lot of little functional scripts. The language itself has no connection to any other functional programming languages and does not include support for object orientation or reflection. The language consists of two parts: a parser (which processes an input file) and a bytecode interpreter, which translates all source code into machine code. Consequently, Go programs tend to compile quickly and run very efficiently because they are mainly independent of the runtime environment and can execute directly on the hardware without being interpreted by some sort of virtual machine first. Additionally, there is no need for a separate interpreter during execution since everything runs natively. The libraries and sources built using the Go programming language provide developers with a straightforward, safe, and extensibility system to build on. We have things like Go Kit, GORM, cli, Vegeta, fuzzy, Authboss, Image, Time, gg, and mgo. These can basically provide pre-built functions and APIs to hook into any old type of service or give a number of things for free. Go was well designed from the outset and while it’s evolved over the years, it hasn’t changed as much as many other languages. with the latest release being Go 1.17. 1.1 came just a couple of months after the initial release to increase how much memory could be used on 64 bit chips by about 10-fold, add detection for race conditions, added the uint for 64 bit integers. Oh and fixed a couple of issues in the compiler. 1.2 also came in 2013 and tweaked how slicing of arrays worked in a really elegant way (almost ruby-like) and allowed developers to call the runtime scheduler for non-inline calls. And added a thread limit, like the ulimit a bash would have, for 10,000 threads. And they doubled the grouting minimum size of the stack. Then the changes got smaller. This happens as every language gets more popular. The more people use it, the more havoc the developers cause when they make breaking changes. Bigger changes are contiguous models of grouting stacks in 1.3, the addition of internal packages in 1.4, a redesigned garbage collector in 1.5 when Go was moved away from C and implemented solely in Go and assembler. And 17 releases later, it’s more popular than ever. While C remains the most popular language today, Go is hovering in the top 10. Imagine, one day saying let’s build a better language for concurrent programming. And then viola; hundreds of thousands of people are using it.
/episode/index/show/thehistoryofcomputing/id/22428557
info_outline
awk && Regular Expressions For Finding Text
03/04/2022
awk && Regular Expressions For Finding Text
Programming was once all about math. And life was good. Then came strings, or those icky non-numbery things. Then we had to process those strings. And much of that is looking for patterns that wouldn’t be a need with integers, or numbers. For example, a space in a string of text. Let’s say we want to print hello world to the screen in bash. That would be the echo command, followed by “Hello World!” Now let’s say we ran that without the quotes then it would simply echo out the word Hello to the screen, given that the interpreter saw the space and ended the command, or looked for the next operator or verb according to which command is being used. Unix was started in 1969 at Bell Labs. Part of that work was The Thompson shell, the first Unix shell, which shipped in 1971. And C was written in 1972. These make up the ancestral underpinnings of the modern Linux, BSD, Android, Chrome, iPhone, and Mac operating systems. A lot of the work the team at Bell Labs was doing was shifting from pure statistical and mathematical operations to connect phones and do R&D faster to more general computing applications. Those meant going from math to those annoying stringy things. Unix was an early operating system and that shell gave them new abilities to interact with the computer. People called files funny things. There was text in those files. And so text manipulation became a thing. Lee McMahon developed sed in 1974, which was great for finding patterns and doing basic substitutions. Another team at Bell Labs that included Finnish programmer Alfred Aho, Peter Weinberger, and Brian Kernighan had more advanced needs. Take their last name initials and we get awk. Awk is a programming language they developed in 1977 for data processing, or more specifically for text manipulation. Marc Rochkind had been working on a version management tool for code at Bell and that involved some text manipulation, as well as a good starting point for awk. It’s meant to be concise and given some input, produce the desired output. Nice, short, and efficient scripting language to help people that didn’t need to go out and learn C to do some basic tasks. AWK is a programming language with its own interpreter, so no need to compile to run AWK scripts as executable programs. Sed and awk are both written to be used as one0line programs, or more if needed. But building in an implicit loops and implicit variables made it simple to build short but power regular expressions. Think of awk as a pair of objects. The first is a pattern followed by an action to take in curly brackets. It can be dangerous to call if the pattern is too wide open.; especially when piping information For example, ls -al at the root of a volume and piping that to awk $1 or some other position and then piping that into xargs to rm and a systems administrator could have a really rough day. Those $1, $2, and so-on represent the positions of words. So could be directories. Think about this, though. In a world before relational databases, when we were looking to query the 3rd column in a file with information separated by some delimiter, piping those positions represented a simple way to effectively join tables of information into a text file or screen output. Or to find files on a computer that match a pattern for whatever reason. Awk began powerful. Over time, improvements have enabled it to be used in increasingly complicated scenarios. Especially when it comes to pattern matching with regular expressions. Various coding styles for input and output have been added as well, which can be changed depending on the need at hand. Awk is also important because it influenced other languages. After becoming part of the IEEE Standard 1003.1, it is now a part of the POSIX standard. And after a few years, Larry Wall came up with some improvements, and along came Perl. But the awk syntax has always been the most succinct and useable regular expression engines. Part of that is the wildcard, piping, and file redirection techniques borrowed from the original shells. The AWK creators wrote a book called The AWK Programming Language for Addison-Wesley in 1988. Aho would go on to develop influential algorithms, write compilers, and write books (some of which were about compilers). Weinberger continued to do work at Bell before becoming the Chief Technology Officer of Hedge Fund Renaissance Technologies with former code breaker and mathematician James Simon and Robert Mercer. His face led to much love from his coworkers at Bell during the advent of digital photography and hopefully some day we’ll see it on the Google Search page, given he now works there. Brian Kernighan was a contributor to the early Multics then Unix work, as well as C. In fact, an important C implementation, K&R C, stands for Kernighan and Ritchie C. He coauthored The C Programming Language ands written a number of other books, most recently on the Go Programming Language. He also wrote a number of influential algorithms, as well as some other programming languages, including AMPL. His 1978 description of how to manage memory when working with those pesky strings we discussed earlier went on to give us the Hello World example we use for pretty much all introductions to programming languages today. He worked on ARPA projects at Stanford, helped with emacs, and now teaches computer science at Princeton, where he can help to shape the minds of future generations of programming languages and their creators.
/episode/index/show/thehistoryofcomputing/id/22329959