Managerial bureaucracy becomes abusive to the engineer class (1940-1970)
As of 1932, the majority of these corporations were, in all practicality, no longer controlled by their majority shareholders, classified by economists as “management-controlled.” The management fad which became known as “separation of ownership and control” spread throughout the major public corporations.
The moral hazards of management-controlled companies became increasingly obvious as the 1930s wore on. Management-controlled companies were run by executives which, despite not owning many shares, eventually achieved “self-perpetuating positions of control” of policies, because they are able to manipulate the boards of directors through proxies and majority shareholder votes. These machinations sometimes created high levels of conflict. In the early 1940s, the idea emerged that this structural divide in the corporate world was being mimicked in the social and political worlds, with a distinct elite “management class” emerging in society.
Institutional economists drew a distinction between the management class and the class of “technical operators” (the people doing the work, in many cases engineers and technicians). The managerial elite consisted of the “analysts” or “specialists” who acted as the bureaucratic planners, budgetary allocators, and non-technical managers.
A strange power dynamic emerged between the analysts and the technical staff in the computer companies which had emerged between 1957 and 1969; this dynamic was studied by industrial economists in both the UK and US. They found that the analysts jockeyed for power, creating conflict. They won favor and influence over the company by expanding their divisions, creating opportunities to hire more direct reports, or to win a new promotion, a tactic known as “‘empire building.” The overall effect on the organization was misallocation of resources and incredible pressure to grow. Sales and development cycles were persistently rushed. The computer analysts’ slogan became, ‘if it works, it’s obsolescent.’” The analysts had ‘a vested interest in change.’”
This dynamic had created dysfunction. Managers used a variety of social tactics to enforce their will and agenda, in spite of technical realities, reflecting Veblen’s observation about “ceremonial” institutions 75 years before. Documented tactics included:
Organizational inertia:
New and threatening ideas are blocked with advice “idea killers" including: "the boss won't like it," "it's not policy," "I don't have the authority," "it's never been tried," "we've always done it that way," and "why change something that works?"
Budget games:
“Foot in the door,” where a new program is sold in modestly, concealing its real magnitude; “Hidden ball,” where a politically unattractive program is concealed within an attractive one; “Divide and conquer,” where approval of a budget request is sought from more than one supervisor; “It's free,” where it is argued that someone else will pay for the project so the organization might as well approve it; “Razzle-dazzle,” where a request is supported with voluminous data, but arranged in such a way that their significance is not clear; “Delayed Buck,” where deliverables are submitted late, with the argument that the budget guidelines require too much detailed calculation; and many others.
These tales from the 1960s anticipate the emergence of the popular cartoon Dilbert in the 1990s, which skewered absurd managerial behavior. Its author, Scott Adams, had worked as a computer programmer and manager at Pacific Bell from 1986 to 1995.
Group identity develops amongst professional technologists (1980-2000)
The dictatorial behavior of the management class belied the true balance of power in technical organizations.
In the 1980s, the entire weight of many industrial giants rested upon its technologists. But their role put them in a strange position, at odds with the rest of their organization. Placed at the margins of the organization, closest to the work, they were removed from the C-suite and its power plays. Not working with executives directly, the technologists identified far less with the heads of the company than the managers, who directly reported to C-suite.
The technologists’ work was enjoyable to them, but opaque to the rest of the organization. A power dynamic emerged between the technical operators and the rest of the company; their projects were difficult to supervise, and proceeded whimsically, in ways that reflected the developers’ own interests.
Their power to work this way originated in their critical skills. These skills act as a wedge within organizations, earning technical operators considerable freedom of direction. The efficacy of this wedge increased when the technical operator provided a skill which was in great demand, affording them job mobility. In this instance, their dependence on the organization was reduced. Company ideology was typically not a strong force amongst technologists, in comparison to “professional ideology,” or the belief in the profession and its norms. The elite technologists were becoming outsiders within their own companies.
Instead of loyalty to company or CEO, technologists developed, as a professional goal, loyalty to the end-user or client. A company’s technologists were focused on the needs of the existing customer, while the analysts and managers (whose work did not deal directly with the end-user) supported more abstract goals like efficiency and growth.
The hacker movement emerges
The hacker movement had originated amongst software-makers at MIT in the 1960s. Perhaps seen as an antidote to the managerial dysfunction inside the older corporate tech companies, the hacker movement’s focus on practical, useful, and excellent software spread rapidly across the country in the 1980s and 1990s. MIT software activist Richard Stallman described hackers as playful but diligent problem-solvers who prided themselves on their individual ingenuity:
“What they had in common was mainly love of excellence and programming. They wanted to make their programs that they used be as good as they could. They also wanted to make them do neat things. They wanted to be able to do something in a more exciting way than anyone believed possible and show ‘Look how wonderful this is. I bet you didn't believe this could be done.’ Hackers don’t want to work, they want to play.”
At a conference in 1984, a hacker who had gone to work at Apple to build the Macintosh described hacker status as follows: “Hackers can do almost anything and be a hacker. It’s not necessarily high tech. I think it has to do with craftsmanship and caring about what you’re doing.”
The hacker movement is not unlike the Luddite movement of the early 19th century, in which cotton and wool artisans in central England rose up to destroy the Jaquard loom which threatened to automate them. Unlike the Luddites, who proposed no better alternative to the loom, hackers came up with another approach to making software which has since produced superior products to their commercial alternatives. By using the Internet to collaborate, groups of volunteer developers have come to produce software that rivaled the products of nation states and corporations.
New Jersey style emerges
The “New Jersey style” of hacking was originated by Unix engineers at AT%story%T in suburban New Jersey. AT%story%T had lost an antitrust settlement in 1956 which precluded it from entering the computer business; thus it was free to circulate the computer operating system it had built, called Unix, to other private companies and research institutions throughout the 1970s. The source code was included, and these institutions regularly modified it to run on their particular minicomputers. Hacking Unix became a cultural phenomenon within R%story%D departments around the US.
Unix was rewritten for personal computers by several groups of developers. Linus Torvalds created his own version, “Linux,” and distributed it for free, just as AT%story%T had done with Unix. (As we will show, Linux has become enormously successful.) The approach taken by Torvalds’ and other Unix hackers uses playfulness as an energizing force to build useful (if difficult) free software projects. The Finnish computer scientist and philosopher Pekka Himanen wrote at the time: “To do the Unix philosophy right, you have to be loyal to excellence. You have to believe that software is a craft worth all the intelligence and passion you can muster.“
R%story%D developers realize “Worse is Better”
Out of New Jersey style, software engineers developed a set of ad-hoc design principles that went against the perfectionism of institutionalized software. The old way said to build “the right thing,” completely and consistently, but this approach wasted time and often led to an over-reliance on theory.
Written during the early 1980s by Richard Gabriel and published by Netscape Navigator engineer Jamie Zawinski in 1991, the “worse-is-better” philosophy boiled down the best of New Jersey style and hacker wisdom. It was seen as a practical improvement on the MIT-Stanford hacker approach. Much like the MIT ethic, worse-is-better values excellence in software. But unlike MIT-Stanford, the worse-is-better approach redefines “excellence” in a way that prioritizes positive real-world user feedback and adoption over theoretical ideals.
Worse-is-better holds that, so long as the design of the initial program is a clear expression of a solution to a specific problem, then it will take less time and effort to implement a “good” version initially, and adapt it to new situations, than it will to build a “perfect” version straight away. Releasing software to users early and improving a program often is sometimes called “iterative” development.
Iterative development allows software to spread rapidly and benefit from real-world reactions from users. Programs released early and improved often become successful long before “better” versions written in the MIT approach have a chance to be deployed. With two seminal papers in 1981 and 1982, the concept of “first-mover advantage” emerged in the software industry around the same time that Gabriel was formalizing his ideas about why, in networked software, “worse is better.”
The logic of worse-is-better prioritizes viral growth over fit and finish. Once a “good” program has spread widely, there will be many users with an interest in improving its functionality and making it excellent. An abbreviated version of the principles of “worse is better” are below. They admonish developers to avoid doing what is conceptually pleasing (“the right thing”) in favor of doing whatever results in practical, functional programs (emphasis added):
Simplicity:
This is the most important consideration in a design.
Correctness:
The design must be a correct solution to the problem. It is slightly better to be simple than correct.
Consistency:
Consistency can be sacrificed for simplicity in some cases, but it is better to drop those parts of the design that deal with less common circumstances than to introduce either implementational complexity or inconsistency.
Completeness:
The design must cover as many important situations as is practical. Completeness can be sacrificed in favor of any other quality. In fact, completeness must be sacrificed whenever implementation simplicity is jeopardized.
These conceptual breakthroughs must have been exciting to the technologists of the early 1980s. But the excitement would soon be disrupted by rapid changes in business.
The shareholders use hostile takeovers to clamp down on everyone
The hacker-centric environment inside universities and large research corporations collapsed, and researchers at places like the MIT AI Lab were poached away by venture capitalists to continue their work, but in a proprietary setting. The hostile take-over trend had begun a decade before in the UK, where clever investors began noticing that many of the family-run businesses were no longer majority owned by their founding families. Financiers like Jim Slater and James Goldsmith quietly bought up shares in these companies, eventually wrestling enough control to break up and sell off units of the company. This became known as “asset stripping,” and we will return to this topic in Section VII of this essay.
In the 1980s, American bankers hit upon a way finance takeovers at massive scale by floating so-called junk bonds, then busting up the target company and reaping massive rewards from the sale of the parts. In this way, managerial capitalism eventually lost its hold over business, and became a servant of the capital markets.
“Activist investors” came to represent shareholder interests, and took action to fire and hire C-suite executives who would maximize share price. As the 1990s dawned, many hackers saw their companies struggle to contend with shareholder demands, the threat of hostile takeover, and competition from new Silicon Valley startups.
As tech companies moved faster, they developed ways for management to enforce policy and resource allocation. Microsoft and others adopted a rigorous “stack ranking” system whereby employees were assigned numerical scores on regular intervals using a “performance review” process, in order to determine promotions, bonuses, and team assignments. A certain percentage of bottom-ranking employees were fired. This system is still used by tech companies today, but Microsoft abandoned it in 2013. Google adopted stack ranking recently to establish eligibility for promotions, but does not fire poorly-scoring employees. Stack ranking systems are widely hated for the uncomfortable power dynamics they create.
Today, investors demand from their companies precise predictions about each quarter’s profitability, and less concern is paid to capital investment. Tesla is one notable technology company which has articulated the way quarterly guidance and short-termism diminish a high-tech company’s long-term prospects. According to the Business Roundtable, a corporate alliance chaired by Chase Bank CEO Jamie Dimon, quarterly guidance has become “detrimental long term strategic investments.”
Summary
In this section, we have looked at the ways that 1940s-era management make life unpleasant for high-tech workers, and how these patterns persisted into the 1990s, disenfranchising technical workers. We’ve shown a strong “guild” identity developed which transcends loyalty to the employer. We’ve associated this identity with the growth of hacker culture and its principles.
Next, we will explore how antipathy towards the management class grew into a wider suspicion of all institutional oversight, and how their struggle to get out from under such oversight acquired a moral dimension. We will examine why hackers looked to cyberspace and cryptography for sanctuary, with a determination to build new tools outside the purview of the management class. We will consider the surprising success of free software tools produced by hackers, and consider the ways that corporate employers have alternately fought, and also tried to emulate, hacker methodology. Finally, we will encounter Bitcoin as the realization of many hacker ambitions in a single network.
Some people say that Ether’s price may reach $0 for the first time, while some claim that it will recover. Some analysts even predict that Ethereum price could rise to $1,000 by the end of the year as developers improve the platform and solve some of its scaling issues.Litecoin, however, uses the scrypt algorithm – originally named as s-crypt, but pronounced as ‘script’. This algorithm incorporates the SHA-256 algorithm, but its calculations are much more serialised than those of SHA-256 in bitcoin. Scrypt favours large amounts of high-speed RAM, rather than raw processing power alone. As a result, scrypt is known as a ‘memory hard problem‘.coinmarketcap bitcoin The Ledger Nano X is the second generation hardware wallet from Ledger, a French company launched in 2014. Ledger’s first cryptocurrency product, the Ledger Nano S, was one of the first hardware wallets on the market and dominated the space for a number of years.lavkalavka bitcoin tether обзор reddit cryptocurrency trezor ethereum clame bitcoin bitcoin central bitcoin x2 faucet cryptocurrency