I was one of those who saw the TV adaptation of Robert J. Sawyer's Flashforward before I read the book.
But I did get to it not long afterward. Sawyer's tale of an international team of scientists at CERN coping with the implications of the totally unexpected event--that two minutes' of unconsciousness in which everyone got a glimpse of their future--was, for me, an appealingly old-fashioned science fiction novel. Reasonably compact, comparatively lacking in the bulkiness and clunkiness that has made me read less and less recent fiction of any kind, it was genuinely interested in its "What if" and straightfoward in its storytelling, as idea-driven fiction generally ought to be. Enough so that Sawyer didn't hesitate to follow his characters' trains of thought about the issue at hand, or permit them to have "explicit dialogues" in the Wellsian (or Shavian) manner. (I, for one, must admit admit I am fond of such dialogue, much fonder of it than the tenth rate Flaubert to which the advocates of "good form" expect us to aspire.) The kind of thing that, with so many people less inclined to it, encouraged me in the view that science fiction was waning as a distinct genre.
Reading the book so short a time after seeing the show I found myself inevitably drawing comparisons between one and the other. On the ABC version we got not an international team working at a particle collider outside Geneva, but a thoroughly Americanized cast of characters and setting, and these turned law enforcement types tackling an international conspiracy, the intellectual interest of the tale cast aside in favor of conventional thriller mechanics and soap opera, and mawkishness about Big Collective Moments like hack journalists write endless amounts of drivel about.
It was predictable that his idea would be forced to fit into the conventions of American prime time network television, and cease to be recognizable in the process. And there have certainly been worse shows. But I preferred the book all the same.
Thursday, July 19, 2018
Remember Meg?
No, probably not.
But I do. I remember it because back in the '90s the terms of the publication of Steve Alten's Meg was the kind of media sensation the rags-to-riches story loving press likes to blow way out of proportion at every opportunity--a first novel by an unknown getting them a seven figure advance (impressive numbers today, even more impressive back in those comparatively uninflated, pre-J.K. Rowling, pre-E.L. James days). The big money was tied up with plans for a big movie, and why not, when Meg, short for the Megalodon that was at the center of the story, looked like it could combine the underwater terror of Jaws with the scale and sci-fi exoticism of Jurassic Park.
But the book's sales did not live up to the blockbuster expectations, while the movie plans went down quick into development hell.
For twenty years.
During which, as Jan de Bont and Guillermo del Toro and lots of other people whose names you know better than Alten's were attached and detached from the project, the summer blockbuster of the monster/disaster type went from being a seasonal treat to a week-to-week, year-round thing, not just on the big screen but the small, where Syfy supplied us with an endless string of deliberately bad movies about the theme (mostly because it's an easier thing to do than deliberately make good movies).
Hearing "This time they're serious" I'd think "I'll believe it when I see it," but last year I heard a report that the film actually was shooting, with a $150 million budget and Jason Statham in the lead (I don't think he's been that before in a megabudget movie like this one) as a Sino-American coproduction.
It seemed real enough this time.
I also heard that it would be coming out in March, which it didn't, instead bumped to August 10, as the commercial I caught on TV last week indicated.
Of course, getting one's release date bumped by five months is not a good sign. Still less is it a good sign when the new date is in August--traditionally a "dump month" where the competition is less intense and not too much expected from the receipts.
But then one might imagine that this Chinese coproduction (due to come out in China the same weekend as in the U.S.) is being timed to take advantage of China's notorious late summer blacking out of its film market, giving its domestic productions a chance to clean up, and that it will help the numbers in that country--by itself, quite enough to make or break many a big production. (And as even the quickest review of the numbers over at Box Office Mojo shows, monster movies certainly play well in the Chinese market. The Monster Hunt franchise, The Great Wall--while Hollywood's own Rampage did much better business in China than at home.1) And here in the States August sometimes produces a winner. (Guardians of the Galaxy proved a surprise hit in that period a few years back.)
What's your guess on how the movie will do?
1. Rampage pulled in $99 million in America (yes, just shy of the $100 million mark), but took in $156 million in China.
But I do. I remember it because back in the '90s the terms of the publication of Steve Alten's Meg was the kind of media sensation the rags-to-riches story loving press likes to blow way out of proportion at every opportunity--a first novel by an unknown getting them a seven figure advance (impressive numbers today, even more impressive back in those comparatively uninflated, pre-J.K. Rowling, pre-E.L. James days). The big money was tied up with plans for a big movie, and why not, when Meg, short for the Megalodon that was at the center of the story, looked like it could combine the underwater terror of Jaws with the scale and sci-fi exoticism of Jurassic Park.
But the book's sales did not live up to the blockbuster expectations, while the movie plans went down quick into development hell.
For twenty years.
During which, as Jan de Bont and Guillermo del Toro and lots of other people whose names you know better than Alten's were attached and detached from the project, the summer blockbuster of the monster/disaster type went from being a seasonal treat to a week-to-week, year-round thing, not just on the big screen but the small, where Syfy supplied us with an endless string of deliberately bad movies about the theme (mostly because it's an easier thing to do than deliberately make good movies).
Hearing "This time they're serious" I'd think "I'll believe it when I see it," but last year I heard a report that the film actually was shooting, with a $150 million budget and Jason Statham in the lead (I don't think he's been that before in a megabudget movie like this one) as a Sino-American coproduction.
It seemed real enough this time.
I also heard that it would be coming out in March, which it didn't, instead bumped to August 10, as the commercial I caught on TV last week indicated.
Of course, getting one's release date bumped by five months is not a good sign. Still less is it a good sign when the new date is in August--traditionally a "dump month" where the competition is less intense and not too much expected from the receipts.
But then one might imagine that this Chinese coproduction (due to come out in China the same weekend as in the U.S.) is being timed to take advantage of China's notorious late summer blacking out of its film market, giving its domestic productions a chance to clean up, and that it will help the numbers in that country--by itself, quite enough to make or break many a big production. (And as even the quickest review of the numbers over at Box Office Mojo shows, monster movies certainly play well in the Chinese market. The Monster Hunt franchise, The Great Wall--while Hollywood's own Rampage did much better business in China than at home.1) And here in the States August sometimes produces a winner. (Guardians of the Galaxy proved a surprise hit in that period a few years back.)
What's your guess on how the movie will do?
1. Rampage pulled in $99 million in America (yes, just shy of the $100 million mark), but took in $156 million in China.
Review: Trojan Odyssey, by Clive Cussler
New York: Putnam, 2003, pp. 496.
WARNING: SPOILERS AHEAD
Clive Cussler's Trojan Odyssey marks the end of an era in many ways. Not only is it the case that this is the last Dirk Pitt novel on which Clive Cussler's name appears without a coauthor. It is also the first novel to include Dirk Jr. and Summer in the adventure, just as the single, adventurous Dirk Sr. puts that part of his life behind him, marrying his longtime girlfriend Loren Smith and taking over the directorship of NUMA from Admiral Sandecker. These changes of life will pretty much keep him out of the field (in both those ways), and pave the way for the principal "Dirk Pitt" of the novels to be the son from this point on.
Naturally I would like to be able to say that the sequence of "Dirk Pitt Sr." novels goes out strong. Alas, the first time I tried to read it I didn't make it past page one hundred and fifty. I suppose that part of the problem was that I had become less forgiving of Cussler's literary weaknesses in the decade since I started the series. However, it is also the case that this part of the book--which consists mostly of an ocean resort and the Pitt twins being inconveniently in the way of a hurricane--comes across as both overfamiliar to readers of his previous books, and agonizingly slow.
Hoping that the rest of the book would be better I later returned to it and pressed on. Having done so I can say that the pace and interest do pick up in the second half of the story. Still, while the book gets better, it does not get very much better. Despite being shorter than many of the preceding Pitt epics, ending on just page 485 in the hardback edition, it felt overlong, with many later bits excessively drawn out.
Moreover, that sense of overfamiliarity remained. Cussler's previous novels are not merely referenced (as would be appropriate in a transitional work like this), but recycled. There is a lot of Sahara here in particular--for instance, in the spread of conspicuous, rapidly spreading, ocean-killing pollution from a mysterious source ("brown crud" in the Caribbean), and the dispatch of Pitt, Al Giordino and Rudi Gunn aboard a disguised high-tech vessel to check it out early in the story. Same as in that other novel, the subsequent adventure involves a river journey, an old colonial fortress, an ecologically destructive high-tech facility incorporating cutting-edge energy technology, action underground, and a party of foreigners enslaved by the baddie.
Where the book does not reuse familiar elements it is often simply hokey--as with the holographic pirates Pitt encounters off the coast of Nicaragua, which as a deception intended to keep people from the area would embarrass the clumsiest Scooby-Doo villain; or the unmasking of a major villain in the middle of a Congressional hearing, which likewise comes across as something out of a Scooby-Doo episode. One does not expect, or usually get, much logic in the schemes of supervillains in novels like these, but the connection of their revolutionary technological breakthrough with a plan to change the climate of much of the Earth seemed especially shaky. At the same time, while the novel's exercise in fringe history is one of Cussler's more intriguing--an alternative explanation of the true history behind the legend of the Trojan War based on the work of Iman Wilkens--we never quite get clarity on the meaning of of the revelations (like why a Druidess is buried in an elaborate tomb in the West Indies).
The handling of these elements also leaves much to be desired, Dirk and Al's Central American adventure paling next to what we saw in earlier installments (not least, the West African adventure Cussler recycled so much of here). When the narration remarks Pitt and Giordino feeling more tired than they can remember ever being, this comes across as a reflection of their aging bodies, rather than the author's topping his earlier efforts. It may also be that advancing age is the reason for the painful slowness of the protagonists to grasp what became obvious to the reader--the secret the villain conceals behind the unusual attire, the source of the brown crud. It does not help that NUMA's next generation--Dirk Jr. and Summer--have little interest in themselves, and what Cussler does with them this time around, at least, is not terribly interesting either. Nor does it contribute meaningfully to the sense of transition the rest of the book seems to strive for. Rather than their being given a chance to come into their own, which would give us a sense of the torch being passed, they come off as marginal and hapless in the tale's bigger events.
Cussler partially redeems himself in the final chapter, which manages to be suitably charming, and takes some of the sting out of the disappointment in what came before--but just as Loren Smith may feel that she has waited longer than she should have to make her longtime relationship with Dirk legal, the reader may come away feeling that this should all have happened a few novels earlier.
WARNING: SPOILERS AHEAD
Clive Cussler's Trojan Odyssey marks the end of an era in many ways. Not only is it the case that this is the last Dirk Pitt novel on which Clive Cussler's name appears without a coauthor. It is also the first novel to include Dirk Jr. and Summer in the adventure, just as the single, adventurous Dirk Sr. puts that part of his life behind him, marrying his longtime girlfriend Loren Smith and taking over the directorship of NUMA from Admiral Sandecker. These changes of life will pretty much keep him out of the field (in both those ways), and pave the way for the principal "Dirk Pitt" of the novels to be the son from this point on.
Naturally I would like to be able to say that the sequence of "Dirk Pitt Sr." novels goes out strong. Alas, the first time I tried to read it I didn't make it past page one hundred and fifty. I suppose that part of the problem was that I had become less forgiving of Cussler's literary weaknesses in the decade since I started the series. However, it is also the case that this part of the book--which consists mostly of an ocean resort and the Pitt twins being inconveniently in the way of a hurricane--comes across as both overfamiliar to readers of his previous books, and agonizingly slow.
Hoping that the rest of the book would be better I later returned to it and pressed on. Having done so I can say that the pace and interest do pick up in the second half of the story. Still, while the book gets better, it does not get very much better. Despite being shorter than many of the preceding Pitt epics, ending on just page 485 in the hardback edition, it felt overlong, with many later bits excessively drawn out.
Moreover, that sense of overfamiliarity remained. Cussler's previous novels are not merely referenced (as would be appropriate in a transitional work like this), but recycled. There is a lot of Sahara here in particular--for instance, in the spread of conspicuous, rapidly spreading, ocean-killing pollution from a mysterious source ("brown crud" in the Caribbean), and the dispatch of Pitt, Al Giordino and Rudi Gunn aboard a disguised high-tech vessel to check it out early in the story. Same as in that other novel, the subsequent adventure involves a river journey, an old colonial fortress, an ecologically destructive high-tech facility incorporating cutting-edge energy technology, action underground, and a party of foreigners enslaved by the baddie.
Where the book does not reuse familiar elements it is often simply hokey--as with the holographic pirates Pitt encounters off the coast of Nicaragua, which as a deception intended to keep people from the area would embarrass the clumsiest Scooby-Doo villain; or the unmasking of a major villain in the middle of a Congressional hearing, which likewise comes across as something out of a Scooby-Doo episode. One does not expect, or usually get, much logic in the schemes of supervillains in novels like these, but the connection of their revolutionary technological breakthrough with a plan to change the climate of much of the Earth seemed especially shaky. At the same time, while the novel's exercise in fringe history is one of Cussler's more intriguing--an alternative explanation of the true history behind the legend of the Trojan War based on the work of Iman Wilkens--we never quite get clarity on the meaning of of the revelations (like why a Druidess is buried in an elaborate tomb in the West Indies).
The handling of these elements also leaves much to be desired, Dirk and Al's Central American adventure paling next to what we saw in earlier installments (not least, the West African adventure Cussler recycled so much of here). When the narration remarks Pitt and Giordino feeling more tired than they can remember ever being, this comes across as a reflection of their aging bodies, rather than the author's topping his earlier efforts. It may also be that advancing age is the reason for the painful slowness of the protagonists to grasp what became obvious to the reader--the secret the villain conceals behind the unusual attire, the source of the brown crud. It does not help that NUMA's next generation--Dirk Jr. and Summer--have little interest in themselves, and what Cussler does with them this time around, at least, is not terribly interesting either. Nor does it contribute meaningfully to the sense of transition the rest of the book seems to strive for. Rather than their being given a chance to come into their own, which would give us a sense of the torch being passed, they come off as marginal and hapless in the tale's bigger events.
Cussler partially redeems himself in the final chapter, which manages to be suitably charming, and takes some of the sting out of the disappointment in what came before--but just as Loren Smith may feel that she has waited longer than she should have to make her longtime relationship with Dirk legal, the reader may come away feeling that this should all have happened a few novels earlier.
Review: The Rise of the Novel, by Ian Watt
Berkeley, CA: The University of California Press, 1957, pp. 319.
It is a commonplace that the literary genre of the novel is distinguished by its telling about the lives of everyday individuals in everyday circumstances, in a manner distinct in two ways--its plainness of style, and its attentiveness to the "inner lives" of its characters.
In his classic discussion of the matter in The Rise of the Novel Ian Watt does far, far more than present that view, instead getting to the root of the matter. As he makes clear the ancient-Medieval tendency had been to think of the "universal" and "timeless" as real, and particularities as ephemeral and meaningless. That earlier logic had authors striving to portray the universal above all, serving up characters who are "characteristic" types down to their names, and time and place often conceived imprecisely, with the authors' ornamentation of the portrayal the principal object and measure of their skill.
By contrast, this era saw writers aspiring to tell the stories of particular individuals, in particular places and times--and indeed, the unfolding of the events of their lives through cause and effect sequences, over time. That shift, in turn, was the reason for the plainness of style--the prioritization of a denotative, descriptive usage of language, with the lavishness in detailing taking the place of the old lavishness of ornament. He stresses, too, how this came together in a genre of private accounts, read privately, permitting an "intimacy" unattainable in other, publicly performed and publicly enjoyed styles of work (plays, poems) that both extended the bounds of what seemed permissible in art, and allowed a new intensity of audience identification with protagonists.
All of these features--the strong sense of particularity, time and cause and effect; the use of language as description rather than ornament; the private, "intimate" novel-reading experience--seem implausible outside the rise of a rationalistic, individualistic, science-touched Modernity that afforded individuals a meaningful range of economic choices (a capitalism, increasingly thoroughgoing and increasingly industrialized), and sanctioned them ideologically (Puritan ideas about salvation, secular Enlightenment thought). And indeed, Watt's classic is best remembered for its stress on the historical context in which the novel emerged, which reflects a good deal of what Mills was to call "the sociological imagination."
Appropriately the book begins with a good deal of historical background, specifically two chapters regarding the philosophical developments underlying the rise of what we think of as "realism," and a reconstruction of the eighteenth century reading public from the concrete facts of the era. In these the territory ranges from the epistemology of John Locke to the nitty-gritty of how many and who would have had the time and money to buy this kind of reading material and a place in which to read it in that private way he described (delving into prices and incomes and the rest). Afterward, when turning to the founding authors and works themselves--Daniel Defoe's Robinson Crusoe and Moll Flanders, Samuel Richardson's Pamela and Clarissa, Henry Fielding's Tom Jones--he continued to draw on historians' understanding of aspects of the period ranging from the rise of Grub Street to the rise of suburbia, and how these factored into the authors' creative work by way of their personal backgrounds, and quite profitably. (Wary as I tend to be of the biographical approach, I found it impossible to come away from this study thinking it accidental that Richardson, a retiring, low-born, Low Church suburbanite earnest about "middle class morality" wrote epistolary novels about domestic themes, and Fielding, a robust Tory squire who attended Eton with William Pitt, served up Jones' picaresque adventure.)
Watt is all the more effective and interesting for his attentiveness to the finer points of form so much more difficult to discuss than content, and the ways in which form, too, was remade by practical imperatives of the writers' business. (That exhaustive descriptiveness, and the need of the writer to make themselves understood to readers of a less certain educational level--both connected with their selling to a wide public rather than catering to an elite patron--combined with per-word pay rates to put a premium on prolixity.)
All this is considerably enriched by a good many observations of wider literary significance. Not the least of these concerned the tendency of writers to essentially write themselves when presenting a protagonist, especially in a first-person narrative. ("Defoe's identification with Moll Flanders was so complete that, despite a few feminine traits, he created a personality that was in essence his own" (115).) Interesting, too, is his quip about an aspect of the style of fiction Pamela helped pioneer far less likely to be acknowledged by any critic in our day wishing to retain their mainstream respectability. ("[T]he direction of the plot . . . outrageously flatters the imagination of the readers of one sex and severely disciplines that of the other" (153-154)--the female and male sexes respectively here.)
Indeed, the breadth of his vision and scope of his interest is such that some of his more interesting remarks have nothing to do with eighteenth century fiction at all. (As Watt observes Shakespeare, the "inventor of the human," is less a modern than a Medieval--similarly different in his thought about time and causality so that these come across as loosely handled, and different in his thought about language, too, so that the propensity for the purple prevails over the comprehensible in his poetry, some of the reasons why he is less accessible and compelling to most of us than we think he is "supposed to be.") Watt is attentive, too, to the trade-offs that writers have to make when they actually create something--like that between plot and character (the one as a practical matter attended to in "inverse proportion" to the other (279)), and many of the pitfalls of literary criticism in his time, and our own. ("Coleridge's enthusiasm" as critic, he remarks in one instance, "may . . . serve to remind us of the danger . . . of seeing too much" in a work (120).)
As the cited passages indicate, Watt, for all his richness in insight, is also extraordinarily accessible, partly because in comparison with our pretentiously and trivially theoretical, jargon-laden contemporary work Watt is clearly focused on his subject and straightforward in his communication, while also exceptionally gifted as a wordsmith himself. Altogether this makes the book not just a key work for anyone trying to understand eighteenth century literature, or the novel that remains the central fictional form of today, but literature in general.
It is a commonplace that the literary genre of the novel is distinguished by its telling about the lives of everyday individuals in everyday circumstances, in a manner distinct in two ways--its plainness of style, and its attentiveness to the "inner lives" of its characters.
In his classic discussion of the matter in The Rise of the Novel Ian Watt does far, far more than present that view, instead getting to the root of the matter. As he makes clear the ancient-Medieval tendency had been to think of the "universal" and "timeless" as real, and particularities as ephemeral and meaningless. That earlier logic had authors striving to portray the universal above all, serving up characters who are "characteristic" types down to their names, and time and place often conceived imprecisely, with the authors' ornamentation of the portrayal the principal object and measure of their skill.
By contrast, this era saw writers aspiring to tell the stories of particular individuals, in particular places and times--and indeed, the unfolding of the events of their lives through cause and effect sequences, over time. That shift, in turn, was the reason for the plainness of style--the prioritization of a denotative, descriptive usage of language, with the lavishness in detailing taking the place of the old lavishness of ornament. He stresses, too, how this came together in a genre of private accounts, read privately, permitting an "intimacy" unattainable in other, publicly performed and publicly enjoyed styles of work (plays, poems) that both extended the bounds of what seemed permissible in art, and allowed a new intensity of audience identification with protagonists.
All of these features--the strong sense of particularity, time and cause and effect; the use of language as description rather than ornament; the private, "intimate" novel-reading experience--seem implausible outside the rise of a rationalistic, individualistic, science-touched Modernity that afforded individuals a meaningful range of economic choices (a capitalism, increasingly thoroughgoing and increasingly industrialized), and sanctioned them ideologically (Puritan ideas about salvation, secular Enlightenment thought). And indeed, Watt's classic is best remembered for its stress on the historical context in which the novel emerged, which reflects a good deal of what Mills was to call "the sociological imagination."
Appropriately the book begins with a good deal of historical background, specifically two chapters regarding the philosophical developments underlying the rise of what we think of as "realism," and a reconstruction of the eighteenth century reading public from the concrete facts of the era. In these the territory ranges from the epistemology of John Locke to the nitty-gritty of how many and who would have had the time and money to buy this kind of reading material and a place in which to read it in that private way he described (delving into prices and incomes and the rest). Afterward, when turning to the founding authors and works themselves--Daniel Defoe's Robinson Crusoe and Moll Flanders, Samuel Richardson's Pamela and Clarissa, Henry Fielding's Tom Jones--he continued to draw on historians' understanding of aspects of the period ranging from the rise of Grub Street to the rise of suburbia, and how these factored into the authors' creative work by way of their personal backgrounds, and quite profitably. (Wary as I tend to be of the biographical approach, I found it impossible to come away from this study thinking it accidental that Richardson, a retiring, low-born, Low Church suburbanite earnest about "middle class morality" wrote epistolary novels about domestic themes, and Fielding, a robust Tory squire who attended Eton with William Pitt, served up Jones' picaresque adventure.)
Watt is all the more effective and interesting for his attentiveness to the finer points of form so much more difficult to discuss than content, and the ways in which form, too, was remade by practical imperatives of the writers' business. (That exhaustive descriptiveness, and the need of the writer to make themselves understood to readers of a less certain educational level--both connected with their selling to a wide public rather than catering to an elite patron--combined with per-word pay rates to put a premium on prolixity.)
All this is considerably enriched by a good many observations of wider literary significance. Not the least of these concerned the tendency of writers to essentially write themselves when presenting a protagonist, especially in a first-person narrative. ("Defoe's identification with Moll Flanders was so complete that, despite a few feminine traits, he created a personality that was in essence his own" (115).) Interesting, too, is his quip about an aspect of the style of fiction Pamela helped pioneer far less likely to be acknowledged by any critic in our day wishing to retain their mainstream respectability. ("[T]he direction of the plot . . . outrageously flatters the imagination of the readers of one sex and severely disciplines that of the other" (153-154)--the female and male sexes respectively here.)
Indeed, the breadth of his vision and scope of his interest is such that some of his more interesting remarks have nothing to do with eighteenth century fiction at all. (As Watt observes Shakespeare, the "inventor of the human," is less a modern than a Medieval--similarly different in his thought about time and causality so that these come across as loosely handled, and different in his thought about language, too, so that the propensity for the purple prevails over the comprehensible in his poetry, some of the reasons why he is less accessible and compelling to most of us than we think he is "supposed to be.") Watt is attentive, too, to the trade-offs that writers have to make when they actually create something--like that between plot and character (the one as a practical matter attended to in "inverse proportion" to the other (279)), and many of the pitfalls of literary criticism in his time, and our own. ("Coleridge's enthusiasm" as critic, he remarks in one instance, "may . . . serve to remind us of the danger . . . of seeing too much" in a work (120).)
As the cited passages indicate, Watt, for all his richness in insight, is also extraordinarily accessible, partly because in comparison with our pretentiously and trivially theoretical, jargon-laden contemporary work Watt is clearly focused on his subject and straightforward in his communication, while also exceptionally gifted as a wordsmith himself. Altogether this makes the book not just a key work for anyone trying to understand eighteenth century literature, or the novel that remains the central fictional form of today, but literature in general.
The Best Words Ever Written About "Human Nature"
Back in tenth grade I was first introduced to F. Scott Fitzgerald, by way of The Great Gatsby. I wasn't all that impressed with the book at the time, but I later came to be more appreciative--in part, I think, because I happened to run across Fitzgerald's earlier book This Side of Paradise.
The book has its quirks and limitations--its at times' ostentatiously Modernist experimentation, for one. Still, I was impressed with some of its more substantial exchanges, which happen to include, as the title of the blog post implies, the last of them, where the protagonist Amory Blaine debates the Social Question with an acquaintance of his dead college friend's father. Said acquaintance, dismissing socialism, speaks of "'certain things which are human nature' . . . with an owl-like look, 'which always have been and always will be, which can't be changed.'"
Amory's response, after astonished looking from one man to the other, that he "can name offhand over one hundred natural phenomena that have been changed by the will of man--a hundred instincts in man that have been wiped out or are now held in check by civilization," and that to imply that such things cannot be is not only false but "a flat impeachment of all that's worth while in human nature" that "negates the efforts of every scientist, statesman, moralist, reformer, doctor, and philosopher that ever gave his life to humanity's service." Indeed, far from being the argument-ender that speaker with the owl-like look thought it was, it was in Amory's view "the last refuge of the associated mutton-heads of the world," such that any individual "over twenty-five years old who makes that statement in cold blood ought to be deprived of the franchise."
Alas, a hundred years later we are still apt to hear that inane use of those two words, "human nature," most dismayingly of all, from ostensible progressives. But at the same time, Fitzgerald's reply to it still stands.
The book has its quirks and limitations--its at times' ostentatiously Modernist experimentation, for one. Still, I was impressed with some of its more substantial exchanges, which happen to include, as the title of the blog post implies, the last of them, where the protagonist Amory Blaine debates the Social Question with an acquaintance of his dead college friend's father. Said acquaintance, dismissing socialism, speaks of "'certain things which are human nature' . . . with an owl-like look, 'which always have been and always will be, which can't be changed.'"
Amory's response, after astonished looking from one man to the other, that he "can name offhand over one hundred natural phenomena that have been changed by the will of man--a hundred instincts in man that have been wiped out or are now held in check by civilization," and that to imply that such things cannot be is not only false but "a flat impeachment of all that's worth while in human nature" that "negates the efforts of every scientist, statesman, moralist, reformer, doctor, and philosopher that ever gave his life to humanity's service." Indeed, far from being the argument-ender that speaker with the owl-like look thought it was, it was in Amory's view "the last refuge of the associated mutton-heads of the world," such that any individual "over twenty-five years old who makes that statement in cold blood ought to be deprived of the franchise."
Alas, a hundred years later we are still apt to hear that inane use of those two words, "human nature," most dismayingly of all, from ostensible progressives. But at the same time, Fitzgerald's reply to it still stands.
"Geography, Technology and the Flux of Opportunity"
Back when I wrote the first edition of The Many Lives and Deaths of James Bond I included in the appendix a brief essay discussing Britain's extraordinary industrial, imperial and military predominance at its nineteenth century peak, and its subsequent passing over the following hundred years. Basically what it came down to was the spread of industrialization as other nations increasingly consolidated. One result was larger-scale industrialized states, another was states which evolved a more sophisticated form of industrialization, and still another were states which combined both features, epitomized by the United States. The balance of economic power shifted in a hurry, the balance of military power with it. Meanwhile the declining acceptability of colonial rule, and the cost of two world wars, accelerated the unraveling of Britain's empire, but that economic change was first and foremost.
Given my historical and other interests I did much more reading and thinking about the issue, unavoidably rethinking what I wrote earlier (just why did others make a bigger success of the "Second" Industrial Revolution, for example?), and writing more. Initially I intended to produce a longer, better-grounded version of the piece in my appendix, later a few short pieces, but it eventually blew up into the paper I have just published through SSRN, "Geography, Technology and Opportunity: The Rise and Decline of British Economic Power" (the second revision of which is now up). About 74,000 words long it is less an appendix than a book in itself--which I suppose offers the same explanation in the end. With (I hope) more rigor, in detail and depth (unexpectedly I found myself coping with matters raging from the comparative phosphorous content of different iron ores to waterway-territory ratios in countries around the world to the finer points of the 1965 National Plan), but nonetheless, the same essential explanation.
If nothing else, I've done a fair job of convincing myself.
Given my historical and other interests I did much more reading and thinking about the issue, unavoidably rethinking what I wrote earlier (just why did others make a bigger success of the "Second" Industrial Revolution, for example?), and writing more. Initially I intended to produce a longer, better-grounded version of the piece in my appendix, later a few short pieces, but it eventually blew up into the paper I have just published through SSRN, "Geography, Technology and Opportunity: The Rise and Decline of British Economic Power" (the second revision of which is now up). About 74,000 words long it is less an appendix than a book in itself--which I suppose offers the same explanation in the end. With (I hope) more rigor, in detail and depth (unexpectedly I found myself coping with matters raging from the comparative phosphorous content of different iron ores to waterway-territory ratios in countries around the world to the finer points of the 1965 National Plan), but nonetheless, the same essential explanation.
If nothing else, I've done a fair job of convincing myself.
Monday, July 16, 2018
Admiral Kirk and Captain Spock Ride The Bus
Recently I ran across Star Trek IV: The Voyage Home again--the bit where Kirk and Spock walk around twentieth century San Francisco and then ride the bus and have their run-in with the punk with the boom box. I was impressed by how much satire (remarking how we treat each other, what passes for literature in our era) was packed into this one little scene, and how gracefully, how entertainingly, how charmingly it plays (for a scene with an obnoxious punk annoying a whole bus and giving our heroes the finger, anyway).
Star Trek IV was, of course, a summer blockbuster in its day. And recently it occurred to me--when was the last time that our bombastic summer blockbusters gave us a scene that worked in such a way? Simple, straightforward filmmaking (two people walking around, getting on a bus, talking), which said so much and amused so much at the same time, and proved so memorable? Certainly I can't remember any such thing in the recent Trek films. Anyone else able to think of one?
Star Trek IV was, of course, a summer blockbuster in its day. And recently it occurred to me--when was the last time that our bombastic summer blockbusters gave us a scene that worked in such a way? Simple, straightforward filmmaking (two people walking around, getting on a bus, talking), which said so much and amused so much at the same time, and proved so memorable? Certainly I can't remember any such thing in the recent Trek films. Anyone else able to think of one?
Of Starhunter
Starhunter appeared at the tail end of the '90s-era science fiction television boom, and came into that long-crowded field quietly, at least in the United States. Rather than a prime time slot on even a cable network, it went to syndication--and in my area, late night syndication. It ran for only one year before getting a major overhaul (it became Starhunter 2300, with the principal cast mostly changed), then lasted for only one year after that during which, even in that period when I followed the scene very closely and caught something of just about every show (in fact, watched just about all of just about every other small screen space opera), I not only didn't see it but was scarcely aware of its existence until it was already off the air, past which it didn't have much of a life in reruns (not in my area, anyway), and only happened to see it years later on DVD. (Remember those?)
The slightness of the show's wiki (which, despite dating back to June 2009, has a mere 31 articles, most of them quite short) does not suggest much has changed since then.
All the same, a continuation of the show, Starhunter: Transformation is reportedly in the works, and perhaps that is why the show has recently been remastered and reissued as Starhunter Redux, while since May airing in prime time reruns on Robert Rodriguez's TV network, El Rey.
Catching one of those reruns recently the show reminded me of another space opera that flew under most people's radars, Lexx. Like Starhunter, Lexx was an internationally financed Canadian project, and relatively low-budget, but there they parted ways. Despite the slenderness of the resources put up for it, Lexx, packed with colorful sets and bright CGI and surprisingly globe-trotting location shooting (from Namibia to Thailand to Iceland the crew went, and it's all up there on the screen) inclined toward the exotic, the weird, the extravagant--the end of the universe in a gray goo-induced Big Crunch (and no, it wasn't just a dream, an alternate timeline or any other such lame cop-out, it really was the end of the universe) a mere season finale.
By contrast Starhunter, while having some hints of something bigger going on in the background (and not always just the background, as the season two cliffhanger shows), tended toward the low-key and small-scale in its plots and its look from episode to episode, the space ships and stations and colonies utilitarian to the point of being bare bones, and most of the episodes taking place in their dark, dusty interiors, which matched the tendency toward the noirisih and gritty in tone. Looking back on it I suspect this probably did not help it win over a broad audience, but it did set it apart from the generally flashier, splashier, zanier fare that characterized the genre then and now.
The slightness of the show's wiki (which, despite dating back to June 2009, has a mere 31 articles, most of them quite short) does not suggest much has changed since then.
All the same, a continuation of the show, Starhunter: Transformation is reportedly in the works, and perhaps that is why the show has recently been remastered and reissued as Starhunter Redux, while since May airing in prime time reruns on Robert Rodriguez's TV network, El Rey.
Catching one of those reruns recently the show reminded me of another space opera that flew under most people's radars, Lexx. Like Starhunter, Lexx was an internationally financed Canadian project, and relatively low-budget, but there they parted ways. Despite the slenderness of the resources put up for it, Lexx, packed with colorful sets and bright CGI and surprisingly globe-trotting location shooting (from Namibia to Thailand to Iceland the crew went, and it's all up there on the screen) inclined toward the exotic, the weird, the extravagant--the end of the universe in a gray goo-induced Big Crunch (and no, it wasn't just a dream, an alternate timeline or any other such lame cop-out, it really was the end of the universe) a mere season finale.
By contrast Starhunter, while having some hints of something bigger going on in the background (and not always just the background, as the season two cliffhanger shows), tended toward the low-key and small-scale in its plots and its look from episode to episode, the space ships and stations and colonies utilitarian to the point of being bare bones, and most of the episodes taking place in their dark, dusty interiors, which matched the tendency toward the noirisih and gritty in tone. Looking back on it I suspect this probably did not help it win over a broad audience, but it did set it apart from the generally flashier, splashier, zanier fare that characterized the genre then and now.
Reconsidering Philo-Fiction
Some years ago the philosopher Terence Blake raised the question of "philo-fiction," fiction which uses philosophy the way science fiction uses science, and whether we might see it come into its own as a genre. Of course, that raises the question of what we mean by "philosophy." My initial thought was that philo-fiction as he used the term (fiction where the fundamental rules of the universe differed so deeply from our own) could be thought of a subset of science fiction, and so, perhaps.
My answer's changed since then. Fiction dependent on such a radical difference, it seems to me, is so demanding for the writer, and the reader, that it could probably never be very prolific--so that while we probably will keep seeing people try their hand at it every now and then, I don't think I could see it becoming a full-blown genre, certainly not on the scale that science fiction has at its peak.
However, I have also found myself thinking about the matter of philo-fiction another way, because I find myself ever less satisfied with the way we delineate "philosophy." After all, all intellectual investigation was known by that name, once. However, what happened was that proponents of a particular philosophical approach--old-fashioned induction and deduction, applied in a materialist, empirical way--was formalized by figures like Francis Bacon into the scientific method, after which it was known not as "philosophy" anymore, but a separate enterprise. This has in fact gone so far that many, maybe even most, of today's scientists actually have little intellectual grasp of the premises of their life's work.
So has it also gone with investigation of the social world. Studying International Relations in college we were exposed to a considerable amount of Plato, Aristotle, Augustine, Aquinas, Hobbes, Spinoza, Locke, Kant, Hegel, St.-Simon, Marx. In time the IR student (especially if they go past the B.A. level) is expected to read at least some of the actual texts deemed more important, which amount to a healthy chunk of the philosophical canon. Yet one is unlikely to read anyone labeled a "philosopher" from after the mid-nineteenth century, not because people stopped doing philosophy relevant to the subject, but because the labeling of those philosophers of obvious and direct relevance to the field changed to "economist" or "sociologist," "political scientist" or "social scientist," because of their use of a particular epistemological approach now labeled not "philosophy," but the science held to be a very different thing.
Now we use "philosophy" to denote inquiry into epistemology and ethics and little else, just those things that "we haven't learned to treat scientifically yet," with the invidious comparison between the rigorous applier of the scientific method and the fuzzy, verbal, non-quantitative folks with the ever-smaller turf not at all subtle.1 I remember, for instance, a scene in Robert Sawyer's book Flashforward where his protagonist Lloyd Simcoe reacts contemptuously to a philosophy professor's remark on the titular event. After reading it he "found himself crumpling up the newspaper page and throwing it across his office" as sighed and thought to himself "A philosophy professor! . . . Couldn't they have gotten a scientist to address this issue? Someone who understands what really constitutes evidence?" And then because he is that emphatic about it, once again "A philosophy professor. Give me a fucking break."
To my knowledge no reader of the book has remarked the scene in a significant and public way--and it does not strain credulity that they have not, because this attitude seems so commonplace. But perhaps that is one reason why it seems to me all the more important to argue that despite the relabeling, science did not stop being philosophy--perhaps the more so because of how invidious the comparisons between philosophy and science can get. Accordingly we may regard speculative fiction depicting or extrapolating from or simply playing with the theories, practice, knowledge gained by the sciences as "philo-fiction." Even if in deference to the irrationality of prevailing usages we only regard fiction which is more narrowly interested in epistemology or ethics as philo-fiction, then science fiction has been doing that too, concerned with what we can know and how, and what we ought to do about it, can go by that name as well, at times more recognizable than others, but by any plausible measure never rare or marginal, and by this point quite prolific. (I certainly would not deny that Isaac Asimov, in those early Robot stories, was dealing with philosophy, in "philo-fiction," even as he was producing some of what we think of as the hardest of hard Campbellian science fiction.) In fact, I will go further and say that, in this sense at least, science fiction has simply been philo-fiction all along.
1. I find myself thinking of other eighteenth century terms similarly narrowed--and impoverished. Take, for example, "manners," which has been reduced from culture to etiquette, or "education," which rather than a whole upbringing seems to mean formal academic training and that alone.
My answer's changed since then. Fiction dependent on such a radical difference, it seems to me, is so demanding for the writer, and the reader, that it could probably never be very prolific--so that while we probably will keep seeing people try their hand at it every now and then, I don't think I could see it becoming a full-blown genre, certainly not on the scale that science fiction has at its peak.
However, I have also found myself thinking about the matter of philo-fiction another way, because I find myself ever less satisfied with the way we delineate "philosophy." After all, all intellectual investigation was known by that name, once. However, what happened was that proponents of a particular philosophical approach--old-fashioned induction and deduction, applied in a materialist, empirical way--was formalized by figures like Francis Bacon into the scientific method, after which it was known not as "philosophy" anymore, but a separate enterprise. This has in fact gone so far that many, maybe even most, of today's scientists actually have little intellectual grasp of the premises of their life's work.
So has it also gone with investigation of the social world. Studying International Relations in college we were exposed to a considerable amount of Plato, Aristotle, Augustine, Aquinas, Hobbes, Spinoza, Locke, Kant, Hegel, St.-Simon, Marx. In time the IR student (especially if they go past the B.A. level) is expected to read at least some of the actual texts deemed more important, which amount to a healthy chunk of the philosophical canon. Yet one is unlikely to read anyone labeled a "philosopher" from after the mid-nineteenth century, not because people stopped doing philosophy relevant to the subject, but because the labeling of those philosophers of obvious and direct relevance to the field changed to "economist" or "sociologist," "political scientist" or "social scientist," because of their use of a particular epistemological approach now labeled not "philosophy," but the science held to be a very different thing.
Now we use "philosophy" to denote inquiry into epistemology and ethics and little else, just those things that "we haven't learned to treat scientifically yet," with the invidious comparison between the rigorous applier of the scientific method and the fuzzy, verbal, non-quantitative folks with the ever-smaller turf not at all subtle.1 I remember, for instance, a scene in Robert Sawyer's book Flashforward where his protagonist Lloyd Simcoe reacts contemptuously to a philosophy professor's remark on the titular event. After reading it he "found himself crumpling up the newspaper page and throwing it across his office" as sighed and thought to himself "A philosophy professor! . . . Couldn't they have gotten a scientist to address this issue? Someone who understands what really constitutes evidence?" And then because he is that emphatic about it, once again "A philosophy professor. Give me a fucking break."
To my knowledge no reader of the book has remarked the scene in a significant and public way--and it does not strain credulity that they have not, because this attitude seems so commonplace. But perhaps that is one reason why it seems to me all the more important to argue that despite the relabeling, science did not stop being philosophy--perhaps the more so because of how invidious the comparisons between philosophy and science can get. Accordingly we may regard speculative fiction depicting or extrapolating from or simply playing with the theories, practice, knowledge gained by the sciences as "philo-fiction." Even if in deference to the irrationality of prevailing usages we only regard fiction which is more narrowly interested in epistemology or ethics as philo-fiction, then science fiction has been doing that too, concerned with what we can know and how, and what we ought to do about it, can go by that name as well, at times more recognizable than others, but by any plausible measure never rare or marginal, and by this point quite prolific. (I certainly would not deny that Isaac Asimov, in those early Robot stories, was dealing with philosophy, in "philo-fiction," even as he was producing some of what we think of as the hardest of hard Campbellian science fiction.) In fact, I will go further and say that, in this sense at least, science fiction has simply been philo-fiction all along.
1. I find myself thinking of other eighteenth century terms similarly narrowed--and impoverished. Take, for example, "manners," which has been reduced from culture to etiquette, or "education," which rather than a whole upbringing seems to mean formal academic training and that alone.
Friday, June 1, 2018
Reflections of a NewSpace Skeptic
NOTE: I penned this piece last summer but have only got around to publishing it now. While some of the details have since dated, it still seems to me worth sharing because its purpose, after all, is reflection on the trend of a decade, unaltered by subsequent events.
I recall many years ago running across a headline about Virgin Galactic's unveiling SpaceShip Two. Thinking that progress on the vehicle had gone a lot faster than I had heard, and expecting to see it rolling out of a hangar before the world press, I clicked on it—and saw instead a picture of a broadly grinning Richard Branson holding a model of what SpaceShip Two would eventually look like when it was built.
Of course, headlines commonly promise much more than they deliver. Still, the particular way in which they do so where technology stories are concerned is especially problematic. A company announces an R & D project, or even some scientist mentions a technical possibility—and the journalist in question presents the technology as if it were already developed, put on the market, and in satisfactory, practical, everyday use. This kind of writing is, unfortunately, evident all across the spectrum of cutting-edge technologies today, sometimes in publications whose editors ought to know better. (I recall one respectable popular science magazine that, six years ago, made it sound as if submarine communications systems were already encrypted with quantum keys.) Moreover, this kind of distorting coverage does not prove that a particular line of research and development is all smoke and no fire.
All the same, following the news regarding space technology, and the space business generally, it seemed that from year to year we were hearing much more about boldly announced initiatives, or even the discussion of mere possibilities, than about evidence of concrete progress—and that expectations generally were overblown, especially given the combination of financial and technical hurdles to be overcome, especially in regard to the more genuinely exciting opportunities. (Space tourism was one thing, space-based solar power of significance on a macroeconomic scale a much more interesting but much more demanding thing.)
It seems to me that much of what has happened has validated my skepticism. Virgin Galactic was, after all, was supposed to take its first suborbital tourists up in 2007. Ten years on neither Virgin, nor any other private company, has flown a single customer. Quite the contrary, the greatest success story for NewSpace in recent years has not been a matter of tourism, but the more established (and less glamorous) business of launching satellites, and that not by means of newfangled spaceplanes, but multi-stage rockets of the sort we have been using since the beginning of the space age.
The protagonist of this story, SpaceX's Falcon 9, has yet to attain even the reliability of existing boosters in completing its mission. Its success rate, recently appraised at 94 percent, compares unfavorably with the 99 percent success rate of the Atlas V, the 97 percent success rate of the much more heavily used Soyuz-U, or even the reusable Space Shuttle. The same goes for the success rate of its recovery for reuse, its achievement of aircraft-like turnaround times, or its holding up under repeated missions—thus far, the Falcon 9 not yet bettering the track record of the much-derided space shuttle. Only the first stage of the rocket has been recoverable so far, and that just 80 percent of the time, its refurbishment for another flight recently took six months (again, no better than the shuttle, even post-Challenger) and even its most heavily used rocket remains a long way from matching Discovery's 39 safely completed flights.
Is this an entirely fair comparison? I freely admit that it is not. The Falcon 9 program, after all, remains not just a work in progress, but at an early point in its history compared with these other programs. Despite that the price charged by the company makes it plausible to claim its improving on the dollars-per-pound-to-orbit calculus (for the first time, a Western satellite-launcher might be approaching the $2,500 a pound-to-orbit mark, however one regards the terms), while its technique of rocket recovery, even if less complete or reliable than might be hoped, is clearly deserving of the history books. Still, it is worth remembering that the concrete progress we have seen did not happen anywhere near so quickly as some hoped. Moreover, the advances the company has made so far are more a matter of incremental improvement of familiar equipment where space vehicles are concerned, and a turn to developments from outside spacecraft-building, narrowly conceived, like 3-D printing (now being used to produce rocket parts).
I think it plausible and even likely that the company will continue to make headway in improving the rocket's performance according to all the relevant metrics—higher reliability in its launchings, more frequent and complete recovery, and more rapid refurbishment, which in turn may bring prices down appreciably. However, it seems likely that this will go on being a matter of incremental progress with familiar technology, and the broader improvement of production methods, and that the journey to $100 a pound-to-orbit will be measured in decades rather than years. Additionally, even the most guarded projection is a very different thing from a done deal—all this not having happened yet. However, I am much more optimistic about the efforts of SpaceX and the whole sector than I have been in a long time.
I recall many years ago running across a headline about Virgin Galactic's unveiling SpaceShip Two. Thinking that progress on the vehicle had gone a lot faster than I had heard, and expecting to see it rolling out of a hangar before the world press, I clicked on it—and saw instead a picture of a broadly grinning Richard Branson holding a model of what SpaceShip Two would eventually look like when it was built.
Of course, headlines commonly promise much more than they deliver. Still, the particular way in which they do so where technology stories are concerned is especially problematic. A company announces an R & D project, or even some scientist mentions a technical possibility—and the journalist in question presents the technology as if it were already developed, put on the market, and in satisfactory, practical, everyday use. This kind of writing is, unfortunately, evident all across the spectrum of cutting-edge technologies today, sometimes in publications whose editors ought to know better. (I recall one respectable popular science magazine that, six years ago, made it sound as if submarine communications systems were already encrypted with quantum keys.) Moreover, this kind of distorting coverage does not prove that a particular line of research and development is all smoke and no fire.
All the same, following the news regarding space technology, and the space business generally, it seemed that from year to year we were hearing much more about boldly announced initiatives, or even the discussion of mere possibilities, than about evidence of concrete progress—and that expectations generally were overblown, especially given the combination of financial and technical hurdles to be overcome, especially in regard to the more genuinely exciting opportunities. (Space tourism was one thing, space-based solar power of significance on a macroeconomic scale a much more interesting but much more demanding thing.)
It seems to me that much of what has happened has validated my skepticism. Virgin Galactic was, after all, was supposed to take its first suborbital tourists up in 2007. Ten years on neither Virgin, nor any other private company, has flown a single customer. Quite the contrary, the greatest success story for NewSpace in recent years has not been a matter of tourism, but the more established (and less glamorous) business of launching satellites, and that not by means of newfangled spaceplanes, but multi-stage rockets of the sort we have been using since the beginning of the space age.
The protagonist of this story, SpaceX's Falcon 9, has yet to attain even the reliability of existing boosters in completing its mission. Its success rate, recently appraised at 94 percent, compares unfavorably with the 99 percent success rate of the Atlas V, the 97 percent success rate of the much more heavily used Soyuz-U, or even the reusable Space Shuttle. The same goes for the success rate of its recovery for reuse, its achievement of aircraft-like turnaround times, or its holding up under repeated missions—thus far, the Falcon 9 not yet bettering the track record of the much-derided space shuttle. Only the first stage of the rocket has been recoverable so far, and that just 80 percent of the time, its refurbishment for another flight recently took six months (again, no better than the shuttle, even post-Challenger) and even its most heavily used rocket remains a long way from matching Discovery's 39 safely completed flights.
Is this an entirely fair comparison? I freely admit that it is not. The Falcon 9 program, after all, remains not just a work in progress, but at an early point in its history compared with these other programs. Despite that the price charged by the company makes it plausible to claim its improving on the dollars-per-pound-to-orbit calculus (for the first time, a Western satellite-launcher might be approaching the $2,500 a pound-to-orbit mark, however one regards the terms), while its technique of rocket recovery, even if less complete or reliable than might be hoped, is clearly deserving of the history books. Still, it is worth remembering that the concrete progress we have seen did not happen anywhere near so quickly as some hoped. Moreover, the advances the company has made so far are more a matter of incremental improvement of familiar equipment where space vehicles are concerned, and a turn to developments from outside spacecraft-building, narrowly conceived, like 3-D printing (now being used to produce rocket parts).
I think it plausible and even likely that the company will continue to make headway in improving the rocket's performance according to all the relevant metrics—higher reliability in its launchings, more frequent and complete recovery, and more rapid refurbishment, which in turn may bring prices down appreciably. However, it seems likely that this will go on being a matter of incremental progress with familiar technology, and the broader improvement of production methods, and that the journey to $100 a pound-to-orbit will be measured in decades rather than years. Additionally, even the most guarded projection is a very different thing from a done deal—all this not having happened yet. However, I am much more optimistic about the efforts of SpaceX and the whole sector than I have been in a long time.
Wednesday, May 30, 2018
Is the Sci-Fi Blockbuster Frozen in the Early '80s?
Those who have paid any attention to film history know how high concept and the action/science fiction blockbuster arrived in Hollywood in the mid-'70s, and in the subsequent three decades virtually swallowed up the market. (Because so many of the suck-up poptimist critics immediately "Nuh-uh!" any such claim, I afford some hard numbers about this here that should turn that pathetic "nuh-uh" into a spluttering "But, but, but . . .")
Looking back, I'm struck by the extent to which not just this broad trend, but the specific franchises date back to that time--and along with them, the common touchstones in discussion of science ficion film. Star Wars, Star Trek, Alien, Blade Runner, The Terminator. (Indeed, it often seems that our hack journalists can't have a single discussion of artificial intelligence or robotics without bringing up an Arnold Schwarzenegger B-movie from 1984. Am I really the only one who finds this pathetic and tiresome?)
This past weekend we got our tenth live-action Star Wars feature film, while the last three years saw brand-new, high-profile, big-budget installments in each and every one of these franchises. Our thirteenth Star Trek feature film in 2016 (while the franchise has also returned to TV--I hear, in debased form as Discovery), our sixth (or if you count the two Alien vs. Predator movies, eighth) Alien film last year, a Blade Runner sequel a few months after that, a fifth Terminator movie back in 2015 with the promise that the franchise had finally come to an end quickly broken when Kathryn Bigelow's ex-husband put off the sequels to the much more profitable Avatar yet again to reteam with his other ex-wife and the septugenarian former governor of California for what will apparently be an all senior citizen reboot of the movie they made back when the author of this post was too little even to know about R-rated movies.
Basically, it seems, Hollywood in that early, post-Star Wars boom period when high-concept and sci-fi action blockbusters were new amassed a certain number of properties and concepts that we can call "creative capital." To a great extent this side of its production has been living off that capital ever since, very little not connected with it in some way, coming from the same people, using its images, following in its footsteps. (Avatar, for instance, was a James Cameron production.) The principal exception would seem to be the comic book superhero-based blockbusters, which also had a crucial precedent in this period--the original Superman, somewhat ahead of its time, but with the slowness to tread the same path more than made up for in the enthusiasm that has left us up to our ears in movies based on comic books that are a half century old or older, with one Marvel movie barely leaving theaters before the next has arrived in them (indeed, the #1 position passed directly from Avengers 3 to Deadpool 2 this month, before passing again to Star Wars this weekend), and DC Comics failing to match Marvel but still taking a big bite out of the market in the process. (That disappointing Justice League movie was still the #10 hit of last year, while Wonder Woman, the champion of the previous summer, was #3.)
This is partly a testimony to how salable all this has been to a public extremely susceptible to brand name and nine figure marketing budgets, and very tolerant of repetition of the same material, even the same CGI imagery, far, far past the point of diminishing returns, but also a testimony to the sheer determination to keep milking an old IP, as the flops show. According to the figures over at Box Office Mojo, the last really impressive commercial performance by an Alien movie was in 1986, when Cameron's Aliens was #7 in its year at the American box office, and a very big hit internationally as well. Alien 3 was only #28 in 1992, Alien: Resurrection #43 in 1997, Prometheus a better but still less than stellar #24 in 2012, and last year's Alien: Covenant just #42, with room for doubt about whether there was any real profit in it. Compared with the colossal success of Terminator 2: Judgment Day, Terminator 3: Rise of the Machines was also a less than stellar performer, while Terminators 4 and especially 5 were real disappointments. (Hence the reboot.) But to high concept-minded executives, hey, following up a string of underperformers or even outright flops with more of the same beats actually giving a new idea a chance. And the fact that government tax breaks, product placement, merchandising and foreign moviegoers to whom the experience of Hollywood's offerings are still more novel helps them get away with this approach by reducing their out-of-pocket expenses and relying less on the readiness of those moviegoers who have seen it all before to fork over twenty bucks to sit in front of a big screen in 3-D glasses.
For now.
Looking back, I'm struck by the extent to which not just this broad trend, but the specific franchises date back to that time--and along with them, the common touchstones in discussion of science ficion film. Star Wars, Star Trek, Alien, Blade Runner, The Terminator. (Indeed, it often seems that our hack journalists can't have a single discussion of artificial intelligence or robotics without bringing up an Arnold Schwarzenegger B-movie from 1984. Am I really the only one who finds this pathetic and tiresome?)
This past weekend we got our tenth live-action Star Wars feature film, while the last three years saw brand-new, high-profile, big-budget installments in each and every one of these franchises. Our thirteenth Star Trek feature film in 2016 (while the franchise has also returned to TV--I hear, in debased form as Discovery), our sixth (or if you count the two Alien vs. Predator movies, eighth) Alien film last year, a Blade Runner sequel a few months after that, a fifth Terminator movie back in 2015 with the promise that the franchise had finally come to an end quickly broken when Kathryn Bigelow's ex-husband put off the sequels to the much more profitable Avatar yet again to reteam with his other ex-wife and the septugenarian former governor of California for what will apparently be an all senior citizen reboot of the movie they made back when the author of this post was too little even to know about R-rated movies.
Basically, it seems, Hollywood in that early, post-Star Wars boom period when high-concept and sci-fi action blockbusters were new amassed a certain number of properties and concepts that we can call "creative capital." To a great extent this side of its production has been living off that capital ever since, very little not connected with it in some way, coming from the same people, using its images, following in its footsteps. (Avatar, for instance, was a James Cameron production.) The principal exception would seem to be the comic book superhero-based blockbusters, which also had a crucial precedent in this period--the original Superman, somewhat ahead of its time, but with the slowness to tread the same path more than made up for in the enthusiasm that has left us up to our ears in movies based on comic books that are a half century old or older, with one Marvel movie barely leaving theaters before the next has arrived in them (indeed, the #1 position passed directly from Avengers 3 to Deadpool 2 this month, before passing again to Star Wars this weekend), and DC Comics failing to match Marvel but still taking a big bite out of the market in the process. (That disappointing Justice League movie was still the #10 hit of last year, while Wonder Woman, the champion of the previous summer, was #3.)
This is partly a testimony to how salable all this has been to a public extremely susceptible to brand name and nine figure marketing budgets, and very tolerant of repetition of the same material, even the same CGI imagery, far, far past the point of diminishing returns, but also a testimony to the sheer determination to keep milking an old IP, as the flops show. According to the figures over at Box Office Mojo, the last really impressive commercial performance by an Alien movie was in 1986, when Cameron's Aliens was #7 in its year at the American box office, and a very big hit internationally as well. Alien 3 was only #28 in 1992, Alien: Resurrection #43 in 1997, Prometheus a better but still less than stellar #24 in 2012, and last year's Alien: Covenant just #42, with room for doubt about whether there was any real profit in it. Compared with the colossal success of Terminator 2: Judgment Day, Terminator 3: Rise of the Machines was also a less than stellar performer, while Terminators 4 and especially 5 were real disappointments. (Hence the reboot.) But to high concept-minded executives, hey, following up a string of underperformers or even outright flops with more of the same beats actually giving a new idea a chance. And the fact that government tax breaks, product placement, merchandising and foreign moviegoers to whom the experience of Hollywood's offerings are still more novel helps them get away with this approach by reducing their out-of-pocket expenses and relying less on the readiness of those moviegoers who have seen it all before to fork over twenty bucks to sit in front of a big screen in 3-D glasses.
For now.
Thoughts on Alien: Covenant
WARNING: SPOILERS
I recently saw Alien: Covenant.
I didn't expect much. I got even less.
To be fair, the Alien series has not been a favorite of mine. I appreciate the place the first two films have in science fiction film history (and I did enjoy the second movie in particular, as a shoot 'em up from a period when that kind of thing was still fresh), but my interest in it was limited all the same, and it soon came to seem yet another instance of a series dragging on way, way too long mostly because Hollywood studios are so adamant about keeping every single established IP going for as long as possible.
Besides, I was dubious about the film's predecessor, Prometheus, for reasons John McCalmont described fairly well at the time of release. It raised fairly commonplace, trite (to me, silly) questions, and then didn't try to answer them, instead giving us a plot that "is really nothing more than a series of doors slammed in characters' faces by a cruelly indifferent universe."1
Rather than "playful," it was hateful.
I grant that this sort of thing might--might--have meant something, once. But after three generations of smug postmodernist "subversion" (itself, really just a recycling of a tradition of misanthropy elites have promulgated for self-serving reasons going back to the ancients), do we really need more of this?
I say that we don't.
But I got a vague idea from some of the discussion of the movie that it had something to say about the mysteries--what the deal was with those alien "Engineers." So I gave it a chance.
Instead we got a typically pretentious opening in a huge white room with a grand piano in it (does no one else notice this cliche?), Billy Crudup's character whining about people of faith being discriminated against in an atheist world (that must be that "liberal Hollywood" at it again), and more Frankenstein complex inanity as yet another robot created in our image decides to turn on us, and once again actors look terrified as pieces of rubber (or were they CGI?) jumped on them and members of the crew splashed blood all over the set, because instead of dropping the xenomorphs from the film, like was apparently discussed at one point (and like I would have preferred), the movie pretty was mostly xenomorphs attacking people, all on the way to a final "twist" that even the dumbest viewer of such movies must have seen coming prior to a pretentious close where David has to declare to us his own choice of soundtrack as he heads off to wreak interplanetary havoc.
After seeing this film I wasn't terribly surprised to see that Transcendence's Jack Paglen had been involved (Frankenstein, Frankenstein, Frankenstein, groan), and that John Logan had his hand in this too (groan again).
Allegedly the moviemakers involved with the next installment (this film wasn't such a moneymaker that such an installment could be taken for granted, but, hey, sequel) are again thinking about not spending so much time on the xenomorphs, but I wouldn't hold my breath for that. Simply recycling the stuff of a forty year old movie is a lot easier than actually doing anything seriously interesting. And I expect that when that movie inevitably comes out, having learned my lesson, I will take a pass on it.
1. I'd link back to McCalmont's original piece on the film, which is well worth a read, but apparently the trolls have driven him to turn Ruthless Culture "private." Yes, what you asswipes do does have consequences.
I recently saw Alien: Covenant.
I didn't expect much. I got even less.
To be fair, the Alien series has not been a favorite of mine. I appreciate the place the first two films have in science fiction film history (and I did enjoy the second movie in particular, as a shoot 'em up from a period when that kind of thing was still fresh), but my interest in it was limited all the same, and it soon came to seem yet another instance of a series dragging on way, way too long mostly because Hollywood studios are so adamant about keeping every single established IP going for as long as possible.
Besides, I was dubious about the film's predecessor, Prometheus, for reasons John McCalmont described fairly well at the time of release. It raised fairly commonplace, trite (to me, silly) questions, and then didn't try to answer them, instead giving us a plot that "is really nothing more than a series of doors slammed in characters' faces by a cruelly indifferent universe."1
Rather than "playful," it was hateful.
I grant that this sort of thing might--might--have meant something, once. But after three generations of smug postmodernist "subversion" (itself, really just a recycling of a tradition of misanthropy elites have promulgated for self-serving reasons going back to the ancients), do we really need more of this?
I say that we don't.
But I got a vague idea from some of the discussion of the movie that it had something to say about the mysteries--what the deal was with those alien "Engineers." So I gave it a chance.
Instead we got a typically pretentious opening in a huge white room with a grand piano in it (does no one else notice this cliche?), Billy Crudup's character whining about people of faith being discriminated against in an atheist world (that must be that "liberal Hollywood" at it again), and more Frankenstein complex inanity as yet another robot created in our image decides to turn on us, and once again actors look terrified as pieces of rubber (or were they CGI?) jumped on them and members of the crew splashed blood all over the set, because instead of dropping the xenomorphs from the film, like was apparently discussed at one point (and like I would have preferred), the movie pretty was mostly xenomorphs attacking people, all on the way to a final "twist" that even the dumbest viewer of such movies must have seen coming prior to a pretentious close where David has to declare to us his own choice of soundtrack as he heads off to wreak interplanetary havoc.
After seeing this film I wasn't terribly surprised to see that Transcendence's Jack Paglen had been involved (Frankenstein, Frankenstein, Frankenstein, groan), and that John Logan had his hand in this too (groan again).
Allegedly the moviemakers involved with the next installment (this film wasn't such a moneymaker that such an installment could be taken for granted, but, hey, sequel) are again thinking about not spending so much time on the xenomorphs, but I wouldn't hold my breath for that. Simply recycling the stuff of a forty year old movie is a lot easier than actually doing anything seriously interesting. And I expect that when that movie inevitably comes out, having learned my lesson, I will take a pass on it.
1. I'd link back to McCalmont's original piece on the film, which is well worth a read, but apparently the trolls have driven him to turn Ruthless Culture "private." Yes, what you asswipes do does have consequences.
Remake, Remake and Remake Again
Hollywood has always been quick to remake movies. Astonishingly it made three versions of Dashiell Hammett's The Maltese Falcon between 1931 and 1941. (It's actually the third, John Huston-Humphrey Bogart movie we generally remember.)
Still, Hollywood was different then. The remakes were part of a far higher output of feature films, a major studio like MGM putting out one movie a week. And film was seen as nearly disposable back then, a bit more like how we view TV than movies, the more so because of how rapid changes were seen as making older material unsalable. With the talkies, "no one" wanted silent movies, while color and widescreen changed the terms yet again. At the same time the more straitlaced "Hayes' Code" meant that a lot of older material made in a freer period was no longer screenable--while if you were going to screen something to which you would have to sell tickets in competition with brand new movies, why not have new stars in it when they were what people wanted to see? All this was reflected in, and itself reflected, the fact that the studios didn't work very hard to old onto older material, much of it literally lost over the tumult of these decades, while the relaxation of censorship later meant that old stories which were presented only in bowdlerized fashion could get more faithful adaptation. (This was, in fact, a justification for the flurry of remakes of noir classics in the '70s and early '80s--The Big Sleep, The Postman Always Rings Twice.)
None of that applies now. The output of feature film is limited in quantity, each film representing a bigger proportion of the whole. The medium is no longer going through such flux as it did earlier, and the same applies for the bounds of censorship. We no longer think of films as disposable--each and every one treated as representing precious Intellectual Property, to be clasped tightly until the end of time. In fact, far from competing for ticket sales because it is the only way that one can see movies, TV, the Internet and the rest mean that audiences have never had cheaper, easier access to older movies. In the process, much of the justification for remakes has disappeared. Accordingly, it was possible to justify three Maltese Falcons over the '30s in a way that it does not seem possible to justify not three Spiderman movies but three Spiderman franchises in a decade of the twenty-first century (2007-2017). That Hollywood insists on doing it anyway is solely a matter of a critical (or is it uncritical?) minimum of people being willing to come in and see the resulting product, as has undeniably been the case. In commercial terms, high concept remains a success. And so long as that remains the case, it too will remain with us, no matter how much film critics and cultural commentators complain.
Still, Hollywood was different then. The remakes were part of a far higher output of feature films, a major studio like MGM putting out one movie a week. And film was seen as nearly disposable back then, a bit more like how we view TV than movies, the more so because of how rapid changes were seen as making older material unsalable. With the talkies, "no one" wanted silent movies, while color and widescreen changed the terms yet again. At the same time the more straitlaced "Hayes' Code" meant that a lot of older material made in a freer period was no longer screenable--while if you were going to screen something to which you would have to sell tickets in competition with brand new movies, why not have new stars in it when they were what people wanted to see? All this was reflected in, and itself reflected, the fact that the studios didn't work very hard to old onto older material, much of it literally lost over the tumult of these decades, while the relaxation of censorship later meant that old stories which were presented only in bowdlerized fashion could get more faithful adaptation. (This was, in fact, a justification for the flurry of remakes of noir classics in the '70s and early '80s--The Big Sleep, The Postman Always Rings Twice.)
None of that applies now. The output of feature film is limited in quantity, each film representing a bigger proportion of the whole. The medium is no longer going through such flux as it did earlier, and the same applies for the bounds of censorship. We no longer think of films as disposable--each and every one treated as representing precious Intellectual Property, to be clasped tightly until the end of time. In fact, far from competing for ticket sales because it is the only way that one can see movies, TV, the Internet and the rest mean that audiences have never had cheaper, easier access to older movies. In the process, much of the justification for remakes has disappeared. Accordingly, it was possible to justify three Maltese Falcons over the '30s in a way that it does not seem possible to justify not three Spiderman movies but three Spiderman franchises in a decade of the twenty-first century (2007-2017). That Hollywood insists on doing it anyway is solely a matter of a critical (or is it uncritical?) minimum of people being willing to come in and see the resulting product, as has undeniably been the case. In commercial terms, high concept remains a success. And so long as that remains the case, it too will remain with us, no matter how much film critics and cultural commentators complain.
E.H. Carr and William Haggard
One of the classics one becomes familiar with studying International Relations is E.H. Carr's The Twenty Years' Crisis, widely considered the starting point for modern "realist" thought.
Appropriately the book is no narrow discussion of billiard-ball-type politics among nations, considering a good deal else, with one issue I find myself returning to every now and then what Carr referred to as the problem of the "Intellectual and the Bureaucrat." The intellectual inclines to theory, reasoning, principles, and what is good and right, and it is from this that their tendency to be involved with radical movements derives.
The bureaucrat--the civil servant--by contrast, "recoils from written constitutions and solemn covenants, and lets himself be guided by precedent, by instinct, by feel for the right thing," a feel guided by experience that leads them to claim "an esoteric understanding of appropriate procedures . . . not accessible even to the most intelligence outsider," and the superiority of bureaucratic experience and training to the most brilliant intellect or refined theoretical understanding in these matters. And whether one sees this as self-serving, obscurantist nonsense or not, it carries carries serious political implications. That "practical practice" by which they set such store "easily degenerates into the rigid and empty formalism of the mandarin," with "politics an end in themselves," adding to the implications inherent in their position. More than just about "any other class of the community," the bureaucrat is "bound up with the existing order, the maintenance of tradition, and . . . precedent as the 'safe' criterion of action.'"
Recently recalling Carr's comment on this "antithesis" I found myself thinking of William Haggard's Colonel Russell novels. I can think of no other works of spy fiction that treat the opposition between the two types so directly or extensively. Nor of any that, in treating that opposition, is so vehement in taking a side.
Where political life is concerned, Haggard is as hostile to intellectuals as anyone you might care to name. In Slow Burn Haggard had the latter to say of scientists:
In The Power House, Haggard bashes a type of intellectual to which one would expect him to be even more hostile, not the physicist who dares have an opinion about politics, but those whose principal concern is the social, economic, political order, in his depiction of the hapless Labour MP Victor Demuth. A "fossil," espousing "doctrines as archaic for a modern left-leaning party as the Divine Right of Kings was now archaic to the Right," Russell held the man's attitudes to be a mark of deep personal failure, resulting from deep defects of what a certain sort of pompous person would call "character." Despite a background of great privilege, which combined with a genuine intelligence and determination "equipped [him] to compete at any level he'd cared to aim at," the Prime Ministership included, a lack of confidence and inclination to "flinch from conflict" made Demuth "slip . . . into the security of protest" instead.1 Ineffectual protest because, as another character in the novel reflects, "The Time of the Left would come perhaps, but it wouldn't be . . . the intellectuals, the professional washed-out rebels, but ruthless and determined men" who made it happen, ruthless and determined men who, whatever else they happened to be, would not be mere intellectuals.
By contrast Haggard's hero, Colonel Russell, is the consummate civil servant, and not merely by virtue of his title or pay grade, but his being an administrator who, unlike most spy chiefs in spy novels, actually administers, and plays his main part in the story by administering. However many times Fleming calls Bond a civil servant, what we usually see is Bond playing commando. And even as he ascended to a fairly senior level in Central Intelligence, Jack Ryan's adventures tended to have him caught up in heroics of some kind or another--in Clear and Present Danger this Acting Deputy Director of the CIA personally flying down to Colombia in a Pave Low and manning a minigun with which he mows down drug cartel soldiers in the course of rescuing an American special forces team inserted into the country. (I repeat: Deputy Director shooting lots and lots of people with a gatling machine gun as if he were Arnold Schwarzenegger.) Such black bag work and gunplay as the Haggard stories offer, however, Russell leaves to others, while he navigates the system, and in doing so "saves the day" in dramas that, like no other, celebrate the Bureaucrat as Hero.
Of course, in continuing his series over the next few decades Haggard did eventually start playing up the action, with Yesterday's Enemy an example of this. Still, this is how they started, and tended to run. In Yesterday's, certainly, what drew Russell into his more conventional spy adventure was in fact his old, legendary reputation as Super-Bureaucrat Extraordinaire.
In hindsight, it is an additional way in which these largely forgotten books were and remain unique within this genre.
1. To round out the right-wing cliche, we also see the upper-class leftist portrayed as a snob and a bigot, disliking his niece's suitor, allegedly, for his being in the casino industry, unintellectual, Catholic and therefore "a reactionary fascist beast."
Appropriately the book is no narrow discussion of billiard-ball-type politics among nations, considering a good deal else, with one issue I find myself returning to every now and then what Carr referred to as the problem of the "Intellectual and the Bureaucrat." The intellectual inclines to theory, reasoning, principles, and what is good and right, and it is from this that their tendency to be involved with radical movements derives.
The bureaucrat--the civil servant--by contrast, "recoils from written constitutions and solemn covenants, and lets himself be guided by precedent, by instinct, by feel for the right thing," a feel guided by experience that leads them to claim "an esoteric understanding of appropriate procedures . . . not accessible even to the most intelligence outsider," and the superiority of bureaucratic experience and training to the most brilliant intellect or refined theoretical understanding in these matters. And whether one sees this as self-serving, obscurantist nonsense or not, it carries carries serious political implications. That "practical practice" by which they set such store "easily degenerates into the rigid and empty formalism of the mandarin," with "politics an end in themselves," adding to the implications inherent in their position. More than just about "any other class of the community," the bureaucrat is "bound up with the existing order, the maintenance of tradition, and . . . precedent as the 'safe' criterion of action.'"
Recently recalling Carr's comment on this "antithesis" I found myself thinking of William Haggard's Colonel Russell novels. I can think of no other works of spy fiction that treat the opposition between the two types so directly or extensively. Nor of any that, in treating that opposition, is so vehement in taking a side.
Where political life is concerned, Haggard is as hostile to intellectuals as anyone you might care to name. In Slow Burn Haggard had the latter to say of scientists:
Take a clever boy . . . and put him into a laboratory for the next seven or eight years. What emerged inevitably was a materialist . . . a man who would assume without question that the methods of science could be applied to human societies.In short, in the eyes of this particular right-winger, they were a bunch of damned crypto-Communists, all too likely to turn traitor. And indeed, one of them (not just a leftist, but "no gentleman" either) did prove to be the traitor Russell spent the novel ferreting out.
In The Power House, Haggard bashes a type of intellectual to which one would expect him to be even more hostile, not the physicist who dares have an opinion about politics, but those whose principal concern is the social, economic, political order, in his depiction of the hapless Labour MP Victor Demuth. A "fossil," espousing "doctrines as archaic for a modern left-leaning party as the Divine Right of Kings was now archaic to the Right," Russell held the man's attitudes to be a mark of deep personal failure, resulting from deep defects of what a certain sort of pompous person would call "character." Despite a background of great privilege, which combined with a genuine intelligence and determination "equipped [him] to compete at any level he'd cared to aim at," the Prime Ministership included, a lack of confidence and inclination to "flinch from conflict" made Demuth "slip . . . into the security of protest" instead.1 Ineffectual protest because, as another character in the novel reflects, "The Time of the Left would come perhaps, but it wouldn't be . . . the intellectuals, the professional washed-out rebels, but ruthless and determined men" who made it happen, ruthless and determined men who, whatever else they happened to be, would not be mere intellectuals.
By contrast Haggard's hero, Colonel Russell, is the consummate civil servant, and not merely by virtue of his title or pay grade, but his being an administrator who, unlike most spy chiefs in spy novels, actually administers, and plays his main part in the story by administering. However many times Fleming calls Bond a civil servant, what we usually see is Bond playing commando. And even as he ascended to a fairly senior level in Central Intelligence, Jack Ryan's adventures tended to have him caught up in heroics of some kind or another--in Clear and Present Danger this Acting Deputy Director of the CIA personally flying down to Colombia in a Pave Low and manning a minigun with which he mows down drug cartel soldiers in the course of rescuing an American special forces team inserted into the country. (I repeat: Deputy Director shooting lots and lots of people with a gatling machine gun as if he were Arnold Schwarzenegger.) Such black bag work and gunplay as the Haggard stories offer, however, Russell leaves to others, while he navigates the system, and in doing so "saves the day" in dramas that, like no other, celebrate the Bureaucrat as Hero.
Of course, in continuing his series over the next few decades Haggard did eventually start playing up the action, with Yesterday's Enemy an example of this. Still, this is how they started, and tended to run. In Yesterday's, certainly, what drew Russell into his more conventional spy adventure was in fact his old, legendary reputation as Super-Bureaucrat Extraordinaire.
In hindsight, it is an additional way in which these largely forgotten books were and remain unique within this genre.
1. To round out the right-wing cliche, we also see the upper-class leftist portrayed as a snob and a bigot, disliking his niece's suitor, allegedly, for his being in the casino industry, unintellectual, Catholic and therefore "a reactionary fascist beast."
Monday, May 28, 2018
Notes on High Concept: Movies and Marketing in Hollywood, by Justin Wyatt
Picking up Justin Wyatt's High Concept I (like most people who pay much attention to this sort of thing these days, I suppose) already had a fairly good idea of what the term denotes. As Wyatt explains it, in great depth and yet concisely, it is a "straightforward, easily communicated and easily comprehended" narrative whose themes and "appeal" to a broad audience are "immediately obvious" (8). The project that sounds good to an executive in a 25-word "elevator pitch"; that they can be persuaded at least has the potential to look good in a 30-second TV spot or even a poster.
The ease of communication and comprehension is simplified in the case of a "pre-sold property" with a "built-in audience," like a sequel to a prior hit--because not only is there a proven past success, but because the job of selling the product has already been done, and all one has to do is remind the audience of it. It helps, too, for the film to have other, pre-sold features--"bankable" stars (however shaky this concept is), a soundtrack capitalizing on already popular hits (sales of which are, in turn, helped by the movie), a compelling look that in itself is the subject of the sale, integrated with and even overwhelming the narrative. (Fast! Flashy! Sleek! Ultra-modern! Maybe there's nothing much to "see" here, but you can't stop "looking" at it, can you?)
A high-concept movie is a movie that looks good in a commercial, or a promotional music video, because in contrast with a classically made movie it is essentially a very long commercial or music video, in part because it was probably made by a director whose background, at any rate, is in directing commercials or music videos. (Wyatt cites Adrian Lyne, and the Scott brothers Ridley and Tony, and one can spend a long time listing those who have entered filmmaking in similar fashion since--David Fincher and Michael Bay and Simon West and Alex Proyas and Spike Jonze and Dominic Sena and Antoine Fuqua and McG and Gore Verbinski and Zack Snyder and Tarsem Singh and and and . . . while by this point such directors have so long dominated the medium that even a director who learned their craft actually studying movies probably can't help being influenced by their practice.)
Still, Wyatt's book did have some surprises for me, the biggest of which was the range of cinematic concepts to which he saw this as being applicable. Looking at the cover's array of images from major films of the '70s and '80s I am unsurprised to recognize a shot from Jaws. But I am surprised to see shots from Flashdance and Saturday Night Fever above Jaws. Where the actual text of the book is concerned, Wyatt begins not with Steven Spielberg and George Lucas, but Grease (and specifically, a comparison of that critical flop and commercial success with the diametrical opposite in All That Jazz).
And this is, I think, a reflection of profound change in the business over the years. We think of the late '70s as the era which saw the rise of the blockbuster, the '80s as the rise of the Hollywood action movie as a staple of the market. Still, the action movie was a comparatively small part of the market back then. Where out of the top ten films of the year in the '80s at the North American box office there were apt to be a couple of action films, in the twenty-first century the figure was more likely to be five movies--half the list--and often more. At the same time the rest of the top ten list was apt to be dominated by a kind of movie which did not make Wyatt's cover at all--the big-budget, Disney or Shrek-style animated movie, which in the 2010s have averaged three of the top ten annually. Which means that between one and the other, they have accounted for eight of the top ten hits year in, year out, with the other movies on the list likely to be of closely related types (the live-action version of the animated Cinderella, for instance). And since action movies and cartoons are what you make if you want a blockbuster, they comprise a much larger share of the market overall than any other one or two such distinctive styles of film ever have before.
As a result, one would not think of many other kinds of movie as high concept (even if musicals, for example, are occasionally popular and profitable). However, as Wyatt shows through a much deeper development of the concept of a film as an ad than I anticipated, movies were not just ad-like in their aesthetic or feel, but ads for a "lifestyle." (Beverly Hills Cop was about the fantasy of what it is to be rich in "Beverly Hills" as much as it was about the adventure of the "Cop.") I do not think that this is quite as prominent in film today, the use of "lifestyle" in it different. Certainly luxury is common currency in today's commercial filmmaking, anything remotely resembling actual middle-class life or working-class life or poverty generally banished from the screen, but a movie, while expected to depict an "attractive lifestyle," gets a lot less mileage out of doing so, enough so that a movie principally selling lifestyle seems unlikely to earn $1 billion in ticket sales and surcharges. (Tony Stark's high-end consumption in the Marvel movies is part of the package, but it is backdrop, relatively less important to the movie than in a comparable film from the '80s that would have luxuriated in it all a great deal more, more noticeably amid the slighter special effects and slower cutting.)
Still, if Wyatt's profuse discussion of blockbuster as lifestyle ad seems a bit dated now, his discussion retains a relevance for other aspects of pop culture, like music--exemplified by Ted Gioia's observation that, in the course of its swallowing up everything else, "Music Criticism Has Degenerated into Lifestyle Reporting."
As Gioia remarks (finally, other people noticing this!), "During the entire year 1967, The Chicago Tribune only employed the word 'lifestyle' seven times," but today "newspapers have full-time lifestyle editors," while coverage of all the rest of life from weather to business is construed in lifestyle terms ("your commute," "your money"), and certainly, pop culture, with music "treated as one more lifestyle accessory, no different from a stylish smartphone" and "music journalism . . . retreated into a permanent TMZ-zone." Indeed,
But that reality doesn't save them from the delusion they do.
Groan, groan and groan again.
The ease of communication and comprehension is simplified in the case of a "pre-sold property" with a "built-in audience," like a sequel to a prior hit--because not only is there a proven past success, but because the job of selling the product has already been done, and all one has to do is remind the audience of it. It helps, too, for the film to have other, pre-sold features--"bankable" stars (however shaky this concept is), a soundtrack capitalizing on already popular hits (sales of which are, in turn, helped by the movie), a compelling look that in itself is the subject of the sale, integrated with and even overwhelming the narrative. (Fast! Flashy! Sleek! Ultra-modern! Maybe there's nothing much to "see" here, but you can't stop "looking" at it, can you?)
A high-concept movie is a movie that looks good in a commercial, or a promotional music video, because in contrast with a classically made movie it is essentially a very long commercial or music video, in part because it was probably made by a director whose background, at any rate, is in directing commercials or music videos. (Wyatt cites Adrian Lyne, and the Scott brothers Ridley and Tony, and one can spend a long time listing those who have entered filmmaking in similar fashion since--David Fincher and Michael Bay and Simon West and Alex Proyas and Spike Jonze and Dominic Sena and Antoine Fuqua and McG and Gore Verbinski and Zack Snyder and Tarsem Singh and and and . . . while by this point such directors have so long dominated the medium that even a director who learned their craft actually studying movies probably can't help being influenced by their practice.)
Still, Wyatt's book did have some surprises for me, the biggest of which was the range of cinematic concepts to which he saw this as being applicable. Looking at the cover's array of images from major films of the '70s and '80s I am unsurprised to recognize a shot from Jaws. But I am surprised to see shots from Flashdance and Saturday Night Fever above Jaws. Where the actual text of the book is concerned, Wyatt begins not with Steven Spielberg and George Lucas, but Grease (and specifically, a comparison of that critical flop and commercial success with the diametrical opposite in All That Jazz).
And this is, I think, a reflection of profound change in the business over the years. We think of the late '70s as the era which saw the rise of the blockbuster, the '80s as the rise of the Hollywood action movie as a staple of the market. Still, the action movie was a comparatively small part of the market back then. Where out of the top ten films of the year in the '80s at the North American box office there were apt to be a couple of action films, in the twenty-first century the figure was more likely to be five movies--half the list--and often more. At the same time the rest of the top ten list was apt to be dominated by a kind of movie which did not make Wyatt's cover at all--the big-budget, Disney or Shrek-style animated movie, which in the 2010s have averaged three of the top ten annually. Which means that between one and the other, they have accounted for eight of the top ten hits year in, year out, with the other movies on the list likely to be of closely related types (the live-action version of the animated Cinderella, for instance). And since action movies and cartoons are what you make if you want a blockbuster, they comprise a much larger share of the market overall than any other one or two such distinctive styles of film ever have before.
As a result, one would not think of many other kinds of movie as high concept (even if musicals, for example, are occasionally popular and profitable). However, as Wyatt shows through a much deeper development of the concept of a film as an ad than I anticipated, movies were not just ad-like in their aesthetic or feel, but ads for a "lifestyle." (Beverly Hills Cop was about the fantasy of what it is to be rich in "Beverly Hills" as much as it was about the adventure of the "Cop.") I do not think that this is quite as prominent in film today, the use of "lifestyle" in it different. Certainly luxury is common currency in today's commercial filmmaking, anything remotely resembling actual middle-class life or working-class life or poverty generally banished from the screen, but a movie, while expected to depict an "attractive lifestyle," gets a lot less mileage out of doing so, enough so that a movie principally selling lifestyle seems unlikely to earn $1 billion in ticket sales and surcharges. (Tony Stark's high-end consumption in the Marvel movies is part of the package, but it is backdrop, relatively less important to the movie than in a comparable film from the '80s that would have luxuriated in it all a great deal more, more noticeably amid the slighter special effects and slower cutting.)
Still, if Wyatt's profuse discussion of blockbuster as lifestyle ad seems a bit dated now, his discussion retains a relevance for other aspects of pop culture, like music--exemplified by Ted Gioia's observation that, in the course of its swallowing up everything else, "Music Criticism Has Degenerated into Lifestyle Reporting."
As Gioia remarks (finally, other people noticing this!), "During the entire year 1967, The Chicago Tribune only employed the word 'lifestyle' seven times," but today "newspapers have full-time lifestyle editors," while coverage of all the rest of life from weather to business is construed in lifestyle terms ("your commute," "your money"), and certainly, pop culture, with music "treated as one more lifestyle accessory, no different from a stylish smartphone" and "music journalism . . . retreated into a permanent TMZ-zone." Indeed,
if you force pop culture insiders to be as precise as possible in articulating the reasons why they favor a band or a singer, it almost always boils down to: "I like fill in the name because they make me feel good about my lifestyle."A still bigger irony would be if most consumers of music actually thought in the same way. After all, they don't have lifestyles to feel good or badly about. They simply can't afford lifestyles.
But that reality doesn't save them from the delusion they do.
Groan, groan and groan again.
Subscribe to:
Posts (Atom)