Obama Year 2: Quo Vadis? Fecking up?

by Michael Blim

Simply-barack-obama Question: The Barack Obama Administration in its first year has been characterized by

  1. fecklessness
  2. inexperience
  3. incompetence
  4. all of the above
  5. none of the above

My friends, as it used to be said in Chicago, vote early and often.

For my part, I’ll vote number 4. After all, incompetence and inexperience can be fixed by learning from mistakes and getting rid of the boobies.

There is no cure for fecklessness, or for the feckless which is probably more to the point.

Still competence and experience might make fecklessness less likely.

Each time I try to I try to sum up and draw the line on this Administration’s losses, with the hope of starting anew, I end up feeling like an idiot. I then stop listening to the TV and talking about politics with my friends while I stanch my losses, and then I try again.

Read more »

The Humanists: Errol Morris’ Gates of Heaven (1978)

Gatef-heaven

by Colin Marshall

Errol Morris is not a humanist. Or at least he's repeatedly claimed not to be one. In fact, he's gone so far as to label himself a misanthrope, a pessimist, a perpetual bemoaner of the human condition and the hopeless, delusional humans eternally walled in by it. But his filmography would seem to give the lie to his assertions. In all of Morris' movies, but especially in this, his first, his attentiveness to and curiosity about humanity, confused or trapped though it may be, is made obvious by every conversation, every shot, every observational choice.

Known as an examiner of the fanciful personal worlds — intellectual cocoons, really — that people build around themselves, he doesn't exempt himself from the diagnosis. Perhaps, then, he's as unable to achieve full self-awareness as his subjects are. Maybe that's why he can insist his own distaste for the human race when the truth is more complex. Gates of Heaven is not the work of a man with an abjectly dim view of humanity; it feels like nothing so much as the product of a questioning mind, a probing demeanor and a highly unorthodox kind of love.

Not every viewer will agree with this. There's a good chance that no other viewer will agree with this, as everyone seems to come away from the film with entirely different ideas about what it thinks, what it argues, what it explains, what it's “about.” By design or by chance, Morris has crafted an artwork that brightly reflects back whatever themes an audience happens to bring into it. The result is a film you can watch over and over again, undergoing a different set of revelations in each screening. Roger Ebert, one of the picture's earliest champions, may have been the first to realize this. “I have seen this film perhaps 30 times, and am still not anywhere near the bottom of it,” he wrote in 1997. “All I know is, it's about a lot more than pet cemeteries.”

But on a concrete level, it's about pet cemeteries. You can frame it as a tale of two of them, one good-intentioned but failed; another successful but, in some ineffable way, bothersome. Through a series of narration-free interviews, Morris spends the film's first half telling the story of Floyd McClure, would-be Los Altos pet cemetery entrepreneur and living definition of the term simpleton. He'd hoped to build a veritable “garden of Eden” to honor the departed domestic critters who have so loved and been loved, but the sales projections didn't pan out. He and his weary investors relate the tale of their clients' dispossessed pet caskets, which had to be exhumed and relocated to the slicker, more solvent Bubbling Well Pet Memorial Park.

Read more »

Wearing rationality badges, popularizing neutrality and saying “I don’t know” to politics: Colin Marshall talks to economist, blogger and rationalist Robin Hanson

Robin Hanson is a professor of economics at George Mason University, research associate at Oxford’s Future of Humanity Institute and chief scientist at Consensus Point. He’s also the thinker behind Overcoming Bias, a popular blog about issues of honesty, signaling, disagreement, forecasting and the far future, around which a large rationality-centric community has developed on the internet. Colin Marshall originally conducted this conversation on the public radio show and podcast The Marketplace of Ideas. [MP3] [iTunes link]


Hanson1 If we are both honest truthseekers, we should not, over the course of this discussion, disagree. Is that correct?

It's more than this discussion. It would be any discussion between any two people who are honest truthseekers on any matter of fact, and it wouldn't have to be by the end of the discussion. It would be at any point in time. I should be able to pick a topic now and guess your next opinion on it. My guess of your next opinion should be the opinion I'm stating to you right now. If I say, “I think this interview will last an hour,” my best guess of what you'll say for the interview lasting should be an hour.


This is going to sound hard to get the mind around for somebody not familiar with what you've written. They'll say, “But people disagree all the time. Humanity is here, essentially, to disagree with one another.” How do you quickly get across to someone like that why there shouldn't theoretically be disagreement?

The whole reason this is interesting is that you have a theory that differs from behavior. It's a normative theory; it says what you should do. It doesn't say what you do do, but it gives you some idea of how we're going wrong. The key idea is that we should be respecting each other's opinions. That is, I don't know how you came to your opinion, I don't know what evidence it's based on, I don't know what reasoning you went through or analysis. I'm sure there's lots of noise and errors in the whole process, but nevertheless I think you were trying to estimate the truth, and that's the key point. When you tell me your opinion, I take that very seriously as a summary of all the things you know and think and have analyzed up to this point on that topic.


What gets in the way of the reality matching the theory? I could probably come out stating an opinion of mine as fact, I could be overstating the probability of some guess I'm making. That's one way this could go off the rails. Why else?

The first thing to notice is that theory and reality do match up, on lots of of ordinary topics we don't care about. It's when our pride or enthusiasm gets hyped up that we start to disagree. If you and I were walking down the street and I said, “I think there's a tree around the corner,” you probably wouldn't disagree with me there. If you said, “No there isn't,” I would say, “Oh, okay.” When our pride isn't on the line or we're working together on a project and we need to achieve something — maybe our job is at stake — we're much more likely to be reasonable. But when we talk about politics or religion or whatever we talk about on these radio shows, that's when we're much more likely to not be reasonable to find it more enjoyable to speak to listen.


Politics, religion — these are topics where people can hold opinions, but when they hold them, they don't actually act on them much of the time, is that correct?

That's true, although it also applies to topics that you do act on but where your pride is on the line. A CEO versus a subordinate director might disagree, or we might disagree about some actual business decision we're making, or about which restaurant we should go to and which is likely to be open and tasty. We disagree on things where we'd rather think that we're right. It's very pleasant and affirming to think we're right and they're wrong. We might rather indulge that feeling than actually be right.


Read more »

Monday Poem

Forty Thousand Two Hundred Eight

I’m out here stacking days as if it were a sport
I’m up to forty thousand two hundred eight

I sweat memory. I’ve taken off my shirt,
I’m feeling great. But as I stack them up
they’re growing short

I tally what till now I’ve done

Not far from a stupa
I eye the spot where I’d begun
near an arbor vitae hedge
in a shade of catalpa

I’m looking for a bona fide antique

On spines of days my curate hands
feel to find the ones with bliss-laced hours
stitched with epiphanic seams

I come upon a few. They’re few
and far between

The sun’s past high. The pallid moon’s
a perfect ghost of round sentinel-still
upon a bald mountain ridge. I think
it might roll down

I breathe honeysuckle and see wisteria
clutch its pole twist up and round

I’d placed the pile with care
so as never to occlude the sun
yet carelessly have thrown
some days upon a previous one
then, too late, gone back to
square them up trying
to undo the done

by Jim Culleny,
January 2010

Basil Beattie: Paintings from the Janus series II 2010

BEATTIE When Night Sidles In (Janus series)2007 213 x 198cm

To find a form that accommodates the mess, that is the task of the artist now.

Samuel Beckett

He was the Roman god of beginnings, the guardian of gates and doors who presided over the first hour of each day, and the first day of each month and, as his name, suggests, January. Depicted on Roman coins with a double faced head, one side bearded, the other clean shaven, Janus represented both sun and moon. A sort of Roman ying and yang he symbolised the light and the dark within human experience. Facing both East and West, the doors of his temple at the Forum marked the beginning and end of each day, whilst many Romans began their morning with prayers to him. Worshipped during the time of planting, he was also evoked during rites of passage such as birth and marriage. Throughout Rome a number of freestanding structures – ceremonial gateways known as jani – were used as thresholds to make symbolically auspicious entrances or exits. Emblematic not only of new beginnings, Janus represented the transition between primitive life and civilisation, between the rural and the urban, youth and old age, whilst having the ability to look back at the past and simultaneously into the future. So what relevance does this obscure Roman god have for a contemporary British painter?
Read more »

Eduardo Mendieta Interviews Jurgen Habermas

Habermas_A2_5-300x277 Over at The Immanent Frame (excerpt can be found here):

EM: Over the last couple of years you have been working on the question of religion from a series of perspectives: philosophical, political, sociological, moral, and cognitive. In your Yale lectures from the fall of 2008, you approached the challenge of the vitality and renewal of religion in world society in terms of the need to rethink the link between social theory and secularization theory. In those lectures, you suggest that we need to uncouple modernization theory from secularization theory. Does this mean that you are taking distance from the dominant trends in social theory in the West, which began with Pareto, continued through Durkheim, and reached their apogee in Weber, and thus also from its explicit and avowed Eurocentrism?

JH: We should not throw out the baby with the bathwater. The debate over the sociological thesis of secularization has led to a revision above all in respect to prognostic statements. On the one hand, the system of religion has become more differentiated and has limited itself to pastoral care, that is, it has largely lost other functions. On the other hand, there is no global connection between societal modernization and religion’s increasing loss of significance, a connection that would be so close that we could count on the disappearance of religion. In the still undecided dispute as to whether the religious USA or the largely secularized Western Europe is the exception to a general developmental trend, José Casanova for example has developed interesting new hypotheses. In any case, globally we have to count on the continuing vitality of world religions.

In view of the consequences of which you speak, I consider the program of the group around Shmuel Eisenstadt and its comparative research on civilizations promising and informative. In the emerging world society, and concerning the social infrastructure, there are, as it were, by now only modern societies, but these appear in the form of multiple modernities because the great world religions have had a great culture-forming power over the centuries, and they have not yet entirely lost this power. As in the West, these “strong” traditions paved the way in East Asia, in the Middle East, and even in Africa for the development of cultural structures that confront each other today—for example, in the dispute over the right interpretation of human rights. Our Western self-understanding of modernity emerged from the confrontation with our own traditions. The same dialectic between tradition and modernity repeats itself today in other parts of the world.

The Playboy Book

Playboy_book Jason Zasky in Failure magazine:

In “Playboy and the Making of the Good Life in Modern America” (Oxford University Press), Elizabeth Fraterrigo—assistant professor of history at Loyola University in Chicago—examines the magazine’s place in postwar America, a time in which sexual mores, gender roles, marriage and family life were evolving, at least in small part due to Playboy’s urging. Though the Playboy brand may be in the midst of a long, slow decline, Fraterrigo’s measured academic analysis (there are 47 pages of footnotes), reminds the reader how Playboy both molded and mirrored American culture in the mid to late twentieth century.

Failure interviewed Fraterrigo by phone to discuss: what readers can expect to learn from her book, what made Playboy so successful, and a handful of Hefner’s pre-Playboy business failures, which in hindsight seem to have been quite fortuitous…

What was Playboy’s agenda in the postwar period?

The kernel of that agenda was there from the very beginning but it really came to fruition by the late 1950s and early ’60s. From the get-go, Hefner was trying to create a magazine that he felt didn’t really exist in American culture. He felt that most of what you saw—whether it be in popular magazines or literature or television—was geared toward families. There wasn’t anything that spoke to the kinds of things in which men were interested. Hefner wanted to create a magazine for adult men but he also wanted to create a vision of a lifestyle that showcased the pleasures one could enjoy in a blooming, blossoming postwar consumer society. So that it wasn’t just about showing up for your job every day and bringing home a paycheck to support a wife and kids. He wanted to have this alternative world where one could indulge in pleasure and materialism.

In the American Grain

Scott McLemee in Inside Higher Ed:

Howard Zinn — whose A People’s History of the United States, first published by Harper & Row in 1980, has sold some two million copies — died last week at the age of 87. His passing has inspired numerous tributes to his role in bringing a radical, pacifist perspective on American history to a wide audience.

It has also provoked denunciations of Zinn as “un-American,” which seems both predictable and entirely to his credit. One of Zinn’s lessons was that protest is a deeply American inclination. The thought is unbearable in some quarters.

One of the most affectionate tributes came from the sports writer Dave Zirin. As with many other readers, he found that reading Zinn changed his whole sense of why you would even want to study the past. “When I was 17 and picked up a dog-eared copy of Zinn’s book,” he writes, “I thought history was about learning that the Magna Carta was signed in 1215. I couldn’t tell you what the Magna Carta was, but I knew it was signed in 1215. Howard took this history of great men in powdered wigs and turned it on its pompous head.” Zirin went on to write A People’s History of Sports (New Press, 2008), which is Zinnian down to its cells.

Another noteworthy commentary comes from Christopher Phelps, an intellectual historian now in the American and Canadian studies program at the University of Nottingham. He assesses Zinn as a kind of existentialist whose perspective was shaped by the experience of the civil rights struggle. (He had joined the movement in the 1950s as a young professor at Spelman College, a historically black institution in Atlanta.)

An existentialist sensibility — the tendency to think in terms of radical commitment, of decision making as a matter of courage in the face the Absurd — was common to activists of his generation. That Phelps can hear the lingering accent in Zinn’s later work is evidence of a good ear.

Zinn “challenged national pieties and encouraged critical reflection on received wisdom,” writes Phelps. “He understood that America’s various radicalisms, far from being ‘un-American,’ have propelled the nation toward more humane and democratic arrangements…. He urged others to seek in the past the inspiration to dispel resignation, demoralization, and deference, the foundations of inertia. The past meant nothing, he argued, if severed from present and future.”

My Escape from Slavery

Frederick Douglass in a moving essay writes:

Frederick-douglass I afterward got steady work at the brass-foundry owned by Mr. Richmond. My duty here was to blow the bellows, swing the crane, and empty the flasks in which castings were made; and at times this was hot and heavy work. The articles produced here were mostly for ship work, and in the busy season the foundry was in operation night and day. I have often worked two nights and every working day of the week. My foreman, Mr. Cobb, was a good man, and more than once protected me from abuse that one or more of the hands was disposed to throw upon me. While in this situation I had little time for mental improvement. Hard work, night and day, over a furnace hot enough to keep the metal running like water, was more favorable to action than thought; yet here I often nailed a newspaper to the post near my bellows, and read while I was performing the up and down motion of the heavy beam by which the bellows was inflated and discharged. It was the pursuit of knowledge under difficulties, and I look back to it now, after so many years, with some complacency and a little wonder that I could have been so earnest and persevering in any pursuit other than for my daily bread. I certainly saw nothing in the conduct of those around to inspire me with such interest: they were all devoted exclusively to what their hands found to do. I am glad to be able to say that, during my engagement in this foundry, no complaint was ever made against me that I did not do my work, and do it well. The bellows which I worked by main strength was, after I left, moved by a steam-engine.

More here.

From Footnote to Fame in Civil Rights History

From The New York Times:

Colvin On that supercharged day in 1955, when Rosa Parks refused to give up her bus seat to a white passenger in Montgomery, Ala., she rode her way into history books, credited with helping to ignite the civil rights movement. But there was another woman, named Claudette Colvin, who refused to be treated like a substandard citizen on one of those Montgomery buses — and she did it nine months before Mrs. Parks. The Rev. Dr. Martin Luther King Jr. made his political debut fighting her arrest. Moreover, she was the star witness in the legal case that eventually forced bus desegregation.

Yet instead of being celebrated, Ms. Colvin has lived unheralded in the Bronx for decades, initially cast off by black leaders who feared she was not the right face for their battle, according to a new book that has plucked her from obscurity. Last week Phillip Hoose won the National Book Award for Young People’s Literature for “Claudette Colvin: Twice Toward Justice,” published by Farrar, Straus & Giroux. The honor sent the little-selling title shooting up 500 spots on Amazon.com’s sales list and immediately thrust Ms. Colvin, 70, back into the cultural conversation. “Young people think Rosa Parks just sat down on a bus and ended segregation, but that wasn’t the case at all,” Ms. Colvin said in an animated interview at a diner near her apartment in the Parkchester section of the Bronx. “Maybe by telling my story — something I was afraid to do for a long time — kids will have a better understanding about what the civil rights movement was about.”

More here.

No Exit

Via John Quiggin at Crooked Timber, Andrew Bacevich in The American Conservative:

A seesawing contest for the Korean peninsula ended in a painfully expensive draw. Kennedy’s Bay of Pigs managed only to pave the way for the Cuban Missile Crisis. Vietnam produced stupendous catastrophe. Jimmy Carter’s expedition to free American hostages held in Iran not only failed but also torpedoed his hopes of winning a second term. Ronald Reagan’s 1983 intervention in Beirut wasted the lives of 241 soldiers, sailors, and Marines for reasons that still defy explanation. Reagan also went after Muammar Qaddafi, sending bombers to pound Tripoli; the Libyan dictator responded by blowing up Pan Am flight 103 over Lockerbie, Scotland—and survived to tell the tale. In 1991, George H.W. Bush portrayed Operation Desert Storm as a great victory sure to provide the basis for a New World Order; in fact the first Gulf War succeeded chiefly in drawing the United States more deeply into the vortex of the Middle East—it settled nothing. With his pronounced propensity for flinging about cruise missiles and precision-guided bombs, Bill Clinton gave us Mogadishu, Haiti, Bosnia, and Kosovo —frenetic activity with little to show in return. As for Bush and his wars in Afghanistan and Iraq, the less said the better.

What are we to make of this record? For Krauthammer, Boot, and Barnes, the lessons are clear: dial up the rhetoric, increase military spending, send in more troops, and give the generals a free hand. The important thing, writes William Kristol in his own assessment of Obama’s Afghanistan decision, is to have a commander in chief who embraces “the use of military force as a key instrument of national power.” If we just keep trying, one of these times things will surely turn out all right.

An alternative reading of our recent military past might suggest the following: first, that the political utility of force—the range of political problems where force possesses real relevance—is actually quite narrow; second, that definitive victory of the sort that yields a formal surrender ceremony at Appomattox or on the deck of an American warship tends to be a rarity; third, that ambiguous outcomes are much more probable, with those achieved at a cost far greater than even the most conscientious war planner is likely to anticipate; and fourth, that the prudent statesman therefore turns to force only as a last resort and only when the most vital national interests are at stake. Contra Kristol, force is an “instrument” in the same sense that a slot machine or a roulette wheel qualifies as an instrument.

To consider the long bloody chronicle of modern history, big wars and small ones alike, is to affirm the validity of these conclusions.

Quiggins' and his readers' comments are also worth reading.

Thus we’re left, appropriately, in suspension

6a00d83452446c69e201127938164028a4-800wi

For any serious French writer who has come of age during the last 30 years, one question imposes itself above all others: what do you do after the nouveau roman? Alain Robbe-Grillet, Claude Simon et compagnie redrew the map of what fiction might offer and aspire to, what its ground rules should be – so much so that some have found their legacy stifling. Michel Houellebecq’s response has been one of adolescent rejection, or, to use the type of psychological language that the nouveaux romanciers so splendidly shun, denial: writing in Artforum in 2008, he claimed never to have finished a Robbe-Grillet novel, since they ‘reminded me of soil cutting’. Other legatees, such as Jean Echenoz, Christian Oster and Olivier Rolin, have come up with more considered answers, ones that, at the very least, acknowledge an indebtedness – enough for their collective corpus to be occasionally tagged with the label ‘nouveau nouveau roman’. Foremost among this group, and bearing that quintessentially French distinction of being Belgian, is Jean-Philippe Toussaint. Born in 1957, Toussaint was out of the blocks quickly: by the age of 35 he’d published four novels. It’s the last of these, the so far untranslated La Réticence, which most blatantly betrays his generation’s haunting by its predecessor. With its setting in an off-season fishing village, its quasi-repeating narrative loops that see an eminently unreliable narrator trace and retrace circuits through the corridors of a hotel or to and from the house of an absent friend-cum-rival whom he may or may not have murdered, its obsessive attention to surfaces and objects, or the geometric pulsing of a lighthouse’s ‘cône fulgurant de clarté’ through the black night, over and over – in all these aspects, the book reads like an apprentice’s studied emulation of Robbe-Grillet’s masterpiece The Voyeur.

more from Tom McCarthy at the LRB here.

water

52023314-04190354

It’s not news to residents of Southern California that the management of water resources has far-reaching economic and political ramifications, but even they may be surprised by the pivotal role journalist Steven Solomon assigns to water throughout human history. It is “Earth’s most potent agent of change,” Solomon asserts in his sweeping book, which begins with the birth of civilization, is midwifed by large-scale, irrigated agriculture, and closes in our current “age of water scarcity, [when] water’s always paramount, but usually discreet role in world history is visibly taking its place at center stage.” Solomon’s intelligent, well-informed assessment of the challenges we face today grows naturally from the comprehensive analysis that precedes it in the first three sections: “Water in Ancient History,” “Water and the Ascendancy of the West,” “Water and the Making of the Modern Industrial Society.” The author astutely synthesizes a vast amount of scholarship to spotlight recurring patterns.

more from Steven Solomon at the LAT here.

Saturday Poem

The Quarter

Maybe the problem is that I got involved with the wrong crowd of gods when I was seven. At first they weren't harmful and only showed themselves as fish, birds, especially herons and loons, turtles, a bobcat and a small bear, but not deer and rabbits who only offered themselves as food. And maybe I spent too much time inside the water of lakes and rivers. Underwater seemed like the safest church I could go to. And sleeping outside that young might have seeped too much dark into my brain and bones. It was not for me to ever recover. The other day I found a quarter in the driveway I lost at the Mecosta County Fair in 1947 and missed out on five rides including the Ferris wheel and the Tilt-A-Whirl. I sat in anger for hours in the bull barn mourning my lost quarter on which the entire tragic history of earth is written. I looked up into the holes of the bulls' massive noses and at the brass rings puncturing their noses which allowed them to be led. It would have been an easier life if I had allowed a ring in my nose, but so many years later I still find the spore of the gods here and there but never in the vicinity of quarters.

by Jim Harrison

from In Search of Small Gods;
Copper Canyon Press

The Troubled (Black) History of the Oscars

From The Root:

Hattie-McDaniels- When Hattie McDaniel, the first African American ever nominated for an Academy Award, arrived at the Ambassador Hotel for the 1940 ceremony, she was seated at a table on the extreme periphery of the auditorium. McDaniel had been nominated for Best Supporting Actress based on her role as Mammy in Gone With the Wind (1939). Though this seating assignment was quite insulting, such slights were not uncommon, as McDaniel had also been forced to miss the film’s Atlanta premiere due to southern Jim Crow laws. McDaniel would go on to win the Academy Award that evening in 1940, becoming the first African American to ever win the prestigious award. It would be 24 years before another African American would be declared an Oscar winner. In the 61-year time span from 1940-2001, only five other African Americans—Sidney Poitier, Lou Gossett, Denzel Washington, Whoopi Goldberg and Cuba Gooding (can we put an asterisk next to this one?)—won the distinct gold statuette in the prestigious acting categories. Of those six total awards, Sidney Poitier is the only one to have won in the Best Actor category for his role in Lilies of the Field (1963); all the others were for supporting roles.

So when Eddie Murphy stood up to present the award for Best Picture at the 1988 awards ceremony, the troubled racial history of Hollywood loomed large. Murphy, a comedian whose persona was generally apolitical, decided that the incongruity of his prominent presence at the 1988 ceremony, set against the lack of recognition for African Americans in Hollywood historically, was just too insulting to ignore. Murphy “went off,” chastising the gathered industry figures regarding Hollywood’s racism. He said that if one looked closely at the scattered history of African Americans receiving Oscars, that at the rate things were going, another African American probably would not be receiving an award until the year 2004. Reports after the show were that the program’s producers, along with some of the celebrities at the event, vocalized their displeasure to Murphy, saying that his critique was both rude and inappropriate. Murphy is said to have responded to these criticisms the way that one would expect to him respond—shall we say, by loudly declaring that he was aggressively indifferent to their objections.

More here.

Some Fun Tonight

From The New York Times:

Richard David Kirby’s brief biography of Little Richard reads the way Richard’s shows of the past few decades have played. It’s an engaging, intermittently exciting but ultimately frustrating mix of assertion, reminiscence, free association, repetition, clowning and showing off, with just enough talent on display to keep you from walking out. Kirby, a poet and critic who teaches English at Florida State University, has plenty to say about the rock ’n’ roll “originator,” as Little Richard likes to call himself. It’s carving the good bits out from everything surrounding them that’s the problem. Kirby would probably argue that his spirited digressions — on the banana pudding in Macon, Ga., Little Richard’s hometown; on Southern stereo­typing in the television series “Friday Night Lights”; on Kirby’s African-American playmates when he was growing up in Louisiana in the 1940s and ’50s — are as essential to understanding the importance of Little Richard as the music itself. A more straightforward discussion of his career, in this view, has ­either been done or would be too staid.

Fair enough — to a degree. Charles White’s “Life and Times of Little Richard,” from which Kirby draws liberally, is a riveting book and would be nearly impossible to displace as the definitive biography. But one of Kirby’s stated motivations for writing “Little Richard” is that his subject, who is 77, is a vastly underrated, if not forgotten, figure. It seems disingenuous to maintain simultaneously that Little Richard has fallen off the cultural radar but is also too well documented to warrant a more comprehensive examination.

More here.