WWI Centennial: U.S. Adopts Draft

Library of Congress
Library of Congress

Erik Sass is covering the events of the war exactly 100 years after they happened. This is the 276th installment in the series. 

 

May 18, 1917: U.S. Adopts Draft

 

After the end of the U.S. Civil War, conscription was swiftly abolished and the American military reverted to its traditional all-volunteer basis, with the U.S. Army bolstered by National Guard units when needed. During the Spanish-American War of 1898, the U.S. Army swelled to around a quarter million, all volunteers and National Guardsmen, and U.S. forces involved in the Philippine-American War of 1899-1902 totaled 126,000. Later the Punitive Expedition against Pancho Villa in northern Mexico in 1916-1917 numbered just 10,000 men, with roughly another 130,000 guarding the border.

 

By the time the United States entered the First World War in April 1917, the U.S. Army tallied 128,000 officers and men, along with 182,535 mobilized National Guardsmen. Needless to say, these figures were laughably small compared to the monstrous machines now locked in a titanic death struggle in Europe. In the spring of 1917 Germany had around 5.5 million men under arms, the British Empire 4.5 million, and France had two million serving on the Western Front alone – and these were just a fraction of the total manpower mobilized over the course of the war (France mobilized a total of 8.3 million men, including around half a million colonial troops, from 1914-1918).

 


Library of Congress

Although America had adopted an unconvincing “preparedness program” in 1916, raising the target size for the National Guard to 450,000 by 1921, this goal was far from being realized, and the Americans would obviously be unable to make more than a symbolic contribution to the Allied war effort in terms of manpower in the near future: in July 1917 just 20,000 Americans were deployed in the American Expeditionary Force in Europe, increasing to 129,000 by the end of the year. 

However the United States was hardly prepared to play a secondary role in the long run, demanding an energetic, decisive intervention to bring Germany to terms and end the war. To accomplish this the country would have to train and equip armed forces numbering four million by the end of 1918 – a massive undertaking which would require months of feverish effort, including the construction of a whole network of training camps and, most importantly, bringing back the draft. 

US Army

President Woodrow Wilson set the ball rolling on May 18, 1917, when he signed into law the Selective Service Act passed by Congress, requiring all men ages 21-30 to register for military service on June 5, 1917, with another round of registration scheduled on June 5, 1918 for those who turned 21 after the previous date. As in other belligerent powers that adopted conscription, young men who were unmarried and had no dependents were the first class to be called. Ultimately 25 million American men would register, and 2.5 million would be drafted, while most of the remaining two million volunteered. 

Beginning in late July 1917 draft committees around the country would conduct random drawings from the first class to be called After being called up, potential recruits underwent medical examinations by boards of doctors to determine whether they were fit enough to serve in the Army. These cursory medical reviews clearly erred on the side of passing candidates as fit: the legendary American folk artist Henry Darger somehow managed to pass a medical review and was briefly drafted into the Army despite glaring emotional and developmental issues, not to mention trouble with his knee and eye (to its credit the Army eventually rejected him anyway). 

Nonetheless they rejected around half a million candidates as unfit for a variety of reasons, and medical records from these examinations leave a mixed portrait of American public health in the first part of the 20th century: top reasons disqualifying individuals for service included bad vision, insufficient height or weight, epilepsy, tuberculosis, venereal disease (especially syphilis), goiter, general “mental deficiency,” bad teeth, and curvature of the spine.

 

 

To induce compliance with the draft the government at all levels unleashed a veritable flood of propaganda, including vivid posters meant to appeal to the patriotism, honor, masculinity and sense of self-worth of young American men. Easily the most memorable image of the propaganda campaign mounted by the new Committee for Public Information featured the iconic image of “Uncle Sam” pointing at the viewer with the caption “I want YOU for the U.S. Army” (top). The poster was originally drawn by illustrator James Montgomery Flagg as a cover for Leslie’s Weekly magazine during the preparedness debate in July 1916, but its popularity exploded in 1917-1918, when over four million copies of the poster were printed. Flagg later claimed he used his own face a model for Uncle Sam.

 

See the previous installment or all entries.

12 Facts About Japanese Internment in the United States

Portrait of internee Tom Kobayashi at Manzanar War Relocation Center, Owens Valley, California, 1943
Portrait of internee Tom Kobayashi at Manzanar War Relocation Center, Owens Valley, California, 1943
Ansel Adams, Library of Congress/Wikimedia Commons // No Known Copyright Restrictions

On February 19, 1942, President Franklin Delano Roosevelt issued Executive Order 9066, which sanctioned the removal of Japanese immigrants and Americans of Japanese heritage from their homes to be imprisoned in internment camps throughout the country.

At the time, the move was sold to the public as a strategic military necessity. Following the attack on Pearl Harbor on December 7, 1941, the government argued that it was impossible to know where the loyalties of Japanese-Americans rested.

Between 110,000 and 120,000 people of Japanese ancestry were relocated to internment camps along the West Coast and as far east as Louisiana. Here are 12 facts about what former first lady Laura Bush has described as "one of the most shameful episodes in U.S. history."

1. The government was already discussing detaining people before the Pearl Harbor attack.

In 1936, President Franklin Roosevelt—who was concerned about Japan’s growing military might—instructed William H. Standley, his chief of naval operations, to clandestinely monitor "every Japanese citizen or non-citizen on the island of Oahu who meets these Japanese ships [arriving in Hawaii] or has any connection with their officers or men" and to secretly place their names "on a special list of those who would be the first to be placed in a concentration camp in the event of trouble."

This sentiment helped lead to the creation of the Custodial Detention List, which would later guide the U.S. in detaining 31,899 Japanese, German, and Italian nationals, separate from the 110,000-plus later interred, without charging them with a crime or offering them any access to legal counsel.

2. Initial studies of the “Japanese problem” proved that there wasn’t one.

In early 1941, Curtis Munson, a special representative of the State Department, was tasked with interviewing West Coast-based Japanese-Americans to gauge their loyalty levels in coordination with the FBI and the Office of Naval Intelligence. Munson reported that there was extraordinary patriotism among Japanese immigrants, saying that "90 percent like our way best," and that they were "extremely good citizen[s]" who were "straining every nerve to show their loyalty." Lieutenant Commander K.D. Ringle’s follow-up report showed the same findings and argued against internment because only a small percentage of the community posed a threat, and most of those individuals were already in custody.

3. The general in charge of Western defense command took nothing happening after Pearl Harbor as proof that something would happen.

Minidoka Relocation Center. Community Store in block 30
National Archives at College Park, Wikimedia Commons // CC BY 3.0

Despite both Munson and Ringle debunking the concept of internment as a strategic necessity, the plan moved ahead—spurred largely by Western Defense Command head General John L. DeWitt. One month after Pearl Harbor, DeWitt created the central ground for mass incarceration by declaring: "The fact that nothing has happened so far is more or less ... ominous in that I feel that in view of the fact that we have had no sporadic attempts at sabotage that there is a control being exercised and when we have it, it will be on a mass basis."

DeWitt, whose ancestors were Dutch, didn’t want anyone of Japanese descent on the West Coast, stating that “American citizenship does not necessarily determine loyalty.”

4. Almost no one protested internment.

Alongside General DeWitt, Wartime Civil Control Administration director Colonel Karl Bendetsen avowed that anyone with even “one drop of Japanese blood” should be incarcerated, and the country generally went along with that assessment. Some newspapers ran op-eds opposing the policy, and the American Baptist Home Mission Societies created pamphlets to push back, but as historian Eric Foner wrote in The Story of American Freedom, "One searches the wartime record in vain for public protests among non-Japanese." Senator Robert Taft was the only congressperson to condemn the policy.

5. Supporting or opposing internment were both matters of economics.

White farmers and landowners on the West Coast had great economic incentives to get rid of Japanese farmers who had come to the area only decades before and found success with new irrigation methods. They fomented deep hatred for their Japanese neighbors and publicly advocated for internment, which is one reason so many of the more than 110,000 Japanese individuals sent to camps came from the West Coast. In Hawaii, it was a different story. White business owners opposed internment, but not for noble reasons: They feared losing their workforce. Thus, only between 1200 and 1800 Japanese-Americans from Hawaii were sent to internment camps.

6. People were tagged for identification.

Children in a drawing class at Minidoka Relocation Center
National Archives at College Park, Wikimedia Commons // CC BY 3.0

Moving entire communities of people to camps in California, Colorado, Texas, and beyond was a gargantuan logistical task. The military assigned tags with ID numbers to families, including the children, to ensure they would be transferred to the correct camp. In 2012, artist Wendy Maruyama recreated thousands of these tags for an art exhibition she titled "The Tag Project."

"The process of replicating these tags using government databases, writing thousands of names, numbers, and camp locations became a meditative process," Maruyama told Voices of San Diego. “And for the hundreds of volunteers, they could, for a minute or two as they wrote the names, contemplate and wonder what this person was thinking as he or she was being moved from the comforts of home to the spare and bare prisons placed in the foreboding deserts and wastelands of America. And could it happen again?”

7. Not everyone went quietly.

Directly combatting the image of the “polite” Japanese-Americans who acquiesced to internment without protest, collections of resistance stories paint a disruptive picture of those who refused to go to the camps or made trouble once inside. Among those who were considered "problematic" were individuals who refused to register for the compulsory loyalty questionnaire, which asked questions about whether the person was a registered voter and with which party, as well as marital status and "citizenship of wife" and "race of wife."

“A broadly understood notion of resistance represents a more complete picture of what happened during World War II,” David Yoo, a professor of Asian American Studies and History and vice provost at UCLA's Institute of American Cultures, told NBC News about collecting these resistance stories. “Because these stories touch upon human rights, they are important for all peoples.”

8. The government converted unused buildings into camp facilities.

For the most part, camps were set against desert scrub land or infertile Ozark hills bordered with barbed wire. Before getting on buses to be transported to their new "homes," detainees had to go through processing centers housed in converted racetracks and fairgrounds, where they might stay for several months. The largest and most noteworthy center was Santa Anita Park, a racetrack in Arcadia, California, which was shut down so that makeshift barracks could be assembled and horse stables could be used for sleeping quarters.

9. Ansel Adams took hundreds of photographs inside the most famous camp, as did an internee with a smuggled camera.

Wooden sign at entrance to the Manzanar War Relocation Center with a car at the gatehouse in the background
Ansel Adams, Library of Congress/Wikimedia Commons // Public Domain

Approximately 200 miles north of Santa Anita Park, at the foot of the Sierra Nevada mountain range, was Manzanar—which, with its 11,000 internees, was perhaps the most famous of America's 10 relocation centers. It was also the most photographed facility. In the fall of 1942, famed photographer Ansel Adams—who was personally outraged by the situation when a family friend was taken from his home and moved halfway across the country—shot more than 200 images of the camp. In a letter to a friend about a book being made of the photos, Adams wrote that, "Through the pictures the reader will be introduced to perhaps 20 individuals ... loyal American citizens who are anxious to get back into the stream of life and contribute to our victory."

While Adams may have successfully offered a small glimpse at life inside Manzanar, Tōyō Miyatake—a photographer and detainee who managed to smuggle a lens and film into the camp, which he later fashioned into a makeshift camera—produced a series of photos that offered a much more intimate depiction of what everyday life was like for the individuals who were imprisoned there between 1942 and 1945. Today, Manzanar is a National Historic Site.

10. Detainees were told they were in camps for their own protection.

Japanese-Hawaiian hula dancers on an improvised stage during one of the frequent talent shows at Santa Anita (California) Assembly Center
U.S. Signal Corps, Library of Congress, Wikimedia Commons // Public Domain

Just as the justification for internment was an erroneous belief in mass disloyalty among a single racial group, the argument given to those incarcerated was that they were better off inside the barbed wire compounds than back in their own homes, where racist neighbors could assault them. When presented with that logic, one detainee rebutted, “If we were put there for our protection, why were the guns at the guard towers pointed inward, instead of outward?”

11. Internees experienced long-term health problems because of the camps, and children had it the worst.

Internment officially lasted through 1944, with the last camp closing in early 1946. In those years, Japanese-Americans did their best to make lives for themselves on the inside. That included jobs and governance, as well as concerts, religion, and sports teams. Children went to school, but there were also dances and comic books to keep them occupied. But the effects of their internment were long-lasting.

There have been multiple studies of the physical and psychological health of former internees. They found those placed in camps had a greater risk for cardiovascular disease and death, as well as traumatic stress. Younger internees experienced low self-esteem, as well as psychological trauma that led many to shed their Japanese culture and language. Gwendolyn M. Jensen’s The Experience of Injustice: Health Consequences of the Japanese American Internment found that younger internees “reported more post-traumatic stress symptoms of unexpected and disturbing flashback experiences than those who were older at the time of incarceration.”

12. A congressional panel called it a “grave injustice" ... 40 years later.

Japanese Americans going to Manzanar gather around a baggage car at the old Santa Fe Station. (April 1942)
Russell Lee, Library of Congress, Wikimedia Commons // Public Domain

It wasn’t until 1983 that a special Congressional commission determined that the mass internment was a matter of racism and not of military strategy. Calling the incarceration a “grave injustice,” the panel cited the ignored Munson and Ringle reports, the absence of any documented acts of espionage, and delays in shutting down the camps due to weak political leadership from President Roosevelt on down as factors in its conclusion. The commission paved the way for President Reagan to sign the Civil Liberties Act, which gave each surviving internee $20,000 and officially apologized. Approximately two-thirds of the more than 110,000 people detained were U.S. citizens.

This list first ran in 2018.

The Disturbing Reason Schools Tattooed Their Students in the 1950s

Kurt Hutton, Hulton Archive/Getty Images
Kurt Hutton, Hulton Archive/Getty Images

When Paul Bailey was born at Beaver County Hospital in Milford, Utah on May 9, 1955, it took less than two hours for the staff to give him a tattoo. Located on his torso under his left arm, the tiny marking was rendered in indelible ink with a needle gun and indicated Bailey’s blood type: O-Positive.

“It is believed to be the youngest baby ever to have his blood type tattooed on his chest,” reported the Beaver County News, cooly referring to the infant as an “it.” A hospital employee was quick to note parental consent had been obtained first.

The permanent tattooing of a child who was only hours old was not met with any hysteria. Just the opposite: In parts of Utah and Indiana, local health officials had long been hard at work instituting a program that would facilitate potentially life-saving blood transfusions in the event of a nuclear attack. By branding children and adults alike with their blood type, donors could be immediately identified and used as “walking blood banks” for the critically injured.

Taken out of context, it seems unimaginable. But in the 1950s, when the Cold War was at its apex and atomic warfare appeared not only possible but likely, children willingly lined up at schools to perform their civic duty. They raised their arm, gritted their teeth, and held still while the tattoo needle began piercing their flesh.

 

The practice of subjecting children to tattoos for blood-typing has appropriately morbid roots. Testifying at the Nuremberg Tribunal on War Crimes in the 1940s, American Medical Association physician Andrew Ivy observed that members of the Nazi Waffen-SS carried body markings indicating their blood type [PDF]. When he returned to his hometown of Chicago, Ivy carried with him a solution for quickly identifying blood donors—a growing concern due to the outbreak of the Korean War in 1950. The conflict was depleting blood banks of inventory, and it was clear that reserves would be necessary.

School children sit next to one another circa the 1950s
Reg Speller, Fox Photos/Getty Images

If the Soviet Union targeted areas of the United States for destruction, it would be vital to have a protocol for blood transfusions to treat radiation poisoning. Matches would need to be found quickly. (Transfusions depend on matching blood to avoid the adverse reactions that come from mixing different types. When a person receives blood different from their own, the body will create antibodies to destroy the red blood cells.)

In 1950, the Department of Defense placed the American Red Cross in charge of blood donor banks for the armed forces. In 1952, the Red Cross was the coordinating agency [PDF] for obtaining blood from civilians for the National Blood Program, which was meant to replenish donor supply during wartime. Those were both measures for soldiers. Meanwhile, local medical societies were left to determine how best to prepare their civilian communities for a nuclear event and its aftermath.

As part of the Chicago Medical Civil Defense Committee, Ivy promoted the use of the tattoos, declaring them as painless as a vaccination. Residents would get blood-typed by having their finger pricked and a tiny droplet smeared on a card. From there, they would be tattooed with the ABO blood group and Rhesus factor (or Rh factor), which denotes whether or not a person has a certain type of blood protein present.

The Chicago Medical Society and the Board of Health endorsed the program and citizens voiced a measure of support for it. One letter to the editor of The Plainfield Courier-News in New Jersey speculated it might even be a good idea to tattoo Social Security numbers on people's bodies to make identification easier.

Despite such marked enthusiasm, the project never entered into a pilot testing stage in Chicago.

Officials with the Lake County Medical Society in nearby Lake County, Indiana were more receptive to the idea. In the spring of 1951, 5000 residents were blood-typed using the card method. But, officials cautioned, the cards could be lost in the chaos of war or even the relative quiet of everyday life. Tattoos and dog tags were encouraged instead. When 1000 people lined up for blood-typing at a county fair, two-thirds agreed to be tattooed as part of what the county had dubbed "Operation Tat-Type." By December 1951, 15,000 Lake County residents had been blood-typed. Roughly 60 percent opted for a permanent marking.

The program was so well-received that the Lake County Medical Society quickly moved toward making children into mobile blood bags. In January 1952, five elementary schools in Hobart, Indiana enrolled in the pilot testing stage. Children were sent home with permission slips explaining the effort. If parents consented, students would line up on appointed tattoo days to get their blood typed with a finger prick. From there, they’d file into a room—often the school library—set up with makeshift curtains behind which they could hear a curious buzzing noise.

When a child stepped inside, they were greeted by a school administrator armed with indelible ink and wielding a Burgess Vibrotool, a medical tattoo gun featuring 30 to 50 needles. The child would raise their left arm to expose their torso (since arms and legs might be blown off in an attack) and were told the process would only take seconds.

A child raises his hand in class circa the 1950s
Vecchio/Three Lions/Getty Images

Some children were stoic. Some cried before, during, or after. One 11-year-old recounting her experience with the program said a classmate emerged from the session and promptly fainted. All were left with a tattoo less than an inch in diameter on their left side, intentionally pale so it would be as unobtrusive as possible.

At the same time that grade schoolers—and subsequently high school students—were being imprinted in Indiana, kids in Cache and Rich counties in Utah were also submitting to the program, despite potential religious obstacles for the region's substantial Mormon population. In fact, Bruce McConkie, a representative of the Church of Jesus Christ of Latter-Day Saints, declared that blood-type tattoos were exempt from the typical prohibitions on Mormons defacing their bodies, giving the program a boost among the devout. The experiment would not last much longer, though.

 

By 1955, 60,000 adults and children had gotten tattooed with their blood types in Lake County. In Milford, health officials persisted in promoting the program widely, offering the tattoos for free during routine vaccination appointments. But despite the cooperation exhibited by communities in Indiana and Utah, the programs never spread beyond their borders.

The Korean conflict had come to an end in 1953, reducing the strain put on blood supplies and along with it, the need for citizens to double as walking blood banks. More importantly, outside of the program's avid boosters, most physicians were extremely reticent to rely solely on a tattoo for blood-typing. They preferred to do their own testing to make certain a donor was a match with a patient.

There were other logistical challenges that made the program less than useful. The climate of a post-nuclear landscape meant that bodies might be charred, burning off tattoos and rendering the entire operation largely pointless. With the Soviet Union’s growing nuclear arsenal—1600 warheads were ready to take to the skies by 1960—the idea of civic defense became outmoded. Ducking and covering under desks, which might have shielded some from the immediate effects of a nuclear blast, would be meaningless in the face of such mass destruction.

Programs like tat-typing eventually fell out of favor, yet tens of thousands of adults consented to participate even after the flaws in the program were publicized, and a portion allowed their young children to be marked, too. Their motivation? According to Carol Fischler, who spoke with the podcast 99% Invisible about being tattooed as a young girl in Indiana, the paranoia over the Cold War in the 1950s drowned out any thought of the practice being outrageous or harmful. Kids wanted to do their part. Many nervously bit their lip but still lined up with the attitude that the tattoo was part of being a proud American.

Perhaps equally important, children who complained of the tattoo leaving them particularly sore received another benefit: They got the rest of the afternoon off.

SECTIONS

arrow
LIVE SMARTER