Grumpy opinions about everything.

Month: August 2025

The Constitutional Foundations

Who Controls Elections?

Donald Trump has repeatedly claimed that the president should have broad authority to change how elections are conducted—particularly when it comes to abolishing mail-in voting and voting machines. As recently as August 2025, Trump pledged to issue an executive order banning mail-in ballots and voting machines ahead of the 2026 midterm elections, insisting that states must comply with his directive because, in his words, “States act merely as ‘agents’ for the Federal Government when it comes to counting and tabulating votes.… They are required to follow what the Federal Government, represented by the President of the United States, instructs them to do, FOR THE GOOD OF OUR COUNTRY”.

But this isn’t the first time he has suggested that he could control the election process.  In March 2025, Trump issued a major executive order titled “Preserving and Protecting the Integrity of American Elections” that aims to expand presidential control over the election process.  The order attempts to direct the Election Assistance Commission (EAC) — an independent, bipartisan agency — to mandate that voters show a passport or other similar document proving citizenship when registering to vote using the federal voter registration form.  The executive order has been the subject of extensive litigation, and several federal judges have issued injunctions against various portions of it.

Amid the COVID-19 pandemic during his first term, President Trump publicly suggested delaying the election. Constitutional scholars and members of Congress quickly pointed out he lacked such authority—the date of federal elections is set by statute, and only Congress could change it.

The U.S. Constitution provides a clear framework for who holds the authority to control elections, and it is not the president.

Article I, Section 4: Congressional and State Authority

The main constitutional authority over U.S. elections is found in Article I, Section 4, commonly called the “Elections Clause.” It states:

“The Times, Places and Manner of holding Elections for Senators and Representatives, shall be prescribed in each State by the Legislature thereof; but the Congress may at any time by Law make or alter such Regulations…”

This language charges state legislatures with defining the details of congressional elections, including logistics and procedures. Importantly, Congress retains the power to override state laws and impose federal rules—such as standardized Election Days or regulations for voter registration and districting.

What does this mean for the president? The Constitution is clear: the president has no direct authority to determine the conduct of congressional elections or to unilaterally change the way federal elections are held. Presidential influence over elections is limited to signing or vetoing congressional legislation, not acting alone.

Article II and the 12th Amendment: Presidential Elections

Presidential elections are regulated by Article II, which created the Electoral College, and by the 12th Amendment .

Article II, Section 1 provides:

“Each State shall appoint, in such Manner as the Legislature thereof may direct, a Number of Electors…”

States arrange how their presidential electors are selected, subject to changes imposed by congressional law. The federal government, through Congress (not the president!), determines the timing of choosing electors and casting electoral votes. The 12th Amendment sets procedures for how electors meet and vote for both president and vice president.

Again, neither Article II nor the 12th Amendment gives the president authority to independently set election rules. At most, the president can recommend reforms, sign laws crafted by Congress, and advocate for certain policies.

Historical Examples of Limits on Presidential Power Over Elections

Even during national crises, presidents have not been able to unilaterally change election rules:

  • 1864 Election (Lincoln): Despite the Civil War, Abraham Lincoln did not postpone or suspend the presidential election. Elections were carried out in the states, including special arrangements for soldiers to vote.
  • 1944 Election (Roosevelt): In the midst of World War II, Franklin Roosevelt stood for re-election. Again, no effort was made by the president to change election laws.

Presidential Powers: What Can the Executive Branch Do?

The president’s responsibilities in elections are more limited than you might expect and are essentially ministerial and ceremonial, not regulatory.

The executive power in Article II invests the president with broad national leadership, command of the military, and responsibility to “take Care that the Laws be faithfully executed”. This can include enforcing voting rights laws and overseeing federal agencies that support election integrity. However, the Constitution and decades of legal precedent restrict the president from directly controlling election rules.

  • The president cannot by executive order change state rules for voting methods (e.g., mail-in voting, voting machines).
  • The president cannot unilaterally suspend or postpone federal elections.
  • The president cannot direct states to alter their voter registration, polling locations, or other administrative details.
  • The president has no role in certifying state results. That function belongs to state officials, with Congress responsible for counting electoral votes.
  • The president can direct federal agencies like the Department of Justice to enforce federal election laws, protect voting rights, and intervene in cases of fraud or intimidation.  The president does not have the authority to direct federal agencies to act in a manner contrary to the law.

When presidents have sought to influence election administration more directly, courts and Congress have reaffirmed the constitutional boundaries. For example, efforts to change the date of an election or prohibit certain voting methods without congressional action have consistently failed in the courts.

Congressional Power: The Real Check on Election Rules

While state legislatures remain the primary manager of elections, Congress retains the final word. The Supreme Court has confirmed that congressional law “preempts” conflicting state rules in matters of federal elections. When Congress acts—through laws like the Voting Rights Act, Help America Vote Act, and the National Voter Registration Act—states must comply, and the president’s role is simply to sign or veto those laws.

Congress has used its power over the years to:

  • Set a uniform national Election Day.
  • Establish protections for disabled voters and overseas citizens.
  • Mandate requirements around voter registration and accessibility.
  • Regulate campaign finance and transparency.

Checks, Balances, and Modern Tensions

Recent political debates have seen calls for presidents to take stronger action on election oversight, especially regarding the use of mail-in ballots or voting machines. However, these calls run up against clear constitutional limits: the president cannot rewrite the rules of elections without Congress or state legislatures.

Any presidential attempt to do so by executive order would face swift legal challenges and almost certainly be invalidated. The intent of the Framers was to divide election power between the states and Congress, with the president largely excluded from direct rule-making authority. This balance—central to federalism—protects elections from potential abuses of executive power and ensures that reforms require broad democratic consensus. While presidents can champion reforms and enforce federal laws supporting fair elections, they are constitutionally forbidden from unilaterally changing election rules.

Conclusion

The framework isn’t perfect—it can create confusion when state and federal authorities clash. But the basic principle remains: states run elections. Congress can regulate them within constitutional bounds, and presidents enforce the resulting laws.

For citizens, lawmakers, and presidents alike, respect for these boundaries secures the foundation of American democracy. The right to vote—and the integrity of how that vote is counted—is protected not by any single leader, but by enduring constitutional principles and the shared power of states and Congress.

Is Fairness a Moral Principle or Just a Feeling?

We live in a culture obsessed with “fairness.” If something doesn’t go our way, we brand it “unfair.” If it does, we call it “fair.” But fairness is inherently subjective: what’s fair to you may feel wildly unfair to me.

When Fairness First Mattered

I first noticed this in my kids. Whenever I told them “No,” they’d cry, “That’s not fair!”— but could only shrug when I asked why. Most of us can’t define fairness any better than a six-year-old: it’s simply a catch-all for “I don’t like this.”

Lately, “unfair” has become a socially acceptable way to say “I disagree.” It even carries moral weight—as if fairness were an absolute like good or evil. But these concepts aren’t easy to pin down.

Is it fair that some people are born with musical talent, while others struggle to clap on beat? Is it fair that some are naturally athletic, while others can’t run a mile without seeing stars? (If it were up to me, I’d call it unfair, since I have neither musical nor athletic ability.)

The Real Inequities

Life is rife with natural inequalities. Those born with an advantage deem it “fair” they benefit; those born without cry “unfair” and demand special considerations. And while I sympathize, everyone roots for the underdog, handouts aren’t the same as opportunity.

The Grumpy Doc’s Take

Here’s what I tell my kids (and myself): There is no absolute fairness. You can’t craft a definition that pleases everyone. Things simply are the way they are:

  • Don’t like it? Roll up your sleeves and change it.
  • Like it? Work to preserve it.

To me, fairness is best understood as this: Everyone should have an equal opportunity to work for what they want. That doesn’t mean we all start in the same place. It does mean we should try to ensure that those who’ve been historically disadvantaged get an even chance—one that accounts for past inequities. And to those who insist you’ve never enjoyed special treatment: what’s your golf handicap?

Embrace the Inequalities

Do I know how to guarantee equal opportunity? No. Others who are far smarter than me will have to figure that out. Am I upset that they’re smarter than me? Not at all. If I were the smartest person around, we’d all be in trouble.

In the end, fairness isn’t a universal yardstick. It’s an invitation to participate: to work, compete, and strive. Because if life were perfectly fair, we’d never need to improve it.

Citizens United: A Supreme Court Decision That Reshaped American Politics

In January 2010, the U.S. Supreme Court issued one of the most consequential and controversial rulings in modern political history: Citizens United v. Federal Election Commission. The 5–4 decision dramatically altered the legal landscape of campaign finance, opening the door for unlimited spending by corporations, unions, and certain nonprofit organizations, potentially giving them disproportionate influence in election campaigning.

Hailed by some as a victory for free expression and condemned by others as unleashing a torrent of special interest cash into politics, Citizens United has continued to define the shape of campaign financing for the past 15 years.

The Origins of the Case

The lawsuit began with a film. In 2008, Citizens United, a conservative nonprofit organization led by David Bossie, produced “Hillary: The Movie,” a 90-minute documentary highly critical of then-Senator Hillary Clinton, who was seeking the Democratic presidential nomination. Citizens United wanted to air the film on cable television through video-on-demand services within 30 days of the Democratic primary elections and planned to run promotional advertisements.

However, federal campaign finance law under the Bipartisan Campaign Reform Act (also known as McCain-Feingold) prohibited corporations from using general treasury funds for “electioneering communications” within specific timeframes before elections.

Fearing civil and criminal penalties, Citizens United sought a court declaration that their film and promotional materials were exempt from these restrictions.

The organization argued that because the documentary didn’t explicitly tell viewers how to vote, it shouldn’t be classified as campaign advocacy subject to corporate spending limits. A federal district court disagreed, ruling unanimously that the film could only be interpreted as telling viewers that Clinton was “unfit for office” and encouraging them to vote against her.

The Supreme Court’s Landmark Ruling

Citizens United appealed, and the Supreme Court ultimately agreed to hear the case, raising questions not just about this particular film, but about the broader constitutionality of limiting corporate and union election expenditures.

In their decision, the justices went far beyond the narrow question in the original case. In a 5-4 decision authored by Justice Anthony Kennedy, and joined by Chief Justice Roberts and Justices Thomas, Scalia, and Alito, the Court ruled that restrictions on independent political expenditures by corporations and unions violated the First Amendment’s free speech protections.

The majority opinion overturned two significant precedents: Austin v. Michigan Chamber of Commerce (1990) and portions of McConnell v. Federal Election Commission (2003). Justice Kennedy wrote that political speech is “indispensable to decision making in a democracy, and this is no less true because the speech comes from a corporation.”

Justice John Paul Stevens, writing in dissent, warned that the decision represented “a rejection of the common sense of the American people, who have recognized a need to prevent corporations from undermining self-government.”

The ruling did not lift limits on direct contributions to candidates; those caps remain in place. But it cleared the way for unlimited spending on independent political advocacy, so long as it is not coordinated with a candidate’s campaign.

The majority decision made two key assumptions: that independent political spending wouldn’t lead to corruption because it would be transparent, and that it would remain truly separate from candidate campaigns. Both assumptions have proven incorrect in practice.

Transforming the Political Landscape

The decision’s impact has been dramatic and far-reaching. Outside spending in federal elections skyrocketed from $730 million at the time of the ruling to $4.5 billion in 2024. The ruling enabled the creation of “super PACs”—political action committees that can raise and spend unlimited amounts as long as they maintain nominal (sometimes fictional) independence from campaigns.

Each election cycle since 2010 has seen a new record in campaign spending, much of it “outside money”—funds raised and spent by organizations not directly affiliated with candidates or parties. The influence of wealthy donors has only grown, with some estimates suggesting that the vast majority of outside election funding comes from a small handful of deep-pocketed interests.

Perhaps more concerning to democracy advocates is the rise of “dark money” — political spending where the funding source remains secret. Dark money expenditures increased from less than $5 million in 2006 to over $1 billion in the 2024 presidential election alone.

The 2024 election exemplified Citizens United‘s influence. Billionaire-backed super PACs helped close substantial fundraising gaps, with groups like those funded by Elon Musk taking on core campaign functions including voter outreach operations, while supposedly remaining independent. This concentration of political influence among ultra-wealthy donors represents a fusion of private wealth and political power unseen since the late 19th century.

Current Political Implications

Today, Citizens United remains deeply unpopular with the American public. A Washington Post – ABC News poll found that 80% of Americans opposed the Citizens United ruling, Including 85% of Democrats, 76% of Republicans, and 81% of independents.

The ruling has created what some campaign finance experts call a “corruption bomb” effect, where wealthy individuals can effectively buy political influence through seemingly independent expenditures. Recent legislative efforts, including the proposed Abolish Super PACs Act introduced in Congress, aim to restore some limits on political spending, though prospects for passage remain unlikely.

The Citizens United decision continues to shape American politics in 2025. Campaigns increasingly rely on outside groups to fund negative advertising, which can deepen political polarization. With no upper limit on independent expenditures, candidates may feel beholden to the interests of big-money backers who can tip the scales in tight races.

Supporters of the ruling argue that it promotes free speech, enabling more voices to be heard in the political arena. They contend that limiting corporate or union spending would amount to government censorship. Opponents counter that equating money with speech effectively drowns out the voices of ordinary voters who cannot match the spending power of corporations or billionaires.

Some reform advocates are pursuing constitutional amendments to overturn Citizens United, though such efforts face steep political and procedural hurdles. Others push for enhanced disclosure laws to ensure voters know who is funding political messages.

As the 2026 midterm elections approach, Citizens United‘s legacy continues to define the relationship between money and political power. It raises fundamental questions about whether democratic governance can function effectively when political speech is increasingly dominated by the ultra-wealthy. Should the First Amendment protect unlimited political spending by corporations and unions, or does such spending distort democracy by giving disproportionate influence to the wealthiest? The Court’s ruling in Citizens United has transformed the way American elections are fought and the consequences of that decision are still unfolding.

One for the Road? How Alcohol Affects Us as We Age

I’ve always enjoyed a cocktail or glass of wine with dinner, but recently I’ve noticed that if I decide to have a second drink it affects me more than it ever has before.  As we age, our relationship with alcohol undergoes significant changes that many people don’t fully understand. What might have been a manageable glass of wine or cocktail with dinner in your thirties can have dramatically different effects in your sixties and beyond. Understanding these changes is crucial for maintaining health and safety as you grow older.

How Aging Changes Alcohol Processing

The human body’s ability to process alcohol diminishes considerably with age, creating a perfect storm of physiological changes that make older adults more vulnerable to alcohol’s effects. These changes begin gradually in middle age and become more pronounced as we enter our senior years.

Decreased Lean Body Mass and Increased Fat: As we age, our bodies naturally lose muscle mass and gain fat tissue. Since alcohol distributes primarily in water-rich lean tissue rather than fat, older adults have less space for alcohol to distribute throughout our bodies. This means the same amount of alcohol that once felt manageable now results in higher blood alcohol concentrations.

Reduced Water Content: Our total body water content decreases significantly with age, dropping from about 60% in young adults to roughly 50% or less in older adults. With less water to dilute alcohol, even modest amounts can lead to higher concentrations in the bloodstream and more pronounced effects.

Slower Metabolism: The liver, our body’s primary alcohol-processing center, becomes less efficient with age. Liver enzymes responsible for breaking down alcohol work more slowly, meaning alcohol stays in the system longer. What once took an hour to metabolize might now take two hours or more, prolonging both the effects and potential for harm.

In sum: Studies now show that chronic and heavy alcohol consumption can actually accelerate the body’s biological aging, making tissues and organs age faster than our calendar years suggest. This effect is especially strong with liquor and binge drinking, both driving up markers of cellular aging more than moderate beer or wine consumption.  A drink that felt “fine” at 40 might lead to wooziness or poor judgment at 70—not because of frailty or weakness, but because of predictable physiological changes.

Physical Effects of Alcohol on the Aging Body

The aging process makes our bodies more susceptible to alcohol’s negative effects across multiple systems. These changes can have serious implications for both immediate safety and long-term health.

Cardiovascular Impact: While moderate alcohol consumption has sometimes been associated with heart benefits in younger adults, aging changes this equation. Older adults are more likely to have existing cardiovascular conditions, and alcohol can exacerbate high blood pressure, irregular heart rhythms, and heart disease. The dehydrating effects of alcohol become more problematic as our bodies become less efficient at maintaining fluid balance.

Brain and Cognitive Effects: The aging brain is particularly vulnerable to alcohol’s effects. Alcohol can worsen age-related cognitive decline and increase the risk of falls due to impaired balance and coordination. Even small amounts can significantly impact reaction time, judgment, and memory in older adults. Regular consumption may accelerate cognitive decline and increase dementia risk.

Bone Health: Alcohol interferes with calcium absorption and bone formation, making older adults more susceptible to osteoporosis and fractures. Since aging already increases fracture risk, alcohol consumption compounds this danger significantly.

Sleep Disruption: While alcohol might initially seem to help with sleep, it actually disrupts sleep quality, particularly in older adults who already face age-related sleep challenges. Poor sleep quality can cascade into numerous other health problems, from weakened immunity to increased fall risk.

Dangerous Drug Interactions

Perhaps the most critical concern for older adults and alcohol consumption is the potential for dangerous interactions with medications. Adults over 65 take an average of four prescription medications, and many of these can have serious interactions with alcohol.

Blood Thinners: Medications like warfarin (Coumadin), apixaban (Eliquis), and even aspirin can have dangerous interactions with alcohol. Alcohol can either increase bleeding risk to dangerous levels or, paradoxically, reduce the medication’s effectiveness, increasing stroke risk.

Diabetes Medications: Alcohol can cause unpredictable blood sugar changes, particularly dangerous for those taking insulin or medications like metformin. The combination can lead to severe hypoglycemia, which can be life-threatening.

Blood Pressure Medications: ACE inhibitors, beta-blockers, and diuretics can all interact with alcohol, potentially causing dangerous drops in blood pressure, dizziness, and increased fall risk.

Pain Medications: The combination of alcohol with opioid pain medications can cause severe respiratory depression with possibly fatal results. Even over-the-counter pain relievers like acetaminophen, when combined with alcohol, can potentially cause liver damage.

Sleep Aids and Anxiety Medications: Benzodiazepines like lorazepam (Ativan) or sleep medications like zolpidem (Ambien) can have amplified sedating effects when combined with alcohol, increasing risk of falls, accidents, and respiratory depression.

Antidepressants: Many older adults take antidepressants, and alcohol can interfere with their effectiveness while increasing side effects like drowsiness and coordination problems.

Key Takeaway: Even small amounts of alcohol can react with medications, increasing accident risk and threatening life. Always consult a healthcare provider about drinking while taking any medication.

Guidelines for Safer Alcohol Consumption

The National Institute on Aging recommends that adults over 65 limit themselves to no more than one drink per day and no more than seven drinks per week, significantly lower than recommendations for younger adults. However, even this guideline may be too liberal for many older adults, particularly those taking medications or with underlying health conditions.

Key Safety Strategies: Always consult with healthcare providers about alcohol consumption and medication interactions. Keep a detailed list of all medications, including over-the-counter drugs and supplements, and review them regularly with doctors and pharmacists. Consider the cumulative effects of multiple medications when making decisions about alcohol consumption.

Timing Matters: If you choose to drink, timing can be crucial. Avoiding alcohol within several hours of taking medications can reduce interaction risks, though some medications require longer intervals.

Recommendations: According to the National Institute on Alcohol Abuse and Alcoholism, adults over 65 should:limit consumption to no more than one standard drink per day;have at least two alcohol-free days per week; avoid alcohol entirely when taking certain medications or managing specific chronic illnesses.  A standard drink is defined as:12 oz of beer (5% alcohol), 5 oz of wine (12% alcohol), or 1.5 oz of distilled spirits (40% alcohol).  Please note that this is smaller than most restaurant or cocktail lounge servings, particularly for distilled spirits that usually average about 3 oz or more per drink.

For some people, the safest level of drinking may be none at all—particularly if falls, liver disease, or cognitive impairment are concerns.

Warning Signs

Alcohol misuse in older adults is often overlooked. Its symptoms can mimic—or be mistaken for—age-related issues like memory loss, depression, or poor balance. Some warning signs include: frequent falls or bruises, unexplained memory lapses or confusion, changes in sleep patterns, neglect of hygiene or nutrition, social withdrawal or irritability, mixing alcohol with medications—intentionally or unintentionally.

Even if alcohol isn’t consumed in large quantities, it can still be harmful if consumed regularly in combination with underlying health conditions or with medications—both prescription and over the counter.

The Bottom Line

Aging fundamentally changes how our bodies process and respond to alcohol, making us more vulnerable to both immediate dangers and long-term health consequences. The combination of physiological changes, increased medication use, and higher baseline health risks means that alcohol consumption strategies that worked in our younger years may no longer be safe or appropriate.

An important step is open communication with healthcare providers about alcohol consumption and its potential interactions with medications and health conditions. For many older adults, reducing alcohol consumption or abstaining entirely becomes the safest choice for maintaining health, independence, and quality of life in their later years. I still enjoy the occasional drink, just not as much or as often as when I was younger—like many things in life.

Resources

The Substance Abuse and Mental Health Services Administration (www.SAMHSA.gov) has a helpline at 1-800-662-HELP (4357).

National Institute on Aging (https://www.nia.nih.gov/health/alcohol-misuse-or-alcohol-use-disorder/facts-about-aging-and-alcohol)

Mayo Clinic (https://www.mayoclinic.org/search/search-results?q=alcohol%20and%20aging).

Centers for Disease Control and Prevention (https://www.cdc.gov/mmwr/preview/mmwrhtml/00001403.htm)

Deborah Sampson: A Revolutionary Soldier

In the story of the American Revolution, the names most often remembered are those of the Founding Fathers and battlefield generals. Yet woven through the familiar narrative are lesser known but extraordinary individuals whose actions defied the norms of their time. One of the most remarkable among them was Deborah Sampson, a Massachusetts woman who disguised herself as a man and served for nearly two years in the Continental Army. Her life reflects not only courage and patriotism, but also the complexity of gender roles in Revolutionary America

A Difficult Early Life

Deborah Sampson was born in Plympton, Massachusetts, in 1760 as the eldest of seven children in a family with deep Pilgrim roots, tracing lineage to Myles Standish and Governor William Bradford. Despite this heritage, her family struggled financially, and she grew up with poverty and abandonment. Her father deserted the family when she was young, leaving her mother with limited resources to care for their children. It was initially thought that he had died at sea, but they later discovered he had actually moved to Maine where he married and raised a second family.

Deborah was still young when her mother died and she was sent to live with a widow, Mary Price Thatcher, then in her 80s. Deborah likely learned to read while living with her.  After Widow Thatcher died, Deborah was bound out as an indentured servant to the Thomas family in Middleborough, Massachusetts, where she worked until she turned 18. This experience exposed her to hard physical labor and taught her skills typically associated with men’s work, including farming and carpentry. During this time, she educated herself and developed a keen intellect that would prove invaluable throughout her life. 

When her term of indenture ended in 1782, Sampson found herself in a precarious position as a young, unmarried woman with few economic opportunities. She intermittently supported herself as a teacher in the summers and a weaver in the winters.

Enlisting in the Army

The Revolutionary War was still raging, and the Continental Army desperately needed recruits. Motivated by both patriotic fervor and economic necessity, Sampson made the audacious decision to enlist in the army disguised as a man. She initially enlisted in 1782 under the name Timothy Taylor and collected a cash enlistment bounty but she failed to report for duty with her company.   She was later recognized as being Taylor and was required to repay what she had not already spent from her enlistment bounty.  No further punishment was made by the civil authorities; however, the Baptist Church withdrew its fellowship until she apologized and asked for forgiveness.

She later made a second enlistment, adopting the name Robert Shurtleff (sometimes spelled Shurtlieff or Shirtliff). This time she followed through and reported for duty.

She bound her chest, cut her hair, and donned men’s clothing to complete her transformation.  Sampson’s physical appearance aided her deception. She was tall for a woman of her era, standing nearly six feet, with a lean build and strong constitution developed through years of manual labor. Her lack of facial hair was not unusual among young male recruits, and she successfully passed the initial examination to join the 4th Massachusetts Regiment in May 1782.

The challenge of maintaining her disguise while living in close quarters with other soldiers required constant vigilance. Sampson developed strategies to protect her secret, including volunteering for guard duty to avoid sleeping arrangements that might expose her, and finding private moments to tend to personal needs. She also had to manage the physical demands of military life while dealing with the unique challenges of being a woman in a male-dominated environment.

Sampson’s military career nearly ended when she was wounded during a skirmish. She received a sword cut to her head and was shot in the thigh. Fearing that medical treatment would reveal her true identity, she initially treated her wounds herself, even digging a musket ball out of her own leg with a knife. Some of the shot remained too deep to remove, leaving her with a lifelong disability.

During her military service, Sampson demonstrated exceptional courage and skill as a soldier. She participated in several skirmishes and battles, including engagements near New York City and in Westchester County. Her fellow soldiers respected her for her dedication, marksmanship, and willingness to volunteer for dangerous scouting missions. She proved herself particularly adept at reconnaissance work, using her intelligence and observational skills to gather valuable information about enemy positions and movements.

Discovery and Discharge

During an epidemic in Philadelphia, she fell seriously ill with a fever and was taken to a hospital, where a physician discovered her secret while treating her. Fortunately, the doctor, Barnabas Binney, chose to protect Sampson rather than expose her. He treated her quietly and helped facilitate her honorable discharge from the army in October 1783. Her commanding officer, General John Paterson, reportedly handled the situation with discretion and respect, recognizing her valuable service to the cause of independence.  Eventually she was discharged by General Henry Knox on October 25, 1783, and was given funds to return home and a Note of Advice, similar to modern discharge papers.

Life After the War

After the war, Sampson returned to Massachusetts, where she married Benjamin Gannet in 1785 and had three children. But like many veterans, she struggled financially and had difficulty obtaining the military pay and benefits she had earned. In 1792, with the help of prominent supporters—including Paul Revere—she successfully petitioned the Massachusetts legislature for back pay and a modest state pension and she later received a pension from the federal government.

Her story didn’t end with domestic life. She became one of the first women in America to go on a speaking tour, traveling throughout New England and New York to share her experiences. Wearing her military uniform, she delivered a combination of storytelling, dramatic performance of military drills, and patriotic appeal.  These lectures, which began in 1802, were groundbreaking for their time, as respectable women rarely spoke publicly before mixed audiences.

A Lasting Legacy

Deborah Sampson’s legacy extends far beyond her military service. She challenged rigid gender roles and demonstrated that women could serve their country with the same valor and effectiveness as men. Her story inspired future generations of women who sought to break barriers and serve in traditionally male-dominated fields.

After she died in 1827, her story continued to gain recognition. In 1838, her husband was awarded a widow’s pension, possibly the first instance in U.S. history that the benefit was granted to a man based on his wife’s military service.

She left behind a legacy of courage, determination, and pioneering spirit that continues to resonate today. In 1983, she was declared the Official Heroine of the Commonwealth of Massachusetts, and in 2020, the U.S. House of Representatives passed the Deborah Sampson Act, expanding healthcare and benefits for female veterans. Statues and memorials, including her gravesite in Sharon, Massachusetts, commemorate her contributions.  Her wartime exploits have been the subject of books, plays, and scholarly research and her story continues to inspire generations as a symbol of courage and the ongoing struggle for gender equality in military service. 

While she was not the only woman to disguise herself and enlist—others like Margaret Corbin and Anna Maria Lane also took up arms—Sampson is among the best documented and celebrated.

Her life represents a crucial chapter in both military history and women’s history, illustrating the complex ways in which the American Revolution created opportunities for individuals to transcend social conventions in service of the greater cause of independence.  Deborah’s journey from indentured servant to Continental Army soldier and national lecturer is a testament to her extraordinary courage and determination. By stepping into a role forbidden to women and excelling under the harshest conditions, she challenged the boundaries of her time and set a precedent for future generations.

Though it is possible that her wartime activities may have been exaggerated—a common practice in biographies of the time—her life remains a powerful reminder of the contributions women have made, often unrecognized, in the shaping of American history.

The illustration at the beginning of this post is from The Female Review: Life of Deborah Sampson, the Female Soldier in the War of Revolution (1916), a reprint of the 1797 biography by Herman Mann.  

The Quasi-War

America’s Undeclared Naval Conflict with France

Following the American Revolution, the newly independent United States found itself caught between the competing imperial ambitions of Britain and France. What began as tensions over trade and neutrality escalated into an undeclared naval war and became the country’s first international crisis. The Quasi-War, as historians have named this conflict, was a pivotal moment in early American history when the republic’s survival hung in the balance. 

The Quasi-War was a limited, undeclared naval conflict fought primarily in the Caribbean and along the American coast. The conflict was “quasi” because it lacked a formal declaration of war and was limited in scope, focusing mainly on naval encounters rather than land-based military campaigns.

Origins of the Conflict

The roots of the Quasi-War began with the disagreements, conflicts, and confusion of international relations that followed the French Revolution. The French declaration of war on Britain in 1793 put the United States in a difficult position. The 1778 Treaty of Alliance with France technically obligated the United States to support France militarily, but President George Washington, recognizing that America was too weak to engage in another major conflict, issued a Proclamation of Neutrality.

This decision to remain neutral infuriated the French government, which viewed it as a betrayal of their wartime alliance.  

Tensions increased when the United States suspended loan repayments to France in 1793 primarily due to the dramatic political upheaval of the French Revolution and uncertainty about the legitimacy of the French government.

When the French Revolution began, the U.S. had outstanding debts to France from loans provided during the American Revolution. By 1793, the situation in France had become extremely volatile — King Louis XVI was executed in January, and various revolutionary factions were competing for power.

The Washington administration faced a dilemma: should they continue making payments to the new revolutionary government, or would doing so constitute recognition of a regime that might not be stable or even legitimate? There were also concerns about whether honoring debts to the revolutionary government might drag the U.S. into France’s expanding wars with European monarchies.

The situation deteriorated even further when the United States signed Jay’s Treaty with Britain in 1794. This commercial agreement resolved several outstanding issues between America and Britain, including the evacuation of British forts in the Northwest Territory and the establishment of limited trade relationships. To the French, Jay’s Treaty represented a clear alignment with their enemy.

The French responded in 1795, by authorizing privateers and naval vessels to attack and capture American merchant ships, particularly those trading with British ports or carrying British goods. By 1797, the French had captured over 300 American vessels, causing significant economic damage to American merchants and threatening the nation’s maritime commerce.

The diplomatic crisis deepened with the infamous XYZ Affair of 1797.  When President John Adams sent envoys to Paris to negotiate a resolution to the mounting tensions, French Foreign Minister Talleyrand’s agents demanded substantial bribes, described as loans, before any negotiations could begin. The American diplomats refused these demands and returned home empty-handed. When news of the attempted extortion became public, it sparked outrage across America and the rallying cry “Millions for defense, but not one cent for tribute!”  In his report to Congress, President Adams referred to the three French officials as X, Y, and Z to protect their actual identity, thus giving the XYZ Affair its name.

The Undeclared War Begins

The failure of diplomatic negotiations and continued French attacks on American shipping pushed the Adams administration toward military action. Adams chose not to declare war, which would have required congressional approval and would have risked drawing in other European powers. Instead, he chose a more limited response, authorizing the United States Navy to capture French vessels found in American waters and to protect American merchant ships on the high seas.

The Adams administration took several significant steps to prepare for this maritime conflict. Congress authorized the creation of a separate Navy Department, expanded the existing naval forces, and approved the construction of new warships. The Marine Corps, which was disbanded in 1783, was also formally reestablished during this period. Additionally, Congress suspended all commercial activities with France and authorized American naval vessels to capture French ships engaged in hostile acts against American commerce.

Naval Encounters and Key Battles

The Quasi-War witnessed numerous naval engagements, ranging from single-ship duels to larger squadron actions. American naval commanders like Thomas Truxtun emerged as heroes during this conflict. Truxtun’s frigate USS Constellation achieved two notable victories, first capturing the French frigate L’Insurgente in February 1799, and later defeating La Vengeance in a fierce night battle in February 1800.

The conflict extended throughout the Caribbean, where American naval squadrons protected merchant convoys and hunted French privateers. The USS Enterprise, USS Experiment, and other smaller American vessels proved particularly effective in these operations, capturing numerous French privateers and recapturing American merchant vessels. The U. S. Navy captured or destroyed over 80 French vessels while having only a single ship captured in combat and that ship was later recaptured.

One of the most significant aspects of the naval war was its impact on American naval development. The conflict provided valuable combat experience for American officers and sailors, many of whom would later serve with distinction in the War of 1812. The success of American naval forces during the Quasi-War also demonstrated the importance of maintaining a strong navy for protecting American commercial interests.

While the war was almost exclusively naval, there were minor land actions such as that which occurred during September 1800, when U.S. Marines landed at Curaçao to drive French forces from two forts.

Fears of a possible French invasion led Congress to authorize a provisional army of 10,000 men. President John Adams appointed George Washington as Commander-in-Chief, with Alexander Hamilton as his second-in-command, though no land battles occurred.

Domestic Political Consequences

The Quasi-War worsened the growing divide between Federalists and Democratic-Republicans. President Adams and the Federalist Party generally supported a strong response to French aggression, viewing it as necessary to protect American honor and commercial interests. The conflict provided Federalists with an opportunity to strengthen the federal government and build up American military forces.

The Alien and Sedition Acts, passed in 1798 during the height of anti-French sentiment, granted the government broad powers to deport foreign nationals and prosecute critics of the administration. These measures were widely seen as attacks on civil liberties and became a major political liability for the Federalists.

Thomas Jefferson and the Democratic-Republicans opposed both the war and the domestic security measures, arguing that Adams was leading the country toward unnecessary conflict and tyranny. They maintained that the French grievances were legitimate and that diplomatic solutions should be pursued more vigorously.

Resolution and the Convention of 1800

By 1799, both sides had grown weary of the costly and disruptive conflict. Napoleon Bonaparte, who had come to power in France, was more interested in European affairs than in continuing a naval war with America. Similarly, President Adams recognized that prolonged conflict would be economically devastating and politically dangerous.

In February 1799, Adams surprised both his own party and the nation by announcing his intention to send new diplomatic envoys to France. This decision split the Federalist Party, with many hawks opposing any negotiations with France. Nevertheless, Adams persisted, believing that peace was in America’s best interest.

The resulting negotiations led to the Convention of 1800, also known as the Treaty of Mortefontaine, signed in September 1800. This agreement effectively ended the Quasi-War by establishing terms for peaceful coexistence based on the principle of “free trade, free goods” between the two nations. The treaty also provided for the mutual restoration of captured vessels, established compensation procedures for maritime losses, and most importantly, formally ended the 1778 Treaty of Alliance between France and the United States.

Long-term Impact and Legacy

The Quasi-War’s conclusion marked a significant turning point in American foreign policy. The conflict demonstrated that the United States could successfully defend its interests against a major European power without formal allies. It also established important precedents for presidential war powers and the use of limited military force without formal declarations of war.  The experience gained during the Quasi-War would prove invaluable in subsequent conflicts, particularly the War of 1812.

The Quasi-War Was the beginning of a long-standing policy of neutrality in European conflicts, that persisted for much of the 19th century and was even echoed in the first half of the 20th century.   It demonstrated to the world that the United States was a viable country that stood ready to defend its sovereignty.

The Quasi-War was America’s first undeclared war.  Although Congress eventually granted limited military authority it was begun at the direction of President Adams. This has influenced American foreign policy and the use of military forces ever since. The Quasi-War was referenced in debates about American involvement in Vietnam and in the Gulf War. The War Powers Act of 1973 was passed an effort to limit the ability of the President to send American troops to combat in foreign countries, but its effectiveness and enforcement have been debated ever since it was passed.

Powered by WordPress & Theme by Anders Norén