Google and US government battle over the future of internet advertising

Google will confront the U.S. government’s latest attempt to topple its internet empire in federal court on Friday as a judge considers how to prevent the abusive tactics that culminated in parts of its digital ad network being branded as an illegal monopoly.

The courtroom showdown in Alexandria, Virginia, will pit lawyers from Google and the U.S. Department of Justice against each other in closing proceedings focused on the complex technology that distributes millions of digital ads across the internet each day.

After a lengthy trial last year, U.S. District Judge Leonie Brinkema ruled in April that pieces of Google’s ad technology had been rigged in a way that made it an illegal monopoly. That set up another 11-day trial earlier this fall to help Brinkema determine how to remedy its anti-competitive practices.

Friday’s closing arguments will give both Google and the Justice Department a final chance to sway Brinkema before she issues a ruling that probably won’t come until early next year.

The Justice Department wants Brinkema to force Google to sell some of the ad technology that it has spent nearly 20 years assembling, contending a breakup is the only way to rein in a company that the agency’s lawyers condemned as a “recidivist monopolist” in filings leading up to Friday’s hearing.

The condemnation refers not only to Google’s practices in digital advertising but also to the illegal monopoly that it unleashed through its dominant search engine. Federal prosecutors also sought a breakup in the search monopoly case, but the judge handling that issue rejected a proposal that would have required Google to sell its popular Chrome web browser.

Although Google is still being ordered to make reforms that it’s resisting, the outcome in the search monopoly case has been widely seen as a proverbial slap on the wrist. The belief that Google got off easy in the search case is the main reason the market value of its parent company Alphabet surged by about $950 billion, or 37%, to nearly $3.5 trillion since U.S. District Judge Amit Mehta’s decision came out in early September.

That setback hasn’t discouraged the Justice Department from arguing for a breakup of an ad tech system that handles 55 million requests per second, according to estimates provided by Google in court filings.

The huge volume of digital ads priced and distributed through Google’s technology is one of the main reasons that the company’s lawyers contend it would be too risky to force a dismantling of the intricate system.

“This is technology that absolutely has to keep working for consumers,” Google argues in documents leading up to Friday’s hearing. The company’s lawyers blasted the Justice Department’s proposal as a package of “legally unprecedented and unsupported divestitures.”

Besides arguing that its own proposed changes will bring more price transparency and foster more competition, Google is also citing market upheaval triggered by artificial intelligence as another reason for the judge to proceed cautiously with her decision.

In his decision in the search monopoly case, Mehta reasoned that AI was already posing more competition to Google.

Keep reading

Microsoft Warns Experimental Windows 11 AI Could Install Malware on Your Computer

Microsoft said in an update on Nov. 17 that Windows 11 users who utilize “agentic features” from its AI services should be cautious because the AI agents could potentially download and install malware.

In an alert, Microsoft warned that its AI models could “occasionally hallucinate” and introduce “novel security risks” such as malware because large language models, a type of AI that processes data and generates human-like text, are susceptible to cyberattacks.

“As these capabilities are introduced, AI models still face functional limitations in terms of how they behave and occasionally may hallucinate and produce unexpected outputs. Additionally, agentic AI applications introduce novel security risks, such as cross-prompt injection (XPIA),” the warning stated. A prompt injection attack is a type of cyberattack where an attacker crafts an input to trick the AI into performing malicious actions.

Microsoft added that in the case of Windows 11’s “experimental” AI services, “malicious content embedded in UI elements or documents can override agent instructions, leading to unintended actions like data exfiltration or malware installation.”

The AI features are turned off by default and operate only after the user opts into them, the company said.

The agentic AI setting “can only be enabled by an administrator user of the device and once enabled, it’s enabled for all users on the device including other administrators and standard users,” Microsoft said of the AI services.

Keep reading

Google Censored Vaccine Info Long Before COVID — Could It Have Anything to Do With Parent Company Alphabet’s Deep Pharma Ties?

Throughout the COVID-19 pandemic, Big Tech companies colluded with the government to silence dissent and criticisms of the lockdowns and a coercive mass vaccination campaign by censoring truthful information that did not align with the political agenda.

The Biden administration’s role in the censorship regime was the subject of a May 2024 congressional report titled “The Censorship-Industrial Complex.”

Three months after that report was released, Meta CEO Mark Zuckerberg admitted in a letter to Congress that Facebook had censored factual information under pressure from the White House.

In September, Google’s parent company, Alphabet, responded to a congressional subpoena with a letter similarly disclosing how the Biden administration had pressured YouTube, owned by Google, to remove videos that didn’t even violate its content policies.

Alphabet called the practice “unacceptable and wrong” while insisting that it withstood the pressure and enforced only its own policies against “misinformation.”

That defense, however, sidesteps the fact that those content guidelines were created in collusion with the same “health authorities” advancing the authoritarian governance — like the Centers for Disease Control and Prevention (CDC) and World Health Organization (WHO).

The result was that true information was censored while government-sanctioned disinformation was allowed to proliferate unchallenged.

As a United Nations official admitted at a September 2022 World Economic Forum (WEF) meeting, Google was helping government authorities to “own the science” in its internet search results.

In its letter to Congress, Alphabet noted that YouTube’s content policies had since evolved. Tacitly admitting how creators had been silenced for telling the truth, Alphabet promised to restore YouTube channels suspended for content no longer deemed misinformative.

So, Alphabet acknowledged the censorship but tried to absolve itself by blaming the White House and public health authorities.

The truth is that Google’s censorship of health-related content, including inconvenient facts about vaccines, predated COVID-19 and continues to this day. Could that be because Alphabet has its own deep financial ties to the pharmaceutical and biotech industries?

Alphabet’s ties to Big Pharma exist through numerous of its subsidiaries, including CalicoDeepMindIsomorphic Labs and Verily Life Sciences. In this article, we will focus on the latter of Google’s sister companies.

Keep reading

Google boss says trillion-dollar AI investment boom has ‘elements of irrationality’

Every company would be affected if the AI bubble were to burst, the head of Google’s parent firm Alphabet has told the BBC.

Speaking exclusively to BBC News, Sundar Pichai said while the growth of artificial intelligence (AI) investment had been an “extraordinary moment”, there was some “irrationality” in the current AI boom.

It comes amid fears in Silicon Valley and beyond of a bubble as the value of AI tech companies has soared in recent months and companies spend big on the burgeoning industry.

Asked whether Google would be immune to the impact of the AI bubble bursting, Mr Pichai said the tech giant could weather that potential storm, but also issued a warning.

“I think no company is going to be immune, including us,” he said.

In a wide-ranging exclusive interview at Google’s California headquarters, he also addressed energy needs, slowing down climate targets, UK investment, the accuracy of his AI models, and the effect of the AI revolution on jobs.

The interview comes as scrutiny on the state of the AI market has never been more intense.

Alphabet shares have doubled in value in seven months to $3.5tn (£2.7tn) as markets have grown more confident in the search giant’s ability to fend off the threat from ChatGPT owner OpenAI.

A particular focus is Alphabet’s development of specialised superchips for AI that compete with Nvidia, run by Jensen Huang, which recently reached a world first $5tn valuation.

As valuations rise, some analysts have expressed scepticism about a complicated web of $1.4tn of deals being done around OpenAI, which is expected to have revenues this year of less than one thousandth of the planned investment.

It has raised fears stock markets are heading for a repeat of the dotcom boom and bust of the late 1990s. This saw the values of early internet companies surge amid a wave of optimism for what was then a new technology, before the bubble burst in early 2000 and many share prices collapsed.

This led to some companies going bust, resulting in job losses. A drop in share prices can also hit the value of people’s savings including their pension funds.

In comments echoing those made by US Federal Reserve chairman Alan Greenspan in 1996, warning of “irrational exuberance” in the market well ahead of the dotcom crash, Mr Pichai said the industry can “overshoot” in investment cycles like this.

“We can look back at the internet right now. There was clearly a lot of excess investment, but none of us would question whether the internet was profound,” he said.

“I expect AI to be the same. So I think it’s both rational and there are elements of irrationality through a moment like this.”

His comments follow a warning from Jamie Dimon, the boss of US bank JP Morgan, who told the BBC last month that investment in AI would pay off, but some of the money poured into the industry would “probably be lost”.

But Mr Pichai said Google’s unique model of owning its own “full stack” of technologies – from chips to YouTube data, to models and frontier science – meant it was in a better position to ride out any AI market turbulence.

Keep reading

Italian Court Orders Google to Restore Banned Catholic Blog

Google has been compelled by the Tribunale di Imperia to restore Messainlatino.it, a major Italian Catholic website that, as you may remember, the company had abruptly taken down from its Blogger platform in July.

The ruling, issued against Google Ireland Limited, the firm’s European branch, also requires payment of approximately €7,000 (about $8,100) in court costs.

The blog’s editor, Luigi Casalini, filed legal action after Google deleted the site without warning, claiming a violation of its “hate speech” rules.

The company’s notification consisted of a short, generic email and provided no explanation or chance to appeal.

For Casalini, whose publication had accumulated over 22,000 articles since 2008 and reached around one million monthly readers, the removal appeared to be less a matter of policy enforcement and more an attempt to silence dissenting religious opinion.

Messainlatino.it was well known for covering issues surrounding traditional Catholic liturgy and had been cited by major outlets.

Following Google’s action, questions were raised in both the European Parliament and Italy’s Chamber of Deputies.

Legislators noted that the deletion “raises serious questions about the respect for freedom of expression, speech and religion” as guaranteed by Article 11 of the EU Charter of Fundamental Rights and Article 10 of the European Convention on Human Rights.

They also pointed to the Digital Services Act (DSA), which, despite being a censorship law, obliges platforms to apply their moderation policies with “due regard” for fundamental rights.

Casalini’s legal case focused on that provision. He argued that Google’s decision breached Article 14 of the DSA, which calls for a balance between policy enforcement and the user’s right to free expression.

As Casalini stated to LifeSiteNews, “Google acted in this way in violation of the Digital Services Act.”

Google responded through five lawyers based in Milan. The company claimed that an interview with Bishop Joseph Strickland, who opposed the ordination of women as deacons, violated its hate speech policy.

When the defense team countered that the post merely reported the bishop’s words and contained no discriminatory content, Google’s attorneys maintained in court documents that “it does not matter the source, more or less authoritative (bishop, Pontiff) of the post, if it violates the Policy.”

Keep reading

Google Sued For Allegedly Using Gemini AI Tool To Track Users’ Private Communications

Google LLC is accused in a civil lawsuit of using its artificial intelligence program Gemini to collect data on users’ private communications in Gmail as well as Google’s instant messaging and video conference programs.

Until around Oct. 10, the Gemini AI assistant required the user to deliberately opt into its feature. After that date, the feature was allegedly “secretly” turned on by Google for all its users’ Gmail, Chat, and Meet accounts by default, enabling AI to track its users’ private data in those platforms “without the users’ knowledge or consent,” according to the complaint filed Nov. 11 in federal court in San Jose.

The class action lawsuit was filed in the U.S. District Court for the Northern District of California, alleging that Google is violating the California Invasion of Privacy Act, a 1967 law that prohibits surreptitious wiretapping and recording of confidential communications without the consent of all parties involved.

Although Google provides a way for users to turn off the feature, it requires users to look for it in the privacy settings to deactivate it, despite never having agreed to it in the first place, the complaint said.

The AI feature is categorized in “Google Workspace smart features” in Google settings. Once turned on, it means the user consents to the program using “Workspace content and activity” across Workspace or in other Google products.

When the feature is turned on, Gemini can “scan, read, and analyze every email (and email attachment), message, and conversation on those services,” according to the complaint.

Technology writer Ruben Circelli wrote in a PCMag article that Gemini is “downright creepy” in diving deep into his personal history, analyzing 16 years’ worth of emails after he signed up for a more advanced pro feature.

In a series of tests by Circelli, Gemini told him one of his character flaws and even knew who his first crush was in elementary school.

“This invasion of privacy wasn’t just disconcerting, though; it was unexpected,” Circelli wrote.

“Google didn’t explain what this integration would do before I signed up for its AI Pro plan, nor did it give me a way to opt out at the start.”

The Epoch Times reached out to Google for comment, but did not receive an immediate response.

“We do not use your Workspace data to train or improve the underlying generative AI and large language models that power Gemini, Search, and other systems outside of Workspace without permission,” the company has stated.

Keep reading

Google Softens Planned Android Sideloading Ban but Keeps Developer ID Verification

Google is slightly relaxing its controversial new Android policy on sideloading, but the shift does little to change its overall direction.

The company confirmed that it will still move ahead with mandatory developer identity verification for nearly all apps while introducing a limited “advanced flow” that lets “experienced users” continue installing software from outside the Play Store.

According to Google, the new system will feature multiple security warnings meant to deter casual users from downloading unverified apps.

“It will include clear warnings to ensure users fully understand the risks involved, but ultimately, it puts the choice in their hands,” the company said.

The process is still being developed, with feedback now underway before finalization in the coming months.

The adjustment follows backlash from developers and Android fans who criticized Google’s original plan to block apps created by unverified developers starting next year.

The community argued that the move would effectively close off Android’s long-standing openness by removing the ability to install software freely.

Despite the new language, Google’s latest policy maintains the same structure.

Developer ID verification will still be required for nearly all app distribution.

Only students and hobbyists will be allowed to share apps with a limited number of devices without providing identification, and businesses deploying internal software will remain exempt.

For everyone else, verification and a $25 registration fee will be mandatory, including for apps distributed outside Google Play. Previously, there was no charge for independent distribution.

The rollout schedule remains the same. Developers who distribute apps outside the Play Store began receiving early-access invitations on November 3, while Play Store developers will get theirs starting November 25.

The early-access period runs through March 2026, after which the verification program will open to all developers. The rules take effect in Brazil, Indonesia, Singapore, and Thailand in September 2026, and globally in 2027.

Google maintains that the new requirements are about security, not control.

Keep reading

Police: Chairman of Pentagon Software Contractor Arrested in Pedophile Sting

A 57-year-old Pittsburgh tech entrepreneur who is the founder and chairman of a software company with federal contracts including with the Pentagon is facing felony charges for allegedly soliciting sexual contact with a young girl in Lebanon County, Pennsylvania.

WTAE in Pittsburgh reports that Eric T. Gillespie, founder and chairman of software company Govini, was arrested on Tuesday after allegedly trying to arrange a meeting with a preteen girl for sexual purposes. The arrest was part of a sting operation conducted by the Pennsylvania Attorney General’s Child Predator Section.

According to Attorney General Dave Sunday, an agent posed as an adult in an online chat platform often used by offenders attempting to arrange meetings with children. The agent uncovered Gillespie, who was using an online pseudonym, lurking on the platform to access children.

Gillespie, a 57-year-old resident of Pittsburgh’s South Side neighborhood, founded Govini to transform how the U.S. government uses AI and data to make decisions. The company recently surpassed $100 million in annual revenue and has been awarded millions of dollars in contracts with federal agencies including the Pentagon and Department of Homeland Security.

In response to Gillespie’s arrest, Govini placed him on administrative leave and stated the company will fully cooperate with law enforcement’s investigation. The company acknowledged the severity of the charges and vowed to hold all employees to the highest ethical standards.

Gillespie is facing four felony counts related to the incident. A judge denied bail, citing flight risk and public safety concerns. The Attorney General’s office said Gillespie denied the allegations against him.

Keep reading

OpenAI asked Trump administration to expand Chips Act tax credit to cover data centers

recent letter from OpenAI reveals more details about how the company is hoping the federal government can support the company’s ambitious plans for data center construction.

The letter — from OpenAI’s chief global affairs officer Chris Lehane and addressed to the White House’s director of science and technology policy Michael Kratsios — argued that the government should consider expanding the Advanced Manufacturing Investment Credit (AMIC) beyond semiconductor fabrication to cover electrical grid components, AI servers, and AI data centers.

The AMIC is a 35% tax credit that was included in the Biden administration’s Chips Act.

“Broadening coverage of the AMIC will lower the effective cost of capital, de-risk early investment, and unlock private capital to help alleviate bottlenecks and accelerate the AI build in the US,” Lehane wrote.

OpenAI’s letter also called for the government to accelerate the permitting and environmental review process for these projects, and to create a strategic reserve of raw materials — such as copper, alumimum, and processed rare earth minerals — needed to build AI infrastructure.

The company first published its letter on October 27, but it didn’t get much press attention until this week, when comments by OpenAI executives prompted broader discussion about what the company wants from the Trump administration.

At a Wall Street Journal event on Wednesday, CFO Sarah Friar said the government should “backstop” OpenAI’s infrastructure loans, though she later posted on LinkedIn that she misspoke:  “OpenAI is not seeking a government backstop for our infrastructure commitments. I used the word ‘backstop’ and it muddied the point.”

Keep reading

Tech billionaires back startup probing gene-edited ‘designer babies’ despite US ban: report

A Silicon Valley startup backed by OpenAI’s Sam Altman and Coinbase’s Brian Armstrong is pursuing research that some fear could lead to the birth of a genetically engineered baby — a step that’s illegal under US law and banned in most countries, a report said.

The company, Preventive, says its goal is to end hereditary disease by editing human embryos before birth, a claim that has ignited fierce debate over safety, ethics and the specter of designer children, according to the Wall Street Journal.

Preventive, founded earlier this year by gene-editing scientist Lucas Harrington, has raised $30 million and set up headquarters in San Francisco, where it is conducting research on modifying embryos to prevent hereditary disease.

The company says its mission is to prove the technology can be made safe and transparent before any attempt to create a baby is made.

Altman and Armstrong are among the firm’s early investors, the Wall Street Journal reported.

Altman’s husband, Oliver Mulherin, said he led their investment, calling it an effort to help families avoid genetic illness.

Armstrong, who has publicly promoted embryo editing, posted that he was “excited” to back Preventive and argued it is far easier to correct a genetic defect in an embryo than to treat disease later in life.

But federal law prohibits the Food and Drug Administration from considering applications for human trials involving genetically edited embryos used to start pregnancies.

Keep reading