JD Vance Slams AI Regulations, Widening U.S.-Europe Divide
At the Artificial Intelligence Action Summit in Paris, U.S. Vice President JD Vance criticized Europe’s stringent AI regulations, arguing they could stifle innovation. His remarks highlighted a growing global divide on AI governance, with the U.S. favoring minimal restrictions, Europe enforcing strict policies, and China expanding state-backed AI. The U.S. notably abstained from signing a global AI pledge signed by over 60 nations, including China, further isolating itself on the issue. Vance’s speech, his first major policy address as vice president, framed AI as a pivotal economic force and emphasized the Trump administration’s commitment to free-market AI development. European leaders, however, defended their approach, citing safety and public trust. Meanwhile, China promoted open-source AI while simultaneously tightening domestic controls. The summit underscored the global power struggle over AI dominance, as well as growing tensions between the U.S. and its allies over regulatory approaches.
Key Takeaways:
U.S. vs. Europe on AI: Vance warned that overregulation could hinder AI innovation, putting the U.S. at odds with Europe’s strict oversight.
China’s AI Strategy: China supported a global AI pledge while expanding state-backed AI at home, further complicating the international landscape.
Global AI Race: The summit revealed intensifying competition between the U.S., Europe, and China to set AI standards and maintain technological leadership.
AI Copyright War: Cohere Faces Lawsuit from Leading Media Companies
A coalition of major publishers, including The Toronto Star, Forbes, The Atlantic, The Guardian, Vox Media, and Politico, has filed a lawsuit against Canadian AI firm Cohere, alleging that the company used copyrighted content without permission to train its generative AI models. The lawsuit, filed in a New York court, demands damages and a permanent injunction to prevent Cohere from reproducing their work. The plaintiffs argue that AI companies profiting from journalism without compensation threaten the media industry, which is already struggling with declining ad revenue. Cohere, valued at $5.5 billion, dismissed the lawsuit as “misguided and frivolous,” asserting that it follows responsible AI training practices. The case is part of a growing legal battle between publishers and AI firms, as media companies push for legal precedents to protect journalistic content from unauthorized AI use.
Key Takeaways:
Publishers vs. AI: Major media companies accuse Cohere of using their content without permission and demand financial compensation.
AI Copyright Challenges: The case reflects broader legal battles as publishers seek to establish protections against unauthorized AI training on their work.
Cohere’s Response: The AI company denies wrongdoing, calling the lawsuit baseless and insisting it adheres to responsible AI practices.
Microsoft Study Warns AI Dependence Erodes Critical Thinking
A new study by Microsoft and Carnegie Mellon University warns that excessive reliance on AI tools can erode critical thinking skills. Researchers found that knowledge workers who trusted AI to complete tasks tended to disengage mentally, particularly with low-stakes work, leading to diminished problem-solving abilities. The study also highlighted that AI-assisted users produced less diverse solutions, raising concerns about creativity and independent thought. Conversely, those skeptical of AI’s accuracy were more likely to engage critically and improve AI-generated content. While AI can enhance efficiency, the study cautions against overdependence, as it may weaken cognitive skills over time.
Key Takeaways:
AI Weakens Critical Thinking: Workers relying heavily on AI became less engaged and struggled to think independently.
Reduced Creativity: AI users generated more homogenous results compared to those working without AI assistance.
Over-Reliance Risk: While AI can improve efficiency, blindly trusting it may lead to cognitive decline and poor decision-making.
Altman Dismisses Musk’s OpenAI Takeover Bid as a ‘Disruption Tactic’
OpenAI CEO Sam Altman has dismissed Elon Musk’s $97.4 billion takeover bid as a “tactic to mess with us,” asserting that the nonprofit controlling OpenAI is not for sale. While Musk’s lawyer claimed the offer was sent to OpenAI’s outside counsel, Altman stated that the board had not officially received or reviewed it but planned to reject it. Musk, who co-founded OpenAI in 2015 before leaving over strategic disagreements, has since launched his own AI company, xAI. The bid comes as OpenAI seeks $40 billion in funding to transition into a for-profit entity, raising legal questions about how nonprofit assets are valued in the shift.
Key Takeaways:
Altman Rejects Musk’s Bid: OpenAI’s CEO dismissed Musk’s $97.4 billion takeover offer as a disruption tactic, insisting the company is not for sale.
Nonprofit to For-Profit Transition: OpenAI’s planned shift to a for-profit model is under legal scrutiny to ensure fair valuation of its nonprofit assets.
Musk’s AI Ambitions: Having left OpenAI, Musk now leads xAI and remains a key player in the AI industry, further complicating his takeover attempt.
Waterloo Researchers Develop Sustainable Graphene Ink for Industry & Environment
Researchers at the University of Waterloo have developed an eco-friendly, 3D-printable graphene ink that could revolutionize applications in healthcare, industry, and environmental science. Unlike traditional graphene powders, which are difficult to manipulate, this water-based ink maintains conductivity without requiring chemical additives or solvents.
Key Takeaways:
Breakthrough in Graphene Printing: Researchers created a 3D-printable, eco-friendly graphene ink that maintains conductivity without chemical additives.
Wide Applications: The ink could be used for wearable sensors, lightweight automotive parts, printed electronics, and environmental cleanup.
Author: Malik Datardina, CPA, CA, CISA. Malik works at Auvenir as a Sr. AI Product Manager who is working to transform the engagement experience for accounting firms and their clients. The opinions expressed here do not necessarily represent UWCISA, UW, Auvenir (or its affiliates), CPA Canada or anyone else. This post was written with the assistance of an AI language model. The model provided suggestions and completions to help me write, but the final content and opinions are my own.
How do you use DeepSeek on your laptop? Check out this video!
In this video, David provides step-by-step guidance on installing LM Studio and DeepSeekR1. The full 400GB model likely won’t run on your laptop (it didn’t on mine), so you’ll need a distilled model. This smaller version is more manageable and easier to use. (See the next article for WSJ’s analysis of distillation).
AI Distillation Explained: How Smaller Labs Are Challenging Big Tech
AI distillation is emerging as a major concern for big tech companies, as it allows smaller AI labs to create powerful models at a fraction of the cost. The technique involves using outputs from advanced AI models (like OpenAI’s GPT) to train cheaper, smaller models that perform nearly as well. Chinese AI startup DeepSeek has been accused of distilling OpenAI’s models, potentially reducing the value of OpenAI’s proprietary technology.
AI Distillation Threatens Big Tech – Smaller AI labs can now replicate powerful AI models cheaply, reducing the competitive advantage of Google, OpenAI, and Meta.
DeepSeek Accused of Distillation – OpenAI suspects DeepSeek used ChatGPT’s outputs to train its own models.
Billions vs. Millions – While big tech spends billions, companies like DeepSeek are proving AI can be developed for far less.
AI Takes Over Research: OpenAI’s Deep Research Automates Knowledge Work
OpenAI has launched Deep Research, an AI-powered agent designed to automate and streamline research tasks. Built on an optimized version of the o3 model, the tool can independently gather, analyze, and synthesize vast amounts of web data into a structured report within 5 to 30 minutes. Deep Research is particularly beneficial for professionals in finance, science, policy, and engineering, who require in-depth, niche information. The reports include clear citations and reasoning breakdowns, allowing users to verify accuracy. While OpenAI claims that Deep Research hallucinates less than previous models, it still requires human oversight. The tool is currently available only to ChatGPT Pro users ($200/month), with plans to expand access to other tiers. OpenAI's move comes amid growing competition, particularly from Google's Gemini Deep Research and Microsoft's Think Deeper features.
AI-Powered Research Automation – Deep Research can generate structured research reports independently, reducing hours of work to minutes.
Targeted at Professionals – Designed for experts in finance, science, and policy who need reliable, high-quality research.
Limited Access & Cost – Available only to ChatGPT Pro users at $200/month, but OpenAI plans to expand access over time.
DeepSeek Crackdown: New US Law Could Mean Prison for Users
A newly proposed US law could impose harsh penalties on individuals and businesses using DeepSeek, a Chinese AI chatbot that recently surged in popularity. Introduced by Senator Josh Hawley, the bill aims to block AI technology from China by prohibiting US individuals and companies from engaging with Chinese-developed AI models. The legislation doesn't mention DeepSeek by name, but its release coincided with the app becoming the most popular AI tool in the US, leading to concerns about national security and data privacy. Those violating the law could face up to 20 years in prison or fines of $1 million for individuals and $100 million for businesses. The US Navy and NASA have already banned DeepSeek on government devices, and Texas became the first state to restrict its use.
Proposed Ban on Chinese AI – A new law aims to block US individuals and businesses from using Chinese-developed AI models like DeepSeek.
Severe Legal Consequences – Violations could lead to 20 years in prison or fines of up to $100 million for businesses.
Security & Privacy Concerns – DeepSeek stores data in China, raising fears over surveillance and national security risks.
Musk’s DOGE Is Building an AI Chatbot for the US Government
Elon Musk’s Department of Government Efficiency (DOGE) is developing GSAi, a custom AI chatbot for the US General Services Administration (GSA) as part of President Donald Trump’s AI-first agenda. The chatbot aims to boost productivity for GSA employees by automating tasks like managing government contracts, analyzing procurement data, and drafting memos. The project was initially considering Google’s Gemini, but Musk’s team opted for an in-house AI system instead. DOGE’s broader AI initiatives are focused on cutting government costs, but have sparked controversy, with critics arguing they lack transparency and security reviews. While Musk’s push aligns with Trump’s goal of US AI dominance, some federal employees and civil rights groups worry about the rapid, unchecked adoption of AI in government operations.
Musk’s Government AI Chatbot – DOGE is building GSAi, an AI assistant for the General Services Administration to streamline government tasks.
Part of Trump’s AI-First Agenda – The project aligns with Trump’s goal of making the US a global leader in AI and reducing government costs.
Controversy & Security Concerns – Critics warn about lack of oversight, while employees worry about AI’s impact on government transparency.
Is Amazon Overspending? AI Race Heats Up with $100B Investment
Amazon has announced plans to spend $100 billion in capital expenditures in 2025, largely focused on AI infrastructure. This is a significant increase from $83 billion in 2024 and aligns with Amazon’s strategy to compete with AI leaders like OpenAI, Google, and Microsoft. The company is investing heavily in data centers, networking hardware, and AI chips to support AWS cloud AI services. CEO Andy Jassy emphasized that AI represents a "once-in-a-lifetime business opportunity", reassuring investors that this spending will drive long-term growth. However, skepticism is rising in the tech industry following the rapid and low-cost success of Chinese AI startup DeepSeek, which developed a competitive AI model for under $6 million. This has raised questions about whether big tech's massive AI spending is truly necessary.
Amazon’s $100B AI Investment – The company is making one of its largest-ever AI investments, focusing on cloud infrastructure and AI development.
Race Against Competitors – Amazon is competing with Google, Microsoft, and OpenAI, all of whom are also spending tens of billions on AI.
DeepSeek Disrupts AI Economics – The low-cost success of DeepSeek’s AI raises concerns about whether such massive AI investments are sustainable.
Author: Malik Datardina, CPA, CA, CISA. Malik works at Auvenir as a Sr. AI Product Manager who is working to transform the engagement experience for accounting firms and their clients. The opinions expressed here do not necessarily represent UWCISA, UW, Auvenir (or its affiliates), CPA Canada or anyone else. This post was written with the assistance of an AI language model. The model provided suggestions and completions to help me write, but the final content and opinions are my own.
With the seismic impact of DeepSeek on AI, the stock market, and geopolitics, we wanted to follow-up our previous post with a deeper exploration of the topic. In this post, we found 5 videos that will help you get up to speed on the unfolding drama.
Vid1: CNBC Covers the Ensuing Market Meltdown
CNBC discusses the impact of China's new AI model, DeepSeek, on the global tech industry. DeepSeek's superior efficiency and performance, even surpassing some American models, is causing a major sell-off in AI-related stocks, particularly impacting companies like Nvidia. The video explores concerns about DeepSeek's potential access to advanced technology and the implications for US technological dominance. The discussion also touches upon the shift towards open-source AI models and the uncertainty surrounding future investments in AI development. Finally, the video highlights the rapid advancement of AI technology and its potential societal impact, comparing the situation to the Sputnik moment of the space race.
Vid2: AI Enthusiast, Matt Wolfe, Gives His Take
Matt Wolfe, who closely follows the AI space, discusses DeepSeek R1, a new Chinese
open-source AI model that has caused significant market reactions. DeepSeek's
impressive performance, achieved with significantly less computing power
than comparable models like GPT-4, is attributed to its efficient training
methods and innovative design. Controversy surrounds DeepSeek's claims regarding
its resource usage, with some suggesting the company downplayed the actual
computational resources employed. Despite this, the video argues the model's
impact may be positive, possibly lowering the barrier to entry for AI
development and increasing overall demand for GPUs. The video also
covers DeepSeek's image generation model, Janice Pro 7B, and provides
instructions on how to access and utilize DeepSeek.
Vid3: A Geopolitical Perspective on the DeepSeek Saga
Here is Cold Fusion’s take on the DeepSeek story. He
discusses the sudden emergence of DeepSeek R1, a free, open-source Chinese AI
model that rivals—and in some ways surpasses—leading American AI models. Its
unexpectedly low development cost and superior efficiency have sent shockwaves
through the US stock market and prompted a reassessment of AI development
strategies. Concerns about intellectual property theft are raised, alongside
geopolitical implications of this technological advancement. The narrative
explores the innovative techniques behind DeepSeek R1's performance and the
competitive landscape it has created, highlighting the resulting cost
reductions and potential for rapid AI progress globally.
Vid4: If you are using DeepSeek, Your Data is Going to China!
Skill Leap AI discusses serious privacy concerns regarding the
DeepSeek website and app, highlighting issues like vague data retention
policies, data storage in China raising compliance issues with international
laws, lack of transparency in data usage, and insufficient age verification.
The creator outlines these issues after reviewing the platform's privacy policy
and terms of service using ChatGPT. To mitigate these risks, the video suggests
using locally installed versions of DeepSeek R1 or utilizing DeepSeek's
integration within the PerplexityAI search engine, a US-based service. Finally,
the video promises a future comparison of DeepSeek R1 and ChatGPT's 01 model.
Vid5: A Video Walkthrough of Dario Amodei's take on DeepSeek's Capabilities
In this video, Matt Berman takes a look at Dario Amodei's
take on the DeepSeek saga. Amodei, the current CEO of OpenAI’s chief rival Anthropic,
wrote an essay discussing the implications of DeepSeek's AI model, R1,
particularly concerning its potential data acquisition from OpenAI and the
resulting impact on the AI industry and geopolitical landscape. The essay
analyzes the three key dynamics of AI development: scaling laws, the
shifting curve, and paradigm shifts, emphasizing the escalating costs and
exponential advancements in AI capabilities. Concerns about China's
access to advanced GPUs and their potential to achieve artificial
general intelligence (AGI) are also highlighted, underscoring the importance of
export controls. Finally, the essay argues that DeepSeek's cost-effective
model, while impressive, does not represent a fundamental shift in AI economics
and that the market's overreaction was unwarranted.
Author: Malik Datardina, CPA, CA, CISA. Malik works at Auvenir as a Sr. AI Product Manager who is working to transform the engagement experience for accounting firms and their clients. The opinions expressed here do not necessarily represent UWCISA, UW, Auvenir (or its affiliates), CPA Canada or anyone else. This post was written with the assistance of an AI language model. The model provided suggestions and completions to help me write, but the final content and opinions are my own.
In my ongoing exploration of emerging AI technologies, I recently encountered a fascinating CNBC deep dive into DeepSeek, a Chinese AI firm that's reshaping the competitive landscape of artificial intelligence. The investigation sparked my interest in understanding how this relatively unknown company has managed to develop AI models that rival industry giants like OpenAI and Google, while maintaining an open-source approach and significantly lower development costs. Through comprehensive research and analysis of DeepSeek's technical innovations, particularly their DeepSeek-V3 and DeepSeek-R1 models, I've uncovered insights into how this disruptor is challenging conventional wisdom about AI development costs and accessibility. This article examines DeepSeek's technological breakthroughs, their implications for the global AI race, and what their emergence means for making advanced AI accessible to all.
Intro
DeepSeek, a relatively unknown Chinese AI firm with roots in the quantitative stock trading firm High-Flyer1 has sent ripples through the AI community with its release of DeepSeek-V3 and DeepSeek-R1, two powerful open-source AI systems. These models are impressive not only for their technical capabilities, which rival those of industry giants like OpenAI and Google, but also for their remarkably low development costs and open-source accessibility. This has sparked considerable discussion about the evolving AI landscape and the intensifying competition between the US and China in AI development.
DeepSeek-V3: A Technical Marvel
With a massive 671 billion parameters, DeepSeek-V3 surpasses even Meta's Llama 3.1 in scale2. However, DeepSeek-V3 distinguishes itself through its innovative Mixture-of-Experts (MoE) architecture. This architecture activates only the necessary neural networks for specific tasks, resulting in significant cost savings and improved efficiency3. Despite its vast number of parameters, DeepSeek-V3 operates with just 37 billion during actual tasks3. This efficient design allows it to achieve high performance with significantly less computational power and cost compared to its peers4.
DeepSeek-V3 excels in various text-based tasks, including coding, translation, and writing2. It has achieved top scores on popular AI benchmarks, such as HumanEval, GSM8K, and MMLU, challenging both open and closed-source models3. Notably, DeepSeek-V3 outperformed Meta's Llama 3.1, OpenAI's GPT-4o, and Anthropic's Claude Sonnet 3.5 in accuracy across various tasks, from complex problem-solving to math and coding6.
However, it's important to acknowledge that DeepSeek-V3 is primarily focused on text-based tasks and does not possess multimodal abilities2. Like many large language models, it may also inherit biases from its training data, requiring careful consideration in real-world applications2.
DeepSeek-R1: Mastering Reasoning
DeepSeek further solidified its position with the release of DeepSeek-R1, a reasoning model designed to tackle complex problems with a focus on logical inference, mathematical problem-solving, and real-time decision-making7. This sets it apart from traditional language models, which primarily focus on text generation and comprehension.
DeepSeek-R1's development began with DeepSeek-R1-Zero, a foundational model trained exclusively via reinforcement learning8. While R1-Zero showed promise in reasoning, it faced challenges with readability and output coherence. DeepSeek addressed these issues in R1 by incorporating cold-start data and a multi-stage reinforcement learning process9.
DeepSeek-R1 has demonstrated remarkable performance on various benchmarks, including AIME (American Invitational Mathematics Examination) and MATH1. While DeepSeek claimed that R1 exceeded the performance of OpenAI's o1 on these benchmarks, independent analysis by The Wall Street Journal found that o1 was faster in solving AIME problems1. Nevertheless, R1's performance remains competitive with leading models in the field.
Feature
DeepSeek-R1
OpenAI o1
Architecture
Mixture-of-Experts (MoE)
-
Parameters
671 billion total, 37 billion active
-
AIME 2024 (Pass@1)
79.8%
79.2%
MATH-500 (Pass@1)
97.3%
96.4%
Codeforces (Percentile)
96.3
96.6
Cost (per million tokens)
$2.19
$60
Key Features
Open-source, transparent reasoning
Chain-of-thought processing
DeepSeek-R1 has also shown promising results in financial analysis, outperforming the S&P 500 and maintaining superior Sharpe and Sortino ratios compared to the market7. Furthermore, it exhibits a unique ability to provide transparent reasoning, offering insights into its decision-making process10. However, it's worth noting that the model tends to align with the official Chinese government position on sensitive political topics10.
DeepSeek-VL: Expanding into Vision-Language
Beyond its language models, DeepSeek is also exploring vision-language (VL) capabilities with DeepSeek-VL11. This model is designed for real-world vision and language understanding applications, with the ability to process logical diagrams, web pages, formulas, scientific literature, and natural images11. DeepSeek-VL showcases the company's commitment to advancing AI research across multiple modalities.
Open-Source and Cost-Effective: AI For All
One of the most significant aspects of DeepSeek's models is their open-source nature. This allows developers and researchers to freely access, modify, and deploy the models, fostering collaboration and innovation within the AI community4. This open approach contrasts with the proprietary models of many US-based companies and has the potential to democratize access to advanced AI technologies12.
DeepSeek has achieved these impressive results with significantly lower development costs. DeepSeek-V3 was reportedly trained in around 55 days at a cost of US$5.58 million, using considerably fewer resources compared to its peers1. This cost-effectiveness challenges the existing paradigm in the AI industry, where high performance has typically been associated with high costs4.
Implications for the US-China AI Race
DeepSeek's emergence has raised concerns in the US about China's growing AI capabilities13. The US government has implemented restrictions on China's access to advanced AI chips, aiming to curb its progress in AI development14. However, DeepSeek has demonstrated that Chinese researchers can develop world-class AI models with limited resources and by leveraging open-source technologies15. This raises questions about the effectiveness of these restrictions and their potential to inadvertently spur innovation in China by forcing researchers to focus on efficiency and alternative approaches14.
DeepSeek's success has been attributed to several factors, including its efficient MoE architecture, innovative training methods, and a focus on maximizing resource utilization16. The company's ability to develop high-performing models at a fraction of the cost of its US counterparts has put pressure on companies like Meta and OpenAI to rethink their AI strategies17.
Expert Opinions and Analysis
Experts in the AI field have recognized DeepSeek's significant contributions to AI development. They highlight the model's impressive performance, cost-effectiveness, and open-source nature as key factors that could reshape the AI landscape7. Some experts suggest that DeepSeek's approach could lead to a democratization of AI, making advanced AI capabilities more accessible to a wider range of developers and researchers12. Others emphasize the potential for DeepSeek to accelerate innovation and competition in the AI industry, potentially leading to breakthroughs in various fields18.
Use Cases and Applications
DeepSeek's models have shown potential in a variety of applications. DeepSeek-R1, for example, has been used to run complex reasoning tasks on smartphones, generate code for rotating objects with collision detection, and even build a clone of the AI-powered conversational search engine Perplexity AI19. The models have also shown promise in areas like software development, business operations, and education3.
Synthesis
DeepSeek's emergence as a major player in the AI landscape has significant implications for the future of AI development. Its open-source approach, combined with its high-performing and cost-effective models, challenges the dominance of US-based companies and has the potential to democratize access to advanced AI technologies. This could foster a more diverse and inclusive AI ecosystem, with wider participation from developers and researchers worldwide.
DeepSeek's success also highlights the growing capabilities of Chinese AI research and the potential for open-source technologies to disrupt the AI industry. As the AI race intensifies, DeepSeek's innovative approach and commitment to accessibility could shape the future of AI development and its impact on the global technology landscape.
Author: Malik Datardina, CPA, CA, CISA. Malik works at Auvenir as a Sr. AI Product Manager who is working to transform the engagement experience for accounting firms and their clients. The opinions expressed here do not necessarily represent UWCISA, UW, Auvenir (or its affiliates), CPA Canada or anyone else. This post was written with the assistance of an AI language model. The model provided suggestions and completions to help me write, but the final content and opinions are my own.