Alphabet is one of the largest tech conglomerates globally, generating around $300 billion annually—equivalent to the economies of Finland or Portugal.
Google remains Alphabet’s most prominent subsidiary, contributing significantly to its influence and financial success.
The visual representation of Alphabet’s Q2 FY23 Income Statement reveals the distribution of revenue and expenses across various segments. Google Search advertising generated USD 42.6 billion, contributing significantly to the USD 58.1 billion in total ad revenue, which also included YouTube ads and Google AdMob. A staggering 57% of the operating profit is attributed to Google Search advertising.
Google has dominated the internet search landscape since the turn of the millennium and — for now — controls about 90% of the market. And search advertising is the core of what Google does as a business.
How Google Makes Money ?
To understand how Google attained its omnipresence, you have to understand its business model. Google primarily makes money through advertising. The majority of Google’s revenue comes from Google Search ads, which appear alongside search results. In 2024, Google led the global digital advertising market with a 25.7% share, followed by Meta Platforms Inc. (19.7%) and Amazon.com Inc. (7.1%). Other contributors include Alibaba, ByteDance, and Tencent, each holding between 2% to 4% of the market.
Google’s revenue model is driven by ads embedded in its search engine results pages (SERPs). This business model is highly profitable, as evidenced by Alphabet’s Q2 2024 income statement, where Google Search advertising generated $48.5 billion in revenue—accounting for a significant portion of Alphabet’s overall profit. The reliance on ad revenue, however, makes Google vulnerable to shifts in user behavior, such as the migration towards ad-free conversational AI platforms.
Meta Platforms founder Mark Zuckerberg famously answered a senator’s question about sustaining a business model without charging users by stating,
Senator, we run ads.
This straightforward statement underscores the business model that not only Facebook but also Google and many other tech giants follow: offering free services funded by advertising. The concentration of digital ad revenue among a few major players highlights both the scale of the opportunity and the potential vulnerability if user preferences change
The Decline of Conventional Search
To understand the disruption caused by LLMs, we first need to evaluate trends in search engine usage over the past few years. According to data from Statista and SimilarWeb, the total number of searches through traditional engines like Google has seen a steady decline of about 5% per year since 2020, while traffic to conversational AI platforms has surged. Users are increasingly seeking more contextual and conversational answers that search engines, despite improvements in natural language processing, fail to provide directly.
PageRank’s Core Formula and Limitations
The PageRank algorithm fundamentally models each webpage’s importance based on its inbound links, weighted by the authority of the linking pages. The formula for PageRank of a page A is:
PR(A) = (1 – d) / N + d * Σ [ PR(T_i) / C(T_i) ]
where:
PR(A): PageRank of page A
d: Damping factor (typically 0.85), representing the likelihood of continuing to click on links
N: Total number of pages in the network
Σ: Sum over all pages that link to A
PR(T_i): PageRank of page T_i (pages that link to A)
C(T_i): Number of outbound links on page T_i
The algorithm iterates this formula across all pages in the graph until the PageRank values stabilize. Lately, many concerning issues have surfaced from this very algorithmic design.
Popularity ≠ Truth
Core Issue: PageRank equates a page’s value with the number and authority of its inbound links, assuming links represent a form of endorsement.
Validation: Links don’t always imply accuracy or quality. In fact, misinformation can achieve high PageRank scores if well-linked by authoritative domains, as observed in studies analysing misinformation spread during news events.
Collusion
PageRank’s trust bias toward certain top-level domains (like .ORG) and established high-authority sites can inadvertently promote content that may not necessarily be the most accurate or high-quality but rather reflects vested interests.
How Collusion Occurs in PageRank – A Case Study
Authority Bias Toward Established Domains: PageRank often favours domains like .ORG or .EDU, as they are presumed to be non-commercial and therefore more trustworthy. This bias gives these sites a higher baseline PageRank authority. Consequently, when such sites link to other content, it passes on high PageRank, regardless of the actual content quality or neutrality.
Content Control via Vested Interests: Established high-authority sites often act as gatekeepers, selectively linking to content that aligns with their backers’ interests. For example,
-> A high-ranking Wikipedia page on “List of Diets” links to kidneyfund.org, a .ORG site with financial ties to pharmaceutical companies.
-> When someone submitted a higher-quality, more comprehensive resource on a .COM site, the link was rejected, not on quality grounds but because of perceived profit motives by being a COM domain.
-> However, the original .ORG link remains, indirectly benefiting pharmaceutical stakeholders while influencing public perception through a seemingly unbiased source.
Hidden Marketing Front Ends: Many .ORG and non-profit sites receive funding from corporate sponsors, effectively acting as subtle marketing channels. This makes them selective in linking to external sites, favoring those aligned with their sponsors’ interests and rejecting those without corporate ties or with competing perspectives.
A significant portion of revenue of KidneyFund.org comes from contributions and grants, for ex a total around $301 million in the year of 2018 came from contributions and grants. This highlights how nonprofits like the American Kidney Fund might rely heavily on corporate or institutional grants, which could influence the content they promote or link to. In March 2024, American Kidney Fund (AKF) announced its latest class of Corporate Members, categorised by their level of support:
Champion Level: Amgen, Inc.; Boehringer Ingelheim and Eli Lilly and Company; GSK plc; Novartis Pharmaceutical Corporation; Travere Therapeutics, Inc.; Vertex Pharmaceuticals, Inc.
Advocate Level: Apellis Pharmaceuticals, Inc.; Aurinia Pharmaceuticals Inc.; Bayer U.S. LLC; Biotechnology Innovation Organization; Hansa BioPharma AB; Human Immunology Biosciences (HI-Bio); Novo Nordisk A/S.
Friend Level: Merck and Co.; PicnicHealth.
There is a possibility that Wikipedia Authors for that page (List of Diets; high traffic backlink source from high DA site like Wikipedia) are paid by KidneyFund.org to "gatekeep" links to KidneyFund.org.
Search Wars !
The large language model (LLM) AI made its mass-market debut with ChatGPT in November 2022. Now, there’s ChatGPT, Claude, Gemini (owned by Google), Perplexity, Bing AI, and more emerging constantly. AI isn’t confined to chatbots either — it’s reshaping nearly every market and industry. The emergence of large language models (LLMs) such as OpenAI’s GPT-4 has brought significant shifts in how people search for information online.
Traditionally, Google Search has dominated the market by providing users with indexed lists of web results based on their queries. However, LLMs offer a paradigm shift: they can provide conversational, direct, and contextualized answers, reducing the need for users to scroll through multiple websites.
This change is starting to disrupt the established search industry—particularly Google’s long-standing monopoly—by offering faster, more natural, and in-depth insights. Let’s explore how this disruption is occurring, the data supporting it, and what the future might hold.
Key Data Insights:
Click-Through Rates (CTR): Click-through rates for search engine result pages (SERPs) have declined as more users receive immediate answers from LLMs. A BrightEdge study found that the overall CTR for Google dropped by 12% between 2020 and 2023, pointing to more users either clicking on fewer links or avoiding them entirely thanks to direct AI responses. Early 2000s: The top organic result often garnered over 50% of clicks, with users primarily focusing on the first few links. By 2024, the first organic result now receives approximately 39.8% of clicks, the second 18.7%, and the third 10.2%
Growth of Chat Platforms: OpenAI’s ChatGPT reached over 100 million active users within its first 8 months. In contrast, Google’s usage growth has stagnated, with a YoY growth rate of only 2% since 2022. This rapid adoption of LLMs underlines how users are turning to chat-like interactions for information.
The Key Advantages of LLMs over Search Engines
LLMs are providing three major improvements over traditional search engines, leading to user migration:
Personalization: Unlike a simple keyword search, LLMs can adjust responses based on user context, preferences, and the conversational flow. This makes it easy for users to get precisely what they need without refining their keywords repeatedly.
Complex Queries: Traditional search engines may struggle with complex, multi-layered questions. LLMs can comprehend and provide structured responses, allowing users to ask deeper questions without breaking them into several smaller searches.
Unified and Direct Information: While a Google search results in links requiring users to browse through different sites, LLMs synthesize multiple sources into a single, comprehensive answer. This eliminates the often-tedious job of cross-referencing between tabs.
SearchGPT
OpenAI has introducedSearchGPT (1 Nov, 2024), an AI-powered search engine integrated into its ChatGPT chatbot, aiming to challenge Google’s dominance in the search market. SearchGPT provides real-time information, including sports scores, stock quotes, news, and weather, by partnering with various news organizations. The last bit explains licensing spree OpenAI has recently gone through.
OpenAI has entered into a series of high-profile content licensing deals from July 2023 to June 2024, partnering with prominent media and content companies. Starting with Shutterstock and AP in July, OpenAI later struck a USD 25M+ deal with Axel Springer in December, covering brands like Politico and Business Insider. Further deals followed with major publishers like Le Monde, PRISA Media, and the Financial Times. In May, OpenAI secured agreements with Stack Overflow, Dotdash Meredith, Reddit, and News Corp (in a $250M+ deal, including the Wall Street Journal and Barron’s). Additional partnerships include Vox Media, The Atlantic, and TIME, indicating OpenAI’s aggressive but strategic expansion in acquiring diverse, quality and realtime content to power SearchGPT to challenge Google’s core business model.
In SearchGPT, users can engage in conversational searches, allowing for follow-up questions and more nuanced interactions. Initially available to ChatGPT Plus and Team users, and those on SearchGPT’s waitlist, it will gradually roll out to other users. OpenAI plans to enhance SearchGPT with improved shopping and travel suggestions and the ability to dynamically create custom web pages in response to search queries
While LLMs bring undeniable advantages, they aren’t without challenges. The following factors are barriers to fully replacing traditional search engines:
Accuracy and Hallucination Issues: LLMs have been known to generate incorrect or misleading information. Google Search’s reliability lies in delivering verified links, whereas LLMs need better integration with fact-checking systems.
Lack of Real-Time Information: Google indexes fresh content every second, providing real-time updates. Current LLMs, unless connected to the internet in real-time, can only respond based on pre-trained data, making them less useful for trending news or events.
Commercialization & Ad Revenue: Google’s model monetizes through ads embedded in SERPs, which fund the cost of their services. For LLMs, an effective monetization model without compromising user experience is still under exploration.
Hallucinate, You Should Not !
To address large language models’ (LLMs) hallucination issues and improve their responsiveness to real-time data, researchers are exploring several innovative approaches. Here’s a summary of some of the latest directions in research:
1. Retrieval-Augmented Generation (RAG) or Retrieval Interspersed Generation (RIG)
Overview: RAG models combine traditional LLMs with real-time retrieval systems. Instead of solely relying on pre-trained knowledge, they query a database or search engine during generation, providing up-to-date and contextually relevant information.
Recent Advancements: Techniques like Google’s Retrieval-Enhanced Transformer (RETRO) and Meta’s Fusion-in-Decoder (FiD) incorporate retrieved passages directly into the generation process. RAG models are effective for reducing hallucinations because they base responses on specific, retrievable data.
2. Fine-Tuning with Real-Time Data Pipelines
Overview: Continuous fine-tuning using real-time or frequently updated datasets can help models stay current. This involves creating data pipelines that constantly stream new information (e.g., news articles, verified knowledge bases).
Challenges: Implementing fine-tuning pipelines is computationally intensive and requires strategies to avoid “catastrophic forgetting” (where new information disrupts previously learned data).
3. Reinforcement Learning from Human Feedback (RLHF) with Real-Time Data
Overview: RLHF has been used to align LLMs with human preferences, but now it’s being extended to incorporate real-time feedback. For example, when users correct or flag hallucinations, the model can learn from these interactions dynamically.
Advancements: Research is exploring adaptive RLHF, where models adjust based on user interactions in near real-time, leveraging feedback loops to reinforce accurate outputs and discourage incorrect ones.
4. Plug-and-Play Models
Overview: Rather than training a monolithic model that “knows everything,” plug-and-play methods allow the LLM to connect with various smaller, specialized models or databases. This enables the main model to pull from an external model or knowledge base when it needs specific information.
Examples: Systems like Toolformer, which Meta introduced, allow the LLM to autonomously decide when to use external tools, APIs, or databases, integrating real-time data more seamlessly and reducing the chances of hallucinations.
5. Hybrid Model Architectures (e.g., Symbolic Reasoning + Neural Networks)
Overview: Combining symbolic AI (rule-based systems) with neural networks allows for more logical consistency and factual accuracy. Symbolic reasoning can verify the output of neural models, flagging or correcting hallucinated information.
Recent Work: Some researchers are creating hybrid models where LLMs generate answers, but symbolic logic layers cross-check outputs against established rules or facts, helping models avoid generating false information.
6. Fact-Checking Layers and Self-Verification Mechanisms
Overview: Researchers are adding fact-checking components to LLMs. After generating a response, a secondary model verifies the output by cross-referencing with factual databases or performing a retrieval-based check.
Examples: Models like “Verifier LMs” and fact-checking agents are being developed to verify responses before finalizing them, reducing hallucinations by ensuring accuracy against real-time data sources.
7. Memory-Augmented Models
Overview: Memory-augmented models store past interactions, facts, or verified information, allowing them to recall and integrate learned facts over multiple sessions. This “memory” can be updated with new data, keeping the model current.
Example: Microsoft’s research on memory-augmented transformers aims to give LLMs dynamic memory, which can be updated with real-time data without extensive retraining.
8. External Knowledge Base Integration (e.g., Wikipedia, Knowledge Graphs)
Overview: Direct integration of structured knowledge bases, such as Wikidata or proprietary databases, allows models to access accurate information in real-time.
Recent Research: By integrating knowledge bases that are updated continuously, researchers aim to mitigate hallucinations by providing a reliable source of truth for LLMs to consult when generating responses.
9. Self-Supervised and Semi-Supervised Learning on Updated Data Streams
Overview: Self-supervised methods that train on continuously updated data streams (e.g., news headlines, social media) allow LLMs to adapt without explicit manual labeling.
Challenges: Although this reduces lag in information, self-supervision on unverified sources can increase the risk of propagating misinformation. Researchers are thus working on filtering mechanisms to maintain data quality.
10. Evaluation Metrics and Training for Factual Consistency
Overview: Developing new metrics to assess factual accuracy and consistency during training can help. Researchers are now using factual consistency benchmarks to guide LLM training and fine-tuning.
Notable Projects: Projects like “TruthfulQA” provide datasets and evaluation metrics specifically designed to measure factual accuracy, guiding models to reduce hallucinations.
Idea: VeritasGPT (by Deb Bose)
This system represents a hybrid approach, blending advanced AI techniques with decentralized data validation, offering a robust and dynamic solution to the challenges of hallucinations, trust, and real-time information access in LLMs. Designing an improved LLM system with dynamic memory, adaptive reinforcement from real-time user feedback, and an autonomous fallback to a decentralized knowledge graph (DKG) requires a multi-layered architecture. This system would balance the benefits of memory-augmented transformers and adaptive RLHF while leveraging a decentralized knowledge base for trust and integrity. Here’s a blueprint for such a system:
1. Dynamic Memory-Augmented Transformer Layer
Purpose: To provide the LLM with a dynamic, updateable memory that can store recent data and context without the need for extensive retraining. This layer allows the LLM to “remember” facts, interactions, and verified data across sessions.
Architecture:
Memory Cells: Set up dedicated memory cells within the transformer model where information can be stored. These cells can hold contextual data, recent events, and frequently updated facts.
Data Access Mechanism: Use a key-value memory addressing system, where keys represent specific facts, entities, or contextual queries, and values store the relevant data. This allows the model to quickly retrieve information by searching for relevant keys.
Memory Update: Implement a “forgetting” mechanism to ensure the memory doesn’t become overloaded. Older, less relevant data can be periodically removed or deprioritized.
Data Quality Control: Memory cells are populated only with high-confidence data verified by trusted sources or validated by the DKG (using Chainlink Oracles). This minimizes the risk of misinformation.
Operation:When generating a response, the transformer first checks its memory for relevant information. If the information is outdated or missing, it moves to retrieve real-time data from the adaptive RLHF or the decentralised knowledge graph.
2. Adaptive RLHF as a Meta-Framework
Purpose: To provide a real-time feedback loop, allowing the model to learn from user interactions and adapt outputs based on user feedback. This framework will help reinforce accurate responses and reduce the likelihood of repeating errors.
Architecture:Feedback Loop System: Use a dynamic reinforcement learning feedback loop where every user interaction provides positive or negative reinforcement. If users correct or flag a response, the model adjusts its weightings to discourage similar future outputs.
Confidence Scoring: Each response is assigned a confidence score. User feedback can adjust this score, reinforcing correct responses with higher confidence levels and downgrading inaccurate ones. The confidence score impacts how often certain data or memory cells are accessed in future responses.
Real-Time Fine-Tuning: Store feedback data in an “experience buffer” and perform lightweight fine-tuning during low-usage times. This allows the model to integrate recent feedback without full-scale retraining.
Operation:After each interaction, the system evaluates user feedback. If feedback is positive, the relevant data or pathway is reinforced. If negative, the feedback is stored, and corrections are applied to prevent similar hallucinations or inaccuracies.Adaptive RLHF also plays a meta-role by deciding when to update the model’s memory layer with new data, based on repeated user confirmations or corrections.
Purpose: To autonomously decide when to fall back on a decentralized knowledge graph (using OriginTrail’s DKG) for validated and tamper-resistant information. This ensures that the model accesses a trustworthy source, especially when dealing with high-stakes or frequently manipulated information.
Architecture:DKG Access Layer: Create an intermediary layer in the model’s architecture that determines when to access the DKG. This layer is responsible for querying the DKG using structured protocols and retrieving trusted data.
Chainlink Oracle Integration: Use Chainlink Oracles to validate external data before it enters the blockchain. This layer ensures that only verified data reaches the DKG, minimizing the risk of corrupted information.
Knowledge Assets: Store critical pieces of information (such as data about entities, events, and relationships) as “knowledge assets” within the DKG. Each asset has a unique identifier and is managed as an ownable container of information. The blockchain infrastructure tracks the integrity and ownership of these assets, making them discoverable and resistant to tampering.
Fallback Decision Mechanism: Implement logic that lets the model autonomously decide when to consult the DKG. For instance, if the LLM encounters a query that triggers certain confidence or trust criteria (e.g., sensitive topics, disputed information, or high-stakes data), it falls back to the DKG.
Operation:The LLM system evaluates the confidence level and sensitivity of each response. If the confidence score is low or if adaptive RLHF flags the topic as potentially high-risk, the model queries the DKG.Through the DKG access layer, the model retrieves validated, up-to-date knowledge assets. For example, if asked for information on a recent or contested event, the model consults the DKG and retrieves verified information, ensuring a higher level of trustworthiness in its response.
4. Workflow Summary
Here’s how the improved LLM system would operate in real time:
User Query: The user asks a question or interacts with the system.
Memory Check: The model first consults its dynamic memory layer to retrieve any relevant, recently stored data.
Adaptive RLHF Analysis: If the memory doesn’t contain the needed information, the adaptive RLHF framework decides whether this query requires fallback to the DKG or whether recent user feedback can guide the response.
DKG Fallback: If the adaptive RLHF indicates a need for high-trust data, the model accesses the DKG through the access layer, querying validated knowledge assets stored within the DKG. Chainlink Oracles ensure that the data feeding into the DKG is accurate.
Response Generation: The model generates a response based on the information retrieved from the memory, RLHF feedback, or DKG.
User Feedback Loop: The user provides feedback on the response. This feedback updates the adaptive RLHF framework, reinforcing or discouraging specific information and updating memory if the feedback is confirmed as consistent.
Future Scenarios
Google has responded to this disruption with its own LLM-based tools such as Bard and by integrating AI-driven snippets into its search results.
However, the competitive edge might lean towards companies that can effectively combine the best of both worlds—LLMs’ conversational depth with search engines’ reliability. The future of search may not entirely replace traditional engines but will likely involve hybrid systems offering both traditional and conversational results depending on query complexity.
Conclusion
LLMs have begun reshaping the search landscape, offering a new way for people to engage with information through personalized, direct, and more nuanced responses. However, challenges such as accuracy, monetization, and real-time indexing remain hurdles. Google, while facing challenges, is also adapting by incorporating LLM technologies into its core product.
The ultimate outcome will depend on how well search giants adapt and how users decide to engage with information—through exploratory browsing or conversational querying. One thing is certain: the landscape of search is transforming, and LLMs are at the heart of that transformation.
NOTE:
This article has originally published at BosonResearch.
AI and Digital Transformation, Chemical Scientist, MBA.
1moIs disrupting search engines the real value proposition?