Patent 9269097
Derivative works
Defensive disclosure: derivative variations of each claim designed to render future incremental improvements obvious or non-novel.
Active provider: Google · gemini-2.5-pro
Derivative works
Defensive disclosure: derivative variations of each claim designed to render future incremental improvements obvious or non-novel.
Defensive Disclosure and Prior Art Derivations for U.S. Patent 9,269,097
Document ID: DPD-2026-0510-001
Publication Date: May 10, 2026
Subject: Derivative Implementations and Obvious Variations of Natural Language Processing Based on Contextual Advertisements
This document describes variations, extensions, and alternative implementations of the system and method disclosed in U.S. Patent 9,269,097. The purpose is to place these concepts into the public domain, thereby establishing prior art against future patent applications claiming these or similar incremental improvements. The core concept involves interpreting a user's natural language utterance, particularly pronouns, by referencing a previously presented advertisement as context.
Analysis of Core Claims
The fundamental principle of US 9,269,097, as articulated in independent claims 1 and 23, is a computer system that:
- Provides an advertisement (the context).
- Receives a subsequent natural language utterance from a user.
- Interprets the utterance by determining if a pronoun within it refers to the product, service, or provider from the advertisement.
The following disclosures build upon this foundation.
Derivative Disclosures
1. Material & Component Substitution
Derivative 1.1: Neuromorphic Processor Implementation
Enabling Description: The method of claim 1 is implemented on a system utilizing a neuromorphic processing unit (NPU), such as Intel's Loihi 2 or a similar spiking neural network (SNN) architecture, instead of a traditional von Neumann CPU/GPU. The advertisement's metadata (product ID, service type, provider name) is encoded as a static or slowly changing neural map within the SNN. When the user's utterance is received, the speech-to-text output is tokenized and fed into the SNN as a series of spikes. The pronoun token triggers a search for the most recently activated and strongly weighted entity in the neural map. The NPU's inherent parallelism and event-driven processing allow for extremely low-latency pronoun resolution, making it suitable for real-time, power-constrained edge devices. The system's memory can be phase-change memory (PCM) to retain the context map even in low-power states.
Mermaid Diagram:
graph TD A[Audio Input: "Buy it"] -- STT --> B{Tokenized Input: ["Buy", "it"]}; C[Ad Context: {Product: 'X', Provider: 'Y'}] -- Encoded --> D[Neuromorphic Core]; subgraph D [Neuromorphic Core] D1[Neural Map: 'X' & 'Y' Nodes] end B -- Spikes --> D; D -- Pattern Match --> E{Pronoun "it" correlates with highly activated 'X' Node}; E --> F[Generate Request: purchase(product='X')];
Derivative 1.2: Federated Learning for Context Models
Enabling Description: The system of claim 23 is implemented as a decentralized, federated learning architecture. Instead of a central server selecting advertisements and interpreting utterances, individual user devices (clients) maintain their own context models. An on-device model, trained using TensorFlow Lite or Core ML, processes the advertisement and the utterance locally. To improve the model without compromising privacy, only the model updates (gradients), not the raw data, are sent to a central aggregation server. This allows the global model to learn common pronoun-antecedent patterns (e.g., "call them" usually refers to the provider) across a population, while the specific context (the ad and the utterance) never leaves the user's device. This substitutes the central server architecture with a distributed, privacy-preserving computational model.
Mermaid Diagram:
sequenceDiagram participant ClientDevice participant AggregationServer ClientDevice->>ClientDevice: Presents Ad, receives utterance "Call them" ClientDevice->>ClientDevice: On-device model resolves "them" to Ad Provider ClientDevice->>ClientDevice: Computes model gradient from this interaction ClientDevice->>AggregationServer: Send encrypted model gradient AggregationServer->>AggregationServer: Aggregates gradients from many clients AggregationServer-->>ClientDevice: Sends updated global model weights
2. Operational Parameter Expansion
Derivative 2.1: High-Frequency Trading (HFT) Ad Auction and Interpretation
Enabling Description: The invention is applied to an HFT environment where ad slots are auctioned and user commands are executed in microseconds. The "advertisement" is a fleeting data feed of a stock price or a market opportunity presented to a trader. The trader's subsequent voice command, e.g., "sell half of it," must be interpreted at sub-millisecond speeds. The system operates on a field-programmable gate array (FPGA) co-located with the trading exchange's servers. The context (the stock symbol) is held in the FPGA's block RAM. The speech-to-text and NLP processes are hardware-accelerated logic circuits. This scales the operation from typical human-interaction speeds to the extreme frequencies and low latencies required for algorithmic trading.
Mermaid Diagram:
graph LR A[Market Data Feed: {Symbol: XYZ, Price: 100.50}] -- Pushed to --> B(FPGA Block RAM); C[Trader Utterance: "Sell half of it"] -- A/D & STT --> D{FPGA Logic}; B -- Context --> D; D -- Pronoun Resolution --> E{Resolved Command: sell(symbol='XYZ', qty=0.5*position)}; E -- Executed in <1ms --> F[Trade Execution Engine];
Derivative 2.2: Nanoscale Molecular Assembly Control
Enabling Description: The concept is scaled down to control nanoscale robotic assemblers. The "advertisement" is a visual feed from an atomic force microscope displaying a specific molecule or binding site. The operator issues a natural language command, such as "apply the catalyst to it." The system, running on a quantum-annealing processor, interprets "it" as the specific molecular coordinate highlighted in the microscope feed. The processor resolves the ambiguity by analyzing the quantum state representing the visual context and the utterance, selecting the most probable antecedent. The output is not a web request but a series of precise voltage changes to manipulate piezoelectric nano-actuators.
Mermaid Diagram:
stateDiagram-v2 [*] --> DisplayingContext DisplayingContext: AFM shows Molecule_A at [x,y,z] DisplayingContext --> WaitingForInput: Operator sees context WaitingForInput --> Processing: Receives command "Activate it" Processing --> Action: System resolves "it" to Molecule_A Action --> [*]: Sends signal to nano-actuator
3. Cross-Domain Application
Derivative 3.1: Aerospace - Cockpit Fault Response
Enabling Description: In an aircraft cockpit, the "advertisement" is a specific warning from the Engine Indicating and Crew Alerting System (EICAS), such as "HYDRAULIC PUMP 1B FAIL." A pilot's subsequent voice command, "run the checklist for it," is interpreted by the avionics system. The system uses the EICAS message as the context to resolve "it" to the "Hydraulic Pump 1B Failure" event. It then automatically displays the correct non-normal checklist on the multi-function display, bypassing the need for manual lookup under high-stress conditions.
Mermaid Diagram:
flowchart TD A[EICAS Alert: "HYDRAULIC PUMP 1B FAIL"] --> B{Avionics Voice Control System}; C[Pilot Utterance: "Run the checklist for it"] --> B; B -- Resolves "it" against EICAS context --> D[Identify Checklist: "QRH HYD 1B PUMP FAIL"]; D --> E[Display checklist on MFD];
Derivative 3.2: AgTech - Precision Irrigation and Pest Control
Enabling Description: An agricultural drone surveying a field identifies a specific patch of crops suffering from dehydration or pest infestation. This visual data, with GPS coordinates, is presented to a remote farm operator as a "context alert." The operator issues the command, "dispatch a sprayer to it." The farm management system interprets "it" as the specific GPS-tagged patch of crops. It then generates a mission plan for an autonomous ground vehicle to navigate to the coordinates and apply the appropriate amount of water or pesticide, directly linking a high-level natural language command to a specific, geolocated agricultural action.
Mermaid Diagram:
sequenceDiagram participant Drone participant Operator participant FarmMgmtSystem participant SprayerBot Drone->>Operator: Pushes Alert: "Pest Infestation at [Lat, Lon]" Operator->>FarmMgmtSystem: Voice Command: "Dispatch a sprayer to it" FarmMgmtSystem->>FarmMgmtSystem: Resolve "it" to [Lat, Lon] FarmMgmtSystem->>SprayerBot: Generate & send mission plan SprayerBot->>SprayerBot: Navigate and execute spraying
Derivative 3.3: Consumer Electronics - Smart Home Ambiance Control
Enabling Description: A smart home hub displays an interactive advertisement for a movie on a connected TV. The ad includes metadata about the movie's genre (e.g., "Horror"). After the ad plays, the user says to a nearby smart speaker, "set the lights for it." The smart home hub uses the advertisement's genre metadata as context to resolve "it." It then executes a pre-defined lighting scene named "Horror Movie," which dims the lights, lowers the color temperature, and perhaps closes the smart blinds, creating an immersive experience without the user needing to specify the exact scene name.
Mermaid Diagram:
graph TD A[TV displays Movie Ad with Genre='Horror' metadata] --> B((Smart Home Hub)); C[Smart Speaker receives "Set the lights for it"] --> B; B -- Resolves "it" using genre context --> D{Action: Execute "Horror Movie" lighting scene}; D --> E[Control Philips Hue Lights]; D --> F[Control Lutron Blinds];
4. Integration with Emerging Tech
Derivative 4.1: AI-Driven Predictive Intent
Enabling Description: The system is integrated with a predictive AI model that analyzes user behavior (gaze tracking, ad dwell time, historical interactions). Before the user even speaks, the system pre-computes a probability distribution of likely pronoun referents. For an advertisement for a car, the AI may determine a 70% probability that "it" will refer to the car, 20% to the dealership (provider), and 10% to the financing offer. This allows the system to pre-fetch information or pre-load application logic. When the utterance "what's the price of it" is received, the system can instantly resolve "it" to the car and provide the answer with minimal latency, because the required data was already cached based on the AI's prediction.
Mermaid Diagram:
flowchart TD subgraph Pre-Utterance A[User views Car Ad] -- Gaze data, history --> B(Predictive AI Model) B -- Generates --> C(Intent Probabilities <br> P(it=Car)=0.7 <br> P(it=Dealer)=0.2) C -- Triggers --> D[Pre-fetch Car Price Data] end subgraph Post-Utterance E[User asks "What's the price of it?"] --> F{NLP Engine} C -- Feeds into --> F F -- Resolves "it" to Car --> G[Retrieve price from cache] G --> H[Synthesize & speak answer] end
Derivative 4.2: IoT Sensor Fusion for Contextual Disambiguation
Enabling Description: The advertisement context is augmented with real-time data from IoT sensors. An ad for a smart thermostat is displayed on a user's phone. Simultaneously, an IoT temperature sensor in the room reports a high ambient temperature. The user says, "turn it down." A simple system might be confused between the product (thermostat) and the service (HVAC). This integrated system fuses the ad context with the IoT data. It infers that the user's command is motivated by the physical environment and correctly resolves "it" to the HVAC service controlled by the thermostat, not just the thermostat device itself, and initiates an action to lower the temperature.
Mermaid Diagram:
erDiagram USER { string utterance } ADVERTISEMENT { string product "Thermostat" string service "HVAC Control" } IOT_SENSOR { int temperature "80F" } NLP_ENGINE { } USER ||--o{ NLP_ENGINE : provides ADVERTISEMENT ||--|{ NLP_ENGINE : provides_context IOT_SENSOR ||--|{ NLP_ENGINE : provides_context NLP_ENGINE }o--|| ACTION : "generates"
Derivative 4.3: Blockchain for Verifiable Ad-Based Transactions
Enabling Description: For high-value transactions initiated via voice command following an ad, the system uses a blockchain for verification. An advertisement for a limited-edition digital art NFT is displayed. The user says, "I'll buy it." The system resolves "it" to the specific NFT. The action generated is a smart contract transaction. The transaction payload immutably records a cryptographic hash of the advertisement content, the resolved user command, and the user's biometric voiceprint signature. This creates an unalterable, auditable record on a public ledger (e.g., Ethereum), proving that the purchase was initiated in direct response to a specific advertisement, preventing disputes about the terms or subject of the sale.
Mermaid Diagram:
sequenceDiagram participant UserDevice participant NLP_Service participant Blockchain UserDevice->>UserDevice: Display NFT Ad (Ad_Hash) UserDevice->>NLP_Service: Utterance: "I'll buy it" + Voiceprint NLP_Service->>NLP_Service: Resolve "it" to NFT_ID from Ad_Hash NLP_Service->>Blockchain: Create Transaction({ad: Ad_Hash, command: "buy", item: NFT_ID, signature: Voiceprint}) Blockchain->>Blockchain: Execute Smart Contract (Transfer NFT & Funds) Blockchain-->>UserDevice: Return Transaction Confirmation
5. The "Inverse" or Failure Mode
Derivative 5.1: Graceful Degradation to Disambiguation Dialog
Enabling Description: The system is designed to fail safely when pronoun resolution confidence is low. An advertisement shows both a product (a smart watch) and a service (a fitness subscription). The user says, "tell me more about it." The system's model computes a low confidence score for the antecedent of "it" (e.g., 55% product, 45% service). Instead of guessing and providing incorrect information, it enters a "limited functionality" mode. It responds with a clarifying question: "Do you mean the smart watch or the fitness subscription?" The system's state machine transitions from
INTERPRETtoDISAMBIGUATE. This ensures that in ambiguous cases, the system fails to a safe, interactive state rather than executing an incorrect action.Mermaid Diagram:
stateDiagram-v2 state "Resolve Pronoun" as Resolve [*] --> Receiving Receiving --> Resolve: Utterance received Resolve --> Action: Confidence > 90% Resolve --> Disambiguate: Confidence <= 90% Disambiguate --> Resolve: User clarifies Action --> [*] state Disambiguate { [*] --> Ask: "Do you mean the product or the service?" Ask --> UserResponse }
Combination Prior Art with Open-Source Standards
Combination 3.1: W3C SRGS (Speech Recognition Grammar Specification)
- Description: The system of US 9,269,097 is combined with the W3C's SRGS standard. After an advertisement for a product is shown, the system dynamically generates and activates an SRGS XML grammar. This grammar includes a specific rule for the advertised product, such as
<item>Smart Speaker</item>. A generic rule for pronouns like<ruleref uri="#pronoun"/>is linked to this item. When the user says "buy it," the speech recognizer uses this active grammar to constrain the recognition, improving accuracy. The parser's output explicitly tags the pronoun's antecedent based on the grammar structure, simplifying the NLP task to a deterministic lookup rather than a probabilistic inference. This combines the patent's method with a well-established open standard for guided speech recognition.
Combination 3.2: Integration with Hugging Face Transformers
- Description: The pronoun resolution module is implemented using a pre-trained language model from the open-source Hugging Face Transformers library, such as BERT (Bidirectional Encoder Representations from Transformers). The text of the advertisement and the user's subsequent utterance are concatenated into a single input sequence, formatted as:
[CLS] ad text [SEP] utterance text [SEP]. This sequence is fed into the BERT model. The model's attention mechanism is inherently trained to understand relationships between tokens. By fine-tuning the model on a dataset of ad/utterance pairs, the system can learn to resolve the pronoun in the utterance to the correct entity in the ad text with high accuracy, leveraging a powerful, open-source NLP foundation.
Combination 3.3: OAuth 2.0 and OpenID Connect for Action Authorization
- Description: When the user's command, resolved in the context of an advertisement, requires an action on a third-party service (e.g., "book it" for a hotel advertised), the system uses the open standards OAuth 2.0 and OpenID Connect to perform the action securely. The system, acting as an OAuth 2.0 client, requests an access token from the third-party provider (e.g., Marriott's booking API). The user authenticates using OpenID Connect. The resulting action (the booking) is executed using an access token with a narrowly defined scope (e.g.,
create_booking), ensuring the system only has permission to perform the requested action and cannot access other user data. This combines the patented NLP method with open standards for secure, delegated authorization.
Generated 5/10/2026, 12:49:41 AM