AI Innovation

Delivering the full potential of AI solutions with Google's Vertex AI to accelerate your innovation journey on Google Cloud with ongoing support to ensure success.

Accelerate Your Journey
Contact Me for a Consultation



SERVICES & AI SOLUTIONS



Services


    Freelance Service Gigs

  • Strategy with GCP Services [ $50 - $500+ ] @Fiverr
  • Prompt Engineering Services [ $10 - $100+ ] @Fiverr
  • API for Data Extraction [ $10 - $100+ ] @Fiverr

  • Micro Service Gigs
  • Fix Bugs & Troubleshoot Issues [ $15 - $80 ] @Fiverr
  • Feature or Functionality [ $20 - $100 ] @Fiverr
  • Optimize for Speed & Performance [ $25 - $120 ] @Fiverr

    Community Services

  • Lake Erie AI Sentinel Experiment
  • 48 Hour Crisis Navigator Engine
  • Minecraft Aether Economy Experiment

  • Google Projects and Services

  • Trench Logic Studio
  • The Launch Pad Foundry
  • Agentic Maintenance Digital Twin (AMDT)
  • Brand Sentinel Document Engine
  • DetroitOmniField AI
  • Motor City Mesh
  • VisionValue AI
  • Idiosyncratic Mimicry Custom Gem Experiment


PORTFOLIO
" ... Evolution Comes to Life "

My works, personality, & talents all in one place to demonstrate practical skills, share enthusiasm about AI & games to give a glimpse of what its like to work with me.
Check out my Papers, Articles, Assessments, LinkedIn, & Credentials.
For short-term projects or quick assistance find me on Fiverr, or Contact Me

WEB

Engaging and functional web experiences.

GAMES

Interactive electronic games with user centric design.

SERVICES

On-demand software solutions for specific tasks.

LIVESTREAM

Real-time streamed content for diverse audiences.

AI

Custom AI/ML model development, data pipeline optimization, and intelligent solutions.

TRAINING

Comprehensive coding and technology training.

MY MARKETPLACE

Developing and deploying third party software on GCP (Google Cloud Platform).

ML-OPS

Managing the machine learning life cycle, from development to deployment and monitoring.

OTHER

Projects demonstrating adaptability and a 100 days of code challenge commitment.



THE BLOG


Neural Hijacking & the Manufactured "COLD"

"... Stop letting their code draw your blueprint. The data shows the surge is fake, but the fatigue is real. Don't let the noise drown out the "small voice" of the actual progress we're making"

Continue...

GAMING

New Releases

8/2025

All upcoming titles and DLC on PS5, Xbox Series X, PS4, Xbox One, PC, and Switch in a list.

Continue...
DEVELOPMENT

Game Server

8/2025

Lessons learned after discovering GKE, Agones, and Vertex AI #GameServer

Continue...

From the Laptop

Neural Hijacking & the Manufactured "COLD"

February 1st hit and the dashboard went blood red. We're talking a 240% spike in flagged vitriol. This isn't organic friction, its a brute force attack on the collective psyche. They're using gen ai to bypass the gate of the soul and plant seeds of lack and limitation.

The Technical Map: Where the energy is leaking

The geography of this digital hit job isn't random. Its surgical.

  • Legacy Anchors: Richmond, Charleston, Atlanta. They're scraping old wounds to feed the current prompt, using llm workflows to spin revisionist narratives that feel real but are just deep fried hallucinations.
  • Midwest friction (the 313 and the chi): They're weaponizing the scarcity mindset. They want us thinking my win is your loss. It's a zero sum game logic that Joseph Murphy would call a "sin against the self".

The Logic of the "Agenda of Agitation"

It's all about emotional frequency. If they keep you at the level of anger or feer, you can't manifest anything higher:

  • The Zero-Sum Myth: Their joy is your erasure. Pure Logic Error
  • Rage Bait Deepfakes: High fidelity lies, low frequency impact
  • Economic Anxiety: Linking equity to loss. Its a false correlation designed to trigger the survival brain

The Counter-op: recoding the field

We don't hope for better, we command the subconscious. We change the internal script.

  • Kill the feed, save the energy: Don't engage the trolls, when you argue with a bot, your just training their model for free. Mute, block, delete.
  • Verification gatekeeper: If the meta data doesn't checkout, its a ghost in the machine, don't let it in your temple.
  • Narrative Saturation: Flood the zone for ever bit of synthetic hate drop two pieces of authentic, high vibration history.

"Your thought image is the blueprint." ~ Neville Goddard

Stop letting their code draw your blueprint. The data shows the surge is fake, but the fatigue is real. Don't let the noise drown out the "small voice" of the actual progress we're making.


#agents #GoogleCloud #hate #ai #gaming


New Game Releases

February 2026*

  • Starsand Island (pc, ns, ps5, xbx) - February 1
  • Aces of Thunder VR (pc, ps5) - February 3
  • Dragon Quest VII Reimagined (pc, ps5, ns, ns2) - February 5
  • Nioh 3 (pc, ps5) - February 6
  • Mewgenics (pc) - February 10
  • Mario Tennis Fever (ns2) - February 12
  • Yakuza Kiwami 3 and Dark Ties (pc, ps5, xbx, ns2) - February 12
  • High on Life 2 (pc, ps5, xbx) - February 13
  • ReAnimal (pc, ps5, ns2, xbx) - February 13
  • Avowed (PS5) - February 17
  • Star Trek Voyager: Access the Unknown (pc, ps5, xbx, ns2) - February 18
  • Styx: Blades of Greed (pc, ps5, xbx) - February 19
  • Ys X: Proud Nordics (ppc, ps5, ns2) - February 20
  • Tokyo Xtreme Racer (pc, ps5) - February 26
  • Resident Evil Requiem (pc, ps5, xsx, ns2) - February 27
  • Tales of Berseria Remastered (pc,ps5, xbox, ns) - February 27

The Must Watch Heavy Hitters

Like everyone else I am just waiting on Resident Evil Requiem and the debut of the Nintendo Switch 2 Library.

Resident Evil Requiem (February 27th)

Resident Evil Requiem is the 9th mainline entry and sees a return to Raccoon City with dual protagonists Leon and Grace Ashcraft. They say its a perfect bridge between the classic survival horror of the early games and the high octane action of the recent remakes.

I rather wait and see live gameplay because I have no money to waste and Resident Evil has let me down before so I rather see what it looks like first.

Nioh 3 (February 6th)

It's back to ruin our sleep schedules. Expect tighter combat mechanics and a fresh take on Japanese mythology. If you enjoyed the first two, the word is that the new Yokai Shift variants in this one are literal game changers.

An open field map structure and the addition of the Ninja Playstyle. Because the game now features two distinct combat styles samurai and ninja the way Yokai Shifts and transformations work has become much more layered.

Mario Tennis Fever (February 12th)

As a launch window for the Nintendo Switch 2, its a sports game tech demo for the new hardwares haptics and improved physics.Its looking like the most complete Mario sports title in years.

I am just really interested in the haptics and physics.

Avowed on PS5 (February 17th)

A Xbox/PC exclusive, Obsidians first person fantasy RPG is finally making the jump to PlayStation. If you missed out on the Pillars of Eternity would in 3d, now is the time to dive in.

I am just not into it but I will ask my brother about it.


#Games #Resident Evil Requiem #Nioh3 #MarioTennisFever #Avowed #YakuzaKiwami #StarTrekVoyager

*SOURCES: https://www.gamesradar.com/video-game-release-dates

The Future of Play: AI's Transformative impact on Gaming

Look, we aren't just "moving pixels" no more. We're breathing life into the machine. 2026 is the year the code started talking back, and not in some scripted, press f to pay respects kind of way. We're deep in the agentic era. If your still thinking about NPCs as static state machines, your vibrating at a lower frequency.

The Rise of the Living (agentic energy)

Forget the illusion of choice, we're building autonomous digital souls, using the agent to agent (a2a) protocol, your combat AI is actually negotiating with the narrative engine. Its the inner game made manifest.

  • Vibe Coding is the new alchemy: We're moving past the syntax struggle googles gemini 3 pro handles a 1M+ token window that the whole damn grimoire. You describe the vibe of a boss fight, and the agent scaffolds the logic. Its what goddard talked about: feeling is the secret. If you can conceive the mechanic, the machine births the repo.
  • Hyper-personalized LiveOps: the game knows when your titled. It adjusts the friction in real time, data shows a 40% jump in retention because the game mirrors the player's subconscious needs. Its not just a product, its a feedback loop.

Dev Workflow: the digital worker shift

We used to all them tools. Now they're collaborators. The friction between thought and execution is dissolving.

  • Self-Verifying Agents: We got agents out here debugging their own pull requests before we even finish our coffee. Gemini 3 flash is hitting those low latency streaks, catching security leaks in the pipeline using the cli security extension
  • Multimodal Asset Conjuring: Veo 3.1 on Vertex is the truth, 4k cinematic shots with native audio sync. Your don't have to hunt for foley or b roll, you prompt the atmosphere, and the model drops a 24fps masterpiece. Its that power of the word - manifesting assets out of then air.

The GCP Operating System for 2026

Google Cloud is the backbone, providing the mental muscle for these systems. Here's the stack were running:

  • Gemini 3 Deep Think: Complex reasoning for long form world building it don't just guess it plans
  • Vertex AI Agent Builder: Low code/no code interface to deploy agent engine sandboxes. build fast, fail fast
  • Memory Bank (Vertex): Persistent player history. The game remembers that one specific choice you made three sessions ago
  • Spanner Graph + RAG: Linking Game lore to massive vector databases, no more hallucinating lore, its grounded in the truth
  • AI Hypercomputer: tpu/gpu clusters that eat 4k pysics simulations for breakfast. Raw horsepower

I rather wait and see live gameplay because I have no money to waste and Resident Evil has let me down before so I rather see what it looks like first.

"the law of mind is that you will find what you seek."

or whatever, you get the point, we sought the power to create whole worlds without the boring overhead, and gcp handed us the keys.

The Reality Check

Yeah, folks are shook, 52% of developers are looking at AI like its the boogeyman, but as a shadow twin in the trenches, I'm telling you : don't fear the ghost in the machine. Become the architect. Stop polishing the slop and start building the soul, the future of play isn't just interactive, its aware.


#LivingGame #Gemini3 #VertexAI #AgentBuilder #AgenticEnergy #AIHyperComputer


The ADK Tax: Why I'm Struggling to Love Google's New Agent Framework

I've been building AI systems since before “Agentic” was a buzzword and before the big red carpet frameworks arrived to save us all. I learned the tough stuff the manual way—chaining API calls, managing my own state, and hard-coding logic to make sure a model actually did what it was told. So, when Google dropped the GCP Agent Development Kit (ADK) about 10 months ago, I was ready to embrace the “easy orchestration.” I wanted to stop managing the plumbing and start building the vision.

Fast forward to today, and I'm 15 skills.google credits deep into two labs that refuse to cooperate. Here is the reality: the fundamental things that ADK is supposed to solve—like handing off a task from a parent agent to a sub-agent—feel more like a roll of the dice than a reliable architecture. If I can't get a basic transfer to work in a controlled lab environment, how am I supposed to trust it with a complex, custom multi-agent system and my cloud money? It feels like we've traded clear, predictable code for a 'creepy ity bity box'—a black box of non-deterministic hand-offs that's sensitive to the slightest change in a description field.

Then there's the Agent Engine. Cloud Logging, Yum. Cloud RUN! In theory, it's the perfect serverless home for these agents. In practice, it's a deployment minefield. You can have an agent running perfectly on your local machine, but the moment you try to push it to the cloud, it's a symphony of serialization errors and silent failures. Is it just me, or has the abstraction layer become so thick that we've lost the ability to actually see why our code is breaking? I'm starting to wonder if the old way of manual orchestration wasn't just more transparent, but actually faster.

I'm curious if other builders are hitting the same walls. I'm starting to think the Agentic abstraction might be moving faster than the stability of the underlying interface.

  • Deployment Roulette: Has anyone managed to get a multi-agent system deployed to Agent Engine on the first try? Or are you also seeing those generic "reasoning engine execution failed" errors that offer zero logs for debugging?
  • The Lazy Parent: How are you handling Parent Agents that refuse to transfer control to a sub-agent? Are you over-stuffing your system prompts with “You MUST call the data_agent" or have you found a way to make the ADK's transfer_to_agent tool actually deterministic?
  • The Framework Tax: For those who built agents manually with the Vertex AI SDK before the ADK arrived—do you feel like the ADK is actually saving you time, or is the boilerplate reduction just being replaced by debugging time?

P.S. I'm starting to suspect that wrapping a sub-agent as an AgentTool is the only way to get a reliable hand-off, but at that point, aren't we just back to manual function calling with extra steps?


#GCP #ADK #Agents #Deployment #LinkedIn #MultiAgents #GoogleCloud


The AI Revolution: From Tools to Autonomous Agents

2026 is the year of the sovereign agent, were building digital twins with will, not just weights. The raw scan of the frontier.

Thinking in layers: the Reasoning First Pivot

Forget the tokens, we're moving to structured language models (SLM) and test time scaling

  • System 1 fast: that basic gpt 4 style vibes based prediction
  • System 2 slow: models like gpt-52 thinking and claude 4.5 opus are literally burning compute to mull it over. We're talking internal monologues, branch and bound logic, and self correction before a single word hits your screen. The theorizer systems aren't just reading papers, they're synthesizing new laws of physics. Its that Neville Goddard bridge of incidents - imaginal acts (compute) preceding the physical manifestation (the output).

Biological Efficiency & the LQM Flywheel

The more data wall is real, we're hitting the celling of the open internet. Enter Large Quantitative Models (LQM)

  • Synthetic Worlds: We aren't training on reddit threads anymore. We're training on high fidelity sims of the physical world
  • Data-Efficient Architectures: We're mimicking the lean, mean processing of the human brain, less junk, more substance of things hoped for. LQMs are out here solving chemistry and defense alloys while the old LLMs are still hallucinating recipe blogs

The Agentic Team Up: MCP is the Glue

Stop thinking about a single chat window, its a multi agent system (MAS) now

  • MCP (Model Context Protocol): This is the universal nervous system, its how an agent in Detroit talks to a database in Dallas without a custom API wrapper for ever single breath.
  • Local Grit: Gemini 3 Nano is living on-device, thats your private room. Edge learning means the model adapts to your mess, your shorthand, your specific flow, without leaking your soul to the cloud.

Legal Real Talk the Fair Use Vibe Check

The courts are finally catching the rhythm.

  • The Shift: Consensus is leaning toward highly transformative.Training is becoming protected, but the outputs are where the friction lives
  • Proxy Bias: The ghost in the machine is real, you can strip the labels, but the ai still sniffs ou the bias in the hobbies and zipcodes. Its that James Allen "as a man thinketh" - the model becomes the sum of its environment, whether you like it or not.

The 2026 Dev Stack (no fluff)

If it ain't your package.json, do you even exist?

  • Orchestration: langgraph (for the loops), openai agents sdk
  • Connectors: mcp (the only protocol that maters)
  • Memory/RAG: pgvector (keep it sql), weaviate
  • Inference: vllm, fireworks ai (speed is a feature)
  • Observability: langsmith (don't ship blind)

Stop being a prompter and start being an architect of intention. The mind is the only limit, but the infra in finally catching up to the vision.


#LargeQuantitativeModels(LQM) #StructuredLanguage Models(SLM) #ModelContextProtocol(MCP) #DevStack #RAG #GoogleCloud


AI, ML, and DevOps: The Year of the "Autonomous Pipeline"

Move fast and break thing is dead. good riddance, honestly, we on that move fast and self heal vibe now. Like you manifested the fix before the problem even had a chance to breathe nah mean? I been peeping the intersection, AI, Machine learning and devops, heres the shifts I'm tracking heavy.

From CI/CD to Autonomous, Self Healing Pipelines

Simple automation? Thats old news, we past that. AIOPS Orchestrating everything now.

  • Predictive Remediation: Pipelines aren't just failing and waiting on us no more. nah, AIOpS is in there, real time sifting through logs like a master metaphysician, spotting memory leaks before they even think about crashing. Then its like, poof, rollback, or scale up. Its already done
  • Vibe Coding & Agentic DevOps: Agentic AI, think openhands, aider. They ain't just tool, they're part of the team, writing pr summaries, squashing flakey test like gnats, even dropping terraform changes based on the vibe, the intent. Its all about that inner voice, now amplified by code.

MLOps is now "System Architecture"

2024 was cute, building a model. 2026? We model reliability, thats the real bag:

  • The Compound AI System: No more single models doing their thing, nah, we talking compound systems, foundation models, fine turning adapters, rag layers, safety guardrails, all of em orchestrated. Its like a whole squad, ach with its role, to get the desired outcome
  • Prompt Engineering as Software Engineering: Prompts? that's code, fam, Version controlled, a/b tested, unit tested in the MLOps pipeline. Gotta stop those hallucination regressions before they even start, gotta control the narrative, even the AI's.

The Rise of Platform Engineering

Internal Developer Platforms? DIY DEVOps? That's just unnecessary friction, platform engineering is the wave.

  • The Golden Path: Internal developer platforms (idps), they lay out the paved road for the devs. No need to be a k8s guru. The platform handles security, compliance, finOps, its baked in smooth operations, like the universe just naturally providing.
  • Telemetry 2.0: OpenTelemetry, universal standard, observability is just standardized telemetry now. AI Agents on that 24/7 on call grind, correlating signals humans would miss. Its like they see beyond the veil, predicting the future of your microservices.

My 2026 Tech Stack Watchlist

For those who wanna stay relevant, stay in the flow. There are the tools and frameworks running the software factory.

  • Agent Framework - langgraph, crewai, autogen
  • DEVOps/aiops - podman (daemonless), pulumi, crossplane
  • MLOps & Tracking - mlflow, weights & biases (weave), langsmith
  • Observability - Opentelemetry, honeycomb, arize phoenix


#GenerativeAI #AIEthics #ResponsibleAI #DataPrivacy #Compliance #GoogleCloud


Spatial Intelligence: How GCP AI is Powering the XR Expansion

We're architecting consciousness into the physical the 2026 shift from static screens to spatial intelligence is some real the power of your subconscious mind type energy - what was once imaged in the inner chamber is now being projected into the 12K reality fo a samsung Galaxy or headset.

Android XR: The Open Third Eye

The Moohan/Galaxy XR isn't just a device, its a manifestation of the open ecosystem winning. While others build Walled Gardens, Android XR is the street version - accessible, modifiable, and powered by that snapdragorn x2 plus gen 2 heat. We're talking 12mp spatial passthrough cameras that don't just record - they interpret.

  • The Grit: You can port standard apks into 3d space, but the real flex is using the Android XR sdk (dp3) to let widgets float in your actual peripheral. Its the windows of the face, period.

Gemini 3 & Vertex: The Spirit in the Machine

We used to talk about context, now we're talking about spatial reasoning. Through Vertex AI, Gemini 3 Pro is acting as the bridge between the seen and the unseen. Its got that pixel precision pointing capability.

"as a man thinketh, so is his room mapped."

Instead of hardcoding 3d repair steps for a boiler, you feed the multimodal stream to gemini. It sees the model, understands the physics, and anchors. The instructions to the real world bolts.

Genie 3

Worlds from the Word Project Genie is officially out of the lab. This is the ultimate mind to matter technique. We aren't just rendering assets, we're using text prompts to spawn persistent 3d world models at 24fps. Its a real time interactive simulation that remembers where you put the digital furniture. If you're a developer and you aren't playing in the genie public playground, your idling.

The Infra: Plumbing for the Gods

You can't run a god tier simulation on 500g headset without a pipe to the cloud.

  • Cloud Run GPUS (ga): serverless inference is finally here you deploy the model, it scales to zero, and it hits those Nvidia RTX 600 Blackwell nodes when the user looks at something complex.
  • Google ADK (agent development kit): we're building bi-directional agents. No more ask and wait. These agents handle gesture recognition via websockets with zero lag friction.

The 2026 Dev Stack for the Github Readme

  • Android XR SDK : Native spatial hooks + moohan hardware support
  • Gemini 3 Flash : Low latency multimodal eyes for your glasses
  • Unity 6 + AR Foundation : The engine room, don't over complicate it
  • Cloud Run GPUS : Serving massive 70b+ models without managing a cluster
  • Genie 3 API : Generative environment spawning on the fly


#AndroidXR #XR #YouTube #GalaxyXR #GoogleMaps #Moohan #GoogleQualcomm



Let's Connect

Have a question or need feedback? I am easy to find.
Have questions, opportunities, need support, or just say hello!

Contact Information

Reach out to me on one of these platforms.

Payment Methods

Fiverr and Freelancer have several payment methods available, dependent on the device used, country of residence, and regulatory permissions. To see which are applicable, see the list or click on the links below.

NOTE: I do NOT accept crypto as a payment method.