Notes on Miscellaneous AI stuff
Can you provide it to me in Markdown with links to corresponding section?
- The Initial Problem: You were deeply involved in reorganizing a massive, highly detailed Markdown document that maps out your custom PC power backup architecture (covering your Dual Parallel UPS strategy, inverter setups, and thermal management).
- The IDE Bottleneck: You attempted to use Gemini Code Assist (GCA) directly inside VS Code to update the document's Table of Contents. However, because you were feeding it such a large amount of text, the VS Code extension choked on the data size, hung up, and you were forced to abort the process.
- The Web Interface Success: You brought that exact same large-document task over to me here on the Gemini Web interface. I was able to process the entire document and output the correctly formatted Markdown ToC almost instantly.
- The "Why" Behind the Tech: This stark contrast in performance prompted you to ask me why GCA struggled so much while I handled it effortlessly, and you questioned whether Google was phasing out GCA. I explained that GCA is still fully supported but is fundamentally bottlenecked by the VS Code Extension Host when it tries to process and render massive text blocks. In contrast, this web interface is an unencumbered pipeline built specifically for raw throughput and heavy lifting.
- Your Workflow Epiphany: You correctly concluded that you need to adopt a "right tool for the right job" approach: keep using GCA for surgical, line-by-line coding and local context within your workspace, but bring massive document rewrites or large data dumps over to the Web interface.
- The Final Task: Recognizing how valuable this operational insight is, you asked me to summarize this technical reality so you could publish it on your blog. Your goal was to document this boundary between the two tools, ensuring you—and others—wouldn't waste time watching an IDE freeze over a task it was never built to handle.
You are correct: both Gemini Code Assist (GCA) and GitHub Copilot are remarkably "chatty" on the network. They don't just send a few bytes of text; they are constantly streaming data back and forth to maintain the "magic" of real-time suggestions.
Why the Data Consumption is So High
It isn't just the final suggestion you see on the screen that consumes data. The background activity is much more intensive:
- Continuous Context Uploading: For a tool like Gemini to give a smart suggestion, it needs to know what you’ve already written. Every few keystrokes or every time you pause, the extension may upload hundreds of lines of "surrounding context" from your current file (and sometimes open tabs) to the cloud model so it understands the logic.
- Frequent "Ghost" Requests: These extensions often trigger requests even if they don't end up showing you a suggestion. If you type a space or a common keyword, a request is fired off. On a fast 4G connection, these hundreds of tiny JSON payloads add up to megabytes very quickly.
- Telemetry and Syncing: Beyond code completion, these extensions are constantly sending telemetry (usage data) and syncing your preferences or indexed "workspace symbols" to improve their internal logic.
- Model "Polling": Some versions of these tools use streaming responses (where the code appears character by character). This keeps a TCP/HTTP connection open and active, which creates more overhead than a simple one-off request.
My View on Your Situation
Given that you are a retired consultant and a spiritual aspirant often working from Puttaparthi (where network stability or data caps might be more of a concern than in a high-speed corporate office), these AI tools can be a "data vampire."
- The "Tax" on 4G: If you are using a mobile hotspot, the overhead of the SSL/TLS handshakes for every completion request is significant.
- Productivity vs. Cost: For a physicist and software veteran like yourself, the boilerplate code these tools provide might not always be worth the massive data drain, especially if you already have a clear mental model of your React Native or PowerShell logic.
- The "Search" Alternative: You might find it more data-efficient to keep these "auto-completion" features off and only use the Chat feature (Ctrl+Shift+I or the sidebar) when you are actually stuck. A single chat query uses far less data than 500 auto-complete checks performed while you're just typing a simple
forloop.
Recommendation
If you want the benefits of AI without the 4G bill:
- Keep Auto-Complete OFF: As you've already done.
- Use Manual Triggers: In VS Code settings, you can often set these extensions to only provide suggestions when you manually trigger them (e.g., pressing
Alt + \orCtrl + Space) rather than "as you type." - Local Alternatives: If you ever find this data usage becomes a dealbreaker, you might look into "Local LLMs" (like Ollama with the Continue.dev extension) which run entirely on your machine's hardware and use zero network data.
Do you find that the "ghost text" suggestions actually speed up your React Native work, or do you prefer writing the logic yourself and just using AI for debugging?
Comments
Post a Comment