Technology behind amplified coding

Context Window Management

The Context Window is a critical component to effectively manage the use of Generative AI models. This consideration becomes particularly significant in software development, as enterprise codebases can be extensive and may rapidly occupy the available Context Window capacity.
The number of tokens a model can process, and the size of its context window vary depending on the specific model. Context windows currently range from 128,000 tokens up to one million, with
advancements occurring rapidly. When choosing a model, it is important to understand its context window to make informed decisions. Also, understanding the context window is important, because the more context is added to it, the slower performance will be.

Tips for Context Window Management

Several strategies can be employed to effectively manage the context window. The following methods have demonstrated strong
results; however, it is important to note that best practices may
evolve over time.

  1. Guide the model in identifying which files are most relevant to the current task
    This approach enables the model to concentrate on the mos relevant files and prevents it from including unrelated files in the current task. Directly referencing files, rather than just their names, guides the model more accurately. Many tools with Generative AI integration in IDEs provide features for referencing files directly using symbols such as @ or /.
  2. Be concise
    Provide only essential information. Excessive context may reduce performance. Clearly guide the model with sufficient, yet not excessive, context to accomplish the task effectively.
  3. Use markdown in your .md files
    This will help the model understand your goals (prompts) better. Also, it is much easier to read and understand for yourself.
  4. Use planning files
    Planning files help not only with providing context to the model, but also with tracking progress. Ask the model to update the planning file for each step. Use a markdown task list ([ ] task) so tasks can be checked off as completed. This allows you to easily resume work if the model gets stuck or you switch contexts.
  5. Start a new chat
    Starting a new chat opens a separate context window. This action can be useful in situations such as when the model is not progressing as expected, when working on a different feature, or implementing changes. If the current chat’s information is no longer necessary, starting a new chat allows work to proceed with a clean slate.
  6. Ask for a summary
    If you’re not using a planning file or have completed tasks in the chat, request a summary of recent actions. You can then copy this summary to a new chat to maintain context.

Memory Management

When working with agents within your IDE, memory management becomes a key aspect to consider. These tools often retain information across sessions by storing important messages
or updates in their so-called memory. This may include design choices, repository-specific details, or naming conventions. While this memory feature can be incredibly helpful, allowing you to avoid repeating context, it’s important to manage it properly. Without active memory management, the AI might rely on outdated or irrelevant information, leading to inconsistent or incorrect outputs. When working with agents within your IDE, memory management becomes a key aspect to consider. These tools often retain information across sessions by storing important messages or updates in their so-called memory. This may include design choices, repository-specific details, or naming conventions. While this memory feature can be incredibly helpful, allowing you to avoid repeating context, it’s important to manage it properly. Without active memory management, the AI might rely on outdated or irrelevant information, leading to inconsistent or incorrect outputs.


Most AI agents offer two types of memory handling:

  • Manual memory updates: You can explicitly tell the agent what to remember.
  • Automatic memory updates: The agent autonomously decides what to store based on what it considers important.

For example, if you’re developing frontend code and the AI observes that you’re using specific button colors, it might remember and reuse that styling. However, if you later change your mind
and choose a different design approach, the AI may continue suggesting the old style, unless you remove or update that memory.
That’s why active memory management is the responsibility of the user. If you change direction or discard earlier choices, make sure to clear or update the AI’s memory accordingly. This ensures the assistant stays aligned with your current goals and avoids introducing outdated suggestions.

Most tools have a dedicated tab or section where memory is stored and can be reviewed, updated, or deleted. It’s worth taking a moment to familiarize yourself with where this is located in your tool of choice. Clear memory visibility makes it easier to keep your AI assistant or agent aligned with your current objectives.

MCP & A2A

Model Context Protocol (MCP) and Agent2Agent (A2A) is a standardized way for Generative AI models and agents to connect with external tools and data sources.
The Model Context Protocol, introduced by AI research company Anthropic, is an open standard for creating secure, two-way links between data sources and AI tools. Developers can use MCP servers to share data or build MCP client applications to connect with these servers.
Agent2Agent, introduced by Google, enables secure and standardized communication between AI agents to share context, delegate tasks, and coordinate actions across complex enterprise workflows.
You can think of MCP as a standardized way to expose functionality to models and agents, similar to how Application Programming Interfaces (APIs) expose functionality to traditional software.
In the past, if you wanted a model to use specific tools or access certain data, you had to manually integrate each API into your application, and also define when and how the model should use it. This process could be repetitive, brittle, and hard to scale.
MCP changes that. Instead of embedding all that logic inside your application or prompt, MCP lets you register available capabilities: like tools, APIs, or functions with an MCP server. The server handles discovery and orchestration.
Now, developers can connect agents or models to this MCP server, and it will automatically know what functionality is available and how to call it, without custom integration each time. It simplifies tool usage, enables reusability, and allows models to dynamically adapt to the tools at their disposal.


Many tools allow integration with others via protocols. For instance, you can link GitHub to your IDE using MCP, giving agents access to repositories and other options. You can also provide up-to-date documentation to ensure models use the latest information rather than potentially outdated training data.

Figure. Model Context Protocol.

In short: MCP decouples the model from the specific tool wiring, much like APIs decouple frontend and backend systems. It provides a shared language and protocol so models and tools can
interact smoothly and at scale.