Generated on: August 22, 2025 Target period: Within the last 24 hours Processing mode: Details Mode Number of updates: 4 items
Published: August 21, 2025 17:00:04 UTC Link: Generally Available: Application Gateway adds MaxSurge support for zero-capacity-impact upgrades
Update ID: 501017 Data source: Azure Updates API
Categories: Launched, Networking, Security, Application Gateway, Features
Summary:
What was updated
Azure Application Gateway now supports MaxSurge during rolling upgrades.
Key changes or new features
MaxSurge enables the provisioning of new gateway instances before decommissioning old ones during version upgrades. This approach ensures zero capacity impact, maintaining full traffic handling capability throughout the upgrade process. It improves upgrade reliability and reduces downtime risks by avoiding any service disruption or performance degradation.
Target audience affected
Developers and IT professionals managing Azure Application Gateway deployments, especially those responsible for maintaining high availability and seamless application delivery during infrastructure updates.
Important notes if any
This feature is generally available and can be leveraged to implement zero-downtime upgrades for Application Gateway. It is particularly beneficial for production environments requiring continuous uptime and consistent capacity during gateway version transitions. Users should review their upgrade strategies to incorporate MaxSurge support for improved operational resilience.
For more details, visit: https://azure.microsoft.com/updates?id=501017
Details:
The recent Azure update announces the general availability of MaxSurge support for Azure Application Gateway, enabling zero-capacity-impact rolling upgrades. This enhancement addresses a critical operational challenge by allowing seamless version transitions without affecting existing gateway capacity or availability.
Background and Purpose
Azure Application Gateway is a web traffic load balancer that manages application-level routing and security. Traditionally, upgrading Application Gateway instances involved sequentially taking existing instances offline to apply updates, which could temporarily reduce capacity and impact application availability. The MaxSurge feature is introduced to eliminate this downtime by provisioning additional instances during upgrades, ensuring continuous traffic handling and maintaining SLA commitments.
Specific Features and Detailed Changes
MaxSurge support allows Azure Application Gateway to temporarily exceed its configured instance count during rolling upgrades. Instead of removing old instances before adding new ones, the system first provisions new instances running the updated version alongside the existing ones. Once the new instances are healthy and operational, the older instances are decommissioned. This process ensures that the total capacity remains stable or even increases during the upgrade window, preventing any capacity degradation.
Key changes include:
Technical Mechanisms and Implementation Methods
The MaxSurge mechanism leverages Azure’s underlying orchestration and scaling infrastructure. During an upgrade, the control plane initiates the creation of additional Application Gateway instances with the new software version. These instances are integrated into the backend pool and undergo health probes to ensure readiness. Traffic routing is gradually shifted to the new instances using Azure’s load balancing algorithms. Only after confirming stable operation are the old instances terminated, maintaining the configured capacity or temporarily exceeding it during the transition.
This approach uses principles similar to Kubernetes’ MaxSurge rolling updates, adapted for the managed Application Gateway service. It requires no user configuration changes; the upgrade process is handled transparently by Azure once MaxSurge support is enabled or default.
Use Cases and Application Scenarios
Important Considerations and Limitations
Integration with Related Azure Services
In summary, the introduction of MaxSurge support for Azure Application Gateway significantly enhances upgrade reliability and availability by enabling additional instances to be provisioned during rolling upgrades. This ensures zero capacity degradation, meeting the high-availability and SLA demands of modern cloud applications while simplifying operational maintenance.
Published: August 21, 2025 16:45:10 UTC Link: Generally, Available: Search Job in Log Analytics - Now Supporting Up to 100 Million Results
Update ID: 500879 Data source: Azure Updates API
Categories: Launched, DevOps, Management and governance, Azure Monitor, Features
Summary:
What was updated
Azure Log Analytics’ Search Job feature is now generally available with enhanced capacity.
Key changes or new features
Search Job enables asynchronous querying across all workspace data, including long-term retention data. The update increases the maximum number of records returned by a Search Job to up to 100 million results. Results are delivered in new Analytics tables, allowing developers and IT professionals to perform deeper analysis and exploration of large datasets efficiently.
Target audience affected
Developers and IT professionals who use Azure Monitor and Log Analytics for large-scale log data analysis, especially those requiring access to extensive historical data and high-volume query results.
Important notes if any
This enhancement supports complex, large-scale investigations and troubleshooting by enabling retrieval and analysis of significantly larger datasets asynchronously. Users should consider query performance and cost implications when working with very large result sets. The feature leverages API access for integration into automated workflows and custom analytics solutions.
Details:
The recent general availability of the Search Job feature in Azure Log Analytics introduces a significant enhancement by supporting up to 100 million results per query, enabling asynchronous, large-scale data exploration across all workspace data, including long-term retention. This update addresses the growing demand for extensive log data analysis in complex IT environments, where traditional synchronous queries and result limits constrained comprehensive investigation and operational insights.
Background and Purpose
Azure Log Analytics is a core component of Azure Monitor, providing powerful querying capabilities over telemetry and log data collected from various Azure resources and on-premises environments. Previously, query results were limited in size and scope, often requiring multiple segmented queries or data exports for large datasets. The Search Job feature was introduced to allow asynchronous execution of queries, enabling users to run extensive searches without timing out or hitting result size caps. The purpose of this update is to scale the maximum number of returned records from a smaller limit to 100 million, thus facilitating deeper and broader data analysis directly within Log Analytics.
Specific Features and Detailed Changes
Technical Mechanisms and Implementation Methods
Search Jobs leverage Azure Monitor’s backend infrastructure to distribute query execution across multiple compute nodes, handling large data volumes efficiently. The asynchronous model uses job IDs to track query progress and retrieve results once available. Results are materialized into internal Analytics tables, which are indexed and optimized for subsequent queries. The system manages resource allocation and throttling to maintain performance and reliability, ensuring that large queries do not degrade service for other users.
Use Cases and Application Scenarios
Important Considerations and Limitations
Integration with Related Azure Services
In summary, the GA release of Search Job in Azure Log Analytics with support for up to 100 million results empowers IT professionals to perform extensive,
Published: August 21, 2025 16:30:27 UTC Link: Generally Available: Azure NetApp Files file access logs
Update ID: 500760 Data source: Azure Updates API
Categories: Launched, Storage, Azure NetApp Files, Features, SDK and Tools, Regions & Datacenters, Services
Summary:
What was updated
Azure NetApp Files has reached general availability for its file access logs feature.
Key changes or new features
The file access logs provide detailed, enterprise-grade visibility into file-level operations on SMB, NFSv4.1, and dual-protocol volumes. This enables tracking of user and application activities such as file reads, writes, and metadata changes. The logs help organizations monitor access patterns, detect anomalies, and support compliance requirements.
Target audience affected
Developers and IT professionals managing Azure NetApp Files deployments, especially those responsible for security, auditing, and compliance in enterprise environments.
Important notes if any
The feature supports all major protocols used in Azure NetApp Files, ensuring comprehensive coverage. Organizations can integrate these logs with their existing security information and event management (SIEM) systems for enhanced monitoring. This update strengthens the security posture by enabling granular access tracking at the file level.
Details:
The Azure NetApp Files file access logs feature has reached general availability, providing enterprise-grade, granular visibility into file-level operations across SMB, NFSv4.1, and dual-protocol volumes. This update addresses the growing need for enhanced security monitoring, compliance auditing, and operational troubleshooting within high-performance file storage environments.
Background and Purpose:
Azure NetApp Files is a high-performance, fully managed file storage service supporting multiple protocols such as SMB and NFS, widely used for enterprise workloads requiring low latency and high throughput. Prior to this update, visibility into file-level access events was limited, constraining organizations’ ability to monitor user activity, detect unauthorized access, and meet compliance requirements. The introduction of file access logs aims to fill this gap by delivering detailed audit trails of file operations, thereby strengthening security posture and operational governance.
Specific Features and Detailed Changes:
The file access logs feature captures detailed metadata about file-level operations including open, read, write, close, and delete events on files and directories. It supports all Azure NetApp Files protocols—SMB, NFSv4.1, and dual-protocol volumes—ensuring comprehensive coverage regardless of client access method. Logs include information such as timestamp, client IP, user identity (where applicable), operation type, file path, and volume details. These logs are generated in near real-time and can be exported to Azure Monitor logs, Event Hubs, or Azure Storage for retention, analysis, and integration with SIEM tools.
Technical Mechanisms and Implementation Methods:
Under the hood, Azure NetApp Files integrates with the underlying storage infrastructure to intercept and record file system operations at the protocol layer. The logging mechanism is designed to minimize performance impact by leveraging asynchronous log processing and efficient data serialization. Administrators enable file access logging at the volume level via the Azure portal, CLI, or ARM templates. Once enabled, logs are streamed to configured destinations using native Azure diagnostic settings. The logs follow a structured JSON schema, facilitating parsing and querying with Kusto Query Language (KQL) in Azure Monitor.
Use Cases and Application Scenarios:
Important Considerations and Limitations:
Integration with Related Azure Services:
File access logs integrate seamlessly with Azure Monitor for centralized log collection, querying, and alerting. Logs can be routed to Azure Sentinel for advanced security analytics and threat detection. Exporting logs to Event Hubs enables integration with third-party SIEM and analytics platforms. Additionally, logs stored in Azure Storage can be archived for long-term retention or compliance auditing. This integration ecosystem empowers organizations to build comprehensive monitoring and security frameworks around Azure NetApp Files.
In summary, the general availability of Azure NetApp Files file access logs equips
Published: August 21, 2025 15:30:02 UTC Link: Generally Available: Azure Functions Flex Consumption plan now supports 512 MB instance size and diagnostic settings
Update ID: 500369 Data source: Azure Updates API
Categories: Launched, Compute, Containers, Internet of Things, Azure Functions, Features
Summary:
What was updated
Azure Functions Flex Consumption plan now generally supports a new 512 MB memory instance size and enhanced diagnostic settings.
Key changes or new features
Developers can now select a smaller 512 MB instance size for Flex Consumption apps, in addition to the existing 2048 MB and 4096 MB options. This allows for finer-grained resource allocation and better cost optimization for lightweight workloads. Additionally, improved diagnostic settings are available to help monitor and troubleshoot function apps more effectively.
Target audience affected
Developers and IT professionals managing Azure Functions apps, especially those focused on cost optimization and performance tuning in serverless environments.
Important notes if any
Choosing the 512 MB instance size is ideal for functions with lower memory demands, enabling more efficient scaling and reduced costs. The enhanced diagnostic settings facilitate better observability, aiding in faster issue resolution and operational insights. Users should evaluate their workload requirements to leverage these new options effectively.
For more details, visit: https://azure.microsoft.com/updates?id=500369
Details:
The recent Azure Functions update introduces general availability support for a 512 MB memory instance size within the Flex Consumption plan, complementing the existing 2048 MB and 4096 MB options, alongside enhanced diagnostic settings capabilities. This update is designed to provide finer-grained resource allocation and improved observability for serverless applications, enabling IT professionals to optimize costs and performance more effectively.
Background and Purpose:
Azure Functions’ Flex Consumption plan offers a serverless hosting model that automatically scales compute resources based on demand, charging users only for the resources consumed. Previously, memory instance sizes were limited to 2 GB and 4 GB, which could lead to over-provisioning for lightweight workloads. The introduction of a 512 MB instance size addresses this inefficiency by allowing smaller, less resource-intensive functions to run on appropriately sized instances, reducing costs and resource waste. Additionally, enhanced diagnostic settings improve monitoring and troubleshooting capabilities, critical for maintaining reliability in production environments.
Specific Features and Detailed Changes:
Technical Mechanisms and Implementation Methods:
The Flex Consumption plan dynamically provisions containerized function instances with specified memory sizes. By supporting a 512 MB option, the underlying infrastructure now allocates smaller container sizes with reduced memory limits, which impacts the function app’s runtime environment and resource quotas. Diagnostic settings integration leverages Azure Monitor’s diagnostic pipeline, enabling users to route telemetry data from the function app to various Azure monitoring and analytics services through standardized diagnostic settings configurations in the Azure portal, CLI, or ARM templates.
Use Cases and Application Scenarios:
Important Considerations and Limitations:
Integration with Related Azure Services:
In summary, the addition of a 512 MB instance size in the Azure Functions Flex Consumption plan, coupled with enhanced diagnostic settings, empowers IT professionals to optimize serverless workloads for cost and performance while gaining improved observability. This update is particularly valuable for lightweight, high-concurrency applications and production environments requiring detailed monitoring, enabling more efficient and manageable serverless architectures
This report was automatically generated - 2025-08-22 03:02:30 UTC