Generative AI Technology Stack Overview – A Comprehensive Guide

Generative AI Technology Stack Overview.

You can translate the content of this page by selecting a language in the select box.

Pass the 2024 AWS Cloud Practitioner CCP CLF-C02 Certification with flying colors Ace the 2024 AWS Solutions Architect Associate SAA-C03 Exam with Confidence

Generative AI Technology Stack Overview – A Comprehensive Guide.

Generative AI (GenAI) is much more than just Large Language Models (LLMs) – it’s an intricate combination of engineering, science, and the business application at hand. Understanding the technology stack behind GenAI solutions is essential because it provides a comprehensive blueprint for building and deploying these powerful AI solutions effectively. The GenAI stack is made up of multiple interrelated layers, each contributing a crucial aspect of functionality, from foundational infrastructure to the final user-facing interface. This one-page guide provides a high-level overview of the technology stack needed to create a production-ready GenAI application.

Listen as a podcast at https://podcasts.apple.com/ca/podcast/generative-ai-technology-stack-overview-generative/id1684415169?i=1000677220601

Generative AI Technology Stack Overview
Generative AI Technology Stack Overview
Generative AI Tech Stack

Layers of the GenAI Technology Stack

The GenAI tech stack can be visualized as a multi-layered structure, each layer serving a unique purpose in the lifecycle of an AI application:

1. Infrastructure

At the base, we have the underlying infrastructure. This layer involves the hardware and cloud services that provide the computational resources needed for AI. Examples include:

  • NVIDIA: Provides the high-performance GPUs required for model training and inference.
  • Cloud Platforms: Platforms like AWS, Google Cloud, Azure, and Together.ai offer scalable infrastructure, providing compute and storage for large-scale AI projects.

2. Foundation Models

Foundation models are pre-trained, large-scale models that provide the base for building specific applications.

  • Examples include models from OpenAI, Anthropic, Cohere, Meta (Mistral), Gemini, and LLaMA. These models can be fine-tuned or used as-is to handle a wide variety of tasks such as text generation, summarization, and more.

3. Retrieval Layer

This layer is crucial for providing efficient and effective access to relevant information. Retrieval can involve several types of data storage and querying mechanisms.

Master AI Machine Learning PRO
Elevate Your Career with AI & Machine Learning For Dummies PRO
Ready to accelerate your career in the fast-growing fields of AI and machine learning? Our app offers user-friendly tutorials and interactive exercises designed to boost your skills and make you stand out to employers. Whether you're aiming for a promotion or searching for a better job, AI & Machine Learning For Dummies PRO is your gateway to success. Start mastering the technologies shaping the future—download now and take the next step in your professional journey!

Download on the App Store

Master AI Machine Learning PRO
Elevate Your Career with AI & Machine Learning For Dummies PRO
Ready to accelerate your career in the fast-growing fields of AI and machine learning? Our app offers user-friendly tutorials and interactive exercises designed to boost your skills and make you stand out to employers. Whether you're aiming for a promotion or searching for a better job, AI & Machine Learning For Dummies PRO is your gateway to success. Start mastering the technologies shaping the future—download now and take the next step in your professional journey!

Download on the App Store

Download the AI & Machine Learning For Dummies PRO App:
iOS - Android
Our AI and Machine Learning For Dummies PRO App can help you Ace the following AI and Machine Learning certifications:



Download the AI & Machine Learning For Dummies PRO App:
iOS - Android
Our AI and Machine Learning For Dummies PRO App can help you Ace the following AI and Machine Learning certifications:

  • Vector Databases: Databases like Pinecone, Weaviate, Qdrant, SingleStore, and Chroma store high-dimensional data representations (embeddings) and allow for efficient similarity search, which is essential for many GenAI use cases.
  • Retrieval approaches can also involve graph databases, keyword-based search, and more, depending on the complexity of the data relationships and querying needs.

4. Runtime/Framework

The frameworks and runtime environments are responsible for orchestrating how the models interact with data, perform inference, and communicate with other components.

  • LangChain: This is a prominent framework that provides useful abstractions for connecting language models with external tools and managing different steps in conversational AI workflows.
  • LlamaIndex and Replicate: Frameworks that are used for indexing and model serving.
  • HuggingFace: Offers a large library of models and tools for deployment, training, and inference, making it ideal for simplifying GenAI workflows.

5. Monitoring and Orchestration

A crucial layer often overlooked, monitoring and orchestration ensure that the models are functioning correctly, performance remains optimal, and the system can handle any issues that arise.

  • This might involve Kubernetes for container orchestration, Prometheus for monitoring, or other specialized tools that keep track of model performance, infrastructure health, and scalability.

6. Frontend Hosting

To make the AI application accessible to users, you need hosting solutions that deliver the frontend interface. While there may be alternative focus areas such as orchestration, frontend hosting plays a vital role in user experience.

  • Platforms like Vercel, Netlify, and GitHub Pages are popular choices for deploying lightweight web-based interfaces that interact with the AI models.

Generative AI (GenAI) Frameworks Overview

Generative AI Technology Stack Overview
Generative AI Technology Stack Overview
Gen AI Framework Overview

The GenAI frameworks provide a diverse set of tools to build advanced AI applications, each with its own strengths and focus areas:

  • LangChain: Excels in creating complex chains of operations, providing diverse integrations and a flexible architecture for language models. It is ideal for building versatile language model applications.
  • LlamaIndex: Specializes in data indexing, efficiently handling structured data, and optimizing queries for large-scale information retrieval. It is particularly suited for data-intensive tasks.
  • Haystack: Known for its robust question-answering capabilities, document search functionality, and production-ready features. It is highly effective for building production-ready search and QA systems.
  • Microsoft Jarvis: Focuses on conversational AI and task automation, seamlessly integrating into the Microsoft ecosystem. It is a strong choice for Microsoft-centric AI solutions.
  • Amazon Bedrock: Provides a comprehensive platform for generative AI, offering deep integration with AWS services and sophisticated model management tools, making it ideal for AWS-integrated generative AI applications.
  • MeshTensorflow: Stands out for its distributed training capabilities, enabling model parallelism and optimizations for Tensor Processing Units (TPUs). It is perfect for high-performance, distributed model training.
  • OpenAI Swarm: Recently introduced and still in the experimental phase, Swarm provides developers with a blueprint for creating interconnected AI networks capable of communicating, collaborating, and tackling complex tasks autonomously. It represents a significant step in making multi-agent systems more accessible to developers.

Each framework has unique strengths:

  • LangChain for versatile language model applications.
  • LlamaIndex for data-intensive tasks.
  • Haystack for production-ready search and QA systems.
  • Microsoft Jarvis for Microsoft-centric AI solutions.
  • Amazon Bedrock for AWS-integrated generative AI.
  • MeshTensorflow for high-performance, distributed model training.
  • OpenAI Swarm for experimental multi-agent systems.

Developers can choose the most suitable framework based on their specific project requirements, infrastructure preferences, and the desired balance between flexibility, performance, and ease of integration.

Why Mastering This Stack Matters

For AI/ML/Data engineers, it’s important to understand not only each layer in isolation but how these layers interact as a cohesive whole. The flow of data across the layers, potential bottlenecks, and optimization strategies are all part of building robust, efficient, and scalable AI solutions. By mastering the GenAI tech stack:

  • Optimized Performance: Engineers can optimize for faster inference, better data management, and improved scalability.
  • Scalable Solutions: The knowledge of each layer’s strengths allows for architecting applications that are scalable and maintainable.
  • Effective Troubleshooting: Understanding the stack enables efficient troubleshooting across all layers, whether the issue lies in data retrieval, model performance, or frontend integration.

Whether you’re building a simple chatbot or a more complex AI system, knowledge of this layered architecture helps create robust and maintainable AI solutions. This understanding is key as GenAI becomes more integrated into business processes.

Genefative AI Tech Stack Implementation

1. Google Cloud Implementation

Google Cloud offers a variety of tools and services that can help you implement the Generative AI technology stack:

  • Infrastructure: Use Google Cloud Compute Engine or Google Kubernetes Engine (GKE) for scalable infrastructure, combined with TPUs for accelerated machine learning tasks.
  • Foundation Models: Leverage Vertex AI to access pre-trained models or fine-tune models using Google’s AI platform.
  • Retrieval Layer: Utilize Cloud Bigtable or Firestore for structured data, and Google Cloud Storage for large datasets and embeddings.
  • Runtime/Framework: Integrate with frameworks like TensorFlow and HuggingFace Transformers, which can be deployed using Google AI services.
  • Monitoring and Orchestration: Use Google Cloud Monitoring and Cloud Logging to manage performance, combined with Google Kubernetes Engine for orchestration.
  • Frontend Hosting: Deploy user-facing applications using Firebase Hosting or Google App Engine.

2. AWS Implementation

Generative AI Technology Stack Overview
Generative AI Technology Stack Overview

Amazon Web Services (AWS) provides a robust ecosystem to support each layer of the Generative AI stack:

  • Infrastructure: Utilize EC2 instances with GPU capabilities or SageMaker for scalable compute resources.
  • Foundation Models: Use Amazon SageMaker to train and deploy models, or access pre-trained models available through AWS.
  • Retrieval Layer: Implement Amazon DynamoDB for fast access to structured data and Amazon OpenSearch for searching across large datasets.
  • Runtime/Framework: Integrate HuggingFace on AWS, with Amazon SageMaker to manage model training and inference workflows.
  • Monitoring and Orchestration: Use CloudWatch for monitoring and logging, and AWS Fargate for orchestrating containerized workloads.
  • Frontend Hosting: Host applications with Amazon S3 and use CloudFront for content delivery.

3. Azure Implementation

Master AI Machine Learning PRO
Elevate Your Career with AI & Machine Learning For Dummies PRO
Ready to accelerate your career in the fast-growing fields of AI and machine learning? Our app offers user-friendly tutorials and interactive exercises designed to boost your skills and make you stand out to employers. Whether you're aiming for a promotion or searching for a better job, AI & Machine Learning For Dummies PRO is your gateway to success. Start mastering the technologies shaping the future—download now and take the next step in your professional journey!

Download on the App Store

Download the AI & Machine Learning For Dummies PRO App:
iOS - Android
Our AI and Machine Learning For Dummies PRO App can help you Ace the following AI and Machine Learning certifications:

Microsoft Azure provides an extensive set of tools to implement the GenAI technology stack effectively:

  • Infrastructure: Use Azure Virtual Machines or Azure Kubernetes Service (AKS) for scalable compute resources, and leverage Azure ML for optimized AI workflows.
  • Foundation Models: Utilize Azure OpenAI Service to access pre-trained language models and build customized AI solutions.
  • Retrieval Layer: Use Azure Cosmos DB for high-performance access to structured data and Azure Blob Storage for large datasets.
  • Runtime/Framework: Integrate frameworks like PyTorch and TensorFlow, and use Azure ML to deploy and manage these models.
  • Monitoring and Orchestration: Use Azure Monitor for monitoring, Log Analytics for insights, and Azure Kubernetes Service for orchestration.
  • Frontend Hosting: Host your frontend with Azure App Service or Static Web Apps for a seamless user experience.

Integrating GenAI into Existing IT Infrastructure

Integrating the GenAI tech stack into an organization’s existing IT infrastructure requires strategic adaptation to leverage existing processes and technologies without a complete overhaul. Here are some ways to include GenAI into your current systems:

1. Incremental Adoption

Organizations can begin by adopting components of the GenAI stack incrementally. For example, instead of moving all workloads to cloud infrastructure, businesses can leverage on-premise GPU resources for specific GenAI tasks, using tools like NVIDIA GPUs or hybrid cloud solutions. Gradual integration reduces disruption and allows the organization to adapt at a comfortable pace.

2. Integration with Existing Data Sources

Instead of replacing existing databases, the retrieval layer of GenAI (such as vector databases) can complement traditional systems. Data pipelines can be designed to pass relevant data to vector databases like Pinecone or Qdrant, while still keeping relational data in existing SQL databases. This approach allows you to add GenAI capabilities without dismantling your current data management systems.

3. Leveraging APIs and Middleware

Many GenAI solutions can be integrated into existing workflows using APIs and middleware. For instance, LangChain or HuggingFace models can be deployed through APIs that interact with your current IT systems, providing AI-enhanced capabilities such as customer service chatbots, while retaining all backend systems. Middleware solutions can further ease integration by connecting GenAI runtime with existing tools and applications.

4. Using Existing Monitoring Tools

To ensure smooth operation of GenAI models, existing monitoring tools such as Prometheus, CloudWatch, or Azure Monitor can be extended to monitor AI components. Integrating GenAI with your current monitoring infrastructure allows your operations team to manage these new components without introducing completely new tools.

5. Cloud Hybrid Solutions

GenAI technology can be deployed in a hybrid cloud model, where some components are run on-premises while others are on the cloud. For example, critical workloads that need lower latency or increased data security can be run locally, while more resource-intensive training processes can be carried out in the cloud using services like AWS SageMaker or Google Vertex AI. This allows organizations to enjoy scalability while keeping sensitive processes within their local infrastructure.

6. Containerization and Orchestration

Using containerized deployments with tools like Docker and Kubernetes makes it easy to deploy GenAI models alongside existing applications. This means GenAI models can be packaged as containers and deployed in the same Kubernetes clusters that are already in use by an organization, reducing the need for changes to existing orchestration processes.

7. Training and Upskilling Staff

Integrating GenAI into existing systems often requires new skill sets. Organizations can bridge this gap by upskilling their IT and development teams through training in GenAI frameworks, cloud infrastructure, and ML lifecycle management. This will ensure that current staff are capable of managing and enhancing GenAI solutions without the need to hire new specialized personnel immediately.

Security and Compliance in GenAI

  • Privacy Concerns: Discuss the data privacy issues that arise with large-scale AI applications. Explain strategies such as data anonymization, federated learning, and encryption to ensure compliance with privacy laws like GDPR.
  • Model Security: Add a section explaining how to secure models against adversarial attacks and data poisoning, emphasizing monitoring, audit trails, and differential privacy techniques.
  • Governance: Address regulatory compliance for AI deployments. Describe best practices for model versioning, auditability, and how to adhere to industry standards.

Implementing Generative AI within an organization’s IT infrastructure requires careful consideration of security and compliance. Ensuring that AI models, data, and the broader system remain secure while adhering to regulatory standards is crucial. Below are the key areas of focus for security and compliance:

1. Privacy Concerns and Data Protection

Generative AI solutions often require large datasets that may include sensitive information. To protect user privacy, organizations must implement measures like data anonymization and encryption. Techniques such as Federated Learning allow AI models to be trained on distributed data without sharing sensitive information between parties. Compliance with regulations such as GDPR or CCPA should be a priority.

2. Model Security and Adversarial Defense

AI models can be susceptible to adversarial attacks, where input data is manipulated to mislead the model. Techniques like adversarial training help make models more robust against such attacks. Additionally, implementing access controls and restricting model access to authorized users can mitigate risks of unauthorized use or model theft.

3. Secure Model Deployment

Secure deployment practices are vital to ensuring GenAI models remain protected from vulnerabilities. Using container security measures, such as scanning images for vulnerabilities, and employing tools like Kubernetes Security Policies can add layers of security. Environments should be segmented to isolate model training, testing, and deployment stages, minimizing the risk of cross-environment contamination.

4. Data Governance and Compliance Monitoring

Compliance monitoring involves continuously checking that AI practices adhere to relevant standards and regulations. This includes maintaining audit trails for data usage and model decisions. Organizations can use tools like Azure PolicyAWS Config, or Google Cloud’s Security Command Center to ensure continuous compliance. Proper data governance also requires documenting the data’s origin, usage, and handling policies.

5. Bias Detection and Mitigation

AI models can inadvertently perpetuate biases present in the training data, leading to unfair or unethical outcomes. Techniques for bias detection and bias mitigation, such as reweighting data samples or using fairness-aware model training, are critical to ensure ethical AI. Regular audits of training data and model outputs can help identify and address bias before deployment.

6. Explainability and Transparency

In many industries, regulations require that AI decisions be explainable. Implementing tools like SHAP (SHapley Additive exPlanations) or LIME (Local Interpretable Model-agnostic Explanations) can help provide insights into how a model arrives at its conclusions. This not only aids in regulatory compliance but also builds user trust in AI solutions.

7. Regulatory Compliance and Best Practices

Different industries have varying requirements for compliance when it comes to AI. For example, healthcare must comply with HIPAA, while financial services need to adhere to standards like SOX or PCI-DSS. Following NIST guidelines for AI security and ensuring adherence to industry-specific regulations are essential to deploying GenAI responsibly and legally.

Optimizing GenAI Stack for Cost Efficiency

  • Cloud Cost Management: Provide strategies for reducing cloud costs when using computationally expensive models, such as serverless deployments, spot instances, and cost monitoring tools.
  • Model Optimization Techniques: Discuss model pruning, quantization, and distillation to reduce model complexity, which in turn lowers computational requirements and costs.

Implementing a Generative AI solution can be expensive due to its computational and storage demands. However, there are strategies to optimize the cost of building and running a GenAI stack without compromising performance. Below are the main approaches to optimize GenAI for cost efficiency:

1. Cloud Cost Management

To optimize cloud-related expenses, it’s essential to leverage cost management tools provided by cloud vendors:

  • Spot Instances and Reserved Instances: AWS, Azure, and Google Cloud offer discounted pricing for long-term or flexible compute instances. Spot instances are great for non-critical batch jobs, while reserved instances can cut costs significantly for long-term workloads.
  • Auto-Scaling and Right-Sizing: Use auto-scaling to automatically adjust resources based on workload demand, which ensures that you are not paying for unused resources. Right-sizing tools offered by cloud vendors can help determine the appropriate instance types.
  • Cost Monitoring and Alerts: Use tools like Google Cloud’s Cost ManagementAWS Cost Explorer, and Azure Cost Management to track expenses and set alerts when costs exceed budget limits.

2. Model Optimization Techniques

Optimizing the models themselves can significantly reduce computational requirements and, therefore, costs:

  • Model Pruning: Remove redundant parameters in a model, which reduces the model’s size and inference time without compromising accuracy.
  • Quantization: Convert the weights of the model from 32-bit to 16-bit or 8-bit precision. This technique decreases memory usage and speeds up computation, leading to lower cloud costs.
  • Knowledge Distillation: Train smaller “student” models to replicate the behavior of larger, complex “teacher” models. The resulting smaller models are cheaper to run while maintaining good performance.

3. Leveraging Serverless Architectures

Adopting serverless solutions can help reduce costs by eliminating the need to manage dedicated servers:

  • Serverless Inference: Platforms like AWS LambdaGoogle Cloud Functions, or Azure Functions can be used to execute inference requests on-demand, which is ideal for workloads that do not require constant uptime.
  • Containerized Serverless: Use tools like Google Cloud Run or AWS Fargate to manage containerized applications without provisioning infrastructure manually, thus avoiding costs related to idle servers.

4. Hybrid Cloud Solutions

Hybrid cloud models help optimize costs by using both on-premises and cloud infrastructure:

  • On-Premises for Inference: If an organization has existing GPU infrastructure, inference tasks can be run on-premises, while more resource-heavy training is performed in the cloud, balancing cost and scalability.
  • Cloud Bursting: During peak demand, workloads can burst to the cloud, allowing organizations to manage costs by only using cloud resources when necessary.

5. Efficient Data Management

Data storage and retrieval are often significant cost drivers in GenAI implementations:

  • Data Tiering: Use different storage tiers for different types of data. For example, frequently accessed data can be stored in high-performance storage, while archival data can be stored in cheaper, long-term storage such as Amazon S3 Glacier.
  • Data Preprocessing: Reduce data size before feeding it into models. Removing unnecessary features, reducing sampling rates, and compressing data can help minimize both storage and computation costs.

6. Using Open-Source Tools

Utilizing open-source tools and frameworks can help avoid the licensing costs associated with proprietary software:

  • TensorFlow, PyTorch, and HuggingFace: These frameworks are open-source and can be run on on-premises or cloud infrastructure without licensing fees.
  • ONNX Runtime: Use ONNX for deploying models across different platforms efficiently. The runtime is optimized for inference, often reducing the cost of operations.

7. Monitoring and Reducing Idle Resources

  • Idle Resource Management: Implement scripts to automatically deallocate unused resources. These can be integrated using cloud-native automation tools like AWS Lambda or Azure Automation to periodically check and terminate idle instances.
  • Scheduling Workloads: Schedule model training and data processing jobs during off-peak hours to take advantage of lower cloud costs (such as discounts during non-business hours).

8. Caching and Reusability

  • Inference Caching: Cache frequently requested responses for popular inference queries, thus avoiding the need to re-run compute-heavy operations for repeated inputs. This can be implemented using Redis or cloud-native caching services like AWS ElastiCache.
  • Reuse of Pre-Processed Data: Store and reuse processed data, embeddings, or intermediate representations to reduce re-computation costs.

9. Optimizing Batch Sizes and Inference Pipeline

  • Batching Requests: Group inference requests to be processed in a single batch to make better use of compute resources, reducing the per-query cost. Batching can be done using tools like TorchServe or custom queue implementations.
  • Pipeline Optimization: Use model inference pipelines to improve the efficiency of the inference process by sharing computations across similar tasks, reducing redundancy and enhancing throughput.

10. Cost Evaluation Metrics

  • Total Cost of Ownership (TCO): Implement methods to evaluate the TCO of different parts of the GenAI stack. Tools like FinOps can provide insights into where your money is being spent and offer strategies to optimize spending.
  • Model Cost-Benefit Analysis: Regularly assess the cost-benefit of maintaining a large model versus utilizing smaller models or open APIs for specific tasks.

Scalability Strategies for GenAI Solutions

Scalability is a crucial factor for GenAI solutions, as these systems often have to handle large datasets, numerous users, or high volumes of requests. A scalable architecture ensures that performance remains consistent, regardless of workload changes. Below are the primary strategies to achieve scalability in GenAI:

1. Horizontal vs. Vertical Scaling

Scalability can be achieved through both horizontal and vertical scaling:

  • Horizontal Scaling: Involves adding more nodes to your system. For GenAI, this might mean adding more servers to handle model training and inference. Tools like Kubernetes are particularly effective for managing clusters of nodes and distributing workloads efficiently.
  • Vertical Scaling: Involves adding more resources (e.g., CPU, GPU, RAM) to a single server. While this may be appropriate for increasing the capacity of a specific workload, it is often limited by hardware constraints and is less cost-effective than horizontal scaling.

2. Containerization and Orchestration

Using containerization tools and orchestration systems can help achieve scalability while maintaining consistency across environments:

  • Docker: By containerizing GenAI components, you ensure that the system is portable and scalable. Each container can be deployed, replicated, or removed based on demand.
  • Kubernetes: Kubernetes can be used to orchestrate containers, automatically scaling up or down based on workload demands. It also allows for efficient load balancing, ensuring no single node becomes overwhelmed.

3. Load Balancing

To efficiently handle multiple requests, load balancing distributes traffic across multiple instances:

  • Cloud Load Balancers: Services such as AWS Elastic Load BalancerAzure Load Balancer, and Google Cloud Load Balancing can be used to manage incoming traffic and distribute it evenly across multiple nodes.
  • Service Mesh: Using tools like Istio or Linkerd for load balancing within microservices-based architecture helps to optimize internal communications and scale smoothly as the number of services grows.

4. Distributed Model Training

GenAI models are often large, making training computationally intensive. Distributed training helps by splitting the workload across multiple resources:

  • Data Parallelism: The dataset is split across multiple nodes, and each node trains on its portion of data. After each training step, updates are shared and combined.
  • Model Parallelism: The model itself is divided across nodes, with each part of the model being trained separately. Tools like Mesh TensorFlow are helpful in this scenario for enabling large-scale, distributed model training.

5. Caching Mechanisms

Caching frequently used outputs can reduce the need for redundant model inference, helping to scale GenAI systems more effectively:

  • Inference Cache: Use tools like Redis or Memcached to store and quickly serve common model responses, thus reducing the need to run expensive computations repeatedly.
  • Embedding Cache: Store embeddings for frequently queried data to avoid recalculating them, which saves time and compute power.

6. Auto-Scaling

Automatically adjusting compute resources based on demand ensures scalability without manual intervention:

  • Cloud Auto-Scaling: Use services like AWS Auto ScalingGoogle Compute Engine Auto Scaler, or Azure Virtual Machine Scale Sets to adjust resources automatically based on traffic patterns.
  • Node Autoscaling in Kubernetes: Configure Kubernetes clusters to add or remove nodes depending on the workload, which helps maintain efficiency during peak and low demand periods.

7. Data Sharding and Replication

Distributing data effectively across multiple databases is essential for scalability:

  • Data Sharding: Split large datasets across multiple database instances to improve query performance. For GenAI, this ensures that high-dimensional vectors or embeddings can be processed in parallel, improving overall throughput.
  • Replication: Create multiple replicas of databases to handle read-heavy workloads. Using MongoDB Atlas or PostgreSQL replication can ensure data is readily available to multiple users without introducing latency.

8. Content Delivery Network (CDN)

Leveraging CDNs helps reduce latency and improve scalability when serving model outputs, particularly for global audiences:

  • Edge Caching: Use CDNs like CloudflareAkamai, or Amazon CloudFront to cache model responses at edge locations, allowing for faster delivery to end-users.
  • Edge Deployment: Where possible, deploy lightweight versions of models to the edge using tools like AWS Greengrass or Google Anthos to bring AI capabilities closer to the user, reducing latency and improving responsiveness.

9. Queueing and Asynchronous Processing

Asynchronous processing can help handle large volumes of requests without blocking system resources:

  • Message Queues: Use tools like RabbitMQApache Kafka, or Amazon SQS to queue incoming requests. This helps manage spikes in traffic by processing requests asynchronously.
  • Batch Processing: Group requests and process them in batches to utilize resources more efficiently, especially during high-traffic periods.

10. Monitoring for Scalability

Monitoring is crucial to ensure that scalability strategies are working effectively:

  • Metrics Collection: Tools like PrometheusGrafana, or Datadog can be used to track system metrics such as CPU usage, memory consumption, and request rates.
  • Scaling Insights: Use these metrics to understand how workloads change over time and proactively scale resources. Predictive scaling, as offered by services like AWS Auto Scaling, helps anticipate demand and scale accordingly.

By implementing these scalability strategies, organizations can ensure that their GenAI solutions maintain high performance, responsiveness, and reliability, regardless of fluctuating user demands or growing datasets. Scalability is not just about handling more users but about doing so efficiently, without compromising on cost or system stability.

User-Centric Design in GenAI Applications

  • User Experience (UX) Considerations: Discuss how to integrate generative AI capabilities into user-facing applications, emphasizing interface design, chatbot responsiveness, and personalization.
  • Human-in-the-Loop Systems: Highlight how integrating human feedback during model inference can improve system reliability, with specific tools for active learning.

Data Management for GenAI Projects

Effective data management is fundamental to the success of Generative AI projects. Since these projects rely on vast amounts of structured, unstructured, and semi-structured data, managing this data efficiently ensures the quality, scalability, and overall performance of GenAI solutions. Below are the key aspects of data management for GenAI:

1. Data Collection and Ingestion

GenAI requires large volumes of data from diverse sources, and efficient data collection and ingestion strategies are vital:

  • Data Integration Tools: Use tools like Apache NiFiFivetran, or Kafka Connect to collect and integrate data from various sources, including databases, APIs, and external data lakes.
  • Batch and Stream Processing: Utilize batch processing for historical data and stream processing for real-time data ingestion using frameworks like Apache Spark or Apache Flink. This hybrid approach ensures up-to-date and historical data are both available for model training and inference.

2. Data Preprocessing and Cleaning

Data preprocessing is a crucial step to ensure that the quality of input data matches the requirements of the AI models:

  • Data Cleaning: Use tools like OpenRefine or Pandas to remove inconsistencies, correct inaccuracies, and deal with missing values.
  • Normalization and Transformation: Convert raw data into a structured format using techniques like tokenization, scaling, and normalization, ensuring that the data is compatible with GenAI models.
  • Data Augmentation: For scenarios involving limited training data, use augmentation techniques like synonym replacement or oversampling to enrich the dataset, particularly for language and vision models.

3. Data Storage Solutions

Data storage solutions should be chosen based on access frequency, performance requirements, and data type:

  • Data Lakes: Use Amazon S3Azure Data Lake, or Google Cloud Storage for storing raw, unstructured, or semi-structured data, which can be used later for model training.
  • Data Warehouses: Structured data that requires fast querying can be stored in data warehouses like SnowflakeAmazon Redshift, or Google BigQuery.
  • Vector Databases: Use vector databases such as Pinecone or Weaviate for storing embeddings generated by models, facilitating efficient retrieval and similarity search.

4. Data Labeling and Annotation

High-quality labeled data is key to supervised learning, which many GenAI models require:

  • Data Annotation Tools: Utilize tools like LabelboxScale AI, or Amazon SageMaker Ground Truth for annotating data. Annotation may include labeling images, transcribing text, or tagging sentiment, depending on the application.
  • Human-in-the-Loop (HITL): Implement HITL workflows where human annotators can verify model outputs and provide corrections, improving the quality of training data iteratively.

5. Data Versioning and Lineage

Data versioning and lineage tracking help maintain transparency and reproducibility:

  • Data Version Control: Use tools like DVC (Data Version Control) or Delta Lake to track changes to datasets over time, ensuring model training can be reproduced with the exact versions of data.
  • Data Lineage Tracking: Tools like Apache Atlas or Amundsen help track the lifecycle of data, showing where data originates, how it changes, and where it is used within GenAI workflows.

6. Data Governance and Compliance

Ensuring compliance with data privacy regulations is crucial in GenAI projects:

  • Access Controls: Implement strict access controls to sensitive data using IAM (Identity and Access Management) tools, ensuring that only authorized users have access.
  • Data Encryption: Encrypt data both at rest and in transit using services like AWS KMSAzure Key Vault, or Google Cloud KMS to prevent unauthorized access.
  • Compliance Management: Use tools like BigID or OneTrust to ensure data handling practices adhere to privacy regulations such as GDPR or CCPA.

7. Data Pipeline Orchestration

Effective orchestration ensures that data flows smoothly from ingestion to model deployment:

  • Orchestration Tools: Use Apache AirflowPrefect, or Azure Data Factory to schedule and monitor data workflows, ensuring data is available where and when it is needed.
  • Real-Time Data Processing: For real-time GenAI applications, use tools like Apache Kafka or Amazon Kinesis to handle continuous data streams.

8. Data Quality and Monitoring

Maintaining high data quality is crucial for reliable model performance:

  • Data Quality Checks: Implement data validation checks using tools like Great Expectations to catch anomalies or inconsistencies in the data pipeline before they impact model training or inference.
  • Data Drift Monitoring: Use monitoring tools to detect data drift, ensuring that the input data distribution remains consistent over time. Services like Evidently AI or WhyLabs can help identify when retraining is needed.

9. Data Access Patterns and Optimization

Optimizing data access helps reduce latency and improves model performance:

  • Indexing: Create indexes for frequently queried data, especially for vector and graph databases, to speed up retrieval times.
  • Partitioning: Partition large datasets to improve query performance. Tools like Hive Partitioning or BigQuery Partitioned Tables can be used to break data into manageable chunks.

By effectively managing data across its lifecycle—from collection to monitoring—organizations can ensure that their GenAI projects are reliable, scalable, and compliant with regulatory standards. Proper data management not only helps in maintaining model accuracy but also in reducing operational complexities and optimizing resource utilization.

Edge Deployment of GenAI

  • Edge AI Use Cases: Illustrate scenarios where GenAI capabilities could be used on edge devices, such as smart home assistants or industrial IoT applications.
  • Frameworks for Edge Deployment: Tools like TensorFlow Lite or ONNX Runtime that enable running models on edge hardware.

Benchmarking and Performance Metrics

  • Evaluating Model Performance: Discuss important metrics such as latency, throughput, and accuracy in the context of generative AI. Suggest using tools like MLPerf for benchmarking.
  • Monitoring User Experience: Methods for tracking user satisfaction, response times, and how well the AI meets expected outcomes in real applications.

Case Studies and Real-World Applications

  • Industry-Specific Implementations: Provide examples of how different sectors—like healthcare, finance, or entertainment—are utilizing GenAI stacks.
  • Lessons Learned from Existing Implementations: Share learnings from companies that have integrated GenAI into their IT landscape, detailing challenges faced and how they were mitigated.

Collaboration and Multi-Agent Systems

  • Swarm and Multi-Agent Systems: Go deeper into OpenAI Swarm and describe how multiple agents can work in tandem for complex workflows. Highlight the use of Reinforcement Learning for enabling such cooperation.
  • Orchestrating Multi-Agent Workflows: Discuss tools like Ray for distributed training and inference, and how they help in deploying multiple generative agents efficiently.

Ethical Considerations and Responsible AI

  • Bias Detection and Mitigation: Explain how bias can be present in foundation models, and the importance of auditing training data and using bias-mitigation techniques.
  • Transparency and Explainability: Address how to achieve explainability in generative models, which is crucial for user trust and regulatory compliance, using tools like SHAP or LIME.

Notes and Future Directions

This tech stack isn’t a rigid blueprint but rather a point of reference. There are many tools and technologies that could fit into each of these layers, depending on your specific needs and constraints.

Moreover, it’s worth noting the importance of a vector database. Vector databases are particularly suited for GenAI applications, as they can handle complex, high-dimensional data while offering efficient querying and retrieval mechanisms. A prime example is SingleStore, which can handle both vector and traditional relational data efficiently, thus offering a flexible solution for AI applications.

In the future, additional layers like advanced monitoring, security, and specialized orchestration tools might become even more crucial to build production-grade GenAI systems.

NVIDIA Full-Stack Generative AI Software Ecosystem
NVIDIA Full-Stack Generative AI Software Ecosystem
NVIDIA Full-Stack Generative AI Software Ecosystem

Master AI Machine Learning PRO
Elevate Your Career with AI & Machine Learning For Dummies PRO
Ready to accelerate your career in the fast-growing fields of AI and machine learning? Our app offers user-friendly tutorials and interactive exercises designed to boost your skills and make you stand out to employers. Whether you're aiming for a promotion or searching for a better job, AI & Machine Learning For Dummies PRO is your gateway to success. Start mastering the technologies shaping the future—download now and take the next step in your professional journey!

Download on the App Store

Download the AI & Machine Learning For Dummies PRO App:
iOS - Android
Our AI and Machine Learning For Dummies PRO App can help you Ace the following AI and Machine Learning certifications:

💪 AI and Machine Learning For Dummies

Djamgatech has launched a new educational app on the Apple App Store, aimed at simplifying AI and machine learning for beginners.

It is a mobile App that can help anyone Master AI & Machine Learning on the phone!

Download “AI and Machine Learning For Dummies ” FROM APPLE APP STORE and conquer any skill level with interactive quizzes, certification exams, & animated concept maps in:

  • Artificial Intelligence
  • Machine Learning
  • Deep Learning
  • Generative AI
  • LLMs
  • NLP
  • xAI
  • Data Science
  • AI and ML Optimization
  • AI Ethics & Bias ⚖️

& more! ➡️ App Store Link: https://apps.apple.com/ca/app/ai-machine-learning-4-dummies/id1611593573

AI Consultation:

We empower organizations to leverage the transformative power of Artificial Intelligence. Our AI consultancy services are designed to meet the unique needs of industries such as oil and gas, healthcare, education, and finance. We provide customized AI and Machine Learning podcast for your organization, training sessions, ongoing advisory services, and tailored AI solutions that drive innovation, efficiency, and growth.

Contact us here (or email us at info@djamgatech.com) to receive a personalized value proposition.

AI Innovations in November 2024


Azure AI Fundamentals AI-900 Exam Preparation

Azure AI Fundamentals AI-900 Exam Prep PRO

You can translate the content of this page by selecting a language in the select box.

Pass the 2024 AWS Cloud Practitioner CCP CLF-C02 Certification with flying colors Ace the 2024 AWS Solutions Architect Associate SAA-C03 Exam with Confidence

Azure AI Fundamentals AI-900 Exam Preparation: Azure AI 900 is an opportunity to demonstrate knowledge of common ML and AI workloads and how to implement them on Azure. This exam is intended for candidates with both technical and non-technical backgrounds. Data science and software engineering experience are not required; however, some general programming knowledge or experience would be beneficial.

Azure AI Fundamentals can be used to prepare for other Azure role-based certifications like Azure Data Scientist Associate or Azure AI Engineer Associate, but it’s not a prerequisite for any of them.

This Azure AI Fundamentals AI-900 Exam Preparation App provides Basics and Advanced Machine Learning Quizzes and Practice Exams on Azure, Azure Machine Learning Job Interviews Questions and Answers, Machine Learning Cheat Sheets.

Download Azure AI 900 on iOs

Download Azure AI 900 on Windows10/11

Azure AI Fundamentals Exam Prep

Azure AI Fundamentals AI-900 Exam Preparation App Features:

– Azure AI-900 Questions and Detailed Answers and References

Master AI Machine Learning PRO
Elevate Your Career with AI & Machine Learning For Dummies PRO
Ready to accelerate your career in the fast-growing fields of AI and machine learning? Our app offers user-friendly tutorials and interactive exercises designed to boost your skills and make you stand out to employers. Whether you're aiming for a promotion or searching for a better job, AI & Machine Learning For Dummies PRO is your gateway to success. Start mastering the technologies shaping the future—download now and take the next step in your professional journey!

Download on the App Store

Master AI Machine Learning PRO
Elevate Your Career with AI & Machine Learning For Dummies PRO
Ready to accelerate your career in the fast-growing fields of AI and machine learning? Our app offers user-friendly tutorials and interactive exercises designed to boost your skills and make you stand out to employers. Whether you're aiming for a promotion or searching for a better job, AI & Machine Learning For Dummies PRO is your gateway to success. Start mastering the technologies shaping the future—download now and take the next step in your professional journey!

Download on the App Store

Download the AI & Machine Learning For Dummies PRO App:
iOS - Android
Our AI and Machine Learning For Dummies PRO App can help you Ace the following AI and Machine Learning certifications:



Download the AI & Machine Learning For Dummies PRO App:
iOS - Android
Our AI and Machine Learning For Dummies PRO App can help you Ace the following AI and Machine Learning certifications:

– Machine Learning Basics Questions and Answers

– Machine Learning Advanced Questions and Answers

– NLP and Computer Vision Questions and Answers

– Scorecard

– Countdown timer

– Machine Learning Cheat Sheets

– Machine Learning Interview Questions and Answers

– Machine Learning Latest News

Azure AI Fundamentals AI-900 Exam Preparation
Azure AI 900 – Machine Learning

This Azure AI Fundamentals AI-900 Exam Prep App covers:

  • ML implementation and Operations,
  • Describe Artificial Intelligence workloads and considerations,
  • Describe fundamental principles of machine learning on Azure,
  • Describe features of computer vision workloads on Azure,
  • Describe features of Natural Language Processing (NLP) workloads on Azure ,
  • Describe features of conversational AI workloads on Azure,
  • QnA Maker service, Language Understanding service (LUIS), Speech service, Translator Text service, Form Recognizer service, Face service, Custom Vision service, Computer Vision service, facial detection, facial recognition, and facial analysis solutions, optical character recognition solutions, object detection solutions, image classification solutions, azure Machine Learning designer, automated ML UI, conversational AI workloads, anomaly detection workloads, forecasting workloads identify features of anomaly detection work, Kafka, SQl, NoSQL, Python, linear regression, logistic regression, Sampling, dataset, statistical interaction, selection bias, non-Gaussian distribution, bias-variance trade-off, Normal Distribution, correlation and covariance, Point Estimates and Confidence Interval, A/B Testing, p-value, statistical power of sensitivity, over-fitting and under-fitting, regularization, Law of Large Numbers, Confounding Variables, Survivorship Bias, univariate, bivariate and multivariate, Resampling, ROC curve, TF/IDF vectorization, Cluster Sampling, etc.
  • This App can help you:
  • – Identify features of common AI workloads
  • – identify prediction/forecasting workloads
  • – identify features of anomaly detection workloads
  • – identify computer vision workloads
  • – identify natural language processing or knowledge mining workloads
  • – identify conversational AI workloads
  • – Identify guiding principles for responsible AI
  • – describe considerations for fairness in an AI solution
  • – describe considerations for reliability and safety in an AI solution
  • – describe considerations for privacy and security in an AI solution
  • – describe considerations for inclusiveness in an AI solution
  • – describe considerations for transparency in an AI solution
  • – describe considerations for accountability in an AI solution
  • – Identify common types of computer vision solution:
  • – Identify Azure tools and services for computer vision tasks
  • – identify features and uses for key phrase extraction
  • – identify features and uses for entity recognition
  • – identify features and uses for sentiment analysis
  • – identify features and uses for language modeling
  • – identify features and uses for speech recognition and synthesis
  • – identify features and uses for translation
  • – identify capabilities of the Text Analytics service
  • – identify capabilities of the Language Understanding service (LUIS)
  • – etc.

Download Azure AI 900 on iOs

Download Azure AI 900 on Windows10/11

Azure AI Fundamentals Breaking News – Azure AI Fundamentals Certifications Testimonials

  • Question about AAD Windows Login Extension
    by /u/Bbrazyy (Microsoft Azure) on March 27, 2025 at 4:50 pm

    submitted by /u/Bbrazyy [link] [comments]

  • Pooled VM with persistence
    by /u/AutisticToasterBath (Microsoft Azure) on March 27, 2025 at 3:47 pm

    At my job, we have a use case for a pooled Azure VM setup. These VMs will only be used around 10–15 hours per week per user, with about 30 users total. We want them to scale up and down as needed using a host pool. The challenge is that we also need them to be persistent. Users might go a few days without using the VMs which would be shut down during this time. I'm trying to figure out how to combine auto-scaling with persistence. Ideally, we want to keep these as shared VMs because of the nature of the use case. Basically these things need to be met. Cheap as possible. Scale up and down based on usage. Pooled resources Persistence between the VMs for the users. I was looking at Fslogic but not sure if that is the right way. submitted by /u/AutisticToasterBath [link] [comments]

  • Problema com ICMP Ping e Recebimento de Dados no Zabbix (Azure vs Local)
    by /u/avueloofc (Microsoft Azure) on March 27, 2025 at 3:22 pm

    Tenho dois servidores Zabbix configurados de forma idêntica para receber dados dos mesmos dispositivos: Servidor Local (Funcionando): Recebe dados corretamente. Servidor Novo (Azure, IP 10.210.0.14): ICMP Ping falha para destinos externos (ex: 8.8.8.8) com fping ("unreachable"), mas funciona para IPs locais. Não recebe dados dos dispositivos, apesar das configurações serem idênticas ao servidor local. Detalhes Técnicos: Ambiente: Servidor Azure: Ubuntu 20.04, Zabbix 6.0, fping com setcap cap_net_raw+ep. Firewall: UFW desativado, iptables permite ICMP. NSG (Azure): Regras Outbound: Liberado para Any (incluindo ICMP). Regras Inbound: Liberadas para Zabbix (10051/TCP, ICMP temporário). Testes Realizados: Conectividade Básica: ping 8.8.8.8 (como root) → OK. fping 8.8.8.8 (como usuário zabbix) → "unreachable". tcpdump mostra que pacotes ICMP não saem da VM. Comunicação com Dispositivos: Servidor Local: Recebe dados via SNMP/Agentes normalmente. Servidor Azure: Não recebe dados, mesmo com configurações idênticas. Verificações Adicionais: sysctl net.ipv4.icmp_echo_ignore_all = 0 (ICMP liberado). curl google.com → OK (conectividade HTTP funciona). Rotas (ip route show): Gateway padrão (10.210.0.1) configurado. Possíveis Causas: Azure Bloqueando Tráfego: NSG ou Azure Firewall bloqueando ICMP ou tráfego SNMP/Agentes. Problema no Gateway/NAT da Azure. Problemas Específicos do Servidor Azure: Configuração de Rede: IP Público, DNS, rotas. SELinux/AppArmor bloqueando fping ou serviços do Zabbix. Timeout de Conexão: Latência alta entre Azure e dispositivos. Diferenças nas Configurações: Arquivos de Configuração do Zabbix (zabbix_server.conf, zabbix_agentd.conf). Versões de Pacotes (SNMP, Zabbix) diferentes entre os servidores. Perguntas para a Comunidade: Azure + ICMP: Alguém já resolveu um problema de fping retornando "unreachable" na Azure, mesmo com NSG liberado? Há configurações ocultas (ex: Azure Policy, Firewall de Camada 7) que possam bloquear ICMP/SNMP? Comunicação com Dispositivos: Por que o servidor Azure não recebe dados dos dispositivos, mesmo com as mesmas configurações do servidor local? Como debugar tráfego SNMP/Agentes na Azure (ferramentas além do tcpdump)? Alternativas: Existe uma forma de substituir o fping por outro método (ex: tcpping) no Zabbix? Devo verificar logs específicos do Zabbix/Azure para identificar o bloqueio? servidor hospedado na azure nao funciona servidor local funciona normal submitted by /u/avueloofc [link] [comments]

  • Anyone got lab on AZURE AI-102?
    by /u/OkDiscount6012 (Microsoft Azure Certifications) on March 27, 2025 at 3:21 pm

    Anyone that got the chance to get a lab on the exam (Azure AI 102) here? I’m just curious what kind of actions do they require you to do on a lab submitted by /u/OkDiscount6012 [link] [comments]

  • At my whit’s end with Microsoft Support. Azure tenant locked out. Hoping someone here has advice.
    by /u/Key-Level-4072 (Microsoft Azure) on March 27, 2025 at 2:18 pm

    I did a really stupid thing with my Azure tenant. I know I was wrong and I know better. This is 100% a result of my hubris. I am a sole admin of my small Azure Tenant and I cannot login to ANY microsoft cloud services because of a conditional access policy that requires Phishing-Resistant MFA. In short, I was testing out passkeys but then decided I didn’t really want to use it further and so I disabled the requirement. Unfortunately, I didn’t do it right. So now, my CA policy requires admins to use a passkey but they’re not allowed to register them in the tenant. It’s a catch 22. I can login and complete MFA just fine, but then Im greeted with the passkey registration user experience flow which fails 100% of the time. I have tried registering it with Microsoft Authenticator. Ive tried using a Yubikey. Ive tried letting MacOS create it. Ive tried letting Bitwarden create it. All avenues result in “Passkey is not accepted by your organization.” I opened a support case in the last week of January. I knew it would take a while for it to get sorted out. I dont have an EA as this is just a small tenant I use for personal stuff and testing new features before we consider implementing them at work. Support has been a nightmare. First, my case was continuously shuffled back and forth between two teams and it was the same person on each team swearing to god that only the other team could fix it. I have explained very clearly exactly what needs to be done so I can login again. But all they do is reset my MFA causing me to have to re-enroll Microsoft Authenticator again after which I am still greeted with the passkey registration flow which fails exactly as it has every step of the way. I asked for escalation but it has not been escalated. I get that these technicians aren’t gods and they cant just do whatever they want and they also have a mountain of tickets to deal with and I shouldn’t expect them to remember every little detail about my particular case. But they keep just doing the same thing that already doesn’t help and then cycling the whole thing back around again. Ive sent so many screenshots of the whole auth flow and experience from my laptop and from my mobile phone but still nothing. Ive reached out to a local Microsoft MVP on LinkedIn who told me he couldnt help if there wasnt an existing delegated tenant relationship on my tenant. Well, I can’t make one if I can’t login so…yeah. Anyway, Im dealing with the Azure Data Protection team who swears they know how to fix this problem but all they do is reset my MFA enrollment and then promise theyre still working on the issue. There HAS to be some magic word or phrase I can add to the conversation in order to get this ticket actually escalated to someone with the power to help me out here. At this point, the only thing I can think of is to call my bank and put a stop payment in place to Microsoft. Then update my DNS to point my mail to a new mail server and let my tenant die. I have two M365-licensed user accounts in there but only one admin and no break glass account (I know, I KNOW!). My other user, who isnt an admin has no issues whatsoever. I can provision other, unlicensed users, to Entra through my AD Synced Active Directory but have no ability to manage licenses or configuration. Am I totally out of options here without an Enterprise Agreement? Or is there some other method Im ignorant of that will get some results? Is there anyone from Microsoft hanging out in here with advice? Or maybe someone has been in this situation before and can tell me what I should expect? submitted by /u/Key-Level-4072 [link] [comments]

  • Accidently deleted a VM, how do I recover it?
    by /u/NYCboredIT (Microsoft Azure) on March 27, 2025 at 2:12 pm

    Documentation seems all over the place and I'm new to Azure. I was able to create a host pool, VMs were fine, everyone is working, great. A few of the VMs got deleted accidently but were recovered. I can get to the machine if I search for it and connect via Bastion, but the VM does not show up in the Host Pool it was originally created in, and the user cannot connect to it. Is there a way to put it back? Any assistance would be much appreciated, please be gentle, thank you. submitted by /u/NYCboredIT [link] [comments]

  • Automated way to resume a paused Azure SQL Database
    by /u/watchoutfor2nd (Microsoft Azure) on March 27, 2025 at 2:01 pm

    We recently move all non prod Azure SQL Databases to serverless with an autopause. This sounds like it will be great from a cost savings perspective, and in my testing the resume is very quick. Now we're looking for a way to resume the database through CLI or automated means. Specifically our deploy pipelines fail because the DB is not reachable. I asked chatgpt and it initially gave me a wrong answer. It suggested Azure powershell command resume-azsqldatabase which sounds EXACTLY like what I want, but the documentation states that this is designed for data warehouses. A second option it gave was to hit an API, so I'm working toward that now, but does anyone have any other ideas/experiences on how to resume a paused Azure SQL Database? submitted by /u/watchoutfor2nd [link] [comments]

  • Azure VPN issues
    by /u/Jaman34 (Microsoft Azure) on March 27, 2025 at 1:51 pm

    I am running into an issue I have never seen before. I have a tunnel between Azure and a FortiGate. When I send traffic over the tunnel from the FortiGate I get the return traffic back with the same source as I initiated the traffic. For Example: Let's say my FG VLAN is 10.10.1.0 and my Azure is 10.20.2.0 the traffic flow would look like this. Src 10.10.1.2 out tunnel dst 10.20.2.2 from the fortigate Src 10.10.1.2 in tunnel dst 10.20.2.2 is what I get back from Azure. It is like Azure is just looping the traffic back to me, and my FG is dropping it to with the src checks to prevent the loop from happening. submitted by /u/Jaman34 [link] [comments]

  • ***100 coupons for DP-900 Practice Tests: Get Ready to Pass! ***
    by /u/TemperatureMoist7636 (Microsoft Azure Certifications) on March 27, 2025 at 1:18 pm

    https://www.udemy.com/course/dp-900-practice-tests-get-ready-to-pass/?couponCode=6B61AD9DA7BFAD6AE07F By DataLearnHub 25 days Starts 03/22/2025 Expires 04/22/2025 Please leave a review if you are satisfied with the course..... I’d truly appreciate your feedback! ⭐⭐⭐⭐⭐ submitted by /u/TemperatureMoist7636 [link] [comments]

  • Blob move advice
    by /u/ade-reddit (Microsoft Azure) on March 27, 2025 at 1:01 pm

    Hoping someone has some advice on the following object storage change I need to make. I have a blob that I need to move to a different subscription in my tenant. Blob Stats: 21,500,000 items, 20.5 TB, 0 Snapshots, 0 Versions, 0 Deleted blobs. The blob is serving files via a VM in Azure. The VM needs to move as well. There is no other copy of the data in the blob. I understand I need to create a new network in the new subscription. I want to minimize risk. I want to minimize downtime. Rough plan is to create the network ahead of time, shutdown the vm and move it. My questions are all about moving the blob storage. (The VM and blob can move at different times if necessary.) Move related questions: How much risk is involved with a move? From what I understand, a blob move is just a metadata change and there's no actual copy - it's more along the lines of repointing? If something goes wrong with the move, am I at risk of losing my data? How long should I expect the move to take? Alternatives? Would I be better off making a copy of the blob? If so, can this be done while the storage is active, and is there a way to keep it in sync? I assume a copy could take several days so I would need to be able either run an incremental or somehow keep it in sync. Ideally, I want another copy of this data, it just can't live in this subscription. Any other advice would be greatly appreciated! submitted by /u/ade-reddit [link] [comments]

  • We’re unable to validate your phone number
    by /u/hardikrsadhu (Microsoft Azure) on March 27, 2025 at 12:21 pm

    I am trying to create azure account but not been able to any body know how to fix this error try using mobile chrome browser and mac windows chrome firefox tried everything still the same issue i only have two number tried both same result. https://preview.redd.it/v9fl4jvx58re1.png?width=1024&format=png&auto=webp&s=ab5bd6b44bb7c9c6ef79cf1cc6020ef3791ec907 submitted by /u/hardikrsadhu [link] [comments]

  • User Cant open Files on Teams after Re creation in Azure
    by /u/ServiceDesk420 (Microsoft Azure) on March 27, 2025 at 11:26 am

    We have a user in our company who was re-created in Active Directory (for specific reasons). The AD sync to Azure was done, so the user was also newly created in Azure. Since then, the user is unable to open files in Teams. From Microsoft support, we were advised to delete the Azure user and manually sync from AD. After doing this, we discovered that the issue only occurs in chats where the user had previous contact with someone before the re-creation. In chats with users he had never contacted before, the issue does not occur. Has anyone encountered a similar problem? (Clearing the Teams cache on both ends does not help.) submitted by /u/ServiceDesk420 [link] [comments]

  • Azure Synapse Workspaces: Multiple workspaces, one git repo?
    by /u/rr-geil-j (Microsoft Azure) on March 27, 2025 at 11:20 am

    Hi! We're planning to manage multiple Synapse Workspaces with one git repo. We're quite clear that we can control this by having separate collaboration branches, but we're not sure if this is also possible with the publish branch(es). Will this work? To illustrate (all workspaces using the same git repo): Workspace 1: Collab branch ws1 Publish branch ws1_publish Workspace 2: Collab branch ws2 Publish branch ws2_publish Workspace 3: Collab branch ws3 Publish branch ws3_publish ... and so on. submitted by /u/rr-geil-j [link] [comments]

  • Azure + Copilot secure setup
    by /u/rickdeaconx (Microsoft Azure) on March 27, 2025 at 11:01 am

    Man, I’m deep into this Copilot setup on Azure, and I’m thinking on how easilyy it could turn into a data faucet. How did this not get rails built in from the start? What I'm thinking: - it's possible this will index and infer many many files that it shouldnt be dumping to just anybody - access controls built in aren't going to stop it as far as I can tell - there has to be a risk with data leakage unless I'm missing something What access controls do YOU guys have in place and what do you recommend? Are file settings sufficient? Any killer Azure tricks or configs to keep it in check? submitted by /u/rickdeaconx [link] [comments]

  • [Certification Thursday] Recently Certified? Post in here so we can congratulate you!
    by /u/AutoModerator (Microsoft Azure) on March 27, 2025 at 11:00 am

    This is the only thread where you should post news about becoming certified. For everyone else, join us in celebrating the recent certifications!!! submitted by /u/AutoModerator [link] [comments]

  • Azure firewall redeployment
    by /u/Greedy-Bid-9581 (Microsoft Azure) on March 27, 2025 at 9:49 am

    Hi, We need to redeploy our azure firewall to enable bring your own ip. The question is, when i delete the firewall, will the policy go as well, or can i just delete/create a firewall, and then attach the same policy? submitted by /u/Greedy-Bid-9581 [link] [comments]

  • Should I pursue Architecture, Network or Security as next certification after Admin/Developer/DevOps Engineer to prepare for Architecture?
    by /u/UnityOfPurpose (Microsoft Azure Certifications) on March 27, 2025 at 9:46 am

    I'm currently thinking of growing to architecture some day): Is it better to go for Azure Architecture, Azure Security Engineer or Azure Network Engineer or aim for another Microsoft certificate? I've completed the following certificates: Azure DevOps Azure Developer Azure Administrator Azure Fundamentals submitted by /u/UnityOfPurpose [link] [comments]

  • First Azure Machine Learning workspace - whats the right set up?
    by /u/SonicBoom_81 (Microsoft Azure) on March 27, 2025 at 9:44 am

    Hi, I'm a data scientist who has mostly worked on local machines and so am expanding into Azure. I am going to work my way through the Azure Certified Data Scientist Associate. As per the first tutorial, I created a Machine Learning workspace. I left it public both for in and out. What I don't want is for the resource to be used by others and my costs to jump. How should I have set this up properly? Also this says that to use the python SDK that I need to have an existing ML workspace. So I guess I don't shut this workspace down when its not in use? In the distant past of setting up spaces for R Studio for AWS, I had to create the workspace new each time. But surely the packages I would download and want to use would be on my workspace and therefore taking up storage. Appreciate the pointers for these basic questions. submitted by /u/SonicBoom_81 [link] [comments]

  • MFA prompts
    by /u/Material-Durian8216 (Microsoft Azure) on March 27, 2025 at 9:41 am

    We've been noticing something strange with MFA prompts for users with admin roles. When opening Office apps like Outlook for the first time in a day they get MFA requests triggered by "Office UWP PWA". They can close the prompt and continue using Outlook normally without reauthentication. Completely closing and reopening doesn’t seem to prompt them again either. Looking at the Conditional Access logs, it traces back to the "Multifactor authentication for admins accessing Microsoft Admin Portals" policy. The weird part? None of these users are actually accessing any admin portal when the prompt appears. Besides that they get an MFA prompt around 2 PM UTC on a daily base. Also triggered by "Office UWP PWA" I can’t put my finger on why this is happening. I’m not planning to exclude some admin roles from the policy due to that sometimes the users do have to access an admin portal (sharepoint). Has anyone else run into this? Any insights would be appreciated. submitted by /u/Material-Durian8216 [link] [comments]

  • Are there architectural templates for Azure Backup strategy and Disaster Recovery Plan?
    by /u/UnityOfPurpose (Microsoft Azure) on March 27, 2025 at 9:39 am

    I wonder if there are templates for these kind of things, architectural templates for azure backup strategy and disaster recovery plan? That can help/guide me a bit? submitted by /u/UnityOfPurpose [link] [comments]

Download Azure AI 900 on iOs

Download Azure AI 900 on Windows10/11