How can I make user defined parameters required inside a pipeline
If I have a parameter that I defined, how can I make it required like this
HDInsight cluster creation- error on configuration+pricing
Hi, I have been trying to create an HDInsight Cluster from Canada, but it fails at the configuration step. I subscribed for PAYG, I tried by selecting different nodes but none of them is working it gives me the error " You have reached your…
How do I find Hive server information
I'm trying to create a pipeline to copy data from csv to a DataBricks table. To do so, I believe I need to set up a HIVE linked service. However, I'm not sure where I can find the necessary information to fill out the LS form - we had a…
Azure HDInsight Spark job is failing with Logger Error
Hello Team, Our jobs are recently failing with this error- ERROR RawSocketSender [MdsLoggerSenderThread]: org.fluentd.logger.sender.RawSocketSender java.net.SocketException: Broken pipe (Write failed) All these pyspark jobs were running fine…
Real Case Scenarios
Hello, Where can I find case scenarios or real life use cases of for example cloud models or high availability and scalability. What I mean is, for example, hybrid cloud is used by banks because they want to control the database and security. …
How to use UA Managed Identity in Data factory On Demand HD Insight Linked Service
When creating an on-demand HD Insight linked service, there's missing detail for how to configure a User Assigned managed identity instead of a service principal. Steps are shown on how to add a UA managed identity to the Data Factory, but what values…
Spark Dataframe writing issue in azure from spark: One of the request inputs is not valid
I am able to read data from azure blob storage but when writing back to azure storage then it throws below error . I am running this program in my local machine. Can someone help me out on this please. Program val config = new SparkConf(); …
How to Add a subqueue in yarn
I already have queues setup on Yarn on HdInsight, they were setup with the Ambari UI. I have a queue for sqoop that takes up 70% of the cluster. However I have a few huge sqoop jobs with a lot of mappers that take up 100% of the queue and block…
HDInsight HBase vs Databricks
Hi, this is probably answered or perhaps a tall question. What would be the difference/benefits between using HDInsight HBase vs Databricks. Azure storage is definitely one. If the aim is to have the convenience of traditional table & sql with…
To use Azure Data Lake Storage Gen2 with Azure HDInsight clusters, do I have to attach the storage to clusters as linked additional storage?
To use Azure Data Lake Storage Gen2 with Azure HDInsight clusters, do I have to attach the storage to clusters as linked additional storage? Or as long as permission is granted to managed identity, spark scala application could access the storage using…
Creating a HDInsight Spark 4.0 cluster with managed identity and a Data Lake Store gen 2 storage account
Hi! I am trying to create a HDI cluster with an ADLS Gen2 storage account as primary storage account. I have created multiple containers inside my storage account, and I want to limit the access of the managed identity to the other containers. …
HDInsights cluster is in the Error status, even when the user assigned Managed Identity is assigned a role as Storage Blob Data Owner.
HDInsights cluster is in the Error status, even when the user assigned Managed Identity is assigned a role as Storage Blob Data Owner.
HD insight Cluster, Worker node, E32_V3 (256 GB), memory issue
HD insight Cluster - Worker node - E32_V3 (256 GB). It is showing 911 GB memory on ambari portal. MS document for E32_v3 of HD insight worker node showing 1600 GB space. Why there is a discrepancy.
Unable to create HDInsight cluster through free azure subscription
There are not enough cores available to support the selected number of nodes. Please adjust the number of nodes selected, pick a different region, or open a support case to request additional HDInsight cores. You have reached your subscription's…
azure hdinsight There are not enough cores available
I want to create HDInsight in my pay as you go subscription, but I get error: There are not enough cores available to support the selected number of nodes. I checked in my subscription usage and quotas for computing and usage is for every processor…
Files not getting saved in Azure blob using Spark in HDInsights cluster
We've setup HDInsights cluster on Azure with Blob as the storage for Hadoop. We tried uploading files to the Hadoop using hadoop CLI and the files were getting uploaded to the Azure Blob. Command used to upload: Hadoop fs -put somefile…
Connect Synapse Spark Pool with Kafka on HDInsight
I have created a Kafka on HDinsight cluster . I have also created Azure Synapse Analytics - Spark Pool on same region as HDinsight. I need guidance on how to consume topics from Kafka into Spark Structured Streaming. Any documentation or steps will be of…
HDInsight - Kafka - Version 3.2
Hi all Is there a roadmap to release a cluster with a higher kafka version than 2.4.1 in the near future? Thanks for the info in advance. Best reagrads, Michael
Can Azure Streaming Analytics read from Kafka on HDInsight and write to Deltalake table on Synapse lake.
Hello I am looking for guidance on building a new event driven platform. The options we are exploring for processing are - Azure Stream Analytics Apache Spark Structured Streaming in Synapse Source is like going to be Kafka on HDInsight …
what is the best way to copy data from my hadoop on prem cluster to the azure hdinsight cluster?
hi experts, what is the best way to copy data from my hadoop on prem cluster to the azure hdinsight cluster? So we recently deployed a new hdinsight cluster and now I would like to copy some data from my onprem cluster to hdinsight. Thanks,