what is rows affected in output and if we write insert statement in script task and if data is there in rows affected, can we consider it is the data inserted into the table ?
@SouravMondal-hb8rzАй бұрын
Can we get the Private Ip of that ADF ( IR) so that we can whitelist that ip
@rahulpanda92563 ай бұрын
Hi Thanks a lot for this explanation. Does this also handle delete operations at the source?
@akshaygidwani92194 күн бұрын
It would not be able to handle delete operation
@kapsingla3 ай бұрын
When you created this setup did you create the publish Branch while configuring the repo in Azure data Factory?
@AshokGupta3 ай бұрын
Adf automatically creates publish branch
@adityavanipenta146 ай бұрын
Hi I need to create a delta table in databricks with scd2 logic
@adityavanipenta146 ай бұрын
Can you help I have the requirements with me
@shanepeck238710 ай бұрын
"promo sm"
@AshokGupta10 ай бұрын
?
@VavilalaPavan10 ай бұрын
voice is first not at all good
@AshokGupta10 ай бұрын
Please watch our other videos we have improved
@AshokGupta10 ай бұрын
Please watch our other videos we have improved
@pavankumarveesam8412 Жыл бұрын
sound is pretty less
@pradeepthanniru2602 Жыл бұрын
The content is good and nice explanation
@AshokGupta Жыл бұрын
Thanks a lot
@PalaniKumar-b7e Жыл бұрын
How to create multiple clusters with same scripts? Also how can we create workflow job clusters in azure pipeline? Please help.
@AshokGupta Жыл бұрын
You can duplicate create cluster script as many times you want.
@tanushreenagar3116 Жыл бұрын
Nice
@paulinexiong3246 Жыл бұрын
for sql database as source, I believe you need to enable CDC on the source database/tables
@guptaashok121 Жыл бұрын
Not required.. it will work based on a timestpam column of table
@arulpalaniappan Жыл бұрын
This is really helpful! Thanks
@swatidorge7133 Жыл бұрын
Well done 👍👍👍👍
@swatidorge7133 Жыл бұрын
Well done 👍👍👍👍
@syamamujuru Жыл бұрын
In Question 36 They mentioned development mode, so Cluster will not shutdown.
@GrowthMindset_J Жыл бұрын
Do you have another set of Databricks practice questions ?
@AshokGupta Жыл бұрын
kzbin.info/www/bejne/haW3kpiKf7mrpdU
@bairagirout9323 Жыл бұрын
In the last step send email, what should i mention in the email body to get the content of the csv in a HTML table. Please explain the send email step also
@coolshrawan3 ай бұрын
just call the output of compose. outputs('Compose_3')
@V-Barah Жыл бұрын
it good video, but you didn't show the required configurations: How to get the token, Cluster ID , and notebook path and all details . Someone new to this area will not understand
@mikebauer9335 Жыл бұрын
"PromoSM"
@tsdhd6275 Жыл бұрын
content is good but voice is so f****d up
@mannykhan7752 Жыл бұрын
Some of the questions here have wrong answers. In the case of Q15 the answer is E as the table has to be updated before the next ingestion. This question also appears on the Databricks Data Engineer official practice test that can be downloaded from their website. That's why I'm sure your answer is incorrect.
@AshokGupta Жыл бұрын
This set of question and answer are from databricks site only. However I agree some answer might be wrong that time. Now they might have corrected it.
@prasadk6507 Жыл бұрын
Nice one, keep updating fraternity 🎉
@ravimishra6792 Жыл бұрын
Very helpful
@AshokGupta Жыл бұрын
Thanks
@Bgmifortimepass Жыл бұрын
what is our source like
@AshokGupta Жыл бұрын
Source can be anything, first it needs to brought to data lake to apply this
@Bgmifortimepass Жыл бұрын
@@AshokGupta if we take our source is data lake and then what about update file(like we need to maintain different file or what?)
@AshokGupta Жыл бұрын
Delta lake supports updates and internally maintains right file version. We need not worry about file
@Bgmifortimepass Жыл бұрын
@@AshokGupta my question is in real time senario how and where we maintain source and updated files in single path or different path
@karlosvaliente Жыл бұрын
Question 15 is A according to databricks. I doubt about it
@AshokGupta Жыл бұрын
Ome of the questions are bit ambiguous. What's right answer according to you
@karlosvaliente Жыл бұрын
@@AshokGupta C, because you have to refresh to avoid getting last cached snapshot from table. However it says "writing cluster" and is confusing for me
@DataTalks2023 Жыл бұрын
Check this out for Databricks SQL -kzbin.info/www/bejne/f6XJc6d6jaapZ9U
@sravankumar17672 жыл бұрын
Superb explanation
@raghvendrapratapsingh79092 жыл бұрын
how to change column sequence in delta table...condition is that i want to use only spark sql not dataframe API please help
@AshokGupta2 жыл бұрын
you can recreate the table, by dropping and re populating.
@vkincanada57812 жыл бұрын
@Ashok - Can you please.. provide one on one tutoring for me? I need support for Azure DevOPS CI/CD for Databrciks, ADF projects. Please provide your means of contact, Will wait for your response .
@AshokGupta2 жыл бұрын
you can join our telegram group, we try to help everyone in community. t.me/AzureDataEngineer
@kiranachanta66312 жыл бұрын
Awesome content!! One question though :) I have built a streaming pipeline. Now let's assume, events/files are getting generated every 3 hrs in my source. How will the data bricks cluster & notebook be invoked every 3 hrs to process the new events? does the cluster should be up and running all the time?
@AshokGupta2 жыл бұрын
You can schedule job for every 3 hour.. on job cluster. It will provision new cluster every time and terminate after its done.
@kiranachanta66312 жыл бұрын
@@AshokGupta Awesome!
@sachinv99232 жыл бұрын
Thank you!
@krish_telugu2 жыл бұрын
Where linux server demo here
@AshokGupta2 жыл бұрын
as long as you are able to authenticate linux server same method will work there as well.
@ViktoriaLessai2 жыл бұрын
The content is perfect, thanks!
@AshokGupta2 жыл бұрын
Glad you think so!
@Prapti_Bisht2 жыл бұрын
What will be the command to access delta table in pyspark?
@AshokGupta2 жыл бұрын
If you are using spark sql, it will be same as any other table. You can just say "select * from deltayablename ". If you are using pyspark API.. you can write. spark.read.format("delta").load("/tmp/delta-table")
@Prapti_Bisht2 жыл бұрын
@@AshokGupta i doubt...then why we use spark.tabel("delta table name")
@AshokGupta2 жыл бұрын
I would recommend to try it once
@niravkothari90712 жыл бұрын
volume is very low
@tanushreenagar31162 жыл бұрын
Nice sir
@AshokGupta2 жыл бұрын
Thank you
@urvxfvdzrnp2 жыл бұрын
Excellent
@AshokGupta2 жыл бұрын
Thank you
@yagnam1232 жыл бұрын
When will you realease next part
@AshokGupta2 жыл бұрын
Soon probably this weekend. Did you like the content? Any feedback is welcome..
@yagnam1232 жыл бұрын
@@AshokGupta Content is good
@msshroff2 жыл бұрын
For question 8, Both options B and C are syntactically correct, But the question says create "regardless of whether a table already exists with this name" So with option C "IF NOT EXISTS", it would fail if table already exists. So option B is the only valid answer.
@AshokGupta2 жыл бұрын
I think you are right. thanks. It will not fail in "If Not Exists" as. well instead it will not do anything.
@yagnam1232 жыл бұрын
Yes it's clearly saying in the question Irrespective of table existing write DDL for create
@anupgupta57812 жыл бұрын
Hi dude is this practice set questions are enough for passing the exam?
@AshokGupta2 жыл бұрын
It will give you idea about quality of questions
@esteban84452 жыл бұрын
promosm 😅
@AshokGupta2 жыл бұрын
I did not get u
@vidyasarathi152 жыл бұрын
Hi , do you have databricks data analyst certification exam dumps?
@AshokGupta2 жыл бұрын
Not really, I have not appeared for that. However, you can see sample questions in site I beleive
@vidyasarathi152 жыл бұрын
@@AshokGupta could you please give me the link.. i don't get to see anything.. 😔
In yml we are creating cluster what is the need for that. As when the jobs will be triggered job cluster will automatically created
@AshokGupta2 жыл бұрын
This yaml will create all purpose cluster
@sravankumar17672 жыл бұрын
Nice explanation 👌 👍 👏
@AshokGupta2 жыл бұрын
Thanks
@subhanivasareddythummapudi38362 жыл бұрын
How INFORMATICA POWERCENTER and ADF is different from each other. Or both work in simple way....? If not, how ADF is better than Informatica.?
@subhanivasareddythummapudi38362 жыл бұрын
How INFORMATICA POWERCENTER and ADF is different from each other. Or both work in simple way....? If not, how ADF is better than Informatica.?
@dcity-b4u2 жыл бұрын
How do you push a file from blob to an ftp site directory?
@AshokGupta2 жыл бұрын
As far as I know, that's not supported directly in ADF.. you can use Azure function instead
@mkumardadhich2 жыл бұрын
In your nested JSON example , It has loaded only one row(One person) in the table whereas in the JSON file there are three Persons. Loading the nested JSON file using copy activity is not the correct way. I think the best option is data flow
@irecommendtv20672 жыл бұрын
Please do you have a CICD pipeline to deploy notebooks to the workspace?