Пікірлер
@vinubendi3843
@vinubendi3843 Ай бұрын
what is rows affected in output and if we write insert statement in script task and if data is there in rows affected, can we consider it is the data inserted into the table ?
@SouravMondal-hb8rz
@SouravMondal-hb8rz Ай бұрын
Can we get the Private Ip of that ADF ( IR) so that we can whitelist that ip
@rahulpanda9256
@rahulpanda9256 3 ай бұрын
Hi Thanks a lot for this explanation. Does this also handle delete operations at the source?
@akshaygidwani9219
@akshaygidwani9219 4 күн бұрын
It would not be able to handle delete operation
@kapsingla
@kapsingla 3 ай бұрын
When you created this setup did you create the publish Branch while configuring the repo in Azure data Factory?
@AshokGupta
@AshokGupta 3 ай бұрын
Adf automatically creates publish branch
@adityavanipenta14
@adityavanipenta14 6 ай бұрын
Hi I need to create a delta table in databricks with scd2 logic
@adityavanipenta14
@adityavanipenta14 6 ай бұрын
Can you help I have the requirements with me
@shanepeck2387
@shanepeck2387 10 ай бұрын
"promo sm"
@AshokGupta
@AshokGupta 10 ай бұрын
?
@VavilalaPavan
@VavilalaPavan 10 ай бұрын
voice is first not at all good
@AshokGupta
@AshokGupta 10 ай бұрын
Please watch our other videos we have improved
@AshokGupta
@AshokGupta 10 ай бұрын
Please watch our other videos we have improved
@pavankumarveesam8412
@pavankumarveesam8412 Жыл бұрын
sound is pretty less
@pradeepthanniru2602
@pradeepthanniru2602 Жыл бұрын
The content is good and nice explanation
@AshokGupta
@AshokGupta Жыл бұрын
Thanks a lot
@PalaniKumar-b7e
@PalaniKumar-b7e Жыл бұрын
How to create multiple clusters with same scripts? Also how can we create workflow job clusters in azure pipeline? Please help.
@AshokGupta
@AshokGupta Жыл бұрын
You can duplicate create cluster script as many times you want.
@tanushreenagar3116
@tanushreenagar3116 Жыл бұрын
Nice
@paulinexiong3246
@paulinexiong3246 Жыл бұрын
for sql database as source, I believe you need to enable CDC on the source database/tables
@guptaashok121
@guptaashok121 Жыл бұрын
Not required.. it will work based on a timestpam column of table
@arulpalaniappan
@arulpalaniappan Жыл бұрын
This is really helpful! Thanks
@swatidorge7133
@swatidorge7133 Жыл бұрын
Well done 👍👍👍👍
@swatidorge7133
@swatidorge7133 Жыл бұрын
Well done 👍👍👍👍
@syamamujuru
@syamamujuru Жыл бұрын
In Question 36 They mentioned development mode, so Cluster will not shutdown.
@GrowthMindset_J
@GrowthMindset_J Жыл бұрын
Do you have another set of Databricks practice questions ?
@AshokGupta
@AshokGupta Жыл бұрын
kzbin.info/www/bejne/haW3kpiKf7mrpdU
@bairagirout9323
@bairagirout9323 Жыл бұрын
In the last step send email, what should i mention in the email body to get the content of the csv in a HTML table. Please explain the send email step also
@coolshrawan
@coolshrawan 3 ай бұрын
just call the output of compose. outputs('Compose_3')
@V-Barah
@V-Barah Жыл бұрын
it good video, but you didn't show the required configurations: How to get the token, Cluster ID , and notebook path and all details . Someone new to this area will not understand
@mikebauer9335
@mikebauer9335 Жыл бұрын
"PromoSM"
@tsdhd6275
@tsdhd6275 Жыл бұрын
content is good but voice is so f****d up
@mannykhan7752
@mannykhan7752 Жыл бұрын
Some of the questions here have wrong answers. In the case of Q15 the answer is E as the table has to be updated before the next ingestion. This question also appears on the Databricks Data Engineer official practice test that can be downloaded from their website. That's why I'm sure your answer is incorrect.
@AshokGupta
@AshokGupta Жыл бұрын
This set of question and answer are from databricks site only. However I agree some answer might be wrong that time. Now they might have corrected it.
@prasadk6507
@prasadk6507 Жыл бұрын
Nice one, keep updating fraternity 🎉
@ravimishra6792
@ravimishra6792 Жыл бұрын
Very helpful
@AshokGupta
@AshokGupta Жыл бұрын
Thanks
@Bgmifortimepass
@Bgmifortimepass Жыл бұрын
what is our source like
@AshokGupta
@AshokGupta Жыл бұрын
Source can be anything, first it needs to brought to data lake to apply this
@Bgmifortimepass
@Bgmifortimepass Жыл бұрын
@@AshokGupta if we take our source is data lake and then what about update file(like we need to maintain different file or what?)
@AshokGupta
@AshokGupta Жыл бұрын
Delta lake supports updates and internally maintains right file version. We need not worry about file
@Bgmifortimepass
@Bgmifortimepass Жыл бұрын
@@AshokGupta my question is in real time senario how and where we maintain source and updated files in single path or different path
@karlosvaliente
@karlosvaliente Жыл бұрын
Question 15 is A according to databricks. I doubt about it
@AshokGupta
@AshokGupta Жыл бұрын
Ome of the questions are bit ambiguous. What's right answer according to you
@karlosvaliente
@karlosvaliente Жыл бұрын
@@AshokGupta C, because you have to refresh to avoid getting last cached snapshot from table. However it says "writing cluster" and is confusing for me
@DataTalks2023
@DataTalks2023 Жыл бұрын
Check this out for Databricks SQL -kzbin.info/www/bejne/f6XJc6d6jaapZ9U
@sravankumar1767
@sravankumar1767 2 жыл бұрын
Superb explanation
@raghvendrapratapsingh7909
@raghvendrapratapsingh7909 2 жыл бұрын
how to change column sequence in delta table...condition is that i want to use only spark sql not dataframe API please help
@AshokGupta
@AshokGupta 2 жыл бұрын
you can recreate the table, by dropping and re populating.
@vkincanada5781
@vkincanada5781 2 жыл бұрын
@Ashok - Can you please.. provide one on one tutoring for me? I need support for Azure DevOPS CI/CD for Databrciks, ADF projects. Please provide your means of contact, Will wait for your response .
@AshokGupta
@AshokGupta 2 жыл бұрын
you can join our telegram group, we try to help everyone in community. t.me/AzureDataEngineer
@kiranachanta6631
@kiranachanta6631 2 жыл бұрын
Awesome content!! One question though :) I have built a streaming pipeline. Now let's assume, events/files are getting generated every 3 hrs in my source. How will the data bricks cluster & notebook be invoked every 3 hrs to process the new events? does the cluster should be up and running all the time?
@AshokGupta
@AshokGupta 2 жыл бұрын
You can schedule job for every 3 hour.. on job cluster. It will provision new cluster every time and terminate after its done.
@kiranachanta6631
@kiranachanta6631 2 жыл бұрын
@@AshokGupta Awesome!
@sachinv9923
@sachinv9923 2 жыл бұрын
Thank you!
@krish_telugu
@krish_telugu 2 жыл бұрын
Where linux server demo here
@AshokGupta
@AshokGupta 2 жыл бұрын
as long as you are able to authenticate linux server same method will work there as well.
@ViktoriaLessai
@ViktoriaLessai 2 жыл бұрын
The content is perfect, thanks!
@AshokGupta
@AshokGupta 2 жыл бұрын
Glad you think so!
@Prapti_Bisht
@Prapti_Bisht 2 жыл бұрын
What will be the command to access delta table in pyspark?
@AshokGupta
@AshokGupta 2 жыл бұрын
If you are using spark sql, it will be same as any other table. You can just say "select * from deltayablename ". If you are using pyspark API.. you can write. spark.read.format("delta").load("/tmp/delta-table")
@Prapti_Bisht
@Prapti_Bisht 2 жыл бұрын
@@AshokGupta i doubt...then why we use spark.tabel("delta table name")
@AshokGupta
@AshokGupta 2 жыл бұрын
I would recommend to try it once
@niravkothari9071
@niravkothari9071 2 жыл бұрын
volume is very low
@tanushreenagar3116
@tanushreenagar3116 2 жыл бұрын
Nice sir
@AshokGupta
@AshokGupta 2 жыл бұрын
Thank you
@urvxfvdzrnp
@urvxfvdzrnp 2 жыл бұрын
Excellent
@AshokGupta
@AshokGupta 2 жыл бұрын
Thank you
@yagnam123
@yagnam123 2 жыл бұрын
When will you realease next part
@AshokGupta
@AshokGupta 2 жыл бұрын
Soon probably this weekend. Did you like the content? Any feedback is welcome..
@yagnam123
@yagnam123 2 жыл бұрын
@@AshokGupta Content is good
@msshroff
@msshroff 2 жыл бұрын
For question 8, Both options B and C are syntactically correct, But the question says create "regardless of whether a table already exists with this name" So with option C "IF NOT EXISTS", it would fail if table already exists. So option B is the only valid answer.
@AshokGupta
@AshokGupta 2 жыл бұрын
I think you are right. thanks. It will not fail in "If Not Exists" as. well instead it will not do anything.
@yagnam123
@yagnam123 2 жыл бұрын
Yes it's clearly saying in the question Irrespective of table existing write DDL for create
@anupgupta5781
@anupgupta5781 2 жыл бұрын
Hi dude is this practice set questions are enough for passing the exam?
@AshokGupta
@AshokGupta 2 жыл бұрын
It will give you idea about quality of questions
@esteban8445
@esteban8445 2 жыл бұрын
promosm 😅
@AshokGupta
@AshokGupta 2 жыл бұрын
I did not get u
@vidyasarathi15
@vidyasarathi15 2 жыл бұрын
Hi , do you have databricks data analyst certification exam dumps?
@AshokGupta
@AshokGupta 2 жыл бұрын
Not really, I have not appeared for that. However, you can see sample questions in site I beleive
@vidyasarathi15
@vidyasarathi15 2 жыл бұрын
@@AshokGupta could you please give me the link.. i don't get to see anything.. 😔
@AshokGupta
@AshokGupta 2 жыл бұрын
Let me try to find
@AshokGupta
@AshokGupta 2 жыл бұрын
@@vidyasarathi15 its there in description
@AshokGupta
@AshokGupta 2 жыл бұрын
@@vidyasarathi15 www.databricks.com/p/thank-you/databricks-certification-preparation-on-demand
@abhijeetsingh9730
@abhijeetsingh9730 2 жыл бұрын
In yml we are creating cluster what is the need for that. As when the jobs will be triggered job cluster will automatically created
@AshokGupta
@AshokGupta 2 жыл бұрын
This yaml will create all purpose cluster
@sravankumar1767
@sravankumar1767 2 жыл бұрын
Nice explanation 👌 👍 👏
@AshokGupta
@AshokGupta 2 жыл бұрын
Thanks
@subhanivasareddythummapudi3836
@subhanivasareddythummapudi3836 2 жыл бұрын
How INFORMATICA POWERCENTER and ADF is different from each other. Or both work in simple way....? If not, how ADF is better than Informatica.?
@subhanivasareddythummapudi3836
@subhanivasareddythummapudi3836 2 жыл бұрын
How INFORMATICA POWERCENTER and ADF is different from each other. Or both work in simple way....? If not, how ADF is better than Informatica.?
@dcity-b4u
@dcity-b4u 2 жыл бұрын
How do you push a file from blob to an ftp site directory?
@AshokGupta
@AshokGupta 2 жыл бұрын
As far as I know, that's not supported directly in ADF.. you can use Azure function instead
@mkumardadhich
@mkumardadhich 2 жыл бұрын
In your nested JSON example , It has loaded only one row(One person) in the table whereas in the JSON file there are three Persons. Loading the nested JSON file using copy activity is not the correct way. I think the best option is data flow
@irecommendtv2067
@irecommendtv2067 2 жыл бұрын
Please do you have a CICD pipeline to deploy notebooks to the workspace?
@AshokGupta
@AshokGupta 2 жыл бұрын
That's explained in video