This is my KZread channel where I explain various topics on Data analytics,Data Engineer, Azure Data engineering ,AWS data engineering , aws glue, aws athena, aws s3, aws emr, Cloud Platform ,AWS, Big data ,ETL, machine learning, deep learning, devops, Linux, GCP,big data engineer, data engineer, and AI with many real-world problem scenarios. My main aim is to make everyone familiar by showing them practical example with real case scenario .Please subscribe and support the channel. As i love new technology, all these videos are free and I promise to make more interesting content as we go ahead.
Mail - [email protected]
Join me t.me/+Cb98j1_fnZs3OTA1
Пікірлер
are these tables delta tables?
Is there any way to have the csv output file name match the source json file name?
Yes I will show
In the query to delete the duplicates the query will also print the record which is duplicate as one of the record is ranked as 1
U can select column what u need
Example for frequent, unfrequent,
i got 3 errors 154 commented memebers not got any single error woow great please help
What is the error
aggregate table creation error: IllegalArgumentException: All week-based patterns are unsupported since Spark 3.0, detected: e, Please use the SQL function EXTRACT instead
Not sure about this error ...which part u are stuck
Thank you! Excited for the course.
You can also follow latest playlist
Covered all the cases, excellent work!!
good explanation thanks but all this are not working in a pipeline can u make a new video for all pramaterization videos (19 ,20,21 videos) by creating new pipeline and triggering it
Sure I will do that
@@learnbydoingit thanks eagerly waiting for it
i have create same as above dataset by going to adf ---author---dataset---its working!! fine 2- when i try to create same dataset via using new pipeline and copy activity---source (dataset) same as above in video and sink as blob 3-when i try to run pipeline im getting error no value provided to parameter db name and table name in case my question is not clear excuse me ! 4- how to do the same activity using pipeline and triger it?
Have u created parameter...if u click on the blank page on copy activity u will see parameter option there have u specified or not ?
@@learnbydoingit yes i have created still im getting error
@@learnbydoingit can u write down steps how to do it in a pipeline n trigger it becoz im struck on this video i need to move on n complete other videos thanks
Very informative bro tnq
can me rename a column drop a column n re arrange the column with derived cloumn as we did in select? if yes then what is the difference between select n derived
If you have to derive new column based on certain expression like do the sum of 2 column and create new column then which one you will use? Hope u will get idea
@@learnbydoingit yes got it thanks with select we can concat the columns
The file '_SUCCESS' may not render correctly as it contains an unrecognized extension triger got sucessful but in container i have recieved file as scuess with 0kb when i open it i get above message without any file to preview or edit
how to delete table from sql via delete activity ?
Table u can't delete through delete activity but, if in sql query u can pass drop table statement and then it will be dropped
this video is helpful for training people who are learning to transition to cloud computing.keep posting
What does KPI means?
KPI stands for Key Performance Indicator, which is a quantifiable metric used to track progress towards a specific business objective.
hi can you explain how to use aliases feature also? thanks for sharing knowledge!
more how many left total for pyspark?
Few more are pending then Project
Thanks Budy
Is it similar to sql window functions?
Yes
I think we can use 'monthdiff' after where in 1st question right? rather than repeating whole datediff() line
Sir, kindly add Azure Synapse videos
im getting below error: wile debuging in get meta data activity no errors and all file copied to output folder evne before trigger then after triggering im getting below erro message Failed to run foreachitrate (Pipeline). {"code":"BadRequest","message":"ErrorCode=InvalidTemplate, ErrorMessage=The template validation failed: 'The 'runAfter' property of template action 'ForEach1Scope' is not valid. The status values for action 'Get Metadata1Scope' must be unique. Found duplicate values: 'Succeeded'","target":"pipeline/foreachitrate/runid/c2f978e0-100a-4235-a1dd-5caa8c62a25e","details":null}
For each are u getting error
how do we know that we have to pass only name in wildcard?
There will be different requirements and usecase and based on that we have to deal
hi what about in a single trigger cant we provide multiple filepath and table name
Thank you brooo
This playlist has 58 videos, is it enough to learn from scratch and get a job of azure data engineer with 3 yrs of experience?? If not , please let us know on what else need to be done to achieve it
Yes and sql also u need to do ....we have another playlist in depth we are covering adf pyspark sql u can follow that too
So grateful for this content. Thank you!
it is nice how you covered all these complex things in such a small timespan
for amazon redshift dataware house, can you upload some tutorial ? and also some data engineering project which covers all these services can you do one video? it will be very useful ?
Sure
bro your content is the best so far :) thankyou so much
Hi sir is your playlist for pyspark enough to completely learn pyspark ?
Yes we are adding more , as well project
error: Dataset is using 'AzureSqlDatabase' linked service with SQLVersion 'Recommended', which is not supported in data flow.
Pls do watch 48-50 video for this error
On second slide you have mentioned ' We have to build one pipeline which will transfer data and run daily' What do you mean by run daily ?
Daily schedule
@@learnbydoingit But how the pipeline will trigger daily ? Because we haven't set any schedule!!
Thanks
This helps df to convert to spark table very useful 🙂
can we take Central India in region in free subscription?
while publishing the trigger im getting below error please help The Microsoft.EventGrid resource provider is not registered in subscription 0bc822f0-35e3-4b16-bc69-bb4d69d152d3. Register the provider in the subscription and retry the operation. Activity id:0cc9f3e7-26da-4503-8e86-9bbca505a7f4, please reply
any update on this please reply thanks
while publishing the trigger im getting below error please help The Microsoft.EventGrid resource provider is not registered in subscription 0bc822f0-35e3-4b16-bc69-bb4d69d152d3. Register the provider in the subscription and retry the operation. Activity id:0cc9f3e7-26da-4503-8e86-9bbca505a7f4,
Hello ji, i have a question that can we publish multiple pipelines at once? please respond on this. There is scenario where I'm creating the If Condition pipeline but in the input container i don't have dept table, so I have created pipeline for copy activity but unable to publish the pipeline since the if condition is incomplete. please help to solve this
If pipeline is incomplete then u will get error
If pipeline is incomplete then u will get error
bhai itne sare playlist of azure data factory kindly delete all confusing and make one playlist in one playlist 72 videos one in 52 which to follow and even datafcatory many playlist and even for projects its getting confused for beginner which playlist to start from?
Both same only bro in 72 we have pyspark also
You can follow 72 vala
Can you please provide these sql data samples resource
Azure sql when u create u will get option to select sample table ...you will get sample data there
Great job bro 🎉